Abstract
Background:
The prefrontal cortices play an essential role in cognitive-emotional and working memory processes through interactions with multiple brain regions.
Methods:
This article further develops a unified neural architecture that explains many recent and classical data about prefrontal function and makes testable predictions.
Results:
Prefrontal properties of desirability, availability, credit assignment, category learning, and feature-based attention are explained. These properties arise through interactions of orbitofrontal, ventrolateral prefrontal, and dorsolateral prefrontal cortices with the inferotemporal cortex, perirhinal cortex, parahippocampal cortices; ventral bank of the principal sulcus, ventral prearcuate gyrus, frontal eye fields, hippocampus, amygdala, basal ganglia, hypothalamus, and visual cortical areas V1, V2, V3A, V4, middle temporal cortex, medial superior temporal area, lateral intraparietal cortex, and posterior parietal cortex. Model explanations also include how the value of visual objects and events is computed, which objects and events cause desired consequences and which may be ignored as predictively irrelevant, and how to plan and act to realise these consequences, including how to selectively filter expected versus unexpected events, leading to movements towards, and conscious perception of, expected events. Modelled processes include reinforcement learning and incentive motivational learning; object and spatial working memory dynamics; and category learning, including the learning of object categories, value categories, object-value categories, and sequence categories, or list chunks.
Conclusion:
This article hereby proposes a unified neural theory of prefrontal cortex and its functions.
Keywords: Category learning, cognitive-emotional interactions, working memory, reinforcement learning, incentive motivational learning, contextual cueing, inferotemporal cortex, orbitofrontal cortex, ventrolateral prefrontal cortex, dorsolateral prefrontal cortex, amygdala, basal ganglia, perirhinal cortex, parahippocampal cortex, frontal eye fields, ventral prearcuate gyrus, ventral bank of the principal sulcus
1. Introduction: towards a mechanistic theoretical understanding of prefrontal functions
1.1. Functional roles of orbitofrontal, ventrolateral, and dorsolateral prefrontal cortex
The prefrontal cortex (PFC) contributes to many of the higher cognitive, emotional, and decision-making processes that define human intelligence, while also controlling the release of goal-oriented actions towards valued goal objects. As noted in the Wikipedia article about PFC,
The most typical psychological term for functions carried out by the prefrontal cortex area is executive function. Executive function relates to abilities to differentiate among conflicting thoughts, determine good and bad, better and best, same and different, future consequences of current activities, working toward a defined goal, prediction of outcomes, expectation based on actions, and social ‘control’ (the ability to suppress urges that, if not suppressed, could lead to socially unacceptable outcomes).
Elliott et al. (2000) discussed how the PFC contributes to generating behaviours that are flexible and adaptive, notably in novel situations, and to suppressing actions that are no longer appropriate, thereby freeing humans and other primates from being forced to respond more reflexively to current sensory inputs. These authors also review the various terms that have been used to describe PFC functions, including planning (Luria, 1966), memory for the future (Ingvar, 1985), executive control (Baddeley, 1986), working memory (Goldman-Rakic, 1987), supervisory attention (Shallice, 1988), and top-down modulation of bottom-up processes (Frith and Dolan, 1997).
Miller and Cohen (2001) reviewed data that are consistent with these concepts, noting that these properties result
from the active maintenance of patterns of activity in the prefrontal cortex that represent goals and the means to achieve them. They provide bias signals to other brain structures whose net effect is to guide the flow of activity along neural pathways that establish the proper mappings between inputs, internal states, and outputs needed to perform a given task. (p. 167)
Said in yet another way, the PFC is involved with predicting future outcomes and enabling animals and humans to respond adaptively to them.
Wise (2008) espoused a similar view that he vividly summarised as follows: ‘The long list of functions often attributed to prefrontal cortex may all contribute to knowing what to do and what will happen when rare risks arise or outstanding opportunities knock’ (p. 599).
Even this brief heuristic summary of some of the multiple functions of the PFC illustrates the challenge facing any theorist who wishes to model this, or indeed any, part of the brain. The challenge is that various functionally distinct parts of the PFC are connected to each other in complex ways, in addition to being widely connected with multiple other brain regions. Broadly speaking, the dorsal PFC is interconnected with brain regions involved with attention, cognition, and action (Goldman-Rakic, 1988), whereas the ventral prefrontal cortex is interconnected with brain regions involved with emotion (Price, 1999). These facts do not, however, explain how these brain circuits give rise to these distinct psychological functions as emergent properties that arise from interactions among brain regions that work together as functional systems.
A critical question for any theorist of mind and brain is thus: How can the divide between brain mechanisms and psychological functions be bridged? How can this be done with sufficient mechanistic precision to explain and predict challenging interdisciplinary data? Section 1.3 summarises a well-established theoretical method whereby the emergent properties of brain mechanisms may be linked to the mental functions that they control.
Before summarising this theoretical method, Section 1.2 will review some of the prefrontal properties that recent experiments have reported and that will be explained in later sections using this theoretical method. In particular, many prefrontal cortical properties can be subsumed under two unifying mechanistic themes: cognitive-emotional dynamics and working memory dynamics. A macrocircuit of the brain regions that embody these processes in a unified predictive Adaptive Resonance Theory (pART), a model that herein explains and predicts many prefrontal cortical data, is shown in Figure 1.
Section 2 will summarise relevant data and models of cognitive-emotional dynamics and Section 3 will summarise relevant data and models of working memory dynamics. Cognitive-emotional dynamics, and models thereof, include how orbitofrontal cortex (OFC) interacts with brain regions like temporal cortex, amygdala, hippocampus, and cerebellum to regulate processes like category learning and adaptively timed motivated attention and action to acquire a valued goal. Working memory dynamics, and models thereof, include how sequences of events are temporarily stored in ventrolateral and dorsolateral prefrontal cortex, how these sequences are unitised, or chunked, into cognitive plans, and how interactions of prefrontal regions with other brain regions like perirhinal cortex (PRC), parahippocampal cortex (PHC), and basal ganglia (BG) enables predictions and actions to be chosen that are most likely to succeed based on sequences of previously rewarded experiences.
These sections will also compare and contrast the neural models that are used herein with other models in the literature and will make testable predictions to further clarify the brain mechanisms that underlie these data.
Section 4 will provide concluding remarks that highlight some of the article’s main themes.
A unified theoretical explanation with such ambitious goals will necessarily be presented in stages. The text will state key data, and outline model explanations of them, at the beginning of each section to help readers to frame subsequent, more detailed, explanations. The text will also attempt to provide a self-contained explanation of all the models that it uses. This explanation cannot be exhaustive because the theoretical and experimental literatures that fall within the scope of the models are vast. The exposition will nonetheless provide enough information for the reader to appreciate how the explanations work, and why they are compelling.
The exposition will also try to deal with another, even more basic, problem: How can any theory penetrate the complexity of a behaving brain? Indeed, a few facts, taken alone, can have multiple explanations. They provide insufficient guidance to rule out plausible, but incorrect, explanations. In contrast, the current theoretical method confronts hundreds of facts about a particular behavioural function, taken from the entire experimental literature on multiple organisational levels from behavioural to cellular, with all known modelling principles, mechanisms, microcircuits, and architectures. When this is done properly, every such fact, and every modelling hypothesis about it, is put under enormous ‘conceptual pressure’ that typically allows only one possible explanation to survive. The text tries to summarise enough of these constraints to clarify how and why the explanations work and to motivate readers to pursue further reading about topics that particularly interest them in related articles.
1.2. Prefrontal desirability, availability, credit assignment, and feature-based attention
These models help to understand and mechanistically explain recent data about the OFC, ventrolateral prefrontal cortex (VLPFC), and dorsolateral prefrontal cortex (DLPFC). The text will first focus upon a striking conclusion that has arisen about roles for OFC and VLPFC that were derived from behavioural experiments in monkeys who had experienced excitotoxic lesions of these brain regions. The conclusion is that OFC and VLPFC are involved in updating a predicted outcome’s desirability versus its availability, respectively (Rudebeck et al., 2017). Recent data have also claimed that DLPFC neurons encode a solution of the credit-assignment problem (Assad et al., 2017). What the concepts of desirability, availability, and credit assignment mean will be operationally defined in the sections where these data are explained.
Some of the results on desirability that are reported by Rudebeck et al. (2017) were predicted, indeed simulated, by a well-established neural model of cognitive-emotional interactions, notably reinforcement learning and motivated attention, in Grossberg et al. (2008). Aspects of the role of VLPFC in updating availability are clarified by a well-established neural model of cognitive working memory and the learning of predictive cognitive plans, or list chunks (Cohen and Grossberg, 1986, 1987; Grossberg, 2017b; Grossberg and Pearson, 2008). This model has been further developed in Huang and Grossberg (2010) to quantitatively simulate data about sequential decisions during contextually cued visual search and in Silver et al. (2011) to simulate prefrontal cortical neurophysiological data about sequences of eye movement decisions. Data about credit assignment are explained by combining the model of cognitive working memory and list chunks with the model of reinforcement learning and motivated attention to show how chunks that lead to successful predictions are amplified, while those that do not are suppressed.
In Sections 2 and 3, the article will first summarise the above kinds of data that are explanatory targets of the article, then review the models that explain them, and go on to use these models to provide explanations and predictions of additional interdisciplinary data. It will then extend them in a consistent way to mechanistically and functionally explain data about prefrontal sources of feature-based attention in monkeys (Bichot et al., 2015) and humans (Baldauf and Desimone, 2014).
The PFC has been studied with multiple methods. Some articles have studied the PFC of humans with functional neuroimaging in normal subjects or clinical patients, while others have studied monkeys or rats with neurophysiological or anatomical methods. Important functional conclusions have also been derived by combining selective lesions with behavioural studies in monkeys. Recent studies have, however, also shown that different lesion methods can yield quite different results. For example, monkeys with selective excitotoxic lesions of the OFC, unlike monkeys who have received aspiration OFC lesions, are unimpaired in learning and reversing object choices based on reward feedback (Rudebeck et al., 2013). Neurotoxic lesions of the amygdala (Izquierdo and Murray, 2007) have also led to results that challenge earlier demonstrations using aspiration and radiofrequency lesions that the amygdala is needed for object reversal learning (Aggleton and Passingham, 1981; Jones and Mishkin, 1972; Spiegler and Mishkin, 1981).
Why do different lesion methods yield such different results? One main reason is that, unlike excitotoxic lesions, other lesion methods, including aspiration and radiofrequency lesions, may also damage fibres of passage to adjacent cortical areas. The fact that OFC activity has been reported during reversal learning (Fellows and Farah, 2003; Morrison et al., 2011; Rolls, 2000; Rolls et al., 1994) suggests that several neuronal regions and pathways may be involved in this behavioural competence.
This picture is complicated further by different definitions of the brain areas that constitute OFC and VLPFC. The conclusions above hold if OFC is understood to consist of areas 11, 13, and 14. If, however, area 12o is also included, which overlaps what Chau et al. (2015) call the lateral OFC (lOFC), then various properties of what Rudebeck et al. (2017) would assign to VLPFC get attributed to OFC. Herein, the convention will be followed that OFC consists of areas 11, 13, and 14. Another caveat is that there appear to be species-specific variations. For example, unlike old world monkeys, excitotoxic lesions in new world monkeys such as marmosets (Roberts, 2006) can impair these animals on reversal tasks. These variations will not be analysed herein.
1.3. A theoretical method for linking brain to mind: method of minimal anatomies
One successful method for linking brain mechanisms to behavioural functions has been developed and applied during the past 60 years. This method has led to neural models that often anticipated psychological and neurobiological data about the PFC, among other brain regions. It continues to do so, as this article will illustrate.
A key theme of this theoretical ‘method of minimal anatomies’ is that one cannot derive a theory of an entire brain in one step. Rather, one does so incrementally in stages. This incremental theoretical method embodies a kind of design evolution whereby each model embodies a certain set of design principles and mechanisms that the evolutionary process has discovered whereby to cope with a given set of environmental challenges. Then, the model is refined, or unlumped, to embody an even larger set of design principles and mechanisms and thereby expands its explanatory and predictive power. This process of evolutionary unlumping continues unabated, leading to current models that can individually explain psychological, anatomical, neurophysiological, biophysical, and biochemical data about a given faculty of biological intelligence.
Using this method, the pART theory of cognitive-emotional and working memory dynamics has been discovered and incrementally elaborated over the past 60 years (Figure 1). The current exposition will emulate the theoretical method by first summarising the simplest models that can explain nontrivial amounts of prefrontal data, before unlumping them to explain even more data. Although, for expository reasons, multiple models will be mentioned, it needs to be understood that there is just one unified theory behind all the explanations that joins together cognitive-emotional dynamics and working memory dynamics.
The theoretical derivation always begins with behavioural data because brain evolution needs to achieve behavioural success. Starting with behavioural data enables models to be derived whose brain mechanisms have been shaped during evolution by behavioural success. Starting with a large database helps to rule out incorrect, but otherwise seemingly plausible, answers.
Such a derivation leads to the discovery of novel design principles and mechanisms with which to explain how an individual, behaving in real time, can generate the behavioural data as emergent properties. This conceptual leap from data to design is the art of modelling. Once derived, despite being based on psychological constraints, the minimal mathematical model that realises the behavioural design principles has always been interpretable in terms of brain mechanisms. Sixty years of modelling have hereby supported the hypothesis that brains look the way that they do because they embody computational designs whereby individuals autonomously adapt to changing environments in real time. The link from behaviour-to-principle-to-model-to-brain has, in addition, often disclosed unexpected functional roles of the derived brain mechanisms that are not clear from neural data alone.
A ‘minimal’ model is one for which if any of the model’s mechanisms is removed, then the surviving model can no longer explain a key set of previously explained data. Once a connection is made top-down from behaviour to brain by such a minimal model, mathematical and computational analysis discloses what data the minimal model, and its variations, can and cannot explain. Such an analysis focuses attention upon design principles that the current model does not yet embody. These new design principles and their mechanistic realisations are then consistently incorporated into the model by unlumping it to generate a more realistic model. If the model cannot be refined in this way, then that is strong evidence that the current model contains a serious error and must be discarded. The unified pART model that is discussed herein, which explains key functional processes in the brain regions depicted in Figure 1, has withstood multiple stages of unlumping.
2. Cognitive-emotional dynamics and the OFC
2.1. Orbitofrontal coding of desirability as probed by selective satiation
This section of the article summarises an explanation of how the desirability of an outcome is computed in our brains. The relevant data will first be reviewed before the model that can explain, indeed anticipated, them is summarised, along with other data explanations and predictions. An outline of the model’s explanation of desirability will first be given to frame the subsequent exposition of the model mechanisms that accomplish this.
As noted above, the experiments of Rudebeck et al. (2017) support the hypothesis that the OFC, but not the VLPFC, plays a necessary role in choices that are based on outcome desirability. In contrast, the VLPFC, but not the OFC, plays a necessary role in choices that are based on outcome availability. What desirability means is explained operationally by an experiment (their Experiment 2) that manipulates the subjective value of different food rewards with a stimulus-based reinforcer devaluation, or satiation, procedure that was earlier used by Málková et al. (1997), while keep the probability and magnitude of reward stable. The monkeys in this experiment were trained with some conditioned stimulus (CS) objects that were associated with Food 1, and others associated with Food 2. The two foods acted like unconditioned stimuli (US) in the experiment. Following the selective satiation procedure, monkeys were presented with pairs of objects, one object associated with Food 1 and the other with Food 2. The effects of devaluation were measured by calculating how much monkeys shifted their choices towards objects associated with a higher value food, relative to baseline choices.
Both unoperated control monkeys and monkeys with excitotoxic VLPFC lesions could update and use the current value of food reward to guide their choices. In contrast, monkeys with excitotoxic OFC lesions chose stimuli that were associated with the sated food at a much higher rate. Various tests led unambiguously to the conclusion that this deficit in monkeys with OFC lesions arose from their inability to link objects with the current value of the food in guiding their choices.
Málková et al. (1997) had earlier used the same devaluation procedure in rhesus monkeys to test whether excitotoxic basolateral amygdala (AMYG) lesions lead to an inability to shift decisions based upon current food value. Their experiments followed up earlier work of Hatfield et al. (1996) and Holland and Straub (1979) in rats. The results on desirability that are reported by Rudebeck et al. (2017) on OFC lesions were earlier predicted and simulated by the MOTIVATOR (Matching Objects To Internal VAlues Triggers Option Revaluations) neural model in Grossberg et al. (2008) as part of an explanation of the Málková et al. (1997) results.
In particular, Grossberg et al. (2008) wrote the following as part of the caption of their Figure 8 that shows data and simulations about reaction time, choice behaviour, and reward value:
Choices made between the two CSs reflect preferences between the different food rewards. Devaluation of a US by food-specific satiety (FSS) shifts the choices of the animal away from cues associated with the devalued rewards (reprinted with permission from Málková et al., 1997). Málková et al. (1997) report the effects of basolateral amygdala lesions using a difference score. The difference score is calculated by measuring the percent of the trials in which the to-be-devalued food is chosen over other foods, before and after FSS. The ‘difference score’ reflects the difference between these two percentages … Using FEF activity to determine cue choice, the intact model (CTL) shows a similar shift in CS preference when the US associated with it is devalued by FSS. Food-specific satiety is implemented by lowering selected DRIVE inputs to the LH … The automatic shifting of visual cue preference when an associated US is devalued by FSS is lost after AMYG lesions (AX) and ORBl lesions (OX). [italics mine]
The devaluation procedure that was used by Málková et al. (1997) and Rudebeck et al. (2017) fed monkeys a lot of Food 1 before testing their choice between Food 1 and an alternative food, Food 2, that has not be devalued. The MOTIVATOR model explains how devaluing Food 1 reduces the drive input (Figure 2(a)) from the lateral hypothalamus (LH) that is needed to activate its value category in the amygdala (AMYG), thereby reducing its ability to compete with the value category of Food 2. The value category of Food 2 can hereby win the competition between value categories, and release an incentive motivational signal to the OFC, which enables the OFC to choose Food 2 with increased probability. Either AMYG or OFC lesions eliminate this pathway to motivated choice of Food 2.
In order to better understand how this decision process is proposed to work, the subsequent text will review three model properties: (1) how the brain can learn different value categories that can be selectively activated by different foods; (2) how internal drive inputs, notably satiety signals, interact with conditioned or unconditioned reinforcing sensory inputs before such sensory-drive combinations compete to determine an incentive motivational output; and (3) how conditioned reinforcer pathways can habituate due to frequent activation by a particular food and thus create progressively smaller inputs to their value categories. The loss of these factors due to an amygdala lesion may prevent an animal from shifting its visual cue preference away from a devalued food.
2.2. Reinforcement learning, motivated attention, resonance, and directed action
The Rudebeck et al. (2017) concept of desirability of predicted outcomes is related to earlier concepts such as the ‘somatic marker hypothesis’ which proposes that decision-making is a process that depends upon emotion (Bechara et al., 1999). The Cognitive-Emotional-Motor, or CogEM, model (Figure 2(a)) that will be described in this section, in several incrementally unlumped versions that include the MOTIVATOR model, provides a mechanistic neural explanation of some aspects of emotionally modulated decision-making by describing different properties of, and interactions between, sensory cortex, amygdala, and OFC in making these decisions (Baxter et al., 2000; Bechara et al., 1999; Schoenbaum et al., 2003; Tremblay and Schultz, 1999). CogEM also clarifies how the OFC contributes in this circuit to the expression of the incentive value of rewards and their sensitivity to reward devaluation (Gallagher et al., 1999), and how acquired value in OFC depends on input from basolateral amygdala (Schoenbaum et al., 2003).
These properties of the CogEM model arise as emergent, or interactive, properties of the neural mechanisms that regulate reinforcement learning among the sensory cortex, amygdala/hypothalamus, and OFC; the allocation of motivated attention to chosen options in the sensory and orbitofrontal cortices as a result of this learning; and the release of motivated actions by the OFC to acquire valued goal objects that can realise these options (Grossberg, 1971, 1972a, 1972b, 1982, 2000b; Grossberg and Levine, 1987; Grossberg and Schmajuk, 1987; Grossberg and Seidman, 2006).
In particular, as summarised in Figure 2(a), the CogEM model explains how invariant object categories, in sensory cortical regions like the anterior inferotemporal cortex (ITa), and object-value categories, in cortical regions like the OFC, interact with value categories, in subcortical emotional centres like amygdala and hypothalamus. These brain regions are linked by a feedback loop which, when activated for sufficiently long time, can generate a cognitive-emotional resonance. Such a resonance can support conscious feelings while conditioned reinforcer learning pathways (from sensory cortex to amygdala; Gore et al., 2015) and incentive motivational learning pathways (from amygdala to OFC; Arana et al., 2003) focus motivated attention upon valued object and object-value representations. These attended object-value representations can, in turn, release commands to perform actions that are compatible with these feelings.
It needs immediately to be noted, however, that the CogEM circuit in Figure 2(a) cannot, by itself, maintain motivated attention during an adaptively timed interval that is sufficiently long to enable reinforcement learning to effectively occur in paradigms where rewards are delayed in time, as happens during trace conditioning and delayed match-to-sample, and to enable a conscious ‘feeling of what happens’ to emerge (Damasio, 1999). The hippocampus is needed to support both of these properties (Figure 2(b)), as sections 2.5 and 2.8 will further discuss.
The next Sections say more about these several types of categories and the learned interactions between them.
2.3. Object, value, and object-value categories
Three different types of learned representations are included in the CogEM circuit of Figure 2(a): invariant object categories respond selectively to objects that are seen from multiple views, positions, and distances from an observer. They are learned by inferotemporal (IT) cortical interactions between anterior IT (ITa), posterior IT (ITp), and prestriate cortical areas like V4 (Figure 1; Desimone, 1998; Gochin et al., 1991; Harries and Perrett, 1991; Mishkin, 1982; Mishkin et al., 1983; Seger and Miller, 2010; Ungerleider and Mishkin, 1982). How such invariant object categories may be learned as the eyes scan a scene is modelled by the ARTSCAN Search neural model that is discussed in Sections 3.28 and 3.29 (Cao et al., 2011; Chang et al., 2014; Fazl et al., 2009).
Value categories are sites of reinforcement learning that control different emotions and incentive motivational output signals. They occur in amygdala and hypothalamus in cells where reinforcing and homeostatic, or internal drive, inputs converge to generate emotional reactions and motivational decisions (Aggleton, 1993; Aggleton and Saunders, 2000; Barbas, 1995, 2007; Bower, 1981; Davis, 1994; Gloor et al., 1982; Halgren et al., 1978; LeDoux, 1993).
Object-value categories respond to converging signals from object categories and value categories. They are proposed to occur in OFC. The properties of object-value categories will be particularly discussed in this exposition. How they interact with representations in other brain regions is also an essential part of such an analysis.
Finally, motor representations (M) control discrete adaptive responses. They include multiple brain regions, including motor cortex and cerebellum (Evarts, 1973; Ito, 1984; Kalaska et al., 1989; Thompson, 1988). More complete models of the internal structure of motor representations and how they generate movement trajectories are described elsewhere (e.g. Bullock et al., 1998; Bullock and Grossberg, 1988; Cisek et al., 1998; Contreras-Vidal et al., 1997; Fiala et al., 1996; Gancarz and Grossberg, 1998, 1999) and can readily be incorporated into CogEM model extensions.
2.4. Conditioned reinforcer, incentive motivational, and motor learning: wanting
Three types of learning are shown in Figure 2(a) between these representations: conditioned reinforcer learning strengthens the pathway from an invariant object category to a value category. Incentive motivational learning strengthens the pathway from a value category to an object-value category. Motor learning enables the performance of an act aimed at acquiring a valued goal object.
Reinforcement learning, such as classical, or Pavlovian, conditioning (Kamin, 1968, 1969; Pavlov, 1927), occurs within conditioned reinforcer pathways (Figure 2(a)). A neutral event is called a CS when it is paired with an emotion-inducing, reflex-triggering event that is called an unconditioned stimulus (US). A CS can become a conditioned reinforcer when its object category is activated sufficiently often just before the value category is activated by an US. As a result of conditioned reinforcer learning, a CS can, on its own, subsequently activate a value category via this learned pathway. When this happens, the CS is called a conditioned reinforcer because it can trigger many of the same reinforcing and emotional effects as an US.
During classical conditioning, incentive motivational learning also occurs from the activated value category to the object-value category that corresponds to the CS, incentive motivational learning enables an active value category to prime, or modulate, the object-value categories of all CSs that have consistently been correlated with it. It is the kind of learning that primes an individual to think about places to eat when feeling hungry.
Motor, or habit, learning enables the sensorimotor maps, vectors, and gains that are involved in sensory-motor control to be adaptively calibrated, thereby enabling a CS to read out correctly calibrated movements via its object-value category.
These conclusions about the CogEM model also hold for operant, or instrumental, conditioning, where rewards or punishments are delivered that are contingent upon particular behaviours (Skinner, 1938). In fact, the CogEM model was introduced to explain data about operant conditioning (Grossberg, 1971). Many reinforcement learning and motivated attentional mechanisms exploit shared neural circuits, even though the experimental paradigms and behaviours that activate these circuits may differ.
The incentive motivational and motor learning pathways contribute to the process that various researchers call wanting. As Berridge et al. (2009) note, ‘By “wanting” we mean incentive salience, a type of incentive motivation that promotes approach towards and consumption of rewards’ (p. 67). See also Smith et al. (2011) for a further discussion of the different brain substrates of pleasure and incentive salience.
2.5. Category learning and memory consolidation: effects of lesions
A fourth kind of learning, category learning, adapts the connections from thalamus to an object category in sensory cortex, and/or from an object category to an object-value category in OFC. This category learning process enables external objects and events in the world to selectively activate object and object-value categories. Category learning was not simulated in the original CogEM model (Figure 2(a)), which focused on reinforcement learning, motivated attention, and the release of actions towards valued goal objects. Category learning does play a key role in extensions of CogEM, such as the neurotrophic Spectrally Timed Adaptive Resonance Theory, or nSTART, model (Figure 2(b)) of Franklin and Grossberg (2017). nSTART augments CogEM to include category learning, as well as adaptively timed learning in the hippocampus that can bridge between CS and US stimuli that are separated in time by an interval that can be hundreds of milliseconds in duration, as can occur during trace conditioning and delayed non-match to sample. Interactions between these two processes, augmented by all the other processes of CogEM, enable nSTART to explain and simulate how memory consolidation of recognition categories may occur after conditioning ends. nSTART supports this explanation by mechanistically explaining and simulating data about the complex pattern of disruptions of memory consolidation that occur in response to early versus late lesions of thalamus, amygdala, hippocampus, and OFC.
2.6. Polyvalent constraints and competition interact to choose the most valued options
Both the CogEM and nSTART circuits in Figure 2 need to have two successive sensory processing stages, an invariant object category stage in the temporal cortex, and an object-value category stage in OFC, in order to ensure that the object-value category can release motivated behaviour most vigorously if both sensory inputs from temporal cortex and motivational inputs from the amygdala are simultaneously delivered to the object-value category. A polyvalent constraint on an object-value category means that it fires most vigorously when it receives input from its invariant object category and from a value category. In other words, an object-value category is amplified when the action that it controls is valued at that time. Only when an object-value category wins a competition with other object-value categories can it trigger an action. After learning occurs, a conditioned reinforcer can, by itself, satisfy this polyvalent constraint by sending a signal from its object category directly to its object-value category, and indirectly via the (conditioned reinforcer)-(incentive motivational) pathway (Figure 2). Converging pathways from sensory cortical areas and amygdala to OFC are well known anatomically (e.g. Barbas, 1995).
The firing of each value category in the amygdala/hypothalamus is also regulated by a combination of polyvalent constraints and competition. Here, the polyvalent constraint (Figure 2(a)) is realised by two converging inputs to each value category: a reinforcing input from an US or conditioned reinforcer CS via a conditioned reinforcer pathway and a sufficiently large internal drive input (e.g. hunger, satiety). Each value category can only become active enough to reliably win the competition with other value categories when it receives sufficiently big converging inputs, and only a winning value category can generate large incentive motivational signals to object-value categories. In particular, even if visual cues such as a familiar food generate a strong conditioned reinforcer inputs to a value category, it cannot fire if its internal drive input is reduced by eating a lot, since then the hunger drive input will decrease and the satiety drive input will increase, thereby preventing its value category from winning the competition.
Because both the value categories and the object-value categories obey polyvalent constraints and compete to determine either incentive motivational or motor outputs, a CogEM circuit tends to choose options for action that are currently the most desired ones. These CogEM dynamics hereby clarify necessary conditions for the computation of desirability by the OFC.
Many issues need to be discussed to better understand how these circuits work in practice. These issues include the following: Why is the amygdala called a value category? How does a value category differ from just an internal drive such as hunger? In particular, can value categories represent specific hungers that can be selectively devalued by eating a lot of a particular food? How is the hypothalamus involved in learning a value category? Finally, are there also pathways for expressing valued goals that do not require the amygdala? Such issues will be discussed as the exposition proceeds.
2.7. Predicting what happens next in a changing world: blocking and unblocking
In the real world, a foraging animal may be confronted with multiple possible sensory cues that predict different kinds of available foods. The case of choosing Food 1 or Food 2 in response to different CSs is a special case of this situation. To more completely understand how a choice of one option over another occurs, along with how its desirability is computed, requires an understanding of how attentional blocking and unblocking occur (Kamin, 1968, 1969; Pavlov, 1927). Blocking and unblocking experiments also clarify how humans and other animals discover what combinations of cues are causal and which are predictively irrelevant. Causal feature combinations tend to be attended and used to control subsequent actions. The following explanation of how this happens can be used to design ecologically more realistic neurobiological studies of desirability.
Blocking and unblocking experiments show that the discovery of true environmental causes is an incremental process. Unless sufficiently many actions based upon unblocked cue combinations are made, and their consequences used to match and mismatch learned expectations, an irrelevant cue can be erroneously thought to be predictive, much as superstitious behaviours may be learned and maintained (Skinner, 1948). Predictive errors hereby play a crucial role in shaping the learning of environmental causes, as popular recent books have noted; for example, Schulz (2010). Section 2.14 discusses how unblocking works in greater detail.
A food’s desirability in the real world, where there may be multiple possible food options to choose from, can only be computed if it is not blocked. For example, suppose that a buzzer sound (CS1) is paired with a food reinforcer (US1) until an animal salivates to this sound in anticipation of eating the food. On subsequent learning trials, suppose that, before the food occurs, the buzzer sounds at the same time that a perceptually equally salient light flashes (CS2). Under these circumstances, the flashing light does not become a source of conditioned salivation because it does not predict, indeed does not cause, any consequence that the buzzer sound alone did not already predict. In other words, the flashing light is predictively irrelevant and is thus attentionally blocked.
On the other hand, suppose that, whenever the flashing light occurs with the buzzer sound, the amount of subsequent food (US2) is much greater than when the buzzer sound occurred alone. Then, the animal does learn to anticipate food in response to the flashing light, with the usual salivation and expectation of subsequent food, because it causally predicts an increase in food. If, instead, the amount of food that is presented when the flashing light and buzzer occur together is much less than when the buzzer sound occurred alone, then a wave of frustration (Amsel, 1962, 1992), which is a negative reinforcer, may be experienced, even though some food, which is a positive reinforcer, has been presented. The process whereby the CS2 becomes predictively relevant is called attentional unblocking, and it may become a source of either a conditioned appetitive response or conditioned frustration, depending on whether the amount of food is more or less ample after the simultaneous cues than in response to the buzzer alone. Unexpected consequences can hereby lead to the discovery of new cue combinations that cause valued consequences. Attention can then be focused on the unblocked cues, which can then be associated with appropriate new responses.
Blocking and unblocking experiments show that humans and many other animals behave like minimal adaptive predictors who can learn to focus their attention upon events that causally predict important affective consequences, while ignoring other events, at least until an unexpected consequence occurs.
The same CogEM dynamics that enable the OFC to release actions aimed at acquiring desirable goal objects also carry out blocking. Blocking can be understood in the CogEM model as a result of how a cognitive-emotional resonance using the feedback pathways between temporal cortex, amygdala, and OFC (Figure 2) triggers competition among the representations in each of these brain regions to choose, and thereby attend to, the object that has the most salient combination of sensory input and motivational feedback from its value category.
For blocking to work properly, the cells in sensory cortex and OFC need to obey the membrane equations of neurophysiology, also called shunting interactions, and to compete with each other using recurrent, or feedback, inhibitory interactions. Such recurrent shunting on-centre off-surround networks tend to conserve, or normalise, the total activity that is shared among their active cells (Grossberg, 1973, 1980, 2013b). Thus, if the activity of one object category gets most amplified by a favourable combination of feedforward and feedback signalling, then the activities of the object categories with which it is competing will be inhibited, leading to attentional blocking. Grossberg and Levine (1987) simulated attentional blocking using these CogEM interactions. An explanation of unblocking requires additional mechanisms that will be described in Section 2.14.
It should be noted that blocking and unblocking experiments share some properties with the Asaad et al. (2017) experiments on credit assignment. Section 3.19 clarifies why this is so.
2.8. Cognitive-emotional resonances, feeling of what happens, and somatic markers
Previous articles review additional psychological and neurobiological data that the CogEM model has explained and predicted, and how it compares with other models of cognitive-emotional dynamics; for example, Grossberg (2013a, 2017b). All of these data explanations are relevant, moreover, to the computation of desirability.
One such comparison relates to the ability of the CogEM model to explain and predict clinical data. Damasio (1999) has derived from clinical data a heuristic version of the CogEM model and used it to describe cognitive-emotional resonances that support ‘the feeling of what happens’. Each processing stage in Damasio’s model (see his Figure 6.1) corresponds to a processing stage in the CogEM circuit of Figure 2(a). In particular, Damasio’s ‘map of object X’ corresponds to the sensory cortical stage where invariant object categories are represented. His ‘map of the proto-self’ becomes the value category and its multiple interactions. His ‘second-order map’ becomes the object-value category. And his ‘map of object X enhanced’ becomes the object category as it is attentively amplified by feedback from the object-value category. Feedback from an object-value category to its object category closes an excitatory feedback loop between object, value, and object-value categories. Persistent activity through this loop – maintained long enough with the help of adaptively timed feedback from sensory cortex to OFC via the hippocampus (Figure 2(b)) – enables the attended object to achieve emotional and motivational significance and to drive the choice of motivated decisions that can trigger context-appropriate actions towards valued goals.
Such sustained activation through a positive feedback loop gives rise to a resonant brain state. In the present instance, it is called a cognitive-emotional resonance because it binds cognitive information in the object category to emotional information in the value category. A resonance is a dynamical state during which neuronal firings across a brain network are amplified and synchronised when they interact via excitatory feedback signals during a matching process that occurs between bottom-up and top-down pathways. Grossberg (2013a, 2017b) explains in detail how resonating cell activities focus attention upon a subset of cells, and how the brain can become conscious of attended events during a resonant state. It is called an adaptive resonance because the resonant state can trigger learning within the adaptive weights, or long-term memory (LTM) traces, that exist at the synapses of these pathways. In a cognitive-emotional resonance, these adaptive weights occur in conditioned reinforcer and incentive motivational pathways, among others. Several different types of adaptive resonances will be described below.
CogEM hereby embodies and anticipated key concepts of the ‘somatic marker hypothesis’ which proposes that decision-making is a process that depends upon emotion, while also providing a mechanistic neural explanation and simulations (e.g. Grossberg et al., 2008) of various different properties of amygdala and OFC in making these decisions (Baxter et al., 2000; Bechara et al., 1999; Schoenbaum et al., 2003).
CogEM has also proposed explanations of additional clinical data, such as the data about memory consolidation that were mentioned above (Franklin and Grossberg, 2017); data about Fragile X syndrome and some types of repetitive behaviours that are found in individuals with autism (Grossberg and Kishnan, 2018); and data about emotional, attentional, and executive deficits that are found in individuals with autism or schizophrenia (Grossberg, 2000b; Grossberg and Seidman, 2006).
2.9. MOTIVATOR: amygdala and basal ganglia dynamics during conditioning
In order to mechanistically explain how devaluation works, and to answer the questions posed at the end of Section 2.6, the MOTIVATOR model, which unlumps the CogEM model, is needed (Dranias et al., 2008; Grossberg et al., 2008; Figure 3(a)). The MOTIVATOR model combines the CogEM model with a model of how the basal ganglia (BG) learns to respond to expected and unexpected rewards (Figure 4(a)–(c); Brown et al., 1999). Thus, in addition to clarifying how value categories are learned, the MOTIVATOR model begins to explain how the amygdala and the BG interact with one another and with the temporal and orbitofrontal cortices to focus motivated attention and trigger choices aimed at acquiring valued goal objects.
The BG need to be discussed along with the amygdala because it plays an important role in both the cognitive-emotional and working memory learning that are relevant to this article’s explanatory goals (Figure 1). Indeed, the amygdala and BG seem to play computationally complementary roles (Grossberg, 2000a), with the BG triggering Now Print learning signals in response to unexpected rewards, and the amygdala learning to activate incentive motivational signals with which to help acquire expected rewards. A Now Print learning signal is a signal that is broadcast broadly to many brain regions where it can modulate learning at all of its recipient neurons (Grossberg, 1974; Harley, 2004; Livingston, 1967; McGaugh, 2003).
As in CogEM, the model amygdala and LH interact to calculate the expected current value of the subjective outcome that the CS predicts, constrained by the current state of deprivation or satiation. The amygdala then relays the expected value information to orbitofrontal cells (ORB in Figure 3(a) and OFC in Figure 1) that receive visual inputs from anterior inferotemporal cells (ITA; Amaral and Price, 1984; Ghashghaei and Barbas, 2002; Öngür and Price, 2000; Reep et al., 1996) and to medial orbitofrontal cells (MORB in Figure 3(a)) that receive gustatory inputs from rhinal cortex (RHIN) (Barbas, 1993, 2000; Barbas et al. 1999; Reep et al., 1996). The activations of these orbitofrontal cells code the expected subjective values of objects. These values guide behavioural choices.
The review of Levy and Glimcher (2012) discusses the OFC computation of expected subjective value from the perspective of neuroeconomic theory. Also of neuroeconomic interest is the Grossberg and Gutowski (1987) exposition of how the CogEM model explains and simulates data of Kahneman and Tversky (1979) about decision-making under risk, thereby explaining Prospect Theory axioms using neural designs that are essential for survival, as well as data that Prospect Theory cannot explain, such as data about preference reversals.
The model BG, or Reward Expectation Filter (Figures 3(a) and 4(a)), detects errors in CS-specific predictions of the value and timing of rewards (Ljungberg et al., 1992; Schultz, 1998; Schultz et al., 1992, 1993, 1995, 1997). Excitatory primary rewarding inputs from the LH reach the substantia nigra pars compacta (SNc), via the pedunculopontine nucleus (PPTN). The SNc also receives adaptively timed inhibitory inputs from model striosomes in the ventral striatum. Mismatches between these signals can trigger widespread dopaminergic burst and dip Now Print signals from cells in SNc (Figure 4(a)) and the ventral tegmental area (VTA in Figure 3(a)). Learning in cortical and striatal regions is strongly modulated by these Now Print signals, with dopamine bursts strengthening conditioned links and dopamine dips weakening them. In particular, such a dopaminergic Now Print signal modulates learning in the two pathways that are activated by a CS in Figure 4(a). After learning occurs, one pathway, via the ventral striatum to the SNc, enables the CS to activate the SNc and generate its own Now Print signals. The other pathway, via the striosomes to the SNc, enables the CS to inhibit responses to an expected US from the LH. This SNc circuit will be important in considering the kinds of reinforcement learning that are spared when amgydala and/or OFC are lesioned.
MOTIVATOR was used to explain and simulate psychological and neurobiological data from tasks that examine food-specific satiety, Pavlovian conditioning, reinforcer devaluation, and simultaneous visual discrimination, while retaining the ability to simulate neurophysiological data about SNc. Model simulations successfully reproduced the neurophysiologically recorded dynamics of hypothalamic cell types, including signals that predict saccadic reaction times and CS-dependent changes in systolic blood pressure.
In order to more fully understand how food-specific satiation occurs, more needs to be said about how value categories in the amygdala are learned as a result of adaptive feedback interactions with the LH, and why, in order to properly regulate reinforcement learning and affective prediction, some of these pathways need to receive internal drive inputs and/or habituative reinforcing inputs. This explanation will be given in Section 2.11.
2.10. Temporal difference models of BG responses to unexpected rewards
The Reward Expectation Filter of the MOTIVATOR model was first published in Brown et al. (1999) who used it to explain and simulate many experiments about how SNc reacts to rewards whose amplitude or timing is unexpected. Several other models have attempted to describe the SNc cell behaviour using a temporal difference (TD) algorithm (Montague et al., 1996; Schultz et al., 1997; Suri and Schultz, 1998). These models suggest that the dopaminergic SNc cells compute a temporal derivative of predicted reward. In other words, they fire in response to the sum of the time-derivative of reward prediction and the actual reward received. These models have not been linked to structures in the brain that might compute the required signals. As Brown et al. (1999) have noted, the Suri and Schultz (1998) model has simulated some of the known dopamine cell data. However, their model can only learn a single fixed interstimulus interval (ISI) that corresponds to the longest-duration timed signal (xlm(t) in their model). If the ISI is shorter than this, dopamine bursts will strengthen all of the active stimulus representations predicting reward at the time of the dopamine burst or later. Thus, their model generates inhibitory reward predictions beyond the primary reward time, and predicts a lasting depression of dopamine firing subsequent to primary reward, which is not found in the data, data that the Brown et al. (1999) model explains.
In contrast to TD models that compute time derivatives immediately prior to dopamine cells, our model uses two distinct pathways: the ventral striatum and PPTN for initial excitatory reward prediction, and the striosomal cells for timed, inhibitory reward prediction. The fast excitation and delayed inhibition are hereby computed by separate structures within the brain, rather than by a single temporal differentiator. This separation avoids the problem of the Suri and Schultz (1998) model by allowing transient rather than sustained signals to cancel the primary reward signal, thereby enabling precisely timed reward-cancelling signals to be trained, and preventing spurious sustained inhibitory signals to the dopamine cells.
2.11. Learning value categories for specific foods and effects of their removal
Figure 3(b) diagrams how the MOTIVATOR model conceptualises the learning of a value category as a result of reciprocal adaptive interactions between the LH and the amygdala (AMYG). This figure summarises how the model embodies a network that calculates the drive-modulated affective value of a food US (Cardinal et al., 2002); notably, how selective responses to different foods can be acquired. Animals have specific hungers that vary inversely with blood levels of metabolites such as sugar, salt, protein, and fat (Davidson et al., 1997). Similarly, the gustatory system has chemical sensitivities to tastes such as sweet, salty, umami, and fatty (Kondoh et al., 2000; Rolls et al., 1999). An AMYG value category (top level in Figure 3(b)) learns to respond to particular LH combinations of these metabolites and tastes (bottom level in Figure 3(b)) in a selective fashion, hence can represent specific hungers.
MOTIVATOR begins its computation of food-specific selectivity with the lower layer of the model’s LH cells in Figure 3(b). These cells perform pairwise multiplications, each involving a taste and its corresponding drive level. They are therefore called taste-drive cells. LH neurons such as glucose-sensitive neurons provide examples of LH cells that are both chemical- and taste-sensitive. Indeed, glucose-sensitive neurons are excited by low glucose levels, inhibited by high glucose levels, and respond to the taste of glucose with excitation (Karadi et al., 1992; Shimizu et al., 1984).
The activation pattern across all these taste-drive cells is projected via converging adaptive pathways to a higher cell layer and summed there by an AMYG value category cell that represents the current value of the specific food US. These cells are therefore also called US-value cells. Such food selective US-value representations can be learned from a competitive learning process (Grossberg, 1976a, 1978a) that associates distributed activation patterns at the taste-drive cells with compressed representations at the US-value cells that survive the competition at the AMYG processing level. The resulting US-value cells in the AMYG help to explain data about neurons in the AMYG that respond selectively to specific foods or associated stimuli in a manner that reflects the expected consumption value of the food (e.g. Nishijo et al., 1988a, 1988b).
Figure 3(a) and (b) illustrates the hypothesis that a visual CS becomes a conditioned reinforcer by learning to activate a US-value representation in the AMYG during CS-US pairing protocols. Despite the fact the CS generates no gustatory inputs to the taste-drive cells and is not actually consumed, the model can use this CS-US association to compute the prospective value of the US, given current drives, during the period between CS onset and the delivery of the food US. The model can do this if the CS-activated US-value representation in the AMYG can, in turn, activate the taste-drive cells in the LH that have activated it in the past, when the US was being consumed.
This is accomplished, as depicted in Figure 3(a) and (b), by adaptive top-down pathways, or learned expectations, from the US-value cells in the AMYG to the taste-drive cells in the LH. The resultant bidirectional adaptive signalling between taste-drive LH cells and integrative US-value AMYG cells can prime the taste-value combinations that are expected in response to the conditioned reinforcer CS. Such reciprocal adaptive interactions have also been shown, as part of Adaptive Resonance Theory (ART), to be capable of stabilising category learning and memory, in whatever brain systems they occur (see Section 2.12; Carpenter and Grossberg, 1987, 1991; Grossberg, 1980, 2013a, 2017b). An ART circuit complements the bottom-up adaptive filter of a competitive learning model with adaptive top-down expectation signals, among other extensions. ART was introduced to overcome the catastrophic forgetting that occurs when only bottom-up learning occurs (Grossberg, 1976a, 1976b). Without top-down expectations to dynamically stabilise the learning of value categories, memory instability could become as great a problem in LH-AMYG dynamics as it would be in the learning of invariant object categories by the inferotemporal cortex (Figures 2 and 3).
These properties of AMYG value categories help to explain how an animal’s behaving changes when its AMYG is lesioned. When the AMYG is lesioned, the ability to selectively respond to specific foods is eliminated. The reduced drive inputs of a satiated food will then not be able to cause a smaller activation of its AMYG value category. Also lost will be the competition among value categories that would determine the choice of a non-satiated food in a normal animal.
If a food was visually presented to a normal animal in order to satiate it, then both reduced internal drive and external cue inputs could contribute to the choice of a non-satiated food (Figures 2 and 3). In particular, as above, eating a lot of food would lead to shrinking appetitive inputs and growing satiety drive inputs to the LH. Seeing the food repeatedly during each eating event could also habituate the conditioned reinforcer inputs that activate the corresponding AMYG value category.
It should immediately be noted that not all responsiveness to reinforcing cues is eliminated by lesions of the AMYG. The BG can still be fully functional. As illustrated by Figure 4(a), LH inputs can still regulate learning of dopaminergic Now Print signals from the SNc to large parts of the brain, including the PFC, in response to unexpected rewards. As will be discussed in Section 3, these Now Print signals can support learning of representations that are sensitive to the probability of reward over a series of previous trials.
2.12. Another mismatch mechanism for correcting disconfirmed behaviour: ART
This BG mechanism for processing an unexpected reinforcing event is not the only way that unexpected events are processed by the brain. Another fundamental process is also at work in perceptual and cognitive processes that greatly influences how the PFC makes predictions that are sensitive to the probability of reward over a series of previous trials. This process enables humans and other primates to rapidly learn new facts without being forced to just as rapidly forget what they already know, even if no one tells them how the rules of each environment differ or change through time. When such forgetting does occur, it is often called catastrophic forgetting.
Grossberg (1980) called the problem whereby the brain learns quickly and stably without catastrophically forgetting its past knowledge the stability-plasticity dilemma. ART was introduced to explain how brains solve the stability-plasticity dilemma. Since its introduction in Grossberg (1976a, 1976b), ART has been incrementally developed into a cognitive and neural theory of how the brain autonomously learns to attend, recognise, and predict objects and events in a changing world, without experiencing catastrophic forgetting. ART currently has the broadest explanatory and predictive range of available cognitive and neural theories. See Grossberg (2013a, 2017b) for reviews.
ART prevents catastrophic forgetting by proposing how top-down expectations focus attention on salient combinations of cues, called critical feature patterns. When a good enough match occurs between a bottom-up input pattern and a top-down expectation, a synchronous resonant state emerges that embodies an attentional focus. Such a resonance is capable of driving fast learning that incorporates the attended critical feature pattern into the bottom-up adaptive filters that activate recognition categories, and the top-down expectations that are read out by them – hence the name adaptive resonance – while suppressing outliers that could have caused catastrophic forgetting.
In contrast, when a bottom-up input pattern represents an unexpected event, it can cause a big mismatch with the currently active top-down expectation. As illustrated in Figure 5, such a mismatch, or disconfirmed expectation, can trigger a burst of nonspecific arousal (‘novel events are arousing’). It is called nonspecific arousal because it affects all category representations equally, since the orienting system that triggers such an arousal burst has no information about which active categories caused the mismatch, and thus must be reset (Figure 5(c)). A nonspecific arousal burst can reset whatever categories caused the mismatch (Figure 5(d)) and initiate a search for a more predictive category. In this way, rapid reset of an object category can occur when it leads to disconfirmed behaviours, and a more predictive category can automatically be chosen in its stead.
Figure 5 illustrates the limiting case in which the category level makes a winner-take-all choice of a single winning cell, or cell population. In general, the category level compresses the distributed patterns on the feature level, but does not necessarily chose winner-take-all categories; for example, Carpenter (1997). The two processing levels in Figure 5 schematise a kind of dynamics that may be repeated in multiple brain regions within the What cortical stream. Figure 18 below illustrates how this can happen in the laminar circuits of neocortex.
How do brain circuits selectively respond to a nonspecific arousal burst in order to inhibit unpredictive categories and trigger a memory search? One mechanism is by causing antagonistic rebounds in opponent processing circuits. Such rebounds can occur in perceptual, cognitive, and affective brain circuits. They will now be explained for the special case of LH-AMYG interactions that are needed to explain how devaluation and reversal learning work.
2.13. Opponent processing by gated dipoles in object and value categories
The LH-AMYG pathways in Figures 2 and 3 include circuits that control opponent emotional and motivational states, such as fear versus relief, and hunger versus satiety. These opponent circuits can trigger an antagonistic rebound in response to two kinds of input changes: changes in the amount of reward or punishment, such as a sudden increase or reduction in the amount of food or shock level; or to an unexpected event, such as the non-occurrence of an expected shock. The rebound that is caused by the non-occurrence of an expected event is triggered by a burst of nonspecific arousal to the dipole (Figure 6(a)) due to a mismatch with the expectation of the non-occurring event (Figure 5(c)). These rebounds can rapidly reset a currently active value category, and the amount of incentive motivation with which it was supporting an ongoing valued behaviour, while simultaneously helping to choose more predictive representations with which to learn and perform the new environmental contingency.
These opponent circuits are modelled by gated dipoles (Dranias et al., 2008; Grossberg, 1972a, 1972b, 1980, 1984a, 2000b; Grossberg et al., 2008; Grossberg and Seidman, 2006). The simplest gated dipole circuit is depicted in Figure 6(a) (Grossberg, 1972b). It has non-recurrent, or feedforward, ON and OFF channels, or pathways. ON and OFF cells within the opponent pathways of a gated dipole can represent an opponent pair of emotional and motivational states. Gated dipoles have helped to explain many data about both classical and operant conditioning, including conditioned extinction, learned escape and avoidance, attentional unblocking, partial reinforcement acquisition effect, gambling behaviours, and self-punitive behaviours – behavioural properties that currently seem to have no other mechanistic neural explanations. Simple mechanisms, occurring in a prescribed order, enable gated dipoles to cause antagonistic rebounds either in response to changes in reinforcer amplitude or to disconfirmations of cognitive expectations of reward. As explained by ART, disconfirmation of a cognitive expectation can cause a nonspecific arousal burst that is broadcast throughout the brain (Figure 5(c)). When such an arousal burst is received by a hypothalamic gated dipole, it can cause an antagonistic rebound. The caption of Figure 6(a) explains how an antagonistic rebound occurs.
Neurophysiological data from hypothalamic ‘opposite cells’ match affective gated dipole properties, including their opponent and rebound properties (Nakamura et al., 1987; Nakamura and Ono, 1986; Ono et al., 1986). These hypothalamic properties have been simulated by the MOTIVATOR model (Figure 3(a); Dranias et al., 2008; Grossberg et al., 2008). The ON and OFF channels within a gated dipole in the LH of the MOTIVATOR model delivers inputs to the AMYG value categories which, in turn, provide incentive motivational signals to OFC object-value categories, and thereby influences what actions are taken to achieve valued goals. Animals with an intact AMYG and OFC can hereby use hypothalamic rebounds to flexibly choose value and object-value categories that can track changing reinforcement contingencies, and thereby update an option’s desirability. Indeed, the primate AMYG contains separate cell populations that respond to positively and negatively valued visual stimuli (Paton et al., 2006), can carry out moment-by-moment tracking of state value (Belova et al., 2008; Morrison and Salzman, 2010), and are modulated by unexpected events (Belova et al., 2007).
Opponent processing circuits with gated dipole rebound properties also occur in perceptual and cognitive brain regions, and reflect the ability of these representations to be rapidly reset, and indeed reversed, when stimulus conditions change. For perceptual examples of how a change in stimulus level can cause a rebound, consider negative aftereffects: Offset of sustained viewing of intersecting radial lines leads to an opponent MacKay negative aftereffect of concentric circles. Offset of sustained viewing of a red surface leads to an opponent green surface aftereffect. Offset of sustained viewing of a video of water flowing downwards leads to an opponent waterfall aftereffect of motion upwards. See Francis and Grossberg (1996) for a model of boundary grouping and surface filling-in that explains and simulates such perceptual aftereffects.
Rebounds also help to carry out tasks in response to changes in perceptual cues. As one example, consider the task for humans of pushing a buzzer, or for pigeons of pecking a key, as fast as possible when a red light shuts off. If the only thing that happened in the brain when the light shuts off was the termination of activity within a category that codes for red (among other features), then there would be no internal signal at stimulus offset to activate the buzzer press. If, however, offset of the ON cell (population) that codes for red triggers an antagonistic rebound in an associated OFF cell (population), then activation of the OFF cell can learn to be associated with the buzzer press, or key peck, command.
2.14. Affective antagonistic rebounds, unblocking, and reversal learning
Antagonistic rebounds facilitate tracking and learning about changing reinforcement contingencies in many other types of situations, including the learning of escape behaviours, attentional unblocking (Section 2.7), and reversal learning. For example, either a sudden reduction of a fearful shock, or the non-occurrence of an expected shock, can cause a relief rebound (Denny, 1971; Masterson, 1970; Reynierse and Rizley, 1970) that can be used to motivate new conditioned responses. The unexpected non-occurrence of food can, in contrast, cause a frustrative rebound (Amsel, 1962, 1992) that can be used to suppress unsuccessful responses. Thus, rebounds can occur from negative to positive affects, such as from fear to relief, or from positive to negative affects, such as from hunger to frustration. These antagonistic rebounds enable the brain to learn responses that quickly adapt to changing reinforcement contingencies.
For example, suppose that an animal who is experiencing a shocked Skinner box floor accidentally pushes a red button that causes the shock to stop and opens a door whereby to escape from the Skinner box. The sudden termination of shock causes a relief rebound within its value category. The object category that is activated by the red button can then be associated with the relief-activated value category, leading to new conditioned reinforcer learning, incentive motivational learning, and motor learning (Figure 2(a)). In this way, conditioned relief can be used to motivate the learned escape behaviour in the future. In the opposite direction, the frustrative rebound that occurs after expected food does not occur can drive forgetting, or extinction, of motivational support for the consummatory actions that no longer lead to food, thereby releasing exploratory behaviours to search for more productive sources of food.
Using the same mechanisms, the unexpected change in reinforcer amplitude in an unblocking experiment, or an unexpected change in reward schedule in a reversal experiment, can cause antagonistic rebounds that immediately modify the net incentive motivation that is controlling ongoing behaviour, while also triggering rapid relearning, via the newly activated value categories, of the conditioned reinforcer, and incentive motivational, and motor learning pathways that will control subsequent motivated choices. These hypothalamic rebounds can also help reversal learning to occur in circuits that do not involve AMYG and OFC, such as the hypothalamic pathway via the SNc that causes Now Print signals to be broadcast to multiple brain regions, including PFC (Figures 3(a) and 4(a)).
In an unblocking experiment, at least two kinds of events, one cognitive and one cognitive-emotional, occur in parallel. The cognitive-emotional event is due to the antagonistic rebounds that have just been described. They cause activation of a new combination of value categories. The cognitive event enables a more predictive set of perceptual features and object categories to be attended, including those that would have remained blocked if there was no change in the expected reinforcer (see Section 2.7). Unblocking happens when a burst of nonspecific arousal is triggered by the unexpected reinforcer and causes new features and categories to be selected and attended (Figure 5(c) and (d)). Then, the new features and categories can be associated with the newly activated value categories to cause new conditioned reinforcer, incentive motivational, and motor learning that can, for example, lead to escape from the shocked Skinner box floor.
When a shock is unexpectedly reduced, it can cause a rebound in the relief channel that can more than cancel the direct effect of the reduced shock input to the fear channel. In this way, an unexpectedly small shock can act as a positive reinforcer, despite the fact that shock is a negative US (Grossberg, 1972b; Section 6). Sudden flips of affective sign can occur in other ways as well in a gated dipole opponent processing circuit. One such effect is called by Robinson and Berridge (2013) an ‘instant transformation of learned repulsion into motivational “wanting”’ (p. 282). In this kind of experiment, a lever is associated with an intensely unpleasant concentration of salty water. The ensuing aversive behaviours after presentation of the lever can be immediately transformed into positive ‘wanting’ responses by injections of deoxycorticosterone and furosemide that mimic sodium deficiency.
Wanting can also be increased in hyperdopaminergic mutant mice who embody a dopamine transporter (DAT) knockdown mutant (Zhuang et al., 2001) that causes 70% higher extracellular dopamine levels in the striatum (Peciña et al., 2003). One way that wanting can be increased in these mice is that larger dopaminergic Now Print signals (Figures 3(a) and 4(a)) during learning may strengthen the conditioned reinforcer, incentive motivation, and motor learning pathways that support stronger ‘wanting’ responses.
2.15. READ circuits: feedback enables stable motivation and secondary conditioning
Once rebound mechanisms of a gated dipole are understood, it is then possible to explain how rebounds can lead to new reinforcement learning, including reversal learning and changes in object desirability. To fully achieve these properties, the non-recurrent dipole of Figure 6(a) needs to be replaced by a recurrent dipole, with positive feedback pathways within the ON and OFF opponent channels. The recurrent gated dipole in Figure 6(b) is called a READ circuit, for REcurrent Associative Dipole (Grossberg and Schmajuk, 1987). READ circuits model the hypothalamic gated dipoles in the MOTIVATOR model. They are omitted in Figure 3(a) for simplicity.
Due to recurrent feedback in both the ON and OFF channels of a READ circuit, activity x7 reactivates x1 in the ON channel, while activity x8 reactivates x2 in the OFF channel. As in Figure 6(a), there are also habituative transmitter gates z1 and z2 in the ON and OFF channels, respectively. In addition, adaptive weights, or LTM traces, wk7 and wk8, within the black hemidisc synapses, sample the ON and OFF channels, respectively, in response to sampling signals Sk that are activated by different CSs. These CSs become conditioned reinforcers when they are associated with reinforcing US inputs at the gated dipole (input J in Figure 6(b)). Any number k of sampling signals, that are activated by different CS signals Sk, can converge on a single READ circuit. In this way, multiple CSs, with different levels of desirability, can all benefit from the same emotional and motivational hypothalamic circuits.
Properties of reversal learning and changes in object desirability build upon the ability of a READ circuit to simulate data about primary and secondary excitatory and inhibitory conditioning, among other important properties. The utility of a recurrent anatomy is vividly illustrated by the case of secondary inhibitory conditioning. Suppose that CS1 is associated with a US shock input J until it becomes a source of conditioned fear. For this to happen, the adaptive weights of CS1 must occur after the position in the ON channel where the US input J is registered, so that they can sample the activity caused by the US. After CS1 has become a source of conditioned fear, suppose that onset of a different CS2 is associated with the offset of CS1 so that it can sample the antagonistic rebound in the relief channel, and thereby learn how to become a source of conditioned relief. For this to happen, CS1 must deliver its input before the habituative gates, so that its offset can cause the rebound in the OFF channel. In contrast, CS2 must occur after the habituative gates, where it can sample the rebound, and thereby achieve secondary inhibitory conditioning.
What does this have to do with recurrent connections? The following sleight of hand makes this clear: this experiment could have been done with any CS1 and CS2. If we now interchange the cues that are used as CS1 and CS2, then by the above argument, each CS must occur both before and after the habituative gates. This can only happen if the network has recurrent connections within the ON and OFF channels.
A READ circuit exhibits several other basic functional properties that enable it to learn and perform motivated behaviours in a naturalistic environment. These properties were demonstrated with computer simulations in Grossberg and Schmajuk (1987). First, it can maintain steady motivation while a behaviour is being performed, even during sufficiently small environmental distractions. This happens because the READ circuit is a specialised recurrent on-centre off-surround network that is capable of storing activity patterns in short-term memory (STM). Section 3.15 explains how storage occurs.
Second, a READ circuit can rapidly switch to support a new behaviour with a different motivation if the distraction, or change in reinforcement contingency, is big enough. This is just a reset of STM in response to external inputs that change enough to overcome the hysteresis caused by the recurrent excitatory feedback.
Third, a READ circuit enables affective learning to remain sensitive to any number of reinforcing events throughout the lifespan; its LTM traces do not saturate. In other words, if conditioning continued to occur over many trials, one could imagine all the LTM traces reaching their maximum values, after which no future conditioning would be possible. LTM saturation is prevented from happening in a READ circuit in the following way: the LTM traces sample dendritic activities (the thick black bars in Figure 6(b)) of cells with activities x7 and x8. These dendrites receive teaching signals in the form of retrograde calcium spikes from the cell bodies (Grossberg, 1975; Magee and Johnston, 1997; Markram et al., 1995, 1997). These teaching signals occur after the ON and OFF channels have undergone opponent competition. Due to opponent competition, if both ON and OFF channels had the same activity, their teaching signals would be inhibited to zero, so that the LTM traces that sample them would also approach zero. When the ON and OFF channels have different activities, the competition computes teaching signals that are normalised net activities because the activities in such a recurrent on-centre off-surround network are normalised. The LTM traces can continue to learn these net values throughout life, without ever saturating.
Fourth, a READ circuit enables affective memories to be preserved for a long time, even years, until reward or punishment schedules change, or cognitive expectations are disconfirmed. Such stable memories help to explain the persistence of instrumental avoidance behaviours and why Pavlovian conditioned inhibitors do not extinguish, among other conditioning data (Grossberg, 1972a; Grossberg and Schmajuk, 1987; Kamin et al., 1963; Lysle and Fowler, 1985; Maier et al., 1969; Miller and Schachtman, 1985; Owren and Kaplan, 1981; Solomon et al., 1953; Witcher, 1978; Zimmer-Hart and Rescorla, 1974). The stable affective memory happens because, when LTM traces are read into STM by gating the signal from a recall probe Sk in Figure 6(b), they instate in STM the normalised net pattern that they have learned. The sampled LTM traces then, in turn, sample this STM pattern and thereby ensure their stability under recall.
2.16. Inverted-U: emotional depression in mental disorders
Gated dipoles have other properties without which reinforcement learning and motivated attention would not be possible. These properties would profitably be more deeply probed by modern neurophysiological methods.
A property of major importance is that the activity of a gated dipole circuit exhibits an Inverted-U as a function of its arousal level (Figure 6(c)). The Inverted-U is a consequence of the same mechanisms that enable a gated dipole to trigger antagonistic rebounds and to thereby quickly adapt to changing reinforcement contingencies. The Inverted-U can be traced to how the state of habituation in the dipole’s transmitter gates (square synapses in Figure 6(a) and (b)) divide the effects of signals through the dipole. This division creates a Weber Law of dipole responsiveness.
In particular, a gated dipole can support normal behavioural dynamics if its tonic arousal level, as distinct from phasic arousal bursts in response to unexpected events, remains within an optimal range that causes peak values of the Inverted-U to occur (Figure 6(c)). These intermediate arousal input sizes generate a Golden Mean of responding that enables sufficient activation of AMYG value categories to occur. Maintaining this optimal range during waking hours is a major achievement of the affective brain. Future experiments are needed to probe how this example of homeostatic plasticity is maintained. Failure to do so is reflected in behavioural symptoms of several mental disorders, including autism and schizophrenia (Grossberg, 2000b; Grossberg and Seidman, 2006), as well as attention deficit hyperactivity disorder, or ADHD (see below).
Gated dipoles play a role in causing symptoms of these mental disorders when the arousal level remains either abnormally small or abnormally large. At either extreme, gated dipole outputs are depressed, but in different ways. Abnormally small arousal causes an underaroused depressive syndrome, whereas abnormally large arousal causes an overaroused depressive syndrome.
In an underaroused gated dipole, the response threshold to inputs is abnormally high but, after input intensity exceeds this elevated threshold, further increments in input intensity lead to hypersensitive emotional responses (Figure 6(c)). This happens because the habituative transmitter that divides dipole responses in abnormally small. In an overaroused gated dipole, the threshold to inputs is abnormally low but, despite the ability of inputs to easily exceed this threshold, all dipole responses are hyposensitive, leading to a flat emotional affect, insufficient incentive motivation, and a hypofrontal condition that is insufficient to support normal executive functions. This happens because the habituative transmitter that divides dipole responses is abnormally large. Underarousal may be one cause of behavioural symptoms in individuals with autism (Baker et al., 2017; Bujnakova et al., 2016) that are explained in Grossberg and Seidman (2006), whereas overarousal may be one cause of behavioural symptoms in schizophrenia (Ban, 1973; Depue, 1974; Haralanova et al., 2011) that are explained in Grossberg (2000b).
Other mental disorders may also reflect these underaroused and overaroused depressive properties. For example, many individuals with ADHD seem to be underaroused; for example, Mayer et al. (2016). The underaroused transmitter has been reported to be dopamine, and it gives rise to the kind of hypersensitivity that gated dipole affective dynamics predict (Sikström and Söderlund, 2007). Moreover, pharmacological ‘uppers’ like Ritalin are often used to bring individuals with ADHD ‘down’ (e.g. Weyandt et al., 2014). In a gated dipole, this happens because such an upper will increase tonic arousal from underaroused hypersensitivity to a Golden Mean of more moderate sensitivity and threshold reactivity (Figure 6(c)).
These clinical symptoms will have predictable effects on reinforcement learning and motivated behaviour, notably on factors like desirability and reversal learning, that the reader can derive using the earlier explanations of these behavioural properties.
2.17. Some other recent models of orbitofrontal functioning
One way to appreciate the explanatory value of the cognitive-emotional processes that have been summarised in Section 2 is by contrast with other recent models of OFC. For example, Wilson et al. (2014) and Schuck et al. (2016) have proposed a model which proposes that OFC is a ‘cognitive map of task space’ during animal and human decision-making. They claim that ‘OFC is critical for representing task states in … partially observable scenarios’ and that ‘OFC is unique in its ability to disambiguate task states that are perceptually similar but conceptually different, for instance by using information from working memory’ (Wilson et al., 2014: 267). One of the types of data that they use the model to simulate is how OFC dysfunction ‘is impaired during reversal learning’ (Wilson et al., 2014: 268). However, as was earlier reviewed in Section 1.2, monkeys with selective excitotoxic lesions of the OFC, unlike monkeys who have received aspiration OFC lesions, are unimpaired in learning and reversing object choices based on reward feedback (Rudebeck et al., 2013). Neurotoxic lesions of the amygdala (Izquierdo and Murray, 2007) have also led to results that challenge earlier demonstrations using aspiration and radiofrequency lesions that the amygdala is needed for object reversal learning (Aggleton and Passingham, 1981; Jones and Mishkin, 1972; Spiegler and Mishkin, 1981).
How did the model lead to the wrong explanation? A basic weakness of the model is that is does not describe any neural mechanism. It does not model the functions of any brain region, including OFC. Instead, it describes abstract state space representations to fit any particular set of data, with different states hypothesised to explain different data sets, and ad hoc assumptions that are tailored to explain each data set about how these abstract states will change due to an OFC lesion. One can go so far to criticise the definition of model states as being circular, designed, and modified until they simulate a small piece of data, even if the data are wrong, as in the case of reversal learning. These states are typically connected by abstract edges in a feedforward network. There are none of the feedback loops of identified neural interactions with critical functional properties that are described herein, for example, Figures 1–8, 10, 12, 15, 17, and 18. A typical feedforward diagram in Wilson et al. (2014) is between one abstract state and two outcomes, or two successive states and an outcome.
Their model proposes no decision-making dynamics. Instead, it is a variant of the kind of formal, discrete time, statistical learning theories that were popular in the 1950s–1970s, such as stimulus sampling theory (Estes, 1950), the Rescorla–Wagner learning rule (Rescorla and Wagner, 1972), and the Luce (1977) choice rule. The Rescorla–Wagner (1972) model was, however, shown long ago to be inadequate even to offer a formal explanation of key classical conditioning data, such as why conditioned excitors extinguish, but conditioned inhibitors do not (Lysle and Fowler, 1985; Owren and Kaplan, 1981; Witcher, 1978; Zimmer-Hart and Rescorla, 1974), which the CogEM model does explain (Grossberg and Schmajuk, 1987). Thus, any model, including the Wilson et al. (2014) model, that attempts to explain data about extinction using this learning law is building on an incorrect foundation. Even if its states happen to be chosen to simulate a fixed piece of data, the learning law by which it does so cannot be trusted.
The second formal rule in the model, the Luce choice rule, is a simple ratio rule that, being a ratio, conveniently represents probabilistic outcomes, but includes no mechanistic explanatory power. This kind of formal rule has been obsolete for 40 years, ever since the self-normalising properties of recurrent shunting on-centre off-surround networks started to be used to explain STM and decision-making in the brain (Douglas et al., 1995; Grossberg, 1973, 1978a, 1978b, 1980; Heeger, 1992; see Section 3.15).
None of the explanations in Wilson et al. (2014) hold up to scrutiny. For example, it is assumed that ‘pressing the lever led to 1 U of reward during conditioning and to −0.2 U in extinction’. None of the real-time dynamics of how an unexpected non-reward counter-conditions a learned action is described. In contrast, the MOTIVATOR model explains in detail how recurrent opponent processes, modelled by gated dipoles, undergo antagonistic rebounds in response to unexpected events (Sections 2.13–2.15), whose matching process is modelled by ART (Figure 5). Each of these model circuits is supported by psychological, anatomical, and neurophysiological data and explains and predicts data that cannot be explained within a Rescorla–Wagner–Luce framework.
Wilson et al. (2014) claim to simulate how immediate versus delayed reward, including delayed alternation (p. 269), influence learned actions, but there is no representation of a delay in their model, unlike the adaptively timed learning in the pART model SNc, VTA, and hippocampus (Figures 2(b), 3(a), and 4(a)) that all share the same substrate of spectral timing which is realised by Ca++-modulated metabotropic glutamate (mGluR) dynamics (Brown et al., 1999). Significantly, our SNc/VTA model provides a real-time neural explanation of dopamine release in response to unexpected rewards. Despite there being no model VTA in the Wilson et al. (2014), the authors nonetheless discuss dopamine firing by VTA neurons (pp. 271–272) in terms of a purely descriptive state diagram that is created to match the experiment.
Schunk et al. (2016) espouses the same model as Wilson et al. (2014), but it primarily reports interesting functional magnetic resonance imaging (fMRI) data from humans who are presented with a series of images, each of which contains a face superimposed on a house. The task is to perform an age judgment of faces or houses on separate blocks of trials:
The age (young or old) of the first trial defined the age of the current ‘mini-block’ and participants were instructed to continue judging the same category as long as the age in that category stayed the same. Upon encountering a trial in which the age in the judged category was different (e.g. a change from ‘young’ to ‘old’), the task rules required participants to switch to judging the age of the other category, starting a new mini-block on the next trial. (p. 1403)
All of their data can be explained by the following pART neural mechanisms, none of which is modelled in their own article. In particular, the first trial in a block can activate an invariant object category for faces or houses (Figures 2(a) and 3(a)). Choosing low vigilance enables the choice of a general category that can match against many faces, or many houses, but not both (Figure 5). This invariant object category can be maintained in STM during the block of trials by, among other circuits, the feedback loop between visual cortex, amygdala, and OFC (Figures 2 and 3). Sustained top-down attentive feedback from the invariant category to lower cortical and thalamic levels is matched against the bottom-up images via the ART Matching Rule (Figure 5(b)). These top-down signals are priming signals that can select face or house features from their overlay in each image, while suppressing mismatched features, thereby creating a match state. When there is a big enough mismatch with the category of the top-down prime, say due to a change in the age of the face in the current image, the mismatch can reset the invariant object category (Figure 5(c)) and the next block can begin.
These category matching events use the invariant object categories in anterior inferotemporal cortex (ITa in Figure 1), which is called sensory cortex in Figures 2 and 3. Each active invariant category reads out a top-down prime against which input exemplars are matched at lower cortical levels, for example, posterior inferotemporal cortex (ITp in Figure 1). In addition, it activates its specific connection from ITa to OFC (Figures 1–3). Because the ITa-to-OFC connections are specific, changes in matching of an active invariant category at ITa are also represented at OFC, where they could be decoded by the support vector machine classifier that the authors applied to their data.
Instead of providing such a mechanistic account, variants of which have explained scores of other experiments about visual search and top-down attentive matching (e.g. Grossberg, 2013a, 2017b), including how low vigilance categories are learned and matched against image morphs (Akrami et al., 2009; Grossberg et al., 2011), the authors show that different choices of abstract states can create fits to the data, thereby illustrating the arbitrary nature of state choices. In particular, the authors write (p. 1404) ‘It is interesting to note that we did not find any evidence that the observable component “current age” could be classified, while its unobservable counterpart “previous age” was decodable in OFC’. This finding can be explained by the fact that the task is to match a current image exemplar against a stored invariant prototype of a face or house category that is young or old. Once that category is stored, all matching throughout a block of trials is against the stored ‘previous age’ until a mismatch causes reset of the active category, leading to choice of a category with which to match exemplars in the next block of trials.
These OFC articles illustrate how all the design principles, mechanisms, circuits, and architectures that are summarised in Section 2 play useful roles in explaining data about OFC and its interactions with multiple other brain regions, as in Figures 1–3.
3. Working memory, chunking, and reinforcement in PFC and related areas
This section will explain how different parts of the PFC (Figure 1) interact to regulate the availability of outcomes (Rudeck et al., 2017), solve the credit assignment problem (Asaad et al., 2017), and control aspects of feature-based attention (Baldauf and Desimone, 2014; Bichot et al., 2015). All the explanations derive from properties of prefrontal working memory circuits, the cognitive plans, or list chunks, which are learned from them, and the regulatory cognitive and emotional machinery that enables prefrontal circuits to predict outcomes that can maximise reward based on previously experienced sequences of events.
3.1. VLPFC lesions cause a deficit in learning probabilistic stimulus-outcome associations
Rudeck et al. (2017) concluded that VLPFC encodes the availability, rather than the desirability, of outcomes. They based this conclusion on experiments in which excitotoxic lesions of the VLPFC led to a profound deficit in the ability of lesioned monkeys to learn probabilistic stimulus-outcome associations. Their Experiment 1 first tested the ability to update likelihood estimates for predicted outcomes by training a group of unoperated control monkeys and a group of monkeys with excitotoxic OFC neurons to perform a three-choice probabilistic learning task. Four of the unoperated control monkeys were tested before they also received excitotoxic VLPFC lesions. This procedure enabled monkeys with OFC lesions to be compared with controls, whereas monkeys with VLPFC lesions were compared with their own preoperative performances.
Each session consisted of 300 trials on which monkeys were presented with three novel stimuli on a touch screen monitor. Monkeys sampled different stimuli over trials to learn which stimulus was associated with the highest probability of reward. Reward delivery was based on one of four different reinforcement schedules, called stable, variable, forward, and backward, which use a predetermined series of reward/no-reward outcomes for each option on each trial over the 300 trial testing session. The likelihood of receiving a reward for choosing an option on each trial was calculated using a moving 20-trial window.
Unoperated controls and monkeys with OFC lesions quickly learned which image predicted the highest probability of reward and could track that image as it changed with the reward schedule on each session. In contrast, the VLPFC-lesioned monkeys were severely impaired on this task, except with the schedule wherein one option had a very high probability of reward compared to the others. Thus, the deficit is greatest when the probabilistic difference between the options is small. Unlike unoperated controls and OFC-lesioned animals, VLPFC-lesioned animals were more likely to change their choices between trials than they were before their lesions. After their lesions, they were more likely to switch choices after a rewarded choice than were controls or OFC-lesioned monkeys.
Rudeck et al. (2017) traced these effects to a reduced effect of the long-term effects of reward history, also called contingent learning. Choices on the five preceding trials were analysed. In monkeys with VLPFC lesions, associations between previous choices and the outcomes that contingently followed had essentially no influence on subsequent choices, except when one option consistently had a very high probability of reward. In this last condition, monitoring which previous sequences of stimuli predicted higher reward is not essential to doing the task.
3.2. Some classical data about sequential dependencies influencing future choices
The Rudeck et al. (2017) experiments contribute to a long history of experiments in psychology and psychobiology that study how probabilistic choices are determined by previous sequences of events. These kind of data are often attributed to the short-term storage of sequences of events and the influence of these stored sequences on current choices. The short-term storage of event sequences is accomplished by a working memory (e.g. Baddeley, 1986, 1996, 2012; Baddeley and Hitch, 1974; Cowan et al., 2005; Engle, 2002; Grossberg, 1978a, 1978b; Grossberg and Pearson, 2008; Oberauer, 2009; Silver et al., 2011). The term ‘working memory’ is used here to describe short-term storage of event sequences, not just the persistent storage of one event. Data and neural models will be summarised below that clarify how multiple items can be simultaneously stored in working memory and use these models to explain the Rudeck et al. (2017) data.
An early discovery was that pupil dilation increases with the number of items that are stored in working memory, for example, Kahneman and Beatty (1966) and Unsworth and Robison (2015). Oddball experiments also activate working memory. In an oddball paradigm (Banquet and Grossberg, 1987; Squires et al., 1975), a subject receives a Bernoulli series of two types of stimulus with complementary probabilities: a frequent distractor stimulus, and a rare target stimulus, with unsignalled switches occurring in the probabilities of these stimuli. The subject has to perform a task such as releasing a motor response to each target stimulus, or counting target stimuli. Various measures indicate that a subject tracks the probabilistic sequential dependencies of distractor and target stimuli, for example, distractors 80% of time and targets 20% of time. These include the P300 event-related potential (ERP; Picton, 1992; Sutton et al., 1965), also called the P3b, whose amplitude tends to vary inversely with target stimulus probability (Duncan-Johnson and Donchin, 1977; Tueting et al., 1971). Longer sequences of distractors elicit larger P300s (Remington, 1969; Squires et al., 1976), and the P300 is amplified by practice as an expectation of sequence structure is learned (Banquet and Grossberg, 1987). Adding an ERP manipulation to neurophysiological recordings can be a useful way to coordinate classical psychophysiological explanations with modern neurophysiological methods.
After an object or event sequence is stored in working memory, it can activate the learning of a cognitive plan, or list chunk, by sending output signals through a bottom-up adaptive filter to a category learning level (Figure 7). The adaptive filter obeys the same laws as the one from the feature level F1 to the category level F2 in the ART circuit of Figure 5(a) and (d). Such an adaptive filter can learn to categorise any spatial pattern of activity across a network of feature detectors. In the case where the activity pattern represents a sequence, or list, of items stored in working memory, the category uses the adaptive filter to learn list chunks that selectively respond to subsequences of these stored items. The dynamics depicted in Figure 7 of these two levels and their interactions will be explained below, including how Masking Fields learn to choose the list chunks that predict the most likely outcomes. For now, it suffices to note that learning by the adaptive weights, or LTM traces (hemidiscs in Figure 7), can learn a list chunk using the same kind of ART dynamics that can learn an object category or a value category, but a list chunk codes sequences, or lists, or events rather than objects or affective values.
3.3. Cognitive working memory and list chunks in VLPFC
This and the next few sections will present experimental and modelling facts that lead to an explanation of how VLPFC working memory and chunking dynamics compute the availability of outcomes. A large cognitive neuroscience experimental literature has implicated the VLPFC in working memory tasks, notably tasks that engage verbal and language working memory properties (e.g. Awh et al., 1996; Schumacher et al., 1996), as well as in temporal sequencing properties that are dissociated from the specific stimulus type (Gelfand and Bookheimer, 2003). Nozari, Mirman, and Thompson-Schill (2016) review a large number of competences that all require such a temporal sequencing property at their core, whatever other control structures also exist in order to convert the stored sequences into observable behaviours.
The theoretical discussion below will propose why so many cognitive competences may exploit a similar temporal sequencing property. It will explain how and why all linguistic, spatial, and motor working memories are predicted to exploit variations and specialisations of a similar circuit design (Section 3.15; Grossberg, 1978a, 1978b; Grossberg and Pearson, 2008; Silver et al., 2011), along with psychological and neurophysiological data that support this hypothesis. Henceforth, the kind of working memory that is found in VLPFC will be called a cognitive working memory to distinguish it from the kind of motor working memory in DLPFC that converts the VLPFC stored sequences into sequences that are monitored to choose and perform properly ordered and timed sequential behaviours (Petrides et al., 2002).
3.4. List chunks and reinforcement interact in a probabilistic choice environment
A list chunk that is active when reinforcement occurs can later be amplified by learned feedback interactions from reinforcement-sensitive midbrain structures. The MOTIVATOR model (Figure 3(a)) has already noted how this can happen due to interactions between OFC (or ORB) and the amygdala and BG. Amplification helps this list chunk to win the competition with other possible list chunks.
In particular, area 12o of macaque monkeys interacts with the LH (Öngür et al., 1998), and this area was spared in the excitotoxic OFC lesions of Rudeck et al. (2017), but not the VLPFC lesion. Interactions with PRC may also play a role (Section 3.26; Petrides and Pandya, 2002). The winning list chunk can then be preferentially associated with a rewarded stimulus by the Now Print signals that are emitted by SNc when this stimulus unexpectedly occurs (Figures 3 and 4). On later trials, output signals from this list chunk can be used to prime, and thus help to choose, the next stimulus to be attended and chosen. List chunks that are active when the most rewards occur will thus be favoured to win the competition and to determine future choices.
On the other hand, when a winning list chunk no longer best represents the correct choice, then its choice of an unexpectedly unrewarded stimulus can cause two parallel processes to occur that reduce the probability of choosing that list chunk in the future. First, the unexpected outcome can cause hypothalamic rebounds that can begin to extinguish the motivational support for the list chunk (Sections 2.14 and 2.15; Figure 6(a)). Second, SNc-mediated dopaminergic dips when a chosen stimulus is unexpectedly unrewarded (Figure 4) can begin to weaken the association between the list chunk and its learned prime. These reinforcing events in response to predictive successes and failures are cumulative across learning and performance trials, thereby flexibly shifting the list chunk that has the most support, and thus the choice that the newly chosen list chunk reads out, across trials. When a list chunk influences future choices based on past event sequences and their reinforcement history, it may be said to function as a cognitive plan.
This article argues that such working memory representations and their list chunks exist in the VLPFC, albeit in different cortical layers (see Figure 10 below; Grossberg and Pearson, 2008), and can be amplified by reinforcing events even if the AMYG and OFC are excitotoxically lesioned. In this way, animals with AMYG or OFC lesions can compute the availability of a valued outcome, even if they can no longer estimate the desirability of such an outcome.
Different probabilistic reinforcement schedules require that an animal be sensitive to, and thus store in working memory, different numbers of preceding rewarded and unrewarded choices. Choosing among lower reinforcement probabilities requires sensitivity to longer preceding sequences, other things being equal. When reinforcement is increased, and in the limit rewards every choice of a particular image, then tracking previous sequences of choices is no longer needed to make a correct choice on the next trial. This sort of consideration clarifies why a VLPFC working memory is not needed to adapt to such a (nearly) deterministic reinforcement schedule.
This consideration also calls attention to a basic problem that the brain needs to solve: how does the brain know how to choose the length of preceding event sequences that a chosen list chunk should encode in order to optimise choices that maximise reward? In particular, how does the brain learn list chunks of stored working memory sequences of variable length? And how do these list chunks compete to choose the most predictive list chunk? A brain design that accomplishes this is called a Masking Field (Cohen and Grossberg, 1987; Grossberg, 1978a; Grossberg and Kazerounian, 2011; Grossberg and Myers, 2000). Masking Fields will be discussed in Sections 3.21–3.23 after the predicted link between working memory and list chunking is explained.
3.5. The predicted link between Item-and-Order working memory and list chunking
In order to explain how Masking Fields work, it is first necessary to summarise how the working memories that input to them are themselves designed. Grossberg (1978a, 1978b) introduced a neural model of working memory which posits that a temporal stream of inputs is stored through time as an evolving spatial pattern of content-addressable item representations (Figure 8). These working memory patterns are, in turn, unitised through learning into list chunk representations that can control context-sensitive behaviours. This working memory model is called an Item-and-Order model because, in it, individual nodes, or cell populations, represent list items, and the temporal order in which the items are presented is stored by an activity gradient across the nodes (Figure 8).
The classical work of Miller (1956) on working memory showed that a fundamental functional unit in speech and language is abstract, namely the chunk, that ‘the memory span is a fixed number of chunks [and] we can increase the number of bits of information that it contains simply by building larger and larger chunks, each chunk containing more information than before’. Chunks can thus be learned from multiple types of acoustic inputs that vary in size. Miller (1956) proposed that approximately 7 ± 2 chunks could be simultaneously stored in working memory. He called this immediate memory span the Magical Number Seven, Plus or Minus Two.
In order to work properly, the networks that use Item-and-Order working memories need to be composed of the correct functional units at successive levels of the brain’s speech and language hierarchy. Some early models proposed levels that process phonemes, letters, and words (e.g. McClelland and Rumelhart, 1981), but these levels do not fit basic psychophysical data and cannot self-organise using unsupervised learning. These problems were solved by the Item-and-Order model levels which represent distributed features, item chunks, and list chunks (Grossberg, 1978a, 1978b, 1984b, 1986). An item chunk selectively responds to one pattern of activity, or a small set of similar patterns of activity, across the distributed feature detectors within a brief time interval (e.g. a phoneme). A list chunk selectively responds to the temporal order of a single sequence of item chunks that are stored in working memory. The properties of these functional units have been used to explain data about word superiority effect, list length effect, and related speech phenomena that are incompatible with alternative processing levels (see Section 3.23).
3.6. Correct temporal order is stored in working memory by a primacy gradient
A primacy gradient stores items in working memory that will be recalled in the correct temporal order. In a primacy gradient, the first item in the sequence activates the corresponding item chunk with the highest activity, the item chunk representing the second item has the second highest activity, and so on, until all items in the sequence are represented (Figure 8). For example, a sequence ‘A-B-C’ of items is transformed into a primacy gradient of activity with cells encoding ‘A’ having the highest activity, cells encoding ‘B’ having the second highest activity, and cells encoding ‘C’ having the least activity.
3.7. Rehearsal and inhibition-of-return
How is a stored spatial pattern in working memory converted to a temporally performed sequence of items during recall? A rehearsal wave that is delivered uniformly, or nonspecifically, from the BG to the entire working memory enables read-out of stored activities (Figure 8). The node with the highest activity is read out fastest because it exceeds its output threshold fastest. As it is read out, it also self-inhibits its working memory representation via a recurrent inhibitory interneuron. This self-inhibitory process mechanises the cognitive concept of inhibition-of-return, which prevents perseverative performance of the most recent item (Posner et al., 1985). Self-inhibition of the last item to be performed is repeated until the entire sequence is performed. These operations may also be influenced by different rehearsal strategies, as illustrated by performance differences during immediate free recall experiments, during which subjects attempt to recall items in any order after a single hearing, versus immediate serial recall (ISR) experiments, during which subjects attempt to recall items in the order that they were heard. These different kinds of data are explained and quantitatively simulated using an Item-and-Order model in Grossberg and Pearson (2008).
3.8. Competitive queuing and primacy models
After Grossberg (1978a, 1978b) introduced the Item-and-Order model, it was used in various forms in a number of other studies (e.g. Boardman and Bullock, 1991; Bohland et al., 2010; Bradski et al., 1994; Bullock and Rhodes, 2003; Grossberg and Pearson, 2008; Houghton, 1990; Page and Norris, 1998). For example, Page and Norris (1998) used a Primacy Model to explain and simulate cognitive data about immediate serial order working memory, including data about word and list length, phonological similarity, and forward and backward recall effects. Properties of the Item-and-Order model were also used in the Competitive Queuing model (Houghton, 1990) which describes how the most active item that is stored in working memory self-inhibits its stored activity when it is performed.
3.9. Supportive psychological and neurophysiological data for Item-and-Order networks
Both psychophysical and neurophysiological data have supported the Item-and-Order predictions that working memories encode item order with relative activity levels and are reset by self-inhibition. For example, Farrell and Lewandowsky (2004) did psychophysical experiments in humans that studied the latency of responses following serial performance errors. They concluded that
Several competing theories of short-term memory can explain serial recall performance at a quantitative level. However, most theories to date have not been applied to the accompanying pattern of response latencies … Data from three experiments … rule out three of the four representational mechanisms. The data support the notion that serial order is represented by a primacy gradient that is accompanied by suppression of recalled items. [italics mine] (p. 115)
Electrophysiological experiments have also directly recorded Item-and-Order working memory properties. For example, recordings in the posterior principal sulcus of the DLPFC of macaque monkeys were done by Averbeck et al. (2002) while monkeys performed learned arm movement sequences that copy geometrical shapes. These authors did extracellular recording from the areas near, but outside the depths of, the caudal portions of the principal sulcus in DLPFC of macaque monkeys during the performance of a sequential motor task. In this task, monkeys used an X-Y joystick to copy concurrently presented geometrical shapes (triangle, square, inverted triangle, trapezoid) on an LCD screen in a prescribed order. Copying proceeded counter-clockwise starting at the top middle of each shape. The recorded cell responses were pooled on the basis of the movement in the sequence with which their firing pattern most correlated. The population response for each movement in the series is shown in Figure 9(a).
The predicted properties of a primacy gradient and a self-inhibitory form of inhibition-of-return are evident in these data. Figure 9(a) depicts stored working memory activities for drawing several different geometrical shapes. In each example, the movement with the largest activity is performed first, before it self-inhibits its activity. Then, the next largest activity reads out its movement while self-inhibiting its activity. After the next-to-last movement is performed, the activity of the final movement grows greater than that of any prior movement command because it has no competitors left that are stored in working memory. How this competition works, and why it exists, is explained in Section 3.15.
These properties were simulated (Figure 9(b)) using an Item-and-Order working memory by the laminar cortical LIST PARSE model of Grossberg and Pearson (2008). This model proposes how a cognitive working memory in VLPFC is converted into a motor working memory in DLPFC (cf. Figure 1) which, when properly monitored, can perform the stored sequence at variable rates that are under volitional control from the BG (Figure 10). In all, the LIST PARSE model shows how a prefrontal linguistic working memory in VLPFC can quantitatively simulate human psychophysical data about ISR, and immediate, delayed, and continuous distractor free recall; whereas a similarly designed prefrontal motor working memory in DLPFC can quantitatively simulate the Averbeck et al. (2002) neurophysiological data about sequential recall of stored motor sequences (Figure 9). This property illustrates the fact that all linguistic, spatial, and motor working memories use variations of the same network design in VLPFC and other prefrontal areas, for reasons that are explained in Section 3.14.
3.10. Bowed and recency gradients during free recall and probabilistic choice behaviours
Before explaining why all working memories are embodied by a similar circuit design, it is important to realise when and why such working memories do not accurately encode or perform temporal order. Some of these data are summarised in this section and the next. Why this happens is explained in Section 3.12, which links this property to the ability to learn and stably remember list chunks. These explanations and predictions would benefit from further neurophysiological studies using longer lists.
Item-and-Order working memories provide principled answers to the following basic questions about working memory: What is the longest list that the brain can store in working memory in the correct temporal order? Why can only relatively short lists be stored with the correct temporal order in vivo? In an Item-and-Order working memory, this question translates into: What is the longest primacy gradient that the working memory can store? How is a primacy gradient altered when longer lists are stored? Free recall data illustrate how primacy gradients change if a longer list is stored. Then, a bowed serial position curve is often observed (Figure 11; for example, Murdock, 1962) such that items at the beginning and the end of the list are performed earliest, and with the highest probability of recall.
Grossberg (1978a, 1978b) noted that these free recall properties can readily be explained if the working memory gradient that stores the list items is also bowed. Then, the first and last items have the largest activities, and items in the middle have less activity. The temporal order of recall is explained as follows: the item with the largest activity is read out first, whether at the list beginning or end, because it exceeds the output threshold first of its output pathway when the rehearsal gate opens (Figure 8). As its output is read out, it then self-inhibits its working memory activity via a recurrent inhibitory interneuron (Figure 8) to prevent preservation. Then, the next largest item will be read out and so on in the order of item relative activity. The probability of recall has the following explanation: items that are stored with larger activities have greater resilience against perturbation by cellular noise. Transpositions of order during recall are explained in the same way because transposed items have similar stored activities.
If even longer lists get stored, then the bow increasingly resembles a recency gradient, such that items at the end of the list are performed earliest, and progressively earlier items are stored with less and less activity (Figure 11). In a probabilistic choice experiment with low probabilities of reward, such a recency gradient may be expected to develop. This working memory property helps to explain the Rudebeck et al. (2017) data that unoperated and OFC-lesioned ‘monkeys were making contingent associations between their specific choices and subsequent outcomes. This effect diminished with increasing distance from the current trial, suggesting that monkeys preferentially used the most recent feedback to guide future choices’ (p. 1211). The way that primacy gradients become bows and then recency gradients as stored list length increases hereby reconciles the different working memory properties that have been reported in experiments like Averbeck et al. (2002) and Rudebeck et al. (2017). It would be interesting to record the transformation from primacy gradient, to bow, to extended recency gradient by systematically changing the number of choices and choice probabilities in this kind of experiment. Another informative kind of experiment could alter the strength of recurrent inhibition across the network, say with a GABA antagonist, and test how it alters the transient and immediate memory spans that will now be explained.
3.11. Magical numbers four and seven: immediate and transient memory spans
What is the longest primacy gradient that can be stored? The classical Magical Number Seven, or immediate memory span, of 7 ± 2 items that is found during free recall (Miller, 1956) estimates the upper bound. Section 3.23 offers an explanation of the Magical Number Seven using an Item-and-Order network. Grossberg (1978a) distinguished between the immediate memory span and the transient memory span. The transient memory span was predicted to be the longest list for which a primacy gradient may be stored in working memory solely as the result of bottom-up inputs (see Figure 11). In contrast, the immediate memory span was predicted to arise from the combined effect of bottom-up inputs and top-down LTM read out from list chunks that could be activated by the working memory (Figure 7). Grossberg (1978a) mathematically proved that the read-out of top-down long-term memories can only increase the maximal primacy gradient that can be stored, and thus that the immediate memory span is longer than the transient memory span. Given an estimated immediate memory span of approximately seven items, it was estimated that the transient memory span should be approximately four items. Cowan (2001) has summarised data that support this prediction by showing that, when the influences of LTM and grouping effects are minimised, there is a working memory capacity limit of 4 ± 1 items. There is thus also a Magical Number Four.
If technically possible, reversible cooling or other method for silencing the list chunk level should therefore lead to a shortening of the longest primacy gradient that the working memory can store.
3.12. LTM Invariance Principle: learning stable list chunks
Why is the transient memory span so short? The proposed answer to this question suggests that neurophysiological experiments be done that combine recordings of working memory storage and list chunk learning. In particular, a network for STM storage of sequences in working memory can only realise its full potential if it can also support the learning and LTM of list chunks. Indeed, without stable list chunk learning and memory, it would be impossible to learn and perform language, motor skills, or navigational routes. Grossberg (1978a, 1978b) predicted that all working memories are designed to enable learning and stable memory of list chunks and showed that two simple postulates imply these properties: the LTM Invariance Principle and the Normalisation rule. Grossberg (1978a, 1978b) also explained how these postulates can be mathematically realised by Item-and-Order working memories and how they generate primacy and bowed gradients to explain free recall data. Since this early derivation, the understanding of how these working memories are realised in vivo has been incrementally refined (e.g. Bradski et al., 1992, 1994), leading to laminar cortical models of how prefrontal circuits realise Item-Order-Rank working memories (Grossberg and Pearson, 2008; Silver et al., 2011) that can store sequences with repeated items in working memory, for example, ABACAD (see Sections 3.16–3.18).
The LTM Invariance Principle implies that novel sequences of items may be stored in working memory and chunked through learning in a way that does not destabilise memories of previously learned chunks. It explains, for example, how a sequence of the item chunks of the longer word MYSELF can be stored in working memory without forcing catastrophic forgetting of previously learned, shorter list chunks for the words MY, SELF, and ELF. Said in another way, the LTM Invariance Principle shows how, if bottom-up inputs activate a previously learned chunk of the word MY, then storage in working memory of the remaining portion SELF of the novel word MYSELF will not erode the previously learned adaptive weights that support activation of the list chunk of MY. Thus, the LTM Invariance Principle begins to explain how sequences of variable length may be stored in working memory and induce learning of list chunks that selectively respond to them. As noted in Section 3.10, longer sequences may need to be stored and selectively chunked when reinforcement probabilities are lower in order to predict the most likely stimulus that will next be rewarded. As Figure 7 illustrates, variable-length list chunks can be learned by a network that is called a Masking Field.
The LTM Invariance Principle is achieved mathematically by preserving the relative activities, or ratios, between previously stored working memory activities as new items are presented and stored in the working memory through time. Newly arriving inputs may, however, alter the total activity of each active cell across the working memory. How does preserving activity ratios help to stabilise the adaptive weights of previously learned categories? These activities send signals to the next processing stage, where the category cells are activated (Figure 7). The signals are multiplied, or gated, by adaptive weights, or LTM traces (in the synaptic knobs with hemidiscs in Figure 7), before the net adaptively gated signals activate their target categories. Multiplicative gating of the bottom-up signals by LTM traces converts the bottom-up pathways into an adaptive filter. The total input to a category thus multiplies a pattern, or vector, of activities times a pattern, or vector, of LTM traces. This kind of multiplication of two vectors is said to carry out an inner product, or dot product, operation. By preserving relative activities of the stored working memory activities, the relative sizes of these total inputs to the category cells do not change through time, and thus nor do the corresponding LTM patterns that track these activities when learning occurs at their category cells.
For example, suppose that bottom-up acoustic inputs are stored in working memory and activate their corresponding list chunks. As these inputs arrive, a chunk such as ‘MY’ may become active once it receives all or most of its expected bottom-up input. If the acoustic inputs are then followed immediately by silence, the chunked representation of ‘MY’ could stably learn from the stored STM pattern of activity that first supported it. On the other hand, if the acoustic inputs are rapidly followed by further acoustic signals (e.g. the inputs corresponding to ‘MYSELF’), then these newly arriving inputs could drastically alter the pattern of activation that represents MY in STM if the LTM Invariance Principle did not hold. If this could happen, then the LTM traces that activate chunk ‘MY’ could change in response to the now altered STM pattern in working memory. The LTM Invariance Principle prevents this from happening, since the newly arriving inputs (corresponding to ‘SELF’) then leave intact the relative pattern of activity in working memory of the already occurring acoustic inputs (corresponding to ‘MY’). A new list chunk for the full word (corresponding to ‘MYSELF’) could then be learned without destabilising the already learned LTM pattern for its subset components (e.g. ‘MY’).
The Normalisation Rule ensures that the total activity that is stored in working memory has an upper bound that tends to be independent of the number of items that are stored. Thus, if more items are stored, then each item tends to be stored with less activity (see Figures 9 and 11). This normalisation property implies the limited capacity of working memory (Baddeley and Hitch, 1974; Grossberg, 1978a, 1978b) by redistributing, rather than adding, activity when new items are stored. Storing more items in working memory thus causes each of them to have less activity. When a sufficiently long list length is reached, trying to store more items will prevent some of them from exceeding the storage threshold.
Consistent with this conclusion, application of a GABA antagonist to DLPFC, and thus a weakening of recurrent inhibition, causes an overall increase in cell activity (Rao et al., 2000). In the opposite direction, excess GABA-mediated inhibition of prefrontal neuronal activity has been identified as a contributor to working memory dysfunction during the first 4 months following a cortical impact injury in rats (Hoskison et al., 2009).
3.13. Bowed gradients for long lists follow from self-stabilising memory
Properties such as the transient memory span were mathematically proved in Grossberg (1978a, 1978b) to follow if both the LTM Invariance Principle and the Normalisation Rule hold. If a list is longer than the transient memory span, then the primacy gradient that is initially stored will evolve into a bowed gradient as more items are stored, as illustrated in Figure 11.
In other words, the ability of a working memory to enable learning and stable memory of stored sequences implies an upper bound on the length of lists that can be temporarily stored in the correct temporal order. The bowed serial position curves of free recall and ISR data can thus be understood as the price paid for being able to rapidly learn, and stably remember, language and sequential spatial and motor skills.
These results about primacy gradients hold when the same amount of attention is paid to each item as it is stored in working memory. If attention is not uniform across items, then multi-modal bows can occur. These Von Restorff (1933) or isolation effects (Hunt and Lamb, 2001) enable items in the middle of a list to be remembered and recalled before other list items.
3.14. Similar circuits for linguistic, spatial, and motor working memories
It can now be seen from the deeper perspective of chunk learning why VLPFC and DLPFC are involved in storing so many types of event sequences, including the probabilistic choice sequences that are activated during the Rudebeck et al. (2017) experiments. Namely, if all working memories obey the LTM Invariance Principle and the Normalisation Rule, then all linguistic, motor, and spatial working memories should have a similar design, and thus should therefore exhibit similar data properties, such as error distributions. Data that support this prediction include the following: Jones et al. (1995) reported similar performance characteristics to those of verbal working memory for a spatial serial recall task in which visual locations were remembered. Agam et al. (2005) reported psychophysical evidence of Item-and-Order working memory properties in humans as they performed sequential copying movements, and Averbeck et al. (2002, 2003a, 2003b) reported neurophysiological evidence for such a working memory in monkeys during performance of sequential copying movements (Figure 9(a)).
Model explanations of working memory data also support the prediction of a universal working memory design for all kinds of input sequences. It has already been noted that the LIST PARSE model of Grossberg and Pearson (2008) has simulated the Averbeck et al. data using a prefrontal motor Item-and-Order working memory (Figure 9(b)) and has used a prefrontal linguistic working memory to quantitatively simulate human psychophysical data about ISR, and immediate, delayed, and continuous distractor free recall. The lisTELOS model of Silver et al. (2011) has, in addition, used a prefrontal Item-Order-Rank spatial working memory to quantitatively simulate neurophysiological data about the learning and planned performance of saccadic eye movement sequences.
The LTM Invariance Principle and Normalisation Rule also imply that there is an intimate connection between the process of storing sequences temporarily in working memory and the learning of list chunks by the next processing stage. Data that support this prediction include the following: Agam et al. (2007) reported data about the formation of list chunks as movement sequences are practiced. Psychophysical experiments on speech perception have also successfully tested this prediction (e.g. Auer and Luce, 2008; Goldinger and Azuma, 2003; Luce and McLennan, 2008; McLennan et al., 2003, 2005; Vitevitch and Luce, 1999).
3.15. Recurrent shunting on-centre off-surround networks embody working memories
What is this shared working memory design? In particular, are postulates such as the LTM Invariance Principle and the Normalisation Rule too sophisticated for evolution to discover? In fact, both the LTM Invariance Principle and the Normalisation Rule are embodied within a ubiquitous neural design, thereby clarifying how such a working memory could arise through evolution: a recurrent on-centre off-surround network (Figure 8) whose cells obey the membrane equations of neurophysiology, also called shunting dynamics. Such networks occur ubiquitously because they enable the brain to process and store distributed patterns of inputs without being degraded by noise – when their inputs are small – or saturation – when their inputs are large, thereby solving the noise-saturation dilemma that is faced by every brain network (Grossberg, 1973, 1980).
How such recurrent shunting networks process ratios (LTM Invariance Principle) and conserve total activity (Normalisation Rule) was mathematically proved in Grossberg (1973). Also see Grossberg (1978a, 1980) for reviews. Bradski et al. (1994) went on to prove theorems about how Item-and-Order recurrent shunting on-centre off-surround networks generate primacy and bowed gradients, among other properties, as a function of network parameters.
The excitatory feedback due to the recurrent on-centre interactions in such a network helps to store an evolving spatial pattern of activities in response to a sequence of inputs. The recurrent shunting off-surround, in concert with the on-centre, helps to preserve the relative activities that are stored. A volitional rehearsal signal from the BG enables the highest stored activity to be read out first, and self-inhibitory feedback prevents perseverative performance of this most highly activated cell population, thereby enabling less active populations to be performed (Figure 8), while the network as a whole gradually renormalises its activity through time. For example, the Normalisation rule clarifies why, in Figure 9, after the next-to-last item that is stored in working memory has been performed, the population that stores the last item is disinhibited. By being freed from inhibitory normalisation, it can reach the highest activity through time of any population that stored the list.
3.16. Storing lists with repeated items: Item-Order-Rank working memory
In a probabilistic learning situation such as the one used by Rudebeck et al. (2017), where a subject chooses one of three images to receive reward, each image may repeat itself over trials. In order to estimate the choice that will most probably lead to reward, an animal needs to keep track of such a contingency. This state of affairs raises the question: How are sequences with repeated items stored in working memory? Can recordings be made during experiments on availability that more completely characterise the predicted circuits whereby repeated items are stored in a PFC working memory? The need for combined prefrontal-parietal recordings of availability in the broadest sense is also suggested by the prediction in Section 3.17 of how PFC working memories that can store repeated occurrences of each experienced option receive inputs from parietal mechanisms for numerically estimating how much of a valued goal object is available during foraging in naturalistic environments.
In its simplest form, an Item-and-Order working memory cannot represent the same item in multiple positions, or ranks, of a list. However, there are many examples in human cognitive data of sensitivity to list position (e.g. Henson, 1998). For example, when presented with the sequence ABC (pause) DEF, exchanges between items at B and E are more common than exchanges between items at B and F. In addition, phonemes or syllables in similar positions in different words may be selectively interchanged. These examples include spoonerisms, for example, ‘hissed my mystery lesson’. Error data in human serial recall experiments also indicate that rank information is available, which some models of serial recall have incorporated (see Grossberg and Pearson, 2008 for a review). In monkeys, some PFC neurons respond, not only to a given item but also to the rank of that item within a sequence of items. Such a cell may respond to a specific target that is presented in a specific list position (e.g. Averbeck et al., 2003a; Barone and Joseph, 1989; Funahashi et al., 1997; Inoue and Mikami, 2006; Kermadi and Joseph, 1995; Ninokura et al., 2004).
3.17. From parietal numerical map to prefrontal Item-Order-Rank working memory
Given that both psychophysical and neurophysiological data also support Item-and-Order models, it remains to explain how rank information may be integrated into such a working memory. Bradski et al. (1994) proposed the first Item-Order-Rank working memory model that incorporated rank-order coding into an Item-and-Order working memory that is capable of storing item repeats at arbitrary list positions, for example, ABACBD. The LIST PARSE model (Grossberg and Pearson, 2008) predicted where in the brain this rank-order coding arises, namely, the analogue spatial representations of numbers that exist in the parietal cortex. The model specifies how an Item-Order-Rank working memory can be created in PFC using parietal-prefrontal projections from this parietal numerical representation. This prediction built upon the Spatial Number Network, or SpaN, model of Grossberg and Repin (2003) which simulated how the analogue map of ordered numerical representations in inferior parietal cortex enables animals and humans to estimate and compare sufficiently small numerical quantities. Such a capability can have life-saving consequences in terrestrial animals who forage for food. For example, choosing a tree with more fruit, or a flower with more honey, illustrates how survival may be enhanced by being able to estimate and compare numerical quantities. The predicted properties of SpaN model parietal neurons were supported by neurophysiological data of Nieder and Miller (2004a), who also studied the prefrontal projections of these parietal numerical representations (Nieder and Miller, 2004b). It would be very interesting to combine the Nieder and Miller experimental manipulations with manipulations of option availability.
In an Item-Order-Rank working memory, a spatial gradient of activity still represents temporal order, with the most active cell population being performed first. The main new idea is that each item representation in an Item-Order-Rank working memory has multiple map positions, or slots, that can store occurrences of the item at multiple list positions. For each item, the cell population that codes all the possible list positions, or ranks, forms a numerical hypercolumn. The parietal number map hereby broadcasts a numerical hypercolumn that is incorporated into multiple PFC item representations using a conjunction of item and numerical information.
This parietal-prefrontal projection enables the correct position in the hypercolumn to be activated when the item occurs with a given rank in the list. The numerical hypercolumn that represents a specific list item can hereby store that item in multiple list positions, just as a positional hypercolumn in the primary visual cortex can selectively respond to multiple orientations at a given position in space (Hubel and Wiesel, 1962, 1963). Thus, to store the list ABAC, item A would be stored in the first and third slots within its hypercolumn, item B would be stored in the second slot within its hypercolumn, and item C would be stored in the fourth slot within its hypercolumn.
A primacy gradient of activity (Figure 11) would still represent the temporal order of a short stored list, whether or not it had repeated items. Working memory gradients would be created in the same way as in an Item-and-Order working memory. The recurrent on-centre off-surround network that stores items in an Item-Order-Rank working memory can still use self-excitatory feedback from each cell population to itself, and a broad off-surround can still equally inhibit all other populations in the working memory, including all the slots of all the items. A similar concept has been used by Davis (2010) to model letter repetitions during visual word identification.
3.18. lisTELOS: an Item-Order-Rank spatial working memory in PFC
How to combine parietal and prefrontal manipulations to get more insight into how option availability is computed may be guided by available models of neural architectures in which Item-Order-Rank working memories play a key role. For example, Silver et al. (2011) implemented an Item-Order-Rank model of spatial working memory in DLPFC that is part of the larger lisTELOS neural architecture (Figure 12) that also includes posterior parietal cortex (PPC), including the lateral intraparietal cortex (LIP), as well as the supplementary eye fields (SEF), frontal eye fields (FEF), BG, thalamus, and superior colliculus (SC). The pART macrocircuit in Figure 1 includes four of these brain regions. The lisTELOS architecture simulates the temporal dynamics of how the brain may control working memory storage, choice, and execution of saccadic eye movement sequences, as they are used to carry out different kinds of tasks. In this spatial working memory, the ‘items’ are target positions to which saccades are commanded to move. An activity gradient across these target positions can store the sequence of target positions to which the eyes will move in the stored temporal order.
The lisTELOS Item-Order-Rank working memory was used to simulate neurophysiological data about SEF and FEF cells that are rank-sensitive (Isoda and Tanji, 2002, 2003). Given that all working memories have a similar network design in order to realise the LTM Invariance Principle and Normalisation Rule, the Item-Order-Rank working memory of the lisTELOS model may be considered a prototype for all linguistic, spatial, and motor working memories in which repeated items may occur.
lisTELOS was also used to learn and perform saccades in five experimental paradigms that did not require storage of a sequence of saccadic target positions (Hikosaka et al., 1989). These paradigms had earlier been simulated by the TELOS model of Brown et al. (2004). Four of these paradigms are shown in Figure 13. A fifth paradigm, the fixation task, requires an animal to learn not to move its eyes from a foveated fixation cue while the cue remains on. This ability is then used to also learn the other four tasks, which all require that the monkey foveate the fixation cue until it turns off, and then to move to an extrafoveal target position. The delayed saccade task, in particular, requires PFC storage of a target light, since the target shuts off 500 ms before the fixation cue does.
Both TELOS and lisTELOS were able to learn all the tasks. After learning was complete, they could perform all the tasks with the model’s learned parameters, which the models used to reproduce, and predict distinct functional roles for, the neurophysiologically recorded dynamics of 17 different cell types, including properties of FEF dynamics and of sustained cell responses in PFC during the delay period. These predicted functional roles remain to be experimentally tested.
Several kinds of simulations tested how the PFC working memory interacts with the other simulated brain regions. One such test simulated the ISR task, which has served as a benchmark for models of working memory for many years. The ISR task can be divided into two phases. First, a sequence of cues is presented that must be remembered in order. Second, the cues are reproduced in the order in which they were presented. In the lisTELOS simulations, a fixation cue was first presented towards which the model must execute a saccade. Then, a sequence of cues was presented at various spatial positions while fixation was maintained. Finally, the fixation cue was removed. Its disappearance acts as a GO signal during which a BG gate opens, thereby initiating saccades to each of the cued positions in order. This task illustrates, among other things, how SEF chooses the next saccade to be performed from a stored sequence, and how three BG loops sequentially open their gates to enable three different kinds of operations to occur: the storage of a target position sequence in working memory, the choice of the next saccade to be performed, and the performance of that saccade. This gating function of the BG is controlled by the substantia nigra pars reticulata (SNr; Alexander and Crutcher, 1990; Alexander et al., 1986; Grahn et al., 2009; Hikosaka and Wurtz, 1983, 1989; see Grossberg, 2016, for a review).
lisTELOS also simulated DLPFC working memory dynamics that quantitatively reproduced behavioural data, including reaction times, from two microstimulation paradigms (Histed and Miller, 2006; Yang et al., 2008). These experiments focused on the SEF as a region that chooses the next saccade to perform from the sequence of target positions that is stored in the DLPFC working memory (Nachev et al., 2005; Parton et al., 2007; So and Stuphorn, 2010; Taylor et al., 2007). Electrode stimulations altered the order in which eye movements were carried out, but not the target positions to which the eyes moved (Figure 14(a) and (b)). Quantitative simulations of how the saccadic order was changed by microstimulation provided strong evidence for Item-Order-Rank coding in a spatial working memory by showing how the relative activities, and thus the order of recall, were altered. The quite different findings of Histed and Miller (2006) and Yang et al. (2008) about how saccade latency responded to miscrostimulation were also given a unified explanation. These miscrostimulation results followed from how the stimulating electrode increased activity-dependent habituation within SEF feedback pathways (Figure 12) near the electrode position and gradually less with increasing distance, as depicted by the spatial gradient from black (most active) through grey to white (least active) in Figure 14(c).
3.19. DLPFC credit assignment by selective and sustained working memory storage
The above examples illustrate some of the working memory properties that are needed to solve the credit assignment problem that Asaad et al. (2017) discuss in the light of their own DLPFC neurophysiological data. As the authors note (p. 6995), ‘credit assignment is the process by which we infer the causes of successes and failures’. In particular, a solution of the temporal credit assignment requires ‘a stable representation of relevant information over time … so that reinforcement received when an outcome becomes apparent can be applied to the same neural ensemble that earlier signaled the causal features’ (p. 6996). A solution of the structural credit assignment problem requires the ability to select causal features when multiple potentially relevant features are simultaneously available.
In the cue learning experiment reported by Asaad et al. (2017) that was used to test credit assignment, rhesus macaques learned which cue was correct by trial and error in blocks of trials, before an unsignalled new block was begun in which a new cue was correct. During each trial, animals were required to maintain central fixation as four cue objects were presented peripherally for 500 ms. The monkeys had to continue maintaining fixation for 1 s after the cues shut off and then to make a choice of the correct cue by making a saccade to the position where the correct cue had appeared. Visual feedback was then presented that indicated whether the choice was correct or not (a green circle or a red X), followed by a juice reward if the choice was correct. Feedback did not reveal which cue had been at the selected position, so the animal had to maintain in working memory a representation of the correct cue against which to match one of the four cues and then saccade to its remembered position. The position of the correct cue was changed across trials within the block.
In the spatial version of the task, the correct choice was determined solely by a cue’s spatial position, and not by the identity of the cue that had previously appeared at that position. In both tasks, an animal’s attention to a particular cue had to be learned from the correct outcomes across trials.
Blocking and unblocking experiments (Section 2.7) also require that an animal infer what cues are predictive based on expected or unexpected consequences that occur after these cues shut off.
A necessary condition for correct credit assignment is thus the ability of the PFC to maintain a representation of the correct cue throughout each trial. Item-Order-Rank working memories accomplish this using the positive feedback in their recurrent shunting on-centre off-surround networks (Section 3.15 and Figure 8). As noted below, such storage is permitted when the appropriate SNr gates open (cf. Figure 11). Another necessary condition is that reinforcing feedback can increase the probability that predictive list chunks will be chosen. This feedback can act either via the amygdala through incentive motivational feedback (Figures 2 and 3) or via the BG through feedback circuits that are strengthened by unexpected rewards. These circuits may, or may not, include the amygdala (Figures 3, 4, and 11).
Reinforcing feedback that enhances the activity of predictive PFC list chunks, combined with SNr gating, are not the only ways that the PFC can selectively choose and store causal contingencies. Sections 3.25 and 3.26 will summarise how the PRC and PHC are proposed to also contribute to this selectivity (Figure 1) by enabling object and spatial contexts, respectively, to influence the storage process. Additional mechanisms regulate whether items will be allowed to be stored in working memory in the first place. Explaining how this occurs requires mechanisms that have not previously been modelled, including functional roles of the ventral bank of the principal sulcus (VPS) and the ventral prearcuate gyrus (VPA; Figure 1). These regions have been shown to enable monkeys to attend and foveate objects that possess task-relevant feature combinations (Bichot et al., 2015). A related role of the inferior frontal junction (IFJ) in humans has also been reported (Baldauf and Desimone, 2014). Here, it will be proposed how these regions may also regulate selective PFC working memory storage.
Before an explanation is offered of how this is proposed to happen, it is useful to summarise some of the data which demonstrate that working memory storage is indeed selective. PFC storage may be prevented on tasks that do not require storage of visual information, consistent with data demonstrating that PFC working memory cells do not fire during such tasks (Fuster, 1973; Kojima and Goldman-Rakic, 1984). Selective storage is also consistent with the observation that, given the presentation of identical stimuli, neural selectivity in PFC depends on subsequent task demands (Warden and Miller, 2010). Awh and Vogel (2008), describing imaging data from McNab and Klingberg (2008), noted that success on working memory tasks was associated with an individual’s ability to selectively identify and store task-related stimuli from a larger sequence of stimuli. Tsushima et al. (2008) showed that subliminal distractors can damage performance in attention tasks, but that making distractors supra-threshold can alleviate performance deficits, perhaps by facilitating the ability to filter them out. Suzuki and Gottlieb (2013) showed similarly that, during a memory saccade task in which a salient distractor was flashed at a variable time and position during the memory delay, responses to the salient distractor were more strongly suppressed and more closely correlated with performance in DLPFC than in LIP. The brain may hereby learn to ‘blacklist’ distracting stimuli before they can be stored in PFC, allowing all other information to be stored. When all of these functions are realised, they embody a solution of the structural and temporal credit assignment problem that the DLPFC has been proposed to solve (Asaad et al., 2017).
3.20. Two processes regulate whether items will be stored in working memory
It is now possible to distinguish two distinct processes that can influence working memory storage, even after all earlier preprocessing has taken place. The first process, as noted above, carries out task-sensitive filtering of individual items before they reach the working memory. This process selects only those items for storage whose feature combinations are compatible with task requirements; for example, only red objects from a sequence of objects with various colours (cf. Egeth et al., 1984; Grossberg et al., 1994; Treisman and Gelade, 1986; Wolfe et al., 1989). This process will be further discussed in Section 3.31 which proposes a mechanistic explanation of the roles of the VPS and the VPA in enabling monkeys to attend and foveate objects that possess task-relevant feature combinations (Bichot et al., 2015) and the related role of the IFJ in humans (Baldauf and Desimone, 2014).
As noted above, the second process enables all the items that get through the filter to be stably stored after they reach the working memory. This corresponds to keeping an SNr gate open during list storage and option prediction (Figure 12). Closing the SNr gate can rapidly reset, or delete, the entire stored sequence from working memory when there is an attention shift to do a different task.
3.21. Masking field working memory chunks variable-length lists
Before discussing the feature-selective gating of objects before they can be stored in VLPFC, it is important to explain how the list chunking networks learn how to selectively categorise sequences of variable length that are stored in working memory; for example, A versus AB versus ABC. How is the most predictive list chunk in a given situation chosen? In particular, in a probabilistic learning situation, how is the most predictive list chunk chosen so that it can read out the best top-down prime with which to predict the most likely stimulus to be reinforced on the next trial, and thereby guide the choice of this stimulus? A Masking Field network has these properties (Figure 7; Cohen and Grossberg, 1986, 1987; Grossberg, 1978a, 1984b, 1986; Grossberg and Kazerounian, 2011; Grossberg and Myers, 2000).
A Masking Field is a specialised type of Item-and-Order working memory that can store chosen list categories through time. As with all Item-and-Order working memories, it is defined by a recurrent on-centre off-surround network whose cells obey the membrane equations of neurophysiology (Section 3.15). The stored ‘items’ of a Masking Field are list chunks that are selectively activated (e.g. chunks representing ‘MY’ and ‘MYSELF’ in Figure 7), via a bottom-up adaptive filter, by prescribed sequences of items that are stored in an Item-and-Order working memory at an earlier processing level (e.g. M, Y, S, E, L, F in Figure 7). The network in Figure 7 thus first compresses spatial patterns of feature detectors into item chunks starting in inferotemporal cortex, and then sequences of the item chunks that are stored in a VLPFC working memory are compressed into list chunks there, among other cortical regions (Figure 1).
3.22. Temporal chunking problem: learning words of variable length
How are the variable-length list chunks of a Masking Field learned, so that they can be used to effectively predict subsequent outcomes? Masking Fields were introduced to solve this Temporal Chunking Problem (Cohen and Grossberg, 1986, 1987; Grossberg, 1978a, 1986) which concerns how a list chunk of an unfamiliar list of familiar speech units – for example, a novel word composed of familiar phonemes or syllables – can be learned under the type of unsupervised learning conditions that are the norm during daily experiences with language.
In particular, before a novel word, or list, can fully activate the adaptive filter, all of its individual item chunks must first be presented. What prevents the familiarity of smaller lists (e.g. MY, ELF, and SELF), which have previously learned to activate their own list chunks, from forcing the novel longer list (e.g. MYSELF) to always be processed as a sequence of these smaller familiar chunks, rather than being able to drive learning of its own list chunk (Figure 7)? How does a not-yet-established word representation overcome the salience of already well-established phoneme, syllable, or word representations to enable learning of the novel word to occur?
3.23. Self-similar growth and competition solve the temporal chunking problem
A Masking Field accomplishes this using cells with multiple cell and receptive field sizes, or scales (Figure 7), that are related to each other by a property of self-similarity; that is, each scale’s properties, including its cell body sizes and their excitatory and inhibitory connection lengths and interaction strengths, are a multiple of the corresponding properties in another scale.
Such a self-similarity property can develop as a result of simple activity-dependent growth laws (Cohen and Grossberg, 1986, 1987). Here is one possible scenario: suppose that item chunk cells in the working memory are endogenously active during a critical period of development. As a result, Masking Field cells that receive inputs from a larger number of item chunk cells receive a larger average total input activity through time. Activity-dependent cell growth causes the Masking Field cell bodies and connections to grow approximately proportionally. This property is called self-similar growth. Cell growth terminates when the cell bodies become large enough to dilute their activities sufficiently in response to their inputs so that they no longer exceed a growth-triggering threshold. Cells that receive more inputs grow larger as a result, so that the effects of individual inputs are smaller on larger cells. In effect, self-similar growth normalises the total effect of all the inputs that converge on a Masking Field cell. Consequently, such a cell fires vigorously only if it receives active inputs from all of its item chunk cells.
Due to self-similar growth, larger list chunks can selectively represent longer lists because they need more inputs, and thus more evidence, to fire. Once they fire, their stronger inhibitory interaction strengths than those of smaller list chunks can inhibit the smaller list chunks more than conversely (Figure 7). This asymmetric competition embodies the intuitive idea that, other things being equal, the longest lists are better predictors of subsequent events than are shorter lists, because a longer list embodies a more unique temporal context.
Asymmetric competition also enables learning of these longer list chunks to occur, because the stronger inhibition from list chunks of longer, but unfamiliar, lists (e.g. MYSELF) enables them to inhibit the chunks that represent shorter, but familiar, sublists (e.g. MY), more than conversely, so that tuning of the LTM traces within the adaptive filter that activates the longer list chunk can occur. Kazerounian and Grossberg (2014) have simulated how variable-length list chunks of a Masking Field can be learned as a list of item chunks is stored in working memory in real time.
Masking Fields have been used to explain cognitive data such as the Magical Number Seven (Miller, 1956), the word length effect in word superiority experiments (Samuel et al., 1982, 1983), phonemic restoration (e.g. Warren and Sherman, 1974), and various other percepts where, in response to a sequence of word inputs, future linguistic context can influence the word sequences that are consciously heard (e.g. Repp et al., 1978).
How does a Masking Field explain the classical Magical Number Seven, or immediate memory span (Section 3.5)? Because a Masking Field can chunk working memory sequences of variable length, it contains cells of variable size. The self-similarity of cell size and asymmetric competition together imply that a larger cell can inhibit a smaller cell more than conversely. This implies that only the largest, and most predictive chunks, will determine Masking Field outputs. Because of self-similarity, the same number of cells of any fixed size can be simultaneously stored. This size will be approximately seven if the transient memory span is chosen to be approximately four (Section 3.11).
How does a Masking Field explain the word length effect in word superiority studies? This effect shows that a letter is progressively better recognised when it is embedded in longer words of lengths from 1 to 4. The word length effect may also seem to follow from self-similarity, since larger chunks are more potent and predictive than smaller chunks. However, self-similarity implies that the list chunk of a familiar multi-letter word can inhibit the list chunk of a shorter word, including a familiar letter (Figure 7), which seems to contradict the property that the word can facilitate perception of its constituent letters, which is the main result of word superiority studies. This problem is resolved in ART systems that use item chunk and list chunk processing levels (Figure 7). In particular, although chunks that represent lists of multiple length compete within the Masking Field that categorises list chunks, the top-down expectations from the list chunk level to the item chunk level are excitatory. By self-similarity, list chunks that represent longer words generate larger inhibitory and excitatory signals (Figure 7). List chunks that represent longer lists will therefore send larger top-down excitatory priming signals to the item chunk level, thereby explaining both how the Magical Number Seven can arise due to asymmetric inhibition among list chunks and how a word length effect in word superiority can arise due to asymmetric top-down excitation from list chunks to the item chunks that activate them.
How future linguistic contexts can influence conscious percepts of previously occurring linguistic items is explained and simulated in articles such as Grossberg et al. (1997) and Grossberg and Kazerounian (2016). Although many neurobiological experiments have shown the importance of activity-dependent plasticity during brain development (e.g. Penn and Shatz, 1999), there seems to be much less experimental evidence that directly studies its effects on neocortical neuron size (e.g. Benders et al., 2015). It would be useful for additional experiments to characterise whether multiple-scale list chunking cells with self-similar properties exist in VLPFC, as predicted by the Masking Field model, particularly in the superficial layers of VLPFC, as suggested by the LIST PARSE model (Grossberg and Pearson, 2008). In particular, if a GABA agonist applied to VLPFC reduces the number of items that can be simultaneously stored there, it should also reduce the longest list that a list chunk can learn and use to predict outcomes from there.
3.24. Visual search: efficient versus inefficient, bottom-up versus top-down
In naturally occurring environments, animals need to search scenes in order to discover the contextual information that can support actions that lead to reward. How this may be achieved will now be discussed, first by reviewing relevant data and then by showing how the pART model explains them, including data about prefrontal mechanisms of feature-based attention and how they may regulate selective storage of items in VLPFC.
Visual attention and eye movements can explore scenes without any goals in mind. Just as often, however, visual searches seek out valued goal objects that are embedded in complex scenes. Common examples include finding a friend in a crowd or locating a menu board in a fast food restaurant. Neurophysiological data from monkeys that illustrate this distinction have been collected by simultaneously recording from multiple electrodes in the parietal and prefrontal cortices (Buschman and Miller, 2007). These experiments used simple stimulus materials to distinguish bottom-up versus top-down processes of attentional control whereby to search a scene. The distinction between a fast automatic bottom-up sweep of activation versus a slower controlled top-down flow of activation has been described in many publications (e.g. Desimone and Duncan, 1995; Grossberg et al., 1994; Hochstein and Ahissar, 2002; Sarter et al., 2001; Treisman and Gelade, 1980). In all conditions of the Buschman and Miller (2007) experiments, a target was randomly located in an array of four stimuli. In the bottom-up, or pop-out, condition, the distractors were identical and differed from the target along the dimensions of colour and orientation. In this case, the target’s salience automatically drew attention to it. In the top-down, or search, condition, each distractor differed independently from the target, and the target matched some of the distractors in each dimension. Memory of the target, rather than its salience, had to be used to find it.
In the pop-out condition, LIP neurons in the PPC were activated first, followed by neurons in the FEF and DLPFC (Figure 1). This kind of search thus proceeded in a primarily bottom-up way. In the search condition, the reverse order of activation was observed, and with a longer latency. Here, search proceeded top-down from prefrontal to lower cortical areas.
In the classical visual search literature, pop-out searches were often called efficient searches. These searches typically yielded zero reaction time (RT) slopes as a function of the number of distractors, hence the term ‘pop-out’ search. The Buschman and Miller (2007) search task illustrates an inefficient search, which in the classical search literature often used targets that are described by a conjunction of features. During an inefficient search, RT increased with the number of distractors (e.g. Treisman and Gelade, 1980). Albeit suggestive, the dichotomy of efficient versus inefficient search based on RT slopes was later shown to be inadequate (e.g. Thornton and Gilden, 2007; Townsend, 1972) because a continuum of flat to steep slopes can be obtained by varying saliency factors (Wolfe, 1998; Wolfe et al., 1989). In particular, search efficiency increases with decreased similarity of targets to distractors and increased similarity between distractors (Duncan and Humphreys, 1989). By proper choice of stimuli, a conjunction search can be rendered efficient, and a feature search can be rendered inefficient, all depending on the degree to which a target can be distinguished from distractors.
Before the eyes move to search a scene, its gist can be rapidly identified (e.g. coast, forest, mountain, countryside; Oliva and Torralba, 2001) if the scene contains enough familiar elements. Gist can be learned as a large-scale texture category in IT (cf. Figure 5) in a manner that is explained and simulated by the ARTSCENE model (Grossberg and Huang, 2009). When the gist texture is supplemented by one to three texture categories of smaller regions of a scene that are learned as the eyes scan the scene, these textures can vote to predict scenic type with up to 91.85% correct on a test set, a benchmark that outperformed alternative models in the literature at that time by 16.5%.
3.25. Object and spatial contexts and reinforcement influence predictive choices
The computation of gist begins the process whereby human observers deploy visual attention using attention shifts and eye movements in a global-to-local and coarse-to-fine manner (Navon, 1977; Schyns and Oliva, 1994), thereby accumulating evidence about a scene, including where a target lies within it (Gold and Shadlen, 2007; Grossberg and Pilly, 2008; Heekeren et al., 2008; Irwin, 1991; Jonides et al., 1982). As part of this evidence accumulation process, object and spatial contexts provide important information that enables working memories and list chunks to carry out more effective visual searches and to more successfully predict which options will be rewarded.
For example, when looking for a friend in a beach picture, our eyes typically fixate the sand at the bottom of the scene before the sky at its top. Such knowledge about the spatial layout of a scene is called spatial contextual cueing (e.g. Chun and Jiang, 1998). Spatial contextual information is not, however, always available in a novel environment. For example, when searching for a beverage in a friend’s refrigerator for the first time, we may not even know where the kitchen is located in the house until we glimpse related objects such as a stove and a sink. In this situation, we may have prior knowledge about which objects may be correlated in a scene like a kitchen, so continue to move towards the room where the stove and sink were glimpsed. However, we may not know the position where the refrigerator is located in this particular kitchen. This is an example of object contextual cueing (e.g. Chun and Jiang, 1999).
Many psychological experiments have described how humans use spatial and object contexts to efficiently search scenes. Such contextual cueing effects are typically measured using the RT for visually searching a familiar scene, subtracted from the RT for more slowly searching a novel scene, as described more fully below. These data have inspired the development of many visual search models (e.g. Backhaus et al., 2005; Brady and Chun, 2007; Grossberg et al., 1994; Itti and Koch, 2000; Torralba et al., 2006; Treisman and Gelade, 1980; Wolfe, 1994). Huang and Grossberg (2010) review how search materials are chosen and how different search models differ.
In general, these models typically try to explain where eye movements fixate to discover targets, and how fixated non-targets lead to the next eye movement. The ARTSCENE Search neural model (Huang and Grossberg, 2010) proposes, in addition, how an eye fixation on an object triggers learning about both its identity and its position, while also matching learned top-down expectations against the object and its position to determine whether it is a target or non-target. Sequences of eye movements also lead to storage of sequences of object and positional representations. These object and spatial contexts are associated through learning with currently fixated objects as the search continues. Associative strength is commensurate with the co-occurrence frequency of the contextual information and the target, the magnitude and frequency of reward of correct target acquisitions, and the attentional valence of both the search target/position and a context object/position. Attentional valence is defined as the degree to which an object attracts attention in response to both bottom-up and top-down factors. For example, a familiar moving object in a scene can attract bottom-up attention due to its activation of transient cells in the Where cortical stream, after which both top-down spatial attention from the PPC and top-down object attention from its learned category in the What cortical stream can further increase its salience, as modelled by Foley et al. (2012). By combining attentionally modulated object and spatial information, each eye movement also helps to accumulate learned contextual evidence about object and spatial contexts that can be used to determine where to look next to most efficiently find the target.
3.26. Perirhinal and parahippocampal cortices store object and spatial contexts
What brain regions carry out these processes? The sequences of scanned objects and their spatial positions are proposed to be stored in object and spatial working memories within the model VLPFC and DLPFC, respectively. Sequences of fixated objects and their spatial positions are also stored in the model PRC and PHC, respectively (Figure 1). Stored PRC and PHC sequences define object and spatial contexts that interact with the VLPFC and DLPFC working memories via bottom-up adaptive filters. The proposed role of PRC and related cortical areas in defining object contexts, and of PHC and related cortical areas in defining spatial contexts, is consistent with neuroimaging data about the dissociation of item and context information by these regions in humans (Aminoff et al., 2007; Diana et al., 2007; Libby et al., 2014).
Learning in the ARTSCENE Search model from a stored object or position in PRC or PHC, respectively, to a stored object or position in VLPFC or DLPFC, respectively, is modulated by a dopamine burst from the model BG (Figures 3(a) and 4(a)) when a target is foveated and reinforced. In this way, predictively successful associations between PRC and VLPFC, and between PHC and DLPFC, can amplify the stored working memory item chunks and list chunks that led to predictive success. The spatial attentional focus can be broadened or narrowed, as task constraints demand, to determine what objects or positions will influence the winning prediction.
Model interactions of IT, PRC, and VLPFC (Figure 1) also clarify neurophysiological data from monkeys that are recorded when they learn, using a delayed match-to-category paradigm, to categorise morphs of image exemplars into two categories; for example, cats versus dogs (Freedman et al., 2001, 2003; Roy et al., 2010). Supervised learning is needed because exemplars that are close to the category boundary, but on opposite sides of it, could be visually more similar than stimuli that belonged to the same category, for example, a cheetah and a housecat. Contextual cueing using sequences of image exemplars, and of a category-predicting discriminative stimulus in cases where it is also presented, help to explain these data. In particular, it was found that IT seems to have properties consistent with ART mechanisms of ITp-ITa category learning (Figure 1), notably attention to critical features of each exemplar (Figure 5), whereas VLPFC seems to have properties consistent with ITa-PRC-VLPFC contextually cued learning (Figure 1), notably sustained activity during the delay period before reward, and greater match/mismatch effects. These studies did not, however, record from PRC, or interactions between PRC and VLPFC. Such additional measures are much to be desired.
3.27. Parietal-prefrontal resonance controls choice of reactive versus planned targets
These last experiments study category learning when stimuli are foveally presented from the start. During daily life, in contrast, an animal typically needs to move its eyes, head, and/or body to foveate important objects. Here, a major design problem needs to be solved: how do planned movements to an extrafoveal object compete with reactive movements to a different extrafoveal object? Rapid reactive movements in response to bottom-up sensory demands are often needed to ensure survival in response to unexpected dangers. Planned movements, that may require PFC executive control and top-down attention, often take longer to select and release. How does the brain prevent reactive movements from being triggered prematurely in situations where a more slowly occurring planned movement would be more adaptive?
A movement gate that is controlled by the SNr is typically tonically active until a movement command inhibits it. Then, the cells that control the corresponding movement can be activated (Figure 15(a)). When a visual cue occurs, the fastest response would be an orienting response to look at it. For this to happen, the cue needs to open the appropriate SNr gate to launch a reactive movement to its position (Figure 15(b)). If the cue is a discriminative cue to do a planned action as quickly as possible, then, as noted above, it may take longer to fully process the features of the discriminative cue to determine the adaptive conditional response. How does the brain know that a plan is being elaborated, even before it is chosen, so that the reactive gate can be kept shut, yet also allow a reactive movement command to open its gate as rapidly as possible when no planned movement command is being formed?
The TELOS model (Brown et al., 2004) explains and simulates how the brain may achieve this balance between reactive and planned movements by predicting how the distribution of excitation and inhibition that converges on the BG when a plan is being elaborated can keep the reactive gate closed (Figure 15(c)). When a movement plan is finally chosen, there is agreement between cells in the FEF and the PPC representation of target position. This agreement is expressed by a synchronous FEF–PPC resonance (Figures 1 and 15(d)) that changes the excitatory–inhibitory balance as it inhibits outlier PPC positions. This resonance is predicted by TELOS to signal attentive consistency between a finally selected movement plan and the target position of the conditional movement. Then, the new balance of excitation and inhibition enables the appropriate FEF-commanded movement gate to open and release the context-appropriate action (Figure 15(d)). As part of their study of bottom-up and top-down attention, Buschman and Miller (2007) reported such prefrontal–parietal resonances, including between FEF and LIP, during movement control, and Pasupathy and Miller (2004) reported that the different time courses of activity in the PFC and BG are consistent with how BG-mediated gating of prefrontal cortical commands are learned in the TELOS model.
3.28. RTs in behavioural data and simulations about object and spatial searches
Computer simulations of contextual cueing data provide considerable support for the ARTSCENE Search model’s proposal of how PRC and PHC help to guide sequential searches. In particular, ARTSCENE Search quantitatively simulates RT data about positive/negative, spatial/object, and local/distant contextual cueing effects during visual search. Figure 16 summarises six of the many experimental conditions about contextual cueing that ARTSCENE Search has successfully simulated. Each panel in the figure depicts RT data (left) and a computer simulation of it (right). These various conditions lead to expectations about what to measure if the neurophysiological methods of Buschman and Miller (2007) are added to traditional psychophysical experiments on contextually cued searches.
Figure 16(a) summarises RT data and a simulation of positive spatial cueing. Positive spatial cueing effects are the RT reductions for search in a familiar spatial context compared to a new context. In this paradigm (Chun and Jiang, 1998), a fixed target position was chosen from a grid search display without replacement and presented in one trial per block. Across blocks of search trials, a target position was accompanied by either a repeated spatial configuration of distractors (Old condition) throughout the entire experiment or a random configuration that was newly generated in each block (New condition). In Figure 16(a) (left panel), the x-axis represents search epochs grouped from blocks of trials, and the y-axis represents search RT for completing a trial. Since the upper and lower curve in each panel of this figure corresponds, respectively, to the New and Old spatial context condition, the separation between these two curves indicates the amount of contextual facilitation in search RT that derives from a regular spatial context. Notice that the RT in the New spatial context condition dropped across epochs, and a further RT reduction in the Old spatial condition also developed as the session progressed.
ARTSCENE Search replicates spatial cueing effects by learning pairwise associations between a context position in PHC and a target position in DLPFC. Specifically, when a search display is presented, it activates model PPC and, from there, both DLPFC and PHC as the eyes search a scene. Each context position that is stored in PHC sends biasing signals to all the target positions with which it has been associated in the past that are currently stored in DLPFC. When the contextual information combines with the intrinsic Masking Field dynamics of DLPFC, the chosen DLPFC list chunks encode the likelihood of seeing a target at each position. Top-down feedback from these list chunks to the FEF (Figure 1) then biases attention and eye movements toward likely target positions given the current scene layout. As a consequence, an eye-scan path becomes more target-based rather than saliency-based. Accordingly, the probability of fixations on salient distractors is reduced, reflected as spatial cueing effects.
Why does the New curve also decrease with increased training? This happens because the strongest pairwise associations learned by the model are typically from a target position to itself due to its perfect self-correlation. Unlike positions where a target never occurs, a target position itself, once it re-appears in a search trial, signifies target presence and strongly attracts overt attention. Therefore, search RT can still decrease during the course of training even if a target position is presented in combination with a new context.
The five other panels in Figure 16 depict other spatial contextual cueing (Figure 16(b) and 16(c)) and object contextual cueing (Figure 16(d)–(f)) paradigms that the model successfully simulates. The object contextual cueing results depend upon model interactions between the model’s IT, VLPFC, and PRC regions (Figure 1). These and other successful simulations of the model strongly support its proposed mechanisms. It should, however, be noted that the simulations compute the decreasing number of eye fixations in the New and Old conditions, rather than absolute RTs. By assigning an RT to each fixation duration, a fit to RTs can also be achieved. However, that would still leave open the question about whether each of these fixation steps takes that amount of time in associated brain regions. Such a mixed experimental and modelling study remains to be done.
3.29. Where-to-What and What-to-Where interactions learn and search for objects
The ability to carry out a search for a desired object requires a solution to a major design problem that sufficiently advanced brains have solved. The What cortical stream learns recognition categories that tend to be increasingly independent of object view, size, and position at higher cortical levels, with ITa cells, among others in the temporal cortex, exhibiting such invariance (Bar et al., 2001; Sigala and Logothetis, 2002; Tanaka et al., 1991). The cortical magnification factor does, however, limit the degree of positional invariance, as reflected by the neurophysiologically recorded trade-off between object selectivity and position tolerance in ITa cells (Zoccolan et al., 2007). Grossberg et al. (2011) have used ART to explain and simulate this trade-off.
Invariant recognition categories avoid a combinatorial explosion in the number of categories that are needed to represent an object. Instead of having to learn a different category for each object view, position, and size, the brain just uses a small population of cells in ITa to code an invariant object category that is significantly invariant under changes in view, size, and position (e.g. Hung et al., 2005). This invariant category can then easily interact with other brain processes, such as reinforcement learning and working memory storage. In particular, an invariant object category in ITa may be attended with higher probability if it receives motivated attention via an ITa-AMYG-OFC cognitive-emotional resonance (Figures 1 and 2).
In becoming positionally invariant, however, ITa recognition categories lose information about the positions in space of the objects that they represent. The Where stream represents target positions and controls actions aimed at acquiring them, but does not represent featural properties of the objects themselves. These What and Where stream properties are computationally complementary (Grossberg, 2000a, 2013a). Interactions between the What stream and the Where stream overcome these complementary computational deficiencies. Using What-to-Where interactions, invariant object categories in the What stream can use Where stream spatial representations to control actions towards desired goals in space. Section 3.31 explains how this is proposed to occur.
In addition to What-to-Where interactions that help to search for an object, the ARTSCAN Search model has simulated how Where-to-What interactions enable invariant object categories to be learned as the eyes freely scan a scene. Along the way, this model has successfully explained and predicted many data about interactions between multiple brain regions, including V1, V2, V3A, V4, ITp, ITa, PPC, LIP, PFC, FEF, and SC (Cao et al., 2011; Chang et al., 2014; Fazl et al., 2009; Foley et al., 2012).
In particular, the model explains how, as the eyes freely scan a scene, learning occurs from prestriate visual cortex (e.g. V4) to posterior inferotemporal cortex (ITp) of categories that combine both featural and positional information, and how multiple ITp categories learn, in turn, to be associated with an emerging invariant ITa category (Figure 1). Category learning from ITp-to-ITa uses ART bottom-up adaptive filters. This learning is dynamically stabilised by learning top-down expectations from ITa-to-ITp (Figures 1 and 5).
For invariant category learning to work, the brain needs to solve a basic View-to-Object Binding Problem: as the eyes scan a scene, two successive eye movements may focus on different parts of the same object or on different objects. How does the brain avoid erroneously classifying views of different objects together into an invariant object category, even before the brain knows what the object is? ARTSCAN Search proposes how spatial attention in the Where stream and object attention in the What stream interact to solve this problem.
The category learning process begins when a view- and position-specific category of a novel object is learned in ITp and activates cells in ITa. These ITa cells will learn to encode an invariant object category as multiple specific ITp categories are associated with it as the eyes explore the object surface. When the object view changes enough, the previously active ITp category gets reset to enable a new one to be activated and learned. The emerging invariant category in ITa cannot, however, get reset because it needs to remain active while it is associated with multiple ITp categories of that object.
Why is the invariant category not reset? An attentional shroud in PPC is predicted to inhibit reset of an invariant object category while the eyes scan different views and positions of the object (Figure 17). An attentional shroud (Tyler and Kontsevich, 1995), or form-fitting distribution of spatial attention, can form pre-attentively even before the brain learns to recognise the surface as representing a particular object. Such a shroud is part of a surface-shroud resonance (Figure 17) that arises due to positive feedback interactions between a surface representation (e.g. in cortical area V4) and spatial attention (e.g. in PPC). A surface-shroud resonance maintains sustained spatial attention upon the object to be learned and triggers the process whereby the attended surface qualia becomes consciously visible (Grossberg, 2017b). While the shroud is active, it inhibits the parietal Category Reset mechanism that would otherwise inhibit ITa (Figure 17) while helping to select eye movement target positions whereby to foveate salient features on the object surface. The process whereby sequential eye movement targets are chosen occurs from V2-to-V3A-to-LIP and FEF (Figure 17). When spatial attention shifts to focus on another object, the reset mechanism is transiently disinhibited, and its burst of activity inhibits the active ITa category (Figure 17), so that a new object can be attended and its invariant category learned. See Grossberg (2013a, 2017b) for extensive discussions of this process and the psychological and neurobiological data that it explains and predicts.
3.30. What working memory filtering and activation of Where target positions
After Where-to-What stream interactions help to learn invariant object categories, What-to-Where stream interactions regulate how to foveate valued target objects in a scene. Both ARTSCAN Search and ARTSCENE Search proposed a minimal anatomy (Section 1.3) that could carry out this function, while also simulating challenging RT data about visual search, for example, Figure 16.
Such an anatomy proposes how an invariant object representation in the What stream can activate a positional representation in the Where stream that can be used to foveate a valued target object in a scene. However, it did not try to solve the problem of how the brain can selectively filter desired targets from a stream that also contains distractors, so that it only attends, stores, and foveates matched targets. The minimal anatomy is summarised here for multiple reasons. First, it was sufficient to quantitatively simulate challenging RT data in many contextual cueing experiments. Second, it may have evolved before the prefrontal mechanisms of selective working memory storage did, may operate in parallel with them, and may be unmasked if the prefrontal mechanisms are lesioned. This last possibility may be worth testing directly.
In the minimal anatomy of ARTSCENE Search, winning VLPFC activities send a top-down attentional prime to ITa. An open SNr gate lets the primed ITa cells fire. These ITa cells can then prime the positionally sensitive categories in ITp with which they were associated when ITa was being learned using resonant bottom-up and top-down interactions (Figure 5). If one of the primed ITp categories also receives a bottom-up input from an object at its position, then it can fire and activate positional representations in LIP and FEF. These positional representations can move the eyes to the position in space that they represent.
3.31. Feature-based attention, saccadic choice, and selective working memory storage
More is needed for VLPFC to also be able to selectively filter desired targets from a stream that also contains distractors, and to enable it to selectively attend, store, and foveate matched targets. Neurophysiological data about the role of VPA as ‘a source for feature-based attention’ (Bichot et al., 2015: 832) may be understood in the light of these additional functional properties, notably how and why cells in VPA selectively match desired combinations of object features, resonate with a target that matches these features, and then rapidly activate a positional representation in FEF that can command a saccade to this target. These properties were recorded in experiments where fixating monkeys were presented with a central cue object that defined the search target, which was followed by a delay during which the monkeys held a representation of the target in memory. Then, an array of eight stimuli appeared that contained the search target and seven distractors. The monkeys found the target using free gaze and were rewarded for maintaining fixation on it for 800 ms.
Bichot et al. (2015) conducted experiments in which they simultaneously recorded from IT, VPA, and FEF in two monkeys, and VPS, VPA, and FEF in two other monkeys (Figure 1). The following summary proposes a mechanistic and functional explanation of how these cells interact together, notably how they enable matched and mismatched objects to be selectively processed in PFC (Figure 1):
Both ITp (TEO) and ITa (TE) project to PFC (Barbas and Pandya, 1989; Tanaka, 1996; Webster et al., 1994).
ITp topographically projects to VPA, whose cells exhibit significant sensitivity to extrafoveal positions (Bichot et al., 2015), as do those in ITp (Tanaka, 1996).
ITa topographically projects to PRC, as in the ARTSCENE Search model, and also to VPS, which in turn projects to VLPFC. As noted in Bichot et al. (2015), VPS had the largest spatial tuning curve of any of the cells that they recorded, consistent with invariance properties of ITa.
VLPFC outputs top-down signals to both VPS and VPA, where they learn modulatory top-down expectations when they are associated with the currently active VPS and VPA cells. These expectations are assumed to obey the ART Matching Rule (Figure 5(b)).
The activity of cells in VPA that are receiving an active VLPFC-to-VPA prime are enhanced when a currently presented extrafoveal object matches target features in their receptive field, and is suppressed when such an object mismatches expected target features, consistent with the ART Matching Rule.
The enhanced VPA activity is sufficient to trigger an output signal to FEF at the corresponding positional representation in FEF (Figure 1), consistent with data of Bichot et al. (2015) showing VPA activating around 20 ms before FEF. FEF can then elicit a saccade to the matched target, leading to it being foveated. By inhibiting inputs from objects that mismatch the VPA expectation, mismatched objects are not foveated.
The activity of cells in VPS that are receiving an active VLPFC-to-VPS prime are enhanced when an invariant object category from ITa matches their receptive field, and are suppressed when such an object mismatches it, consistent with the ART Matching Rule. When a match occurs, a synchronous resonance develops that enables the category to be stored in VLPFC (Figure 1). This resonance propagates through multiple cortical areas, in the manner described in Section 3.32, and supports conscious recognition of the object.
The mapping between VPA and FEF positions is assumed to have been learned in response to series of objects that have, in the past, activated the What and Where streams in parallel (Figure 1). The kind of learning that can associate corresponding VPA and FEF positions has previously been simulated in the FACADE (Form-And-Color-And-DEpth) model of 3D vision and figure-ground perception, where it was used to associate corresponding positions in the boundary and surface representations within the interblob and blob cortical streams, respectively, through V1, V2, and V4 (Figure 1) as they were activated in parallel by a series of objects (Grossberg et al., 2002). Once these cortical streams were associatively linked, output signals from oriented boundaries to the colours with which they were associated in the surface stream provided an explanation of many properties of the McCollough effect, a striking long-term, oriented, chromatic aftereffect, among other percepts (McCollough, 1965).
The pART macrocircuit in Figure 1 contains two cortical areas, VPA and VPS, that have not previously been simulated. However, interactions and functional roles for all the other brain regions in Figure 1, and some that are not included in this Figure (cf. Figures 3–5, 12, 15, and 17), have been extensively simulated in earlier models, and used to explain large psychological and neurobiological data bases. These models provide a secure foundation for including VPA and VPS in a unified model of prefrontal dynamics. Models of cognitive-emotional dynamics, such as CogEM, MOTIVATOR, and nSTART, have already been mentioned above, as have What stream invariant category learning and search models such as ART, ARTSCAN Search, and ARTSCENE Search. In addition to these models, all the Where stream brain regions in Figure 1 have been extensively modelled, including a model of motion processing, form-to-motion interactions, and directional attention (3D FORMOTION model); motion-based decision-making (MODE model); visual navigation using heading, obstacle avoidance, and route selection (ViSTARS model); and target tracking by combinations of saccadic and smooth pursuit eye movements (SAC-SPEM model) that together simulate properties of Where stream cortical areas V1, middle temporal cortex (MT), medial superior temporal area (MST), PPC/LIP, FEF, and PFC, and areas with which they interact (Baloch and Grossberg, 1997; Berzhanskaya et al., 2007; Browning et al., 2009a, 2009b; Elder et al., 2009; Grossberg et al., 1999, 2001, 2012; Grossberg and Pilly, 2008; Pack et al., 2001; Srihasam et al., 2009).
3.32. Synchronisation of multiple cortical regions for feature-based attention
The proposed role of VPA for ‘feature-based attention’ should not be conflated with the ‘feature-based attention’ that supports conscious seeing and knowing about a familiar object. In this regard, ART and ARTSCAN predicted – and thereby explained a lot of data – about how percepts of visual qualia may become conscious due to surface-shroud resonances that are triggered between V4 and PPC (Figures 1 and 17), before propagating both bottom-up and top-down to other cortical areas; how familiar objects may be recognised due to a feature-category resonance that is triggered between V4 and IT (Figures 1 and 17), before propagating both bottom-up and top-down to other cortical areas; and how an observer may consciously see and know about a familiar object when these two types of resonances synchronise (Grossberg, 2013a, 2017b).
VPA processing carries out a type of top-down ‘feature-based attention’ in a strict mechanistic sense because its circuit seems to embody the ART Matching Rule (Figure 5(b)), as do multiple stages of feature-based attention (Grossberg, 2013a, 2017b). Multiple cortical stages that compute ‘feature-based attention’ can synchronise during a match state, as illustrated by MEG and fMRI data of Baldauf and Desimone (2014) in humans. See also Buschman and Miller (2007), Engel et al. (2001), Gregoriou et al. (2009), and Pollen (1999).
The LAMINART (Laminar Adaptive Resonance Theory) model (Grossberg, 1999; Grossberg and Raizada, 2000; Raizada and Grossberg, 2001, 2003) clarifies how multiple cortical stages can synchronise (Figure 18). This model proposes how all granular neocortical areas may combine bottom-up, horizontal, and top-down interactions that embody variations of the same canonical laminar cortical circuitry. Because of this shared circuitry across cortical areas, the ART Matching Rule circuit in Figure 5(b) may be realised using a similar circuit design at multiple stages of cortical processing. For example, in Figure 18, the top-down attention pathway from layer 6 in V2 projects to layer 6 in V1, which sends bottom-up signals to layer 4. These bottom-up signals are sent via a modulatory on-centre (note the balanced excitatory and inhibitory pathways to layer 4) surrounded by a driving off-surround network. The top-down signals from V2 are hereby ‘folded’ at layer 6 in V1 in order to reach layer 4. This property shows that object attention is carried by a top-down, modulatory on-centre, off-surround network whose off-surround is activated by a folded feedback network.
Figure 18 illustrates how a top-down, task-selective priming signal from PFC can propagate through multiple lower cortical areas via their layer 6, which can then activate their layer 6-to-4 modulatory on-centre, off-surround networks via folded feedback. In this way, an entire cortical hierarchy, including multiple stages of ‘feature-based attention’, may get primed to process incoming bottom-up signals to accommodate the bias imposed by the prime. When a matched bottom-up target is received by this cortical hierarchy, multiple processing stages can rapidly go into gamma synchrony, as simulated by Grossberg and Versace (2008), and can support recognition of the target.
In summary, as Figure 1 illustrates, section 3 has summarised how selective, context-sensitive storage of event sequences in cognitive and motor working memories occurs, and how learning and performance of the most predictive cognitive plans, or list chunks, builds upon the event sequences that are stored in working memory. Selectivity is achieved both by feature-based attention and SNr BG gating. Attentive matching can trigger orienting movements to foveate a target object and can cause a feature-category resonance to propagate through multiple cortical levels and to support conscious event recognition and working memory storage. Opening an SNr gate enables storage in working memory to occur, and closing the gate inhibits storage and prepares the working memory to store future sequences without bias. Choice of list chunks that can best predict valued outcomes is supported by a combination of Masking Field competitive dynamics (Figure 7), PRC and PHC contextual associations (Figure 1), and reinforcement learning mechanisms that operate via both the AMYG and OFC (Figures 1–3) and via SNc Now Print signals in response to unexpected rewards and non-rewards (Figures 3(a), 4(a), and 12).
4. Concluding remarks
This article describes an emerging neural architecture, called pART (Figure 1), that embodies explanations and predictions about the neural mechanisms and functional properties of several prefrontal cortical areas, including OFC, VLPFC, DLPFC, VPS, VPA, and FEF as they carry out cognitive-emotional and cognitive working memory dynamics. The explanations describe how these prefrontal areas may realise their functional properties as emergent properties due to interactions among themselves and with other brain regions, including the amygdala, BG, cerebellum, V1, V2, V3A, V4, ITp, ITa, MT, MST, LIP, PPC, and SC. These functional properties include the computation of desirability by OFC and availability by VLPFC (Rudebeck et al., 2017), a solution of the credit assignment problem by DLPFC (Asaad et al., 2017), and how feature-based attention by VPS and VPA may filter expected versus unexpected objects and thereby both help to direct saccadic eye movements to expected objects (Baldauf and Desimone, 2014; Bichot et al., 2015) and select items for storage in the VLPFC cognitive working memory. Cognitive-emotional interactions, including reinforcement learning and incentive motivational learning; object and spatial working memory dynamics; and category learning, including the learning of object categories, value categories, object-value categories, and sequence categories, or list chunks, are among the processes that are carried out by this architecture. Several functionally distinct types of attention (prototype, surface, and motivated attention) help to dynamically stabilise these learning processes as well as to predictively prime their target representations. Prototype attention focuses upon the critical feature patterns that are attended during the feature-category resonances that support object recognition (Figure 5); surface attention focuses on an object surface during the surface-shroud resonances that support conscious seeing of its visual qualia (Figure 17); and motivated attention focuses on valued objects during conscious cognitive-emotional resonances and supports conscious feelings about them (Figures 2 and 3; Grossberg, 2013a, 2017b).
Many of the pART explanations in this article can be viewed as confirmed predictions, because the model mechanisms that are used to explain the data anticipated the experiments that reported them. The article also suggests multiple additional predictions whereby to further test these explanations. Some of these predictions are stated explicitly in the text, but every model interaction and emergent property implicitly embodies additional predictions. In particular, the article’s unified explanation of cognitive-emotional and cognitive working memory dynamics enables each laboratory to imagine new kinds of experimental tests that would be amenable to its own particular laboratory capabilities.
It should also be emphasised that, once models sufficiently explicate the actual design principles, mechanisms, circuits, and architectures that are used by our brains, the models can lead to explanations and predictions in totally different areas of psychology and neuroscience than those that were their original explanatory targets. That is how, for example, explanations have been discovered showing how particular breakdowns in normal brain functioning may give rise to behavioural symptoms of mental disorders that afflict millions of individuals, including Alzheimer’s disease (Grossberg, 2017a), autism and Fragile X syndrome (Grossberg, 2013a; Grossberg and Kishnan, 2018; Grossberg and Seidman, 2006), schizophrenia (Grossberg, 2000b), and medial temporal amnesia (Carpenter and Grossberg, 1993; Grossberg, 2013a). Here too, multiple predictions are made for which additional experimental tests would be of great value.
Footnotes
Declaration of conflicting interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding: The author(s) received no financial support for the research, authorship, and/or publication of this article.
References
- Agam Y, Bullock D, Sekuler R. (2005) Imitating unfamiliar sequences of connected linear motions. Journal of Neurophysiology 94(4): 2832–2843. [DOI] [PubMed] [Google Scholar]
- Agam Y, Galperin H, Gold BJ, et al. (2007) Learning to imitate novel motion sequences. Journal of Vision 7(5): 1.1–1.17. [DOI] [PubMed] [Google Scholar]
- Aggleton JP. (1993) The contribution of the amygdala to normal and abnormal emotional states. Trends in Neurosciences 16(8): 328–333. [DOI] [PubMed] [Google Scholar]
- Aggleton JP, Passingham RE. (1981) Syndrome produced by lesions of the amygdala in monkeys (Macaca mulatta). Journal of Comparative and Physiological Psychology 95(6): 961–977. [DOI] [PubMed] [Google Scholar]
- Aggleton JP, Saunders RC. (2000) The amygdale – What’s happened in the last decade? In: Aggleton JP. (ed.) The Amygdala (2nd edn). New York: Oxford University Press, pp. 1–30. [Google Scholar]
- Akrami A, Liu Y, Treves A, et al. (2009) Converging neuronal activity in inferior temporal cortex during the classification of morphed stimuli. Cerebral Cortex 19(4): 760–776. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alexander GE, Crutcher MD. (1990) Functional architecture of basal ganglia circuits: Neural substrates of parallel processing. Trends in Neurosciences 13(7): 266–271. [DOI] [PubMed] [Google Scholar]
- Alexander GE, DeLong M, Strick PL. (1986) Parallel organization of functionally segregated circuits linking basal ganglia and cortex. Annual Review of Neuroscience 9: 357–381. [DOI] [PubMed] [Google Scholar]
- Amaral DG, Price JL. (1984) Amygdalo-cortical projections in the monkey (Macaca fascicularis). Journal of Comparative Neurology 230(4): 465–496. [DOI] [PubMed] [Google Scholar]
- Aminoff E, Gronau N, Bar M. (2007) The parahippocampal cortex mediates spatial and nonspatial associations. Cerebral Cortex 17(7): 1493–1503. [DOI] [PubMed] [Google Scholar]
- Amsel A. (1962) Frustrative nonreward in partial reinforcement and discriminative learning: Some recent history and a theoretical extension. Psychological Review 69(4): 306–328. [DOI] [PubMed] [Google Scholar]
- Amsel A. (1992) Frustration Theory: An Analysis of Dispositional Learning and Memory. Cambridge: Cambridge University Press. [Google Scholar]
- Arana FS, Parkinson JA, Hinton E, et al. (2003) Dissociable contributions of the human amygdala and orbitofrontal cortex to incentive motivation and goal selection. Journal of Neuroscience 23(29): 9632–9638. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Assad WF, Lauro PM, Perge JA, et al. (2017) Prefrontal neurons encode a solution to the credit-assignment problem. Journal of Neuroscience 37(29): 6995–7007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Auer ET, Jr, Luce PA. (2008) Probabilistic phonotactics in spoken word recognition. In: Pisoni DB, Remez RE. (eds) The Handbook of Speech Perception. Wiley Online Library; Available at: 10.1002/9780470757024.ch25 [DOI] [Google Scholar]
- Averbeck BB, Chafee MV, Crowe DA, et al. (2002) Parallel processing of serial movements in prefrontal cortex. Proceedings of the National Academy of Sciences of the United States of America 99(20): 13172–13177. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Averbeck BB, Crowe DA, Chafee MV, et al. (2003. a) Neural activity in prefrontal cortex during copying geometrical shapes. I. Single cells encode shape, sequence, and metric parameters. Experimental Brain Research 150(2): 127–141. [DOI] [PubMed] [Google Scholar]
- Averbeck BB, Crowe DA, Chafee MV, et al. (2003. b) Neural activity in prefrontal cortex during copying geometrical shapes. II. Decoding shape segments from neural ensembles. Experimental Brain Research 150(2): 142–153. [DOI] [PubMed] [Google Scholar]
- Awh E, Vogel EK. (2008) The bouncer in the brain. Nature Neuroscience 11(1): 5–6. [DOI] [PubMed] [Google Scholar]
- Awh E, Jonides J, Smith EE, et al. (1996) Dissociation of storage and rehearsal in verbal working memory. Psychological Science 7(1): 25–31. [Google Scholar]
- Backhaus A, Heinke D, Humphreys GW. (2005) Contextual learning in the selective attention for identification model (CL-SAIM): Modeling contextual cueing in visual search tasks. In: Proceedings of the 2005 IEEE Computer Society conference on computer vision and pattern recognition-workshops, San Diego, CA, 21–23 September, p. 87 New York: IEEE. [Google Scholar]
- Baddeley A. (1986) Working Memory. Oxford: Clarendon Press. [Google Scholar]
- Baddeley A. (1996) Exploring the central executive. The Quarterly Journal of Experimental Psychology 49(1): 5–28. [Google Scholar]
- Baddeley A. (2012) Working memory: Theories, models, and controversies. Annual Review of Psychology 63: 1–29. [DOI] [PubMed] [Google Scholar]
- Baddeley A, Hitch G. (1974) Working memory. Psychology of Learning and Motivation 8: 47–89. [Google Scholar]
- Baker JK, Fenning RM, Erath SA, et al. (2017) Sympathetic under-arousal and externalizing behavior problems in children with autism spectrum disorder. Journal of Abnormal Child Psychology. Epub ahead of print 24 July DOI: 10.1007/s10802-017-0332-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Baldauf D, Desimone R. (2014) Neural mechanisms of object-based attention. Science 344(6182): 424–427. [DOI] [PubMed] [Google Scholar]
- Baloch AA, Grossberg S. (1997) A neural model of high-level motion processing: Line motion and formotion dynamics. Vision Research 37(21): 3037–3059. [DOI] [PubMed] [Google Scholar]
- Ban TA. (1973) Recent Advances in the Biology of Schizophrenia. Springfield, IL: Charles C Thomas Publisher Ltd. [Google Scholar]
- Banquet JP, Grossberg S. (1987) Probing cognitive processes through the structure of event-related potentials during learning: An experimental and theoretical analysis. Applied Optics 26: 4931–4946. [DOI] [PubMed] [Google Scholar]
- Bar M, Tootell RBH, Schacter DL, et al. (2001) Cortical mechanisms specific to explicit visual object recognition. Neuron 29(2): 529–535. [DOI] [PubMed] [Google Scholar]
- Barbas H. (1993) Organization of cortical afferent input to orbitofrontal areas in the rhesus monkey. Neuroscience 56(4): 841–864. [DOI] [PubMed] [Google Scholar]
- Barbas H. (1995) Anatomic basis of cognitive-emotional interactions in the primate prefrontal cortex. Neuroscience and Biobehavioral Reviews 19(3): 499–510. [DOI] [PubMed] [Google Scholar]
- Barbas H. (2000) Connections underlying the synthesis of cognition, memory, and emotion in primate prefrontal cortices. Brain Research Bulletin 52(5): 319–330. [DOI] [PubMed] [Google Scholar]
- Barbas H. (2007) Flow of information for emotions through temporal and orbitofrontal pathways. Journal of Anatomy 211(2): 237–249. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Barbas H, Pandya DN. (1989) Architecture and intrinsic connections of the prefrontal cortex in the rhesus monkey. Journal of Comparative Neurology 286(3): 353–375. [DOI] [PubMed] [Google Scholar]
- Barbas H, Ghashghaei H, Dombrowski SM, et al. (1999) Medial prefrontal cortices are unified by common connections with superior temporal cortices and distinguished by input from memory-related areas in the rhesus monkey. Journal of Comparative Neurology 410(3): 343–367. [DOI] [PubMed] [Google Scholar]
- Barone P, Joseph JP. (1989) Prefrontal cortex and spatial sequencing in macaque monkey. Experimental Brain Research 78(3): 447–464. [DOI] [PubMed] [Google Scholar]
- Baxter MG, Parker A, Lindner CC, et al. (2000) Control of response selection by reinforcer value requires interaction of amygdala and orbital prefrontal cortex. Journal of Neuroscience 20(11): 4311–4319. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Behara A, Damasio AR. (1999) The somatic marker hypothesis: A neural theory of economic decision. Games and Economic Behavior 52: 336–372. [Google Scholar]
- Bechera A, Damasio H, Damasio AR, et al. (1999) Different contributions of the human amygdala and ventromedial prefrontal cortex to decision-making. Journal of Neuroscience 19(13): 5473–5481. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Belova MA, Paton JJ, Salzman D. (2008) Moment-to-moment tracking of state value in the amygdala. Journal of Neuroscience 28(40): 10023–10030. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Belova MA, Paton JJ, Morrison SE, et al. (2007) Expectation modulates neural responses to pleasant and aversive stimuli in primate amygdala. Neuron 55(6): 970–984. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Benders MJ, Palmu K, Menache C, et al. (2015) Early brain activity relates to subsequent brain growth in premature infants. Cerebral Cortex 25(9): 3014–3024. [DOI] [PubMed] [Google Scholar]
- Berridge KC, Robinson TE, Aldridge JW. (2009) Dissecting components of reward: ‘liking’, ‘wanting’, and learning. Current Opinion in Pharmacology 9(1): 65–73. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Berzhanskaya J, Grossberg S, Mingolla E. (2007) Laminar cortical dynamics of visual form and motion interactions during coherent object motion perception. Spatial Vision 20(4): 337–395. [DOI] [PubMed] [Google Scholar]
- Bichot NP, Heard MT, DeGennaro EM, et al. (2015) A source for feature- based attention in the prefrontal cortex. Neuron 88(4): 832–844. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Boardman I, Bullock D. (1991) A neural network model of serial order recall from short- term memory. In: Proceedings of the international joint conference on neural networks, Seattle, WA, 8–12 July, pp. 879–884. New York: IEEE. [Google Scholar]
- Bohland J, Bullock D, Guenther F. (2010) Neural representations and mechanisms for the performance of simple speech sequences. Journal of Cognitive Neuroscience 22(7): 1504–1529. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bower GH. (1981) Mood and memory. American Psychologist 36(2): 129–148. [DOI] [PubMed] [Google Scholar]
- Bradski G, Carpenter GA, Grossberg S. (1992) Working memory networks for learning temporal order with application to 3-D visual object recognition. Neural Computation 4: 270–286. [Google Scholar]
- Bradski G, Carpenter GA, Grossberg S. (1994) Store working memory networks for storage and recall of arbitrary temporal sequences. Biological Cybernetics 71(6): 468–480. [Google Scholar]
- Brady TF, Chun MM. (2007) Spatial constraints on learning in visual search: Modeling contextual cueing. Journal of Experimental Psychology: Human Perception & Performance 33(4): 798–815. [DOI] [PubMed] [Google Scholar]
- Brown J, Bullock D, Grossberg S. (1999) How the basal ganglia use parallel excitatory and inhibitory learning pathways to selectively respond to unexpected rewarding cues. Journal of Neuroscience 19(23): 10502–10511. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Brown JW, Bullock D, Grossberg S. (2004) How laminar frontal cortex and basal ganglia circuits interact to control planned and reactive saccades. Neural Networks 17(4): 471–510. [DOI] [PubMed] [Google Scholar]
- Browning A, Grossberg S, Mingolla M. (2009. a) Cortical dynamics of navigation and steering in natural scenes: Motion-based object segmentation, heading, and obstacle avoidance. Neural Networks 22(10): 1383–1398. [DOI] [PubMed] [Google Scholar]
- Browning A, Grossberg S, Mingolla M. (2009. b) A neural model of how the brain computes heading from optic flow in realistic scenes. Cognitive Psychology 59(4): 320–356. [DOI] [PubMed] [Google Scholar]
- Bujnakova I, Ondrejka I, Mestanik M, et al. (2016) Autism spectrum disorder is associated with autonomic underarousal. Physiological Research 65(Suppl. 5): S673–S682. [DOI] [PubMed] [Google Scholar]
- Bullock D, Grossberg S. (1988) Neural dynamics of planned arm movements: Emergent invariants and speed-accuracy properties during trajectory formation. Psychological Review 95(1): 49–90. [DOI] [PubMed] [Google Scholar]
- Bullock D, Rhodes B. (2003) Competitive queuing for planning and serial performance. In: Arbib M. (ed.) Handbook of Brain Theory and Neural Networks. Cambridge, MA: MIT Press, pp. 241–244. [Google Scholar]
- Bullock D, Cisek P, Grossberg S. (1998) Cortical networks for control of voluntary arm movements under variable force conditions. Cerebral Cortex 8(1): 48–62. [DOI] [PubMed] [Google Scholar]
- Buschman TJ, Miller EK. (2007) Top-down versus bottom-up control of attention in the prefrontal and posterior parietal cortices. Science 315(5820): 1860–1862. [DOI] [PubMed] [Google Scholar]
- Cao Y, Grossberg S, Markowitz J. (2011) How does the brain rapidly learn and reorganize view-invariant and position-invariant object representations in the inferotemporal cortex? Neural Networks 24(10): 1050–1061. [DOI] [PubMed] [Google Scholar]
- Cardinal RN, Parkinson JA, Hall J, et al. (2002) Emotion and motivation: The role of the amygdala, ventral striatum, and prefrontal cortex. Neuroscience and Biobehavioral Reviews 26(3): 321–352. [DOI] [PubMed] [Google Scholar]
- Carpenter GA. (1997) Distributed learning, recognition, and prediction by ART and ARTMAP neural networks. Neural Networks 10(8): 1473–1494. [DOI] [PubMed] [Google Scholar]
- Carpenter GA, Grossberg S. (1987) A massively parallel architecture for a self-organizing neural pattern recognition machine. Computer Vision, Graphics, and Image Processing 37(1): 54–115. [Google Scholar]
- Carpenter GA, Grossberg S. (1991) Pattern Recognition by Self-Organizing Neural Networks. Cambridge MA: MIT Press. [Google Scholar]
- Carpenter GA, Grossberg S. (1993) Normal and amnesic learning, recognition and memory by a neural model of cortico-hippocampal interactions. Trends in Neurosciences 16(4): 131–137. [DOI] [PubMed] [Google Scholar]
- Chang H-C, Grossberg S, Cao Y. (2014) Where’s Waldo? How perceptual, cognitive, and emotional brain processes cooperate during learning to categorize and find desired objects in a cluttered scene. Frontiers in Integrative Neuroscience 8: 43. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chau BK, Sallet J, Papageorgiou GK, et al. (2015) Contrasting roles for orbitofrontal cortex and amygdala in credit assignment and learning in Macaques. Neuron 87(5): 1106–1118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chun MM. (2000) Contextual cueing of visual attention. Trends in Cognitive Sciences 4(5): 170–178. [DOI] [PubMed] [Google Scholar]
- Chun MM, Jiang Y. (1998) Contextual cueing: implicit learning and memory of visual context guides spatial attention. Cognitive Psychology 36(1): 28–71. [DOI] [PubMed] [Google Scholar]
- Chun MM, Jiang Y. (1999) Top-down attentional guidance based on implicit learning of visual covariation. Psychological Science 10(4): 360–365. [Google Scholar]
- Cisek P, Grossberg S, Bullock D. (1998) A cortico-spinal model of reaching and proprioception under multiple task constraints. Journal of Cognitive Neuroscience 10(4): 425–444. [DOI] [PubMed] [Google Scholar]
- Cohen MA, Grossberg S. (1986) Neural dynamics of speech and language coding: Developmental programs, perceptual grouping, and competition for short term memory. Human Neurobiology 5(1): 1–22. [PubMed] [Google Scholar]
- Cohen MA, Grossberg S. (1987) Masking fields: A massively parallel neural architecture for learning, recognizing, and predicting multiple groupings of patterned data. Applied Optics 26(10): 1866–1891. [DOI] [PubMed] [Google Scholar]
- Contreras-Vidal JL, Grossberg S, Bullock D. (1997) A neural model of cerebellar learning for arm movement control: Cortico-spino-cerebellar dynamics. Learning and Memory 3(6): 475–502. [DOI] [PubMed] [Google Scholar]
- Cowan N. (2001) The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences 24: 87–185. [DOI] [PubMed] [Google Scholar]
- Cowan N, Elliott EM, Scott Saults J, et al. (2005) On the capacity of attention: Its estimation and its role in working memory and cognitive aptitudes. Cognitive Psychology 51(1): 42–100. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Damasio AR. (1999) The Feeling of What Happens: Body and Emotion in the Making of Consciousness. Boston, MA: Houghton Mifflin Harcourt. [Google Scholar]
- Davidson TL, Altizer AM, Benoit SC, et al. (1997) Encoding and selective activation of ‘metabolic memories’ in the rat. Behavioral Neuroscience 111(5): 1014–1030. [DOI] [PubMed] [Google Scholar]
- Davis CJ. (2010) The spatial coding model of visual word identification. Psychological Review 117(3): 713–758. [DOI] [PubMed] [Google Scholar]
- Davis M. (1994) The role of the amygdala in emotional learning. International Review of Neurobiology 36: 225–265. [DOI] [PubMed] [Google Scholar]
- Denny MR. (1971) Relaxation theory and experiments. In: Brush FR. (ed.) Aversive Conditioning and Learning. New York: Academic Press. [Google Scholar]
- Depue RA. (1974) The specificity of response interference to schizophrenia. Journal of Abnormal Psychology 83(5): 529–532. [DOI] [PubMed] [Google Scholar]
- Desimone R. (1998) Visual attention mediated by biased competition in extrastriate visual cortex. Philosophical Transactions of the Royal Society of London 353(1373): 1245–1255. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Desimone R, Duncan J. (1995) Neural mechanisms of selective visual attention. Annual Review of Neuroscience 18: 193–222. [DOI] [PubMed] [Google Scholar]
- Diana RA, Yonelinas AP, Ranganath C. (2007) Imaging recollection and familiarity in the medial temporal lobe: A three-component model. Trends in Cognitive Sciences 11(9): 379–386. [DOI] [PubMed] [Google Scholar]
- Douglas RJ, Koch C, Mahowald M, et al. (1995) Recurrent excitation in neocortical circuits. Science 269(5226): 981–985. [DOI] [PubMed] [Google Scholar]
- Dranias M, Grossberg S, Bullock D. (2008) Dopaminergic and non-dopaminergic value systems in conditioning and outcome-specific revaluation. Brain Research 1238: 239–287. [DOI] [PubMed] [Google Scholar]
- Duncan J, Humphreys GW. (1989) Visual search and stimulus similarity. Psychological Review 96(3): 433–458. [DOI] [PubMed] [Google Scholar]
- Duncan-Johnson CC, Donchin E. (1977) On quantifying surprise: the variation in event-related potentials with subjective probability. Psychophysiology 14(5): 456–467. [DOI] [PubMed] [Google Scholar]
- Egeth H, Virzi RA, Garbart H. (1984) Searching for conjunctively defined targets. Journal of Experimental Psychology: Human Perception and Performance 10(1): 32–39. [DOI] [PubMed] [Google Scholar]
- Elder D, Grossberg S, Mingolla M. (2009) A neural model of visually guided steering, obstacle avoidance, and route selection. Journal of Experimental Psychology: Human Perception & Performance 35(5): 1501–1531. [DOI] [PubMed] [Google Scholar]
- Elliott R, Dolan RJ, Frith CD. (2000) Dissociable functions in the medial and lateral orbitofrontal cortex: Evidence from human neuroimaging studies. Cerebral Cortex 10: 308–317. [DOI] [PubMed] [Google Scholar]
- Engel AK, Fries P, Singer W. (2001) Dynamics predictions: Oscillations and synchrony in top-down processing. Nature Reviews Neuroscience 2: 704–716. [DOI] [PubMed] [Google Scholar]
- Engle RW. (2002) Working memory capacity as executive attention. Current Directions in Psychological Science 11(1): 19–23. [Google Scholar]
- Estes WK. (1950) Toward a statistical theory of learning. Psychological Review 57(2): 94–107. [Google Scholar]
- Evarts EV. (1973) Motor cortex reflexes associated with learned movement. Science 179(4072): 501–503. [DOI] [PubMed] [Google Scholar]
- Farrell S, Lewandowsky S. (2004) Modeling transposition latencies: Constraints for theories of serial order memory. Journal of Memory and Language 51(1): 115–135. [Google Scholar]
- Fazl A, Grossberg S, Mingolla E. (2009) View-invariant object category learning, recognition, and search: How spatial and object attention are coordinated using surface-based attentional shrouds. Cognitive Psychology 58: 1–48. [DOI] [PubMed] [Google Scholar]
- Fellows LK, Farah MJ. (2003) Ventromedial frontal cortex mediates affective shifting in humans: evidence from a reversal learning paradigm. Brain 126(Pt 8): 1830–1837. [DOI] [PubMed] [Google Scholar]
- Fiala JC, Grossberg S, Bullock D. (1996) Metabotropic glutamate receptor activation in cerebellar Purkinje cells as substrate for adaptive timing of the classically conditioned eye-blink response. Journal of Neuroscience 16(11): 3760–3774. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Foley NC, Grossberg S, Mingolla E. (2012) Neural dynamics of object-based multifocal visual spatial attention and priming: Object cueing, useful-field-of-view, and crowding. Cognitive Psychology 65(1): 77–117. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Francis G, Grossberg S. (1996) Cortical dynamics of boundary segmentation and reset: Persistence, afterimages, and residual traces. Perception 25(5): 543–567. [DOI] [PubMed] [Google Scholar]
- Franklin DJ, Grossberg S. (2017) A neural model of normal and abnormal learning and memory consolidation: Adaptively timed conditioning, hippocampus, amnesia, neurotrophins, and consciousness. Cognitive, Affective, and Behavioral Neuroscience 17(1): 24–76. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Freedman DJ, Riesenhuber M, Poggio T, et al. (2001) Categorical representation of visual stimuli and the primate prefrontal cortex. Science 291(5502): 312–316. [DOI] [PubMed] [Google Scholar]
- Freedman DJ, Riesenhuber M, Poggio T, et al. (2003) A comparison of primate prefrontal and inferior temporal cortices during visual categorization. Journal of Neuroscience 23(12): 5235–5246. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Frith CD, Dolan RJ. (1997) Brain mechanisms associated with top-down processes in perception. Philosophical Transactions of the Royal Society of London B 352(1358): 1221–1230. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Funahashi S, Inoue M, Kubota K. (1997) Delay-period activity in the primate prefrontal cortex encoding multiple spatial positions and their order of presentation. Behavioral Brain Research 84(1–2): 203–223. [DOI] [PubMed] [Google Scholar]
- Fuster JM. (1973) Unit activity in prefrontal cortex during delayed-response performance: neuronal correlates of transient memory. Journal of Neurophysiology 36(1): 61–78. [DOI] [PubMed] [Google Scholar]
- Gallagher M, McMahan RW, Schoenbaum G. (1999) Orbitofrontal cortex and representation of incentive value in associative learning. Journal of Neuroscience 19(15): 6610–6614. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gancarz G, Grossberg G. (1998) A neural model of the saccade generator in the reticular formation. Neural Networks 11(7–8): 1159–1174. [DOI] [PubMed] [Google Scholar]
- Gancarz G, Grossberg G. (1999) A neural model of saccadic eye movement control explains task-specific adaptation. Vision Research 39(18): 3123–3143. [DOI] [PubMed] [Google Scholar]
- Gelfand JR, Bookheimer SY. (2003) Dissociating neural mechanisms of temporal sequencing and processing phonemes. Neuron 38(5): 831–842. [DOI] [PubMed] [Google Scholar]
- Ghashghaei HT, Barbas H. (2002) Pathways for emotion: Interactions of prefrontal and anterior temporal pathways in the amygdala of the rhesus monkey. Neuroscience 115(4): 1261–1279. [DOI] [PubMed] [Google Scholar]
- Gloor P, Olivier A, Quesney LF, et al. (1982) The role of the limbic system in experiential phenomena of temporal lobe epilepsy. Annals of Neurology 12(2): 129–144. [DOI] [PubMed] [Google Scholar]
- Gochin PM, Miller EK, Gross CG, et al. (1991) Functional interactions among neurons in inferior temporal cortex of the awake macaque. Experimental Brain Research 84(3): 505–516. [DOI] [PubMed] [Google Scholar]
- Gold JI, Shadlen MN. (2007) The neural basis of decision making. Annual Review of Neuroscience 30: 535–574. [DOI] [PubMed] [Google Scholar]
- Goldinger SD, Azuma T. (2003) Puzzle-solving science: The quixotic quest for units in speech perception. Journal of Phonetics 31(3–4): 305–320. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Goldman-Rakic PS. (1987) Circuitry of primate prefrontal cortex and regulation of behavior by representational memory. In: Mountcastle VB, Plum G. (eds) Handbook of Physiology, volume 5, part 1. Bethesda, MD: American Physiological Society, pp. 373–417. [Google Scholar]
- Goldman-Rakic PS. (1988) Topography of cognition: Parallel distributed networks in primate association cortex. Annual Review of Neuroscience 11: 137–156. [DOI] [PubMed] [Google Scholar]
- Gore F, Schwartz EC, Brangers BC, et al. (2015) Neural representations of unconditioned stimuli in basolateral amygdala mediate innate and learned responses. Cell 162(1): 134–145. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Grahn JA, Parkinson JA, Owen AM. (2009) The role of the basal ganglia in learning and memory: Neuropsychological studies. Behavioural Brain Research 199(1): 53–60. [DOI] [PubMed] [Google Scholar]
- Gregoriou GG, Gotts SJ, Zhou H, et al. (2009) High-frequency, long-range coupling between prefrontal and visual cortex during attention. Science 324(5931): 1207–1210. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Grossberg S. (1971) On the dynamics of operant conditioning. Journal of Theoretical Biology 33(2): 225–255. [DOI] [PubMed] [Google Scholar]
- Grossberg S. (1972. a) A neural theory of punishment and avoidance, I: Qualitative theory. Mathematical Biosciences 15(1–2): 39–67. [Google Scholar]
- Grossberg S. (1972. b) A neural theory of punishment and avoidance, II: Quantitative theory. Mathematical Biosciences 15(3–4): 253–285. [Google Scholar]
- Grossberg S. (1973) Contour enhancement, short-term memory, and constancies in reverberating neural networks. Studies in Applied Mathematics 52(3): 213–257. [Google Scholar]
- Grossberg S. (1974) Classical and instrumental learning by neural networks. In: Rosen R, Snell F. (eds) Progress in Theoretical Biology, vol. 3 New York: Academic Press, pp. 51–141. [Google Scholar]
- Grossberg S. (1975) A neural model of attention, reinforcement, and discrimination learning. International Review of Neurobiology 18: 263–327. [DOI] [PubMed] [Google Scholar]
- Grossberg S. (1976. a) Adaptive pattern classification and universal recoding: I. Parallel development and coding of neural feature detectors. Biological Cybernetics 23(3): 121–134. [DOI] [PubMed] [Google Scholar]
- Grossberg S. (1976. b) Adaptive pattern classification and universal recoding: II. Feedback, expectation, olfaction, and illusions. Biological Cybernetics 23(4): 187–202. [DOI] [PubMed] [Google Scholar]
- Grossberg S. (1978. a) A theory of human memory: Self-organization and performance of sensory-motor codes, maps, and plans. In: Rosen R, Snell F. (eds) Progress in Theoretical Biology, vol. 5 New York: Academic Press, pp. 233–374. [Google Scholar]
- Grossberg S. (1978. b) Behavioral contrast in short-term memory: Serial binary memory models or parallel continuous memory models? Journal of Mathematical Psychology 17(3): 199–219. [Google Scholar]
- Grossberg S. (1980) How does a brain build a cognitive code? Psychological Review 87(1): 1–51. [DOI] [PubMed] [Google Scholar]
- Grossberg S. (1982) Processing of expected and unexpected events during conditioning and attention: A psychophysiological theory. Psychological Review 89(5): 529–572. [PubMed] [Google Scholar]
- Grossberg S. (1984. a) Some psychophysiological and pharmacological correlates of a developmental, cognitive, and motivational theory. In: Karrer R, Cohen J, Tueting P. (eds) Brain and Information: Event Related Potentials. New York: New York Academy of Sciences, pp. 58–142. [DOI] [PubMed] [Google Scholar]
- Grossberg S. (1984. b) Unitization, automaticity, temporal order, and word recognition. Cognition and Brain Theory 7(3–4): 263–283. [Google Scholar]
- Grossberg S. (1986) The adaptive self-organization of serial order in behavior: Speech, language, and motor control. In: Schwab EC, Nusbaum HC. (eds) Pattern Recognition by Humans and Machines, Volume 1: Speech Perception. New York: Academic Press, pp. 187–294. [Google Scholar]
- Grossberg S. (1999) How does the cerebral cortex work? Learning, attention and grouping by the laminar circuits of visual cortex. Spatial Vision 12(2): 163–186. [DOI] [PubMed] [Google Scholar]
- Grossberg S. (2000. a) The complementary brain: Unifying brain dynamics and modularity. Trends in Cognitive Sciences 4(6): 233–246. [DOI] [PubMed] [Google Scholar]
- Grossberg S. (2000. b) The imbalanced Brain: From normal behavior to schizophrenia. Biological Psychiatry 48(2): 81–98. [DOI] [PubMed] [Google Scholar]
- Grossberg S. (2013. a) Adaptive Resonance Theory: How a brain learns to consciously attend, learn, and recognize a changing world. Neural Networks 37: 1–47. [DOI] [PubMed] [Google Scholar]
- Grossberg S. (2013. b) Recurrent neural networks. Scholarpedia. Available at: http://www.scholarpedia.org/article/Recurrent_neural_networks
- Grossberg S. (2016) Neural dynamics of the basal ganglia during perceptual, cognitive, and motor learning and gating. In: Soghomonian J-J. (ed.) The Basal Ganglia: Novel Perspectives on Motor and Cognitive Functions. Berlin: Springer, pp. 457–512. [Google Scholar]
- Grossberg S. (2017. a) Acetylcholine neuromodulation in normal and abnormal learning and memory: Vigilance control in waking, sleep, autism, amnesia, and Alzheimer’s disease. Frontiers in Neural Circuits. Available at: 10.3389/fncir.2017.00082 (accessed 2 November 2017). [DOI] [PMC free article] [PubMed]
- Grossberg S. (2017. b) Towards solving the Hard Problem of Consciousness: The varieties of brain resonances and the conscious experiences that they support. Neural Networks 87: 38–95. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Gutowski WE. (1987) Neural dynamics of decision making under risk: Affective balance and cognitive-emotional interactions. Psychological Review 94(3): 300–318. [PubMed] [Google Scholar]
- Grossberg S, Huang T-R. (2009) ARTSCENE: A neural system for natural scene classification. Journal of Vision 9(4): 6.1–6.19. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Kazerounian S. (2011) Laminar cortical dynamics of conscious speech perception: Neural model of phonemic restoration using subsequent context in noise. Journal of the Acoustical Society of America 130(1): 440–460. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Kazerounian S. (2016) Phoneme restoration and empirical coverage of Interactive Activation and Adaptive Resonance models of human speech processing. Journal of the Acoustical Society of America 140(2): 1130. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Kishnan D. (2018) Neural dynamics of autistic repetitive behaviors and Fragile X syndrome: Basal ganglia movement gating and mGluR-modulated adaptively- timed learning. Frontiers in Psychology: Psychopathology 9: 269 Availabe at: 10.3389/fpsyg.2018.00269 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Grossberg S, Levine DS. (1987) Neural dynamics of attentionally modulated Pavlovian conditioning: Blocking, inter-stimulus interval, and secondary reinforcement. Applied Optics 26(23): 5015–5030. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Myers C. (2000) The resonant dynamics of speech perception: Interword integration and duration-dependent backward effects. Psychological Review 107: 735–767. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Pearson L. (2008) Laminar cortical dynamics of cognitive and motor working memory, sequence learning and performance: Toward a unified theory of how the cerebral cortex works. Psychological Review 115: 677–732. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Pilly P. (2008) Temporal dynamics of decision-making during motion perception in the visual cortex. Vision Research 48(12): 1345–1373. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Raizada R. (2000) Contrast-sensitive perceptual grouping and object-based attention in the laminar circuits of primary visual cortex. Vision Research 40(10–12): 1413–1432. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Repin D. (2003) A neural model of how the brain represents and compares multi-digit numbers: spatial and categorical processes. Neural Networks 16(8): 1107–1140. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Schmajuk NA. (1987) Neural dynamics of attentionally modulated Pavlovian conditioning: Conditioned reinforcement, inhibition, and opponent processing. Psychobiology 15: 195–240. [Google Scholar]
- Grossberg S, Seidman D. (2006) Neural dynamics of autistic behaviors: Cognitive, emotional, and timing substrates. Psychological Review 113(3): 483–525. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Versace M. (2008) Spikes, synchrony, and attentive learning by laminar thalamocortical circuits. Brain Research 1218: 278–312. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Boardman I, Cohen M. (1997) Neural dynamics of variable-rate speech categorization. Journal of Experimental Psychology: Human Perception and Performance 23(2): 481–503. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Bullock D, Dranias M. (2008) Neural dynamics underlying impaired autonomic and conditioned responses following amygdala and orbitofrontal lesions. Behavioral Neuroscience 122(5): 1100–1125. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Hwang S, Mingolla E. (2002) Thalamocortical dynamics of the McCollough effect: Boundary-surface alignment through perceptual learning. Vision Research 42(10): 1259–1286. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Markowitz J, Cao Y. (2011) On the road to invariant recognition: Explaining tradeoff and morph properties of cells in inferotemporal cortex using multiple-scale task-sensitive attentive learning. Neural Networks 24(10): 1036–1049. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Mingolla E, Pack C. (1999) A neural model of motion processing and visual navigation by cortical area MST. Cerebral Cortex 9(8): 878–895. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Mingolla E, Ross WD. (1994) A neural theory of attentive visual search: Interactions of boundary, surface, spatial, and object representations. Psychological Review 101(3): 470–489. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Mingolla E, Viswanathan L. (2001) Neural dynamics of motion integration and segmentation within and across apertures. Vision Research 41(19): 2521–2553. [DOI] [PubMed] [Google Scholar]
- Grossberg S, Srihasam K, Bullock D. (2012) Neural dynamics of saccadic and smooth pursuit eye movement coordination during visual tracking of unpredictably moving targets. Neural Networks 27: 1–20. [DOI] [PubMed] [Google Scholar]
- Halgren E, Walter RD, Cherlow DG, et al. (1978) Mental phenomena evoked by electrical stimulation of the human hippocampal formation and amygdala. Brain 101(1): 83–117. [DOI] [PubMed] [Google Scholar]
- Haralanova E, Haralanov S, Veraldi A, et al. (2011) Subjective emotional over-arousal to neutral social scenes in paranoid schizophrenia. European Archives of Psychiatry and Clinical Neuroscience 262(1): 59–68. [DOI] [PubMed] [Google Scholar]
- Harley CW. (2004) Norepinephrine and dopamine as learning signals. Neural Plasticity 11(3–4): 191–204. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Harries MH, Perrett DI. (1991) Visual processing of faces in temporal cortex: Physiological evidence for a modular organization and possible anatomical correlates. Journal of Cognitive Neuroscience 3(1): 9–24. [DOI] [PubMed] [Google Scholar]
- Hatfield T, Han J-S, Conley M, et al. (1996) Neurotoxic lesions of basolateral, but not central, amygdala interfere with Pavlovian second-order conditioning and reinforcer devaluation effects. Journal of Neuroscience 16(16): 5256–5265. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Heeger DJ. (1992) Normalization of cell responses in cat striate cortex. Visual Neuroscience 9(2): 181–197. [DOI] [PubMed] [Google Scholar]
- Heekeren HR, Marrett S, Ungerleider LG. (2008) The neural systems that mediate human perceptual decision making. Nature Reviews Neuroscience 9: 467–479. [DOI] [PubMed] [Google Scholar]
- Henson RNA. (1998) Short-term memory for serial order: The start-end model. Cognitive Psychology 36(2): 73–137. [DOI] [PubMed] [Google Scholar]
- Hikosaka O, Wurtz RH. (1983) Visual and oculomotor functions of monkey substantia nigra pars reticulata. IV. Relation of substantia nigra to superior colliculus. Journal of Neurophysiology 49(5): 1285–1301. [DOI] [PubMed] [Google Scholar]
- Hikosaka O, Wurtz RH. (1989) The basal ganglia. In: Wurtz R, Goldberg M. (eds) The Neurobiology of Sacccadic Eye Movements. Amsterdam: Elsevier, pp. 257–281. [Google Scholar]
- Hikosaka O, Sakamoto M, Usui S. (1989) Functional properties of monkey caudate neurons. I. activities related to saccadic eye movements. Journal of Neurophysiology 61(4): 780–798. [DOI] [PubMed] [Google Scholar]
- Histed MH, Miller EK. (2006) Microstimulation of frontal cortex can reorder a remembered spatial sequence. PLoS Biology 4(5): e134. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hochstein S, Ahissar M. (2002) View from the top: Hierarchies and reverse hierarchies in the visual system. Neuron 36(5): 791–804. [DOI] [PubMed] [Google Scholar]
- Holland PC, Straub JJ. (1979) Differential effects of two ways of devaluing the unconditioned stimulus after Pavlovian appetitive conditioning. Journal of Experimental Psychology 5(1): 65–78. [DOI] [PubMed] [Google Scholar]
- Hollerman J, Schultz W. (1998) Dopamine neurons report an error in the temporal prediction of reward during learning. Nature Neuroscience 1(4): 304–309. [DOI] [PubMed] [Google Scholar]
- Hoskison MM, Moore AN, Hu B, et al. (2009) Persistent working memory dysfunction following traumatic brain injury: Evidence for a time-dependent mechanism. Neuroscience 159(2): 483–491. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Houghton G. (1990) The problem of serial order: A neural network model of sequence learning and recall. In: Dale R, Mellish C, Zock M. (eds) Current Research in Natural Language Generation. London: Academic Press, pp. 287–319. [Google Scholar]
- Huang T-R, Grossberg S. (2010) Cortical dynamics of contextually cued attentive visual learning and search: Spatial and object evidence accumulation. Psychological Review 117(4): 1080–1112. [DOI] [PubMed] [Google Scholar]
- Hubel DH, Wiesel TN. (1962) Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. Journal of Physiology 160(1): 108–154. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hubel DH, Wiesel TN. (1963) Shape and arrangement of columns in cat’s striate cortex. Journal of Physiology 165(3): 559–568. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hung CP, Kreiman G, Poggio T, et al. (2005) Fast readout of object identity from Macaque inferior temporal cortex. Science 310(5749): 863–866. [DOI] [PubMed] [Google Scholar]
- Hunt RR, Lamb CA. (2001) What causes the isolation effect? Journal of Experimental Psychology: Learning, Memory and Cognition 27(6): 1359–1366. [PubMed] [Google Scholar]
- Ingvar DH. (1985) Memory of the future: An essay on the temporal organisation of conscious awareness. Human Neurobiology 4: 127–136. [PubMed] [Google Scholar]
- Inoue M, Mikami A. (2006) Prefrontal activity during serial probe reproduction task: encoding, mnemonic, and retrieval processes. Journal of Neurophysiology 95(2): 1008–1041. [DOI] [PubMed] [Google Scholar]
- Irwin DE. (1991) Information integration across saccadic eye movements. Cognitive Psychology 23(3): 420–456. [DOI] [PubMed] [Google Scholar]
- Isoda M, Tanji J. (2002) Cellular activity in the supplementary eye field during sequential performance of multiple saccades. Journal of Neurophysiology 88(6): 3541–3545. [DOI] [PubMed] [Google Scholar]
- Isoda M, Tanji J. (2003) Contrasting neuronal activity in the supplementary and frontal eye fields during temporal organization of multiple saccades. Journal of Neurophysiology 90(5): 3054–3065. [DOI] [PubMed] [Google Scholar]
- Ito M. (1984) The Cerebellum and Neural Control. New York: Raven Press. [Google Scholar]
- Itti L, Koch C. (2000) A saliency-based search mechanism for overt and covert shifts of visual attention. Vision Research 40(10–12): 1489–1506. [DOI] [PubMed] [Google Scholar]
- Izquierdo A, Murray EA. (2007) Selective bilateral amygdala lesions in rhesus monkeys fail to disrupt object reversal learning. Journal of Neuroscience 27(5): 1054–1062. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jiang Y, Chun MM. (2001) Selective attention modulates implicit learning. The Quarterly Journal of Experimental Psychology. A, Human Experimental Psychology 54(4): 1105–1124. [DOI] [PubMed] [Google Scholar]
- Jones B, Mishkin M. (1972) Limbic lesions and the problem of stimulus – Reinforcement associations. Experimental Neurology 36(2): 362–377. [DOI] [PubMed] [Google Scholar]
- Jones D, Farrand P, Stuart G, et al. (1995) Functional equivalence of verbal and spatial information in serial short-term memory. Journal of Experimental Psychology: Learning Memory and Cognition 21(4): 1008–1018. [DOI] [PubMed] [Google Scholar]
- Jonides J, Irwin DE, Yantis S. (1982) Integrating visual information from successive fixations. Science 215(4529): 192–194. [DOI] [PubMed] [Google Scholar]
- Kahneman D, Beatty J. (1966) Pupil diameter and load on memory. Science 154(3756): 1583–1585. [DOI] [PubMed] [Google Scholar]
- Kahneman D, Tversky A. (1979) Prospect theory: An analysis of decision under risk. Econometrica 47(2): 263–291. [Google Scholar]
- Kalaska JF, Cohen DAD, Hyde ML, et al. (1989) A comparison of movement direction-related versus load direction-related activity in primate motor cortex, using a two-dimensional reaching task. Journal of Neuroscience 9(6): 2080–2102. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kamin LJ. (1968) ‘Attention-like’ processes in classical conditioning. In: Jones MR. (ed.) Miami Symposium on the Prediction of Behavior: Aversive Stimulation Miami, OH: University of Miami Press. [Google Scholar]
- Kamin LJ. (1969) Predictability, surprise, attention and conditioning. In: Campbell BA, Church RM. (eds) Punishment and Aversive Behavior. New York: Appleton-Century-Crofts. [Google Scholar]
- Kamin LJ, Brimer CJ, Black AH. (1963) Conditioned suppression as a monitor of fear in the course of avoidance training. Journal of Comparative and Physiological Psychology 56(3): 497–501. [DOI] [PubMed] [Google Scholar]
- Karadi Z, Oomura Y, Nishino H, et al. (1992) Responses of lateral hypothalamic glucose-sensitive and glucose-insensitive neurons to chemical stimuli in behaving rhesus monkeys. Journal of Neurophysiology 67(2): 389–400. [DOI] [PubMed] [Google Scholar]
- Kazerounian S, Grossberg S. (2014) Real-time learning of predictive recognition categories that chunk sequences of items stored in working memory. Frontiers in Psychology: Language Sciences 5: 1053. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kermadi I, Joseph JP. (1995) Activity in the caudate nucleus of monkey during spatial sequencing. Journal of Neurophysiology 74(3): 911–933. [DOI] [PubMed] [Google Scholar]
- Kojima S, Goldman-Rakic PS. (1984) Functional analysis of spatially discriminative neurons in prefrontal cortex of rhesus monkey. Brain Research 291(2): 229–240. [DOI] [PubMed] [Google Scholar]
- Kondoh T, Mori M, Ono T, et al. (2000) Mechanisms of umami taste preference and aversion in rats. Journal of Nutrition 130(4S Suppl.): 966S–970S. [DOI] [PubMed] [Google Scholar]
- LeDoux JE. (1993) Emotional memory systems in the brain. Behavioural Brain Research 58(1–2): 69–79. [DOI] [PubMed] [Google Scholar]
- Levy DJ, Glimcher PW. (2012) The root of all value: A neural common currency for choice. Current Opinion in Neurobiology 22(6): 1027–1038. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Libby LA, Hannula DE, Ranganath C. (2014) Medial temporal lobe coding of item and spatial information during relational binding in working memory. Journal of Neuroscience 34(43): 14233–14242. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Livingston RB. (1967) Brain circuitry relating to complex behavior. In: Quarton JC, Melneckuk T, Schmitt FO. (eds) The Neurosciences: A Study Program. New York: The Rockefeller University Press, pp. 499–515. [Google Scholar]
- Ljungberg T, Apicella P, Schultz W. (1992) Responses of monkey dopamine neurons during learning of behavioral reactions. Journal of Neurophysiology 67(1): 145–163. [DOI] [PubMed] [Google Scholar]
- Lleras A, Von Mühlenen A. (2004) Spatial context and top-down strategies in visual search. Spatial Vision 17(4–5): 465–482. [DOI] [PubMed] [Google Scholar]
- Luce PA, McLennan CT. (2008) Spoken word recognition: The challenge of variation. In: Pisoni DB, Remez RE. (eds) The Handbook of Speech Perception. Wiley Online Library: Availabe at: 10.1002/9780470757024.ch24 [DOI] [Google Scholar]
- Luce RD. (1977) The choice axiom after twenty years. Journal of Mathematical Psychology 15(3): 215–233. [Google Scholar]
- Luria AR. (1966) Higher Cortical Functions in Man. New York: Basic Books. [Google Scholar]
- Lysle DT, Fowler H. (1985) Inhibition as a ‘slave’ process: Deactivation of conditioned inhibition through extinction of conditioned excitation. Journal of Experimental Psychology: Animal Behavior Processes 11(1): 71–94. [DOI] [PubMed] [Google Scholar]
- McClelland JL, Rumelhart DE. (1981) An interactive activation model of context effects in letter perception: 1. An account of basic findings. Psychological Review 88(5): 375–407. [PubMed] [Google Scholar]
- McCollough C. (1965) Color adaptation of edge-detectors in the human visual system. Science 149(3688): 1115–1116. [DOI] [PubMed] [Google Scholar]
- McGaugh JL. (2003) Memory and Emotion: The Making of Lasting Memories. New York: Columbia University Press. [Google Scholar]
- McLennan CT, Conor T, Luce PA. (2005) Examining the time course of indexical specificity effects in spoken word recognition. Journal of Experimental Psychology: Learning, Memory, and Cognition 31(2): 306–321. [DOI] [PubMed] [Google Scholar]
- McLennan CT, Luce PA, Charles-Luce J. (2003) Representation of lexical form. Journal of Experimental Psychology: Learning, Memory, and Cognition 29(4): 539–553. [DOI] [PubMed] [Google Scholar]
- McNab F, Klingberg T. (2008) Prefrontal cortex and basal ganglia control access to working memory. Nature Neuroscience 11(1): 103–107. [DOI] [PubMed] [Google Scholar]
- Magee JC, Johnston D. (1997) A synaptically controlled, associative signal for Hebbian plasticity in hippocampal neurons. Science 275(5297): 209–213. [DOI] [PubMed] [Google Scholar]
- Maier SF, Seligman MEP, Solomon RL. (1969) Pavlovian fear conditioning and learned helplessness effects on escape and avoidance behavior of (a) the CS-US contingency and (b) the independent of the US and voluntary responding. In: Campbell BA, Church RM. (eds) Punishment and Aversive Behavior. New York: Appleton. [Google Scholar]
- Málková L, Gaffan D, Murray EA. (1997) Excitotoxic lesions of the amygdala fail to produce impairment in visual learning for auditory secondary reinforcement but interfere with reinforcer devaluation effects in rhesus monkeys. Journal of Neuroscience 17(15): 6011–6020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Markram H, Helm PJ, Sakmann B. (1995) Dendritic calcium transients evoked by single back-propagating action potentials in rat neocortical pyramidal neurons. Journal of Physiology 485(Pt 1): 1–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Markram H, Lübke J, Frorscher M, et al. (1997) Regulation of synaptic efficacy by coincidence of postsynaptic APs and EPSPs. Science 275(5297): 213–215. [DOI] [PubMed] [Google Scholar]
- Masterson FA. (1970) Is termination of a warning signal an effective reward for the rat? Journal of Comparative and Physiological Psychology 72(3): 471–475. [Google Scholar]
- Mayer K, Wyckoff SN, Stehl U. (2016) Underarousal in adult ADHD: How are peripheral and cortical arousal related? Clinical EEG and Neuroscience 47(3): 171–179. [DOI] [PubMed] [Google Scholar]
- Miller EK, Cohen JD. (2001) An integrative theory of prefrontal cortex function. Annual Review of Neuroscience 24: 167–202. [DOI] [PubMed] [Google Scholar]
- Miller GA. (1956) The magical number seven plus or minus two. Psychological Review 63(2): 81–97. [PubMed] [Google Scholar]
- Miller RR, Schachtman TR. (1985) Conditioning context as an associative baseline: Implications for response generation and the nature of conditioned inhibition. In: Miller RR, Spear NE. (eds) Information Processing in Animals: Conditioned Inhibition. Hillsdale, NJ: Erlbaum. [Google Scholar]
- Mishkin M. (1982) A memory system in the monkey. Philosophical Transactions of the Royal Society of London. Series B, Biological Sciences 298: 85–95. [DOI] [PubMed] [Google Scholar]
- Mishkin M, Ungerleider LG, Macko KA. (1983) Object vision and spatial vision: Two cortical pathways. Trends in Neurosciences 6: 414–417. [Google Scholar]
- Montague P, Dayan P, Sejnowski T. (1996) A framework for mesencephalic dopamine systems based on predictive Hebbian learning. Journal of Neuroscience 16(5): 1936–1947. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Morrison SE, Salzman CD. (2010) Re-valuing the amygdala. Current Opinion in Neurobiology 20(2): 221–230. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Morrison SE, Saez A, Lau B, et al. (2011) Different time courses for learning-related changes in amygdala and orbitofrontal cortex. Neuron 71(6): 1127–1140. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Murdock BB. (1962) The serial position effect of free recall. Journal of Experimental Psychology 64(5): 482–488. [Google Scholar]
- Nachev P, Rees G, Parton A, et al. (2005) Volition and conflict in human medial frontal cortex. Current Biology 15(2): 122–128. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nakamura K, Ono T. (1986) Lateral hypothalamus neuron involvement in integration of natural and artificial rewards and cue signals. Journal of Neurophysiology 55(1): 163–181. [DOI] [PubMed] [Google Scholar]
- Nakamura K, Ono T, Tamura R. (1987) Central sites involved in lateral hypothalamus conditioned neural responses to acoustic cues in the rat. Journal of Neurophysiology 58(5): 1123–1148. [DOI] [PubMed] [Google Scholar]
- Nieder A, Miller E. (2004. a) Analog numerical representations in rhesus monkeys: Evidence for parallel processing. Journal of Cognitive Neuroscience 16(5): 889–901. [DOI] [PubMed] [Google Scholar]
- Nieder A, Miller EK. (2004. b) A parieto-frontal network for visual numerical information in the monkey. Proceedings of the National Academy of Sciences of the United States of America 101(19): 7457–7462. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ninokura Y, Mushiaske H, Tanji J. (2004) Integration of temporal order and object information in the monkey lateral prefrontal cortex. Journal of Neurophysiology 91(1): 555–560. [DOI] [PubMed] [Google Scholar]
- Nishijo H, Ono T, Nishino H. (1988. a) Single neuron responses in amygdala of alert monkey during complex sensory stimulation with affective significance. Journal of Neuroscience 8(10): 3570–3583. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nishijo H, Ono T, Nishino H. (1988. b) Topographic distribution of modality-specific amygdalar neurons in alert monkey. Journal of Neuroscience 8(10): 3556–3569. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nozari N, Mirman D, Thompson-Schill SL. (2016) The ventrolateral prefrontal cortex facilitates processing of sentential context to locate referents. Brain and Lanuage 157–158: 1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Oberauer K. (2009) Design for a working memory. In: Ross BH. (ed.) The Psychology of Learning and Motivation, vol. 51 San Diego, CA: Elsevier Academic Press, pp. 45–100. [Google Scholar]
- Oliva A, Torralba A. (2001) Modeling the shape of the scene: A holistic representation of the spatial envelope. International Journal of Computer Vision 42(3): 145–175. [Google Scholar]
- Olson IR, Chun MM. (2002) Perceptual constraints on implicit learning of spatial context. Visual Cognition 9(3): 273–302. [Google Scholar]
- Öngür D, An X, Price JL. (1998) Prefrontal cortical projections to the hypothalamus in Macaque monkeys. Journal of Comparative Neurology 401(4): 480–505. [PubMed] [Google Scholar]
- Öngür D, Price JL. (2000) The organization of networks within the orbital and medial prefrontal cortex of rats, monkeys and humans. Cerebral Cortex 10(3): 206–219. [DOI] [PubMed] [Google Scholar]
- Ono T, Nakamura K, Nishijo H, et al. (1986) Hypothalamic neuron involvement in integration of reward, aversion, and cue signals. Journal of Neurophysiology 56(1): 63–79. [DOI] [PubMed] [Google Scholar]
- Owren MJ, Kaplan PS. (1981) On the failure to extinguish Pavlovian conditioned inhibition: A test of a reinstatement hypothesis. Paper presented to the meeting of the Midwestern Psychological Association, Detroit, April. [Google Scholar]
- Pack C, Grossberg S, Mingolla E. (2001) A neural model of smooth pursuit control and motion perception by cortical area MST. Journal of Cognitive Neuroscience 13(1): 102–120. [DOI] [PubMed] [Google Scholar]
- Page MPA, Norris D. (1998) The primacy model: A new model of immediate serial recall. Psychological Review 105(4): 761–781. [DOI] [PubMed] [Google Scholar]
- Parton A, Nachev P, Hodgson T, et al. (2007) Role of the human supplementary eye field in the control of saccadic eye movements. Neuropsychologia 45(5–4): 997–1008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pasupathy A, Miller EK. (2004) Different time courses of learning-related activity in the prefrontal cortex and striatum. Nature 433: 873–876. [DOI] [PubMed] [Google Scholar]
- Paton JJ, Belova MA, Morrison SE, et al. (2006) The primate amygdala represents the positive and negative value of visual stimuli during learning. Nature 439: 865–870. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pavlov IP. (1927) Conditioned Reflexes. London: Constable and Company; (Reprinted by Dover Publications, 1960). [Google Scholar]
- Peciña S, Cagniard B, Berridge KC, et al. (2003) Hyperdopaminergic mutan mice have higher ‘wanting’ but not ‘liking’ for sweet rewards. Journal of Neuroscience 23(28): 9395–9402. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Penn AA, Shatz CJ. (1999) Brain waves and brain wiring: The role of endogenous and sensory-driven neural activity in development. Pediatric Research 45(4 Pt 1): 447–458. [DOI] [PubMed] [Google Scholar]
- Petrides M, Alivisatos V, Frey S. (2002) Differential activation of the human orbital, mid-ventrolateral, and mid-dorsolateral prefrontal cortex during the processing of visual stimuli. Proceedings of the National Academy of Sciences of the United States of America 99(8): 5649–5654. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Petrides M, Pandya DN. (2002) Comparative cytoarchitectonic analysis of the human and the macaque ventrolateral prefrontal cortex and corticocortical connection patterns in the monkey. European Journal of Neuroscience 16(2): 291–310. [DOI] [PubMed] [Google Scholar]
- Picton WT. (1992) The P300 wave of the human event-related potential. Journal of Clinical Neurophysiology 9(4): 456–479. [DOI] [PubMed] [Google Scholar]
- Pollen DA. (1999) On the neural correlates of visual perception. Cerebral Cortex 9(1): 4–19. [DOI] [PubMed] [Google Scholar]
- Posner MI, Rafal RD, Choate LS, et al. (1985) Inhibition of return: Neural basis and function. Cognitive Neuropsychology 2(3): 211–228. [Google Scholar]
- Price JL. (1999) Prefrontal cortical networks related to visceral function and mood. Annals of the New York Academy of Sciences 877: 383–396. [DOI] [PubMed] [Google Scholar]
- Raizada R, Grossberg S. (2001) Context-sensitive binding by the laminar circuits of V1 and V2: A unified model of perceptual grouping, attention, and orientation contrast. Visual Cognition 8(3–5): 431–466. [Google Scholar]
- Raizada R, Grossberg S. (2003) Towards a theory of the laminar architecture of cerebral cortex: Computational clues from the visual system. Cerebral Cortex 13(1): 100–113. [DOI] [PubMed] [Google Scholar]
- Rao SG, Williams GV, Goldman-Rakic PS. (2000) Destruction and creation of spatial tuning by disinhibition: GABA(A) blockade of prefrontal cortical neurons engaged by working memory. Journal of Neuroscience 20(1): 485–494. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Reep RL, Corwin JV, King V. (1996) Neuronal connections of orbital cortex in rats: Topography of cortical and thalamic afferents. Experimental Brain Research 111(2): 215–232. [DOI] [PubMed] [Google Scholar]
- Remington RJ. (1969) Analysis of sequential effects in choice reaction times. Journal of Experimental Psychology 82(2): 250–257. [DOI] [PubMed] [Google Scholar]
- Repp BH, Liberman AM, Eccardt T, et al. (1978) Perceptual integration of acoustic cues for stop, fricative, and affricate manner. Journal of Experimental Psychology: Human Perception and Performance 4(4): 621–637. [DOI] [PubMed] [Google Scholar]
- Rescorla RA, Wagner AR. (1972) A theory of Pavlovian conditioning: Variations in the effectiveness of reinforcement and nonreinforcement. In: Black AH, Prokasy WF. (eds) Classical Conditioning II: Current Research and Theory. New York: Appleton-Century-Crofts, pp. 64–99. [Google Scholar]
- Reynierse JH, Rizley RC. (1970) Relaxation and fear as determinants of maintained avoidance in rats. Journal of Comparative and Physiological Psychology 72(2): 223–232. [DOI] [PubMed] [Google Scholar]
- Roberts AC. (2006) Primate orbitofrontal cortex and adaptive behaviour. Trends in Cognitive Sciences 10(2): 83–90. [DOI] [PubMed] [Google Scholar]
- Robinson MJF, Berridge KC. (2013) Instant transformation of learned repulsion into motivational ‘wanting’. Current Biology 23(4): 282–289. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rolls ET. (2000) The orbitofrontal cortex and reward. Cerebral Cortex 10(3): 284–294. [DOI] [PubMed] [Google Scholar]
- Rolls ET, Critchley HD, Browning AS, et al. (1999) Responses to the sensory properties of fat of neurons in the primate orbitofrontal cortex. Journal of Neuroscience 19(4): 1532–1540. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rolls ET, Hornak J, Wade D, et al. (1994) Emotion-related learning in patients with social and emotional changes associated with frontal lobe damage. Journal of Neurology, Neurosurgery, and Psychiatry 57(12): 1518–1524. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Roy JE, Riesenhuber M, Poggio T, et al. (2010) Prefrontal cortex activity during flexible categorization. Journal of Neuroscience 30: 8519–8528. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rudebeck PH, Saunders RC, Prescott AT, et al. (2013) Prefrontal mechanisms of behavioral flexibility, emotion regulation and value updating. Nature Neuroscience 16(8): 1140–1145. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rudebeck PH, Suanders RC, Lundgren DA, et al. (2017) Specialized representations of value in the orbital and ventrolateral prefrontal cortex: Desirability versus availability of outcomes. Neuron 95(5): 1208–1220. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Samuel AG, van Santen JPH, Johnston JC. (1982) Length effects in word perception: We is better than I but worse than you or them. Journal of Experimental Psychology: Human Perception and PerFormance 8: 91–105. [DOI] [PubMed] [Google Scholar]
- Samuel AG, van Santen JPH, Johnston JC. (1983) Reply to Matthei: We really is worse than you or them, and so are ma and pa. Journal of Experimental Psychology: Human Perception and Performance 9: 321–322. [DOI] [PubMed] [Google Scholar]
- Sarter M, Givens B, Bruno JP. (2001) The cognitive neuroscience of sustained attention: Where top-down meets bottom-up. Brain Research Reviews 35(2): 146–160. [DOI] [PubMed] [Google Scholar]
- Schoenbaum G, Setlow B, Saddoris MP, et al. (2003) Encoding predicted outcome and acquired value in orbitofrontal cortex during cue sampling depends upon input from basolateral amygdala. Neuron 39(5): 855–867. [DOI] [PubMed] [Google Scholar]
- Schuck NW, Cai MB, Wilson RC, et al. (2016) Human orbitofrontal cortex represents a cognitive map of state space. Neuron 91: 1402–1412. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schultz W. (1998) Predictive reward signal of dopamine neurons. Journal of Neurophysiology 80(1): 1–27. [DOI] [PubMed] [Google Scholar]
- Schultz W, Apicella P, Ljungberg T. (1993) Responses of monkey dopamine neurons to reward and conditioned stimuli during successive steps of learning a delayed response task. Journal of Neuroscience 13(3): 900–913. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schultz W, Apicelli P, Scarnati E, et al. (1992) Neuronal activity in monkey ventral striatum related to the expectation of reward. Journal of Neuroscience 12(12): 4595–4610. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schultz W, Dayan P, Montague P. (1997) A neural substrate of prediction and reward. Science 275(5306): 1593–1598. [DOI] [PubMed] [Google Scholar]
- Schultz W, Romo R, Ljungberg T, et al. (1995) Reward related signals carried by dopamine neurons. In: Houk J, Davis J, Beiser D. (eds) Models of Information Processing in the Basal Ganglia. Cambridge, MA: MIT Press, pp. 11–27. [Google Scholar]
- Schulz K. (2010) Being Wrong: Adventures in the Margin of Error. New York: HarperCollins. [Google Scholar]
- Schumacher EJ, Lauber E, Awh E, et al. (1996) PET evidence for an amodal verbal working memory system. Neuroimage 3(2): 79–88. [DOI] [PubMed] [Google Scholar]
- Seger CA, Miller EK. (2010) Category learning in the brain. Annual Review of Neuroscience 33: 203–219. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shallice T. (1988) From Neuropsychology to Mental Structure. Cambridge: Cambridge University Press. [Google Scholar]
- Shimizu N, Oomura Y, Sakata T. (1984) Modulation of feeding by endogenous sugar acids acting as hunger or satiety factors. American Journal of Physiology: Regulatory, Integrative, and Comparative Physiology 246(4 Pt 2): 542–550. [DOI] [PubMed] [Google Scholar]
- Sigala N, Logothetis NK. (2002) Visual categorization shapes feature selectivity in the primate temporal cortex. Nature 415(6869): 318–320. [DOI] [PubMed] [Google Scholar]
- Sikström S, Söderlund G. (2007) Stimulus-dependent dopamine release in attention-deficit/hyperactivity disorder. Psychological Review 114(4): 1047–1075. [DOI] [PubMed] [Google Scholar]
- Silver MR, Grossberg S, Bullock D, et al. (2011) A neural model of sequential movement planning and control of eye movements: Item-order-rank working memory and saccade selection by the supplementary eye fields. Neural Networks 26: 29–58. [DOI] [PubMed] [Google Scholar]
- Skinner BF. (1938) The Behavior of Organisms: An Experimental Analysis. Oxford: Appleton-Century. [Google Scholar]
- Skinner BF. (1948) Superstition in the pigeon. Journal of Experimental Psychology 38(2): 168–172. [DOI] [PubMed] [Google Scholar]
- Smith KS, Berridge KC, Aldridge JW. (2011) Disentangling pleasure from incentive salience and learning signals in brain reward circuitry. Proceedings of the National Academy of Sciences of the United States of America 108(27): E255–E264. [DOI] [PMC free article] [PubMed] [Google Scholar]
- So N, Stuphorn V. (2010) Supplementary eye field encodes option and action value for saccades with variable reward. Journal of Neurophysiology 104(5): 2634–2653. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Solomon RL, Kamin LJ, Wynne LC. (1953) Traumatic avoidance learning: The outcome of several extinction procedures with dogs. Journal of Abnormal Social Psychology 48(2): 291–302. [DOI] [PubMed] [Google Scholar]
- Spiegler BJ, Mishkin M. (1981) Evidence for the sequential participation of inferior temporal cortex and amygdala in the acquisition of stimulus-reward associations. Behavioral Brain Research 3(3): 303–317. [DOI] [PubMed] [Google Scholar]
- Squires KC, Wickens C, Squires NK, et al. (1976) The effect of stimulus sequence on the waveform of the cortical event-related potential. Science 193(4258): 1142–1146. [DOI] [PubMed] [Google Scholar]
- Squires NK, Squires KC, Hillyard SA. (1975) Two varieties of long-latency positive waves evoked by unpredictable auditory stimuli in man. Electroencephalography and Clinical Neurophysiology 38(4): 387–401. [DOI] [PubMed] [Google Scholar]
- Srihasam K, Bullock D, Grossberg S. (2009) Target selection by the frontal cortex during coordinated saccadic and smooth pursuit eye movements. Journal of Cognitive Neuroscience 21: 1611–1627. [DOI] [PubMed] [Google Scholar]
- Suri R, Schultz W. (1998) Learning of sequential movements by neural network model with dopamine-like reinforcement signal. Experimental Brain Research 121(3): 350–354. [DOI] [PubMed] [Google Scholar]
- Sutton S, Braren M, Zubin J, et al. (1965) Evoked-potential correlates of stimulus uncertainty. Science 150(3700): 1187–1188. [DOI] [PubMed] [Google Scholar]
- Suzuki M, Gottlieb J. (2013) Distinct neural mechanisms of distractor suppression in the frontal and parietal lobe. Nature Neuroscience 16(1): 98–104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tanaka K. (1996) Inferotemporal cortex and object vision. Annual Review of Neuroscience 19: 109–139. [DOI] [PubMed] [Google Scholar]
- Tanaka K, Saito H, Fukada Y, et al. (1991) Coding visual images of objects in the inferotemporal cortex of the macaque monkey. Journal of Neurophysiology 66(1): 170–189. [DOI] [PubMed] [Google Scholar]
- Taylor P, Nobre A, Rushworth M. (2007) Subsecond changes in top down control exerted by human medial frontal cortex during conflict and action selection: a combined transcranial magnetic stimulation electroencephalography study. Journal of Neuroscience 27(42): 11343–11353. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thompson RF. (1988) The neural basis of basic associative learning of discrete behavioral responses. Trends in Neurosciences 11(4): 152–155. [DOI] [PubMed] [Google Scholar]
- Thornton TL, Gilden DL. (2007) Parallel and serial processes in visual search. Psychological Review 114(1): 71–103. [DOI] [PubMed] [Google Scholar]
- Torralba A, Oliva A, Castelhano M, et al. (2006) Contextual guidance of eye movements and attention in real-world scenes: The role of global features in object search. Psychological Review 113(4): 766–786. [DOI] [PubMed] [Google Scholar]
- Townsend JT. (1972) Some results concerning the identifiability of parallel and serial processes. British Journal of Mathematical and Statistical Psychology 25(2): 168–199. [Google Scholar]
- Treisman A, Gelade G. (1980) A feature-integration theory of attention. Cognitive Psychology 12(1): 97–136. [DOI] [PubMed] [Google Scholar]
- Tremblay L, Schultz W. (1999) Relative reward preference in primate orbitofrontal cortex. Nature 398: 704–708. [DOI] [PubMed] [Google Scholar]
- Tsushima Y, Seitz AR, Watanabe T. (2008) Task-irrelevant learning occurs only when the irrelevant feature is weak. Current Biology 18(12): R516–R517. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tueting P, Sutton S, Zubin J. (1971) Quantitative evoked potential correlates of the probability of events. Psychophysiology 7(3): 385–394. [DOI] [PubMed] [Google Scholar]
- Tyler CW, Kontsevich LL. (1995) Mechanisms of stereoscopic processing: stereoattention and surface perception in depth reconstruction. Perception 24: 127–153. [DOI] [PubMed] [Google Scholar]
- Ungerleider LG, Mishkin M. (1982) Two cortical visual systems: Separation of appearance and location of objects. In: Ingle DL, Goodale MA, Mansfield RJW. (eds) Analysis of Visual Behavior. Cambridge, MA: MIT Press, pp. 549–586. [Google Scholar]
- Unsworth N, Robison MK. (2015) Individual differences in the allocation of attention to items in working memory: Evidence from pupillometry. Psychological Bulletin and Review 22(3): 757–765. [DOI] [PubMed] [Google Scholar]
- Vitevitch MS, Luce PA. (1999) Probabilistic phonotactics and neighborhood activation in spoken word recognition. Journal of Memory and Language 40(3): 374–408. [Google Scholar]
- Von Restorff H. (1933) Über die Wirkung von Bereichsbildungen im Spurenfeld [The effects of field formation in the trace field]. Psychologie Forschung 18: 299–234. [Google Scholar]
- Warden MR, Miller EK. (2010) Task-dependent changes in short-term memory in the prefrontal cortex. Journal of Neuroscience 30(47): 15801–15810. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Warren R, Sherman A. (1974) Phonemic restorations based on subsequent context. Perception & Psychophysics 16(1): 150–156. [Google Scholar]
- Webster MJ, Bachevalier J, Ungerleider LG. (1994) Connections of inferior temporal areas TEO and TE with parietal and frontal cortex in macaque monkeys. Cerebral Cortex 4: 470–483. [DOI] [PubMed] [Google Scholar]
- Weyandt LL, Oster DR, Marraccini ME, et al. (2014) Pharmacological interventions for adolescents and adults with ADHD: Stimulant and nonstimulant medications and misuse of prescription stimulants. Psychology Research and Behavior Management 7: 223–249. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wilson RC, Takahashi YK, Schoenbaum G, et al. (2014) Orbitofrontal cortex as a cognitive map of task space. Neuron 81: 267–279. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wise SP. (2008) Forward frontal fields: Phylogeny and fundamental function. Trends in Neurosciences 31(12): 599–608. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Witcher ES. (1978) Extinction of Pavlovian conditioned inhibition. Unpublished Doctoral Dissertation, University of Massachusetts Amherst, Amherst, MA. [Google Scholar]
- Wolfe JM. (1994) Guided search 2.0: A revised model of visual search. Psychonomic Bulletin & Review 1(2): 202–238. [DOI] [PubMed] [Google Scholar]
- Wolfe JM. (1998) What do 1,000,000 trials tell us about visual search? Psychological Science 9(1): 33–39. [Google Scholar]
- Wolfe JM, Cave KR, Franzel SL. (1989) Guided search: An alternative to the feature integration model for visual search. Journal of Experimental Psychology: Human Perception and Performance 15(3): 419–433. [DOI] [PubMed] [Google Scholar]
- Yang S, Heinen S, Missal M. (2008) The effects of microstimulation of the dorsomedial frontal cortex on saccade latency. Journal of Neurophysiology 99(4): 1857–1870. [DOI] [PubMed] [Google Scholar]
- Zhuang M, Oosting RS, Jones SR, et al. (2001) Hyperactivity and impaired response habituation in hyperdopaminergic mice. Proceedings of the National Academy of Sciences of the United States of America 98(4): 1982–1987. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zimmer-Hart CL, Rescorla RA. (1974) Extinction of Pavlovian conditioned inhibition. Journal of Comparative and Physiological Psychology 86(5): 837–845. [DOI] [PubMed] [Google Scholar]
- Zoccolan D, Kouh M, Poggio T, et al. (2007) Trade-off between object selectivity and tolerance in monkey inferotemporal cortex. Journal of Neuroscience 27(45): 12292–12307. [DOI] [PMC free article] [PubMed] [Google Scholar]