Skip to main content
Philosophical Transactions of the Royal Society B: Biological Sciences logoLink to Philosophical Transactions of the Royal Society B: Biological Sciences
. 2014 Nov 5;369(1655):20130483. doi: 10.1098/rstb.2013.0483

The why, what, where, when and how of goal-directed choice: neuronal and computational principles

Paul F M J Verschure 1,2,, Cyriel M A Pennartz 3, Giovanni Pezzulo 4
PMCID: PMC4186236  PMID: 25267825

Abstract

The central problems that goal-directed animals must solve are: ‘What do I need and Why, Where and When can this be obtained, and How do I get it?' or the H4W problem. Here, we elucidate the principles underlying the neuronal solutions to H4W using a combination of neurobiological and neurorobotic approaches. First, we analyse H4W from a system-level perspective by mapping its objectives onto the Distributed Adaptive Control embodied cognitive architecture which sees the generation of adaptive action in the real world as the primary task of the brain rather than optimally solving abstract problems. We next map this functional decomposition to the architecture of the rodent brain to test its consistency. Following this approach, we propose that the mammalian brain solves the H4W problem on the basis of multiple kinds of outcome predictions, integrating central representations of needs and drives (e.g. hypothalamus), valence (e.g. amygdala), world, self and task state spaces (e.g. neocortex, hippocampus and prefrontal cortex, respectively) combined with multi-modal selection (e.g. basal ganglia). In our analysis, goal-directed behaviour results from a well-structured architecture in which goals are bootstrapped on the basis of predefined needs, valence and multiple learning, memory and planning mechanisms rather than being generated by a singular computation.

Keywords: goal-directed behaviour, distributed adaptive control, computational modelling, embodied cognition, reward, decision-making

1. Introduction

Instrumental actions can be either habitual or goal-directed. In order to label an action goal-directed, it must satisfy two requirements. First, the agent must display knowledge of the causal efficacy of its own actions and their outcomes given the current state or context. Second, the agent must select and regulate its behaviour using goal representations, e.g. internally generated representations of desired action outcomes. In other words, such deliberate action is directed towards specific states of the external world, i.e. goals, motivated by states of the internal environment and mediated by internal representations of these goals. Goal-oriented action expresses agency, mental states and is intentional following the nineteenth century philosopher Franz Brentano. A fundamental nature challenge in psychology and neuroscience is to develop a coherent explanation of goal-directed action. This has turned out to be difficult as the history of the study of behaviour illustrates. During behaviourism, constructs such as ‘intention’ or ‘goal’ with their associated teleology were seen as unscientific and were thus eliminated. With the cognitive revolution, these notions were again incorporated in the terminology of mind and brain [1] but the mechanistic functioning and neuronal underpinnings of goal-directed behaviour remained largely unknown. In the perspective of symbolic artificial intelligence (AI), knowledge and goals are organized following a principle of rationality: ‘… if the system wants to attain goal G and knows that to do act A will lead to attaining G, then it will do A. This law is a simple form of rationality that an agent will operate in its own best interest according to what it knows’ [2, p. 49]. In this case, however, the problem was that ‘goals’, ‘knowledge’ and ‘actions’ were all defined a priori and the explanation of cognition limited itself to the operations that could be performed on these symbolic representations. No commitments were made to their etiology and the question thus becomes how an agent can autonomously acquire or define these core elements of cognition [3].

The explanation of goal-oriented behaviour is not only of interest because of the complexity of the associated phenomena but also because it is at the edge of phenomenology and science. As much as there is a ‘hard problem’ in the explanation of consciousness [4,5], we can argue that a similar explanatory gap exists in understanding goal-oriented behaviour, it also critically depends on inferences by the observer on the beliefs an organism entertains with respect to its goals and tasks. The fundamental challenge is to explain this phenomenon while acknowledging the full richness of its intentional aspects, avoiding the behaviourist fallacy. Conversely, our explanation should include the genesis of goals as opposed to assuming them a priori to avoid the trap of symbol grounding of traditional AI. This paper tries to directly answer these challenges by addressing the notion of goals and their role in cognition and action from an integrative perspective building on three components: (i) an evolutionarily motivated hypothesis on brain function and action; (ii) an embodied theory of mind and brain called distributed adaptive control (DAC), and (iii) a detailed system-level analysis of the neuronal substrates of goal-oriented behaviour in the rodent. In this integration, we take as our specific question whether goals should be seen as single scalar functions against which behaviour is optimized or whether they are multi-dimensional processes that emerge from the interaction across a number of perceptual, affective, cognitive and motor systems (see also [6]). We advocate the latter interpretation and show that goals and their impact on ongoing and future behaviour should be seen as resulting from a process that plays out at multiple levels of the neuraxis following distinct principles.

Goal-directed behaviour is unique and distinct from other forms of control such as innate reflexes and habits in the sense that it does not prescribe a specific operation or procedure but rather the end state that an operation should achieve. Goal-directed choice permits an agent to escape from stereotyped interactions with a predictable environment and flexibly and rapidly adapt to complex and dynamic internal (e.g. motivational) and environmental conditions using abstract allocentric procedures and prospection. This fact is epitomized in experimental procedures such as reward devaluation [7]. Here, a rat is first trained to press a lever to obtain a food reward. After reaching the learning criterion, the rat is given the same reward but now ‘devaluated’, e.g. coupled to a nausea inducing treatment. When the rat is subsequently placed again in front of the lever, it can either display the habit of the previously acquired lever-pressing action or display goal-oriented (in)action by avoiding to press the lever. Rats are able to suppress the acquired response. This example illustrates the two hallmarks of goal-directed choice: agency and intentionality. It also shows that goal-directed choice is a systems property that depends on a number of tightly coupled processes including perception, motivation, emotion, cognition and action. It cannot be localized to a central ‘goal nucleus’ in the brain but rather depends on the interplay of a number of mechanisms realized in several brain areas. As a result, understanding goal-directed choice requires a systems-level architectural treatment and recent advances have clarified the possible contributions of specific components of this architecture [8,9]. However, it is less clear how the contributions of these different brain areas are orchestrated to generate actions that lead to goal achievement in the real world and how goal-directed mechanisms coexist with other non-goal-based ones. Here, we propose a systems-level solution to the challenge of goal-directed choice from combined experimental neurophysiological and theoretical perspectives.

In order to create structure in the tangle of neuronal processes and sub-processes that make up the brain and their multi-level organization, we need to define unambiguously what the overall function of this system is. Here, we follow Claude Bernard and Ivan Pavlov in defining the brain as a control system that maintains a metastable balance between the internal world of the body and the external world through action. The question thus becomes: ‘does it take to act?’ We propose that in order to act in the external world, the brain needs to optimize a specific set of objectives which are captured in answering the questions: ‘Why do I need to act? What do I need? Where and When can this be obtained and How do I get it?’ These questions harbour a complex set of computational challenges that can be defined as the H4W problem [10]. In short, an animal needs to determine a behavioural procedure to achieve a goal state (the How of action), which in turn requires defining the ‘Why’ (the motivation for action in terms of needs, drives and goals), ‘What’ (the objects and their affordances in the world that pertain to these goals), ‘Where’ (the location of objects in the world, the spatial configuration of the task domain and the location and confirmation of the self) and ‘When’ (the sequencing and timing of action relative to the dynamics of the world and self). We propose that goal-oriented action in the physical world emerges from the interplay of the different processes subserving H4W. Action in a social world would also require processing ‘Who’, but this aspect will not be dealt with here [11].

Each of the Ws can be seen as a specific objective that the brain must satisfy, to which it designates a large set of sub-objectives of varying complexity, which are laid out in parallel across different levels and scales of organization of the central nervous system. At a first level, the brain must assess the motivational states derived from homeostatic self-essential variables. These motivational states in turn need to be prioritized so that goals can be set: this is the ‘Why’ problem, requiring the modulation of associated behaviour systems. Next, a second layer of control is called for to classify, categorize and valuate states of the world, to identify the spatial layout of the task, including the agent itself, and the dynamics of the task and its affordances: ‘What’, ‘Where’, ‘When’. Lastly, these labelled multi-modal states are grouped in sequences around prioritized goals; for example, in a rodent navigation set-up, to go towards and push a lever, placed at the northeast corner of the environment, given that the cue signal has appeared. At this stage, the ‘How’ has been generated and expressed. Using the accumulated spatio-temporal knowledge of the task and the self in which goal pursuit is framed, a procedural motor strategy (‘How’) can be composed and its elements selected from the set of available options to achieve a goal state. We propose that the H4W framework outlined above is an exclusive set of processes that captures the essential brain mechanisms that mediate and control instrumental interaction with the physical world. We will now show how H4W maps onto a biologically grounded cognitive architecture that will further facilitate to show how we can make sense of the intricate neuronal substrates of goal-oriented behaviour.

(a). The distributed adaptive control theory of mind and brain

To map the functional H4W hypothesis to potential neuronal substrates of goal-oriented behaviour, we first decompose it in terms of a hypothetical neuronal architecture of the mammalian brain called DAC1 (figure 1, see [10] for a review). The DAC theory emphasizes that perception, emotion, cognition and action are realized through an integrated control system that is both embodied and situated in the real world. In this perspective, goal-directed choice is not a monolithic component but depends critically on the interaction between several layers of control, each of which uses specific information (e.g. motivational and sensory versus memory and prospection) to generate and maintain goal representations at varying levels of abstraction that cooperate and compete for the control of action [13]. Thus, from the DAC perspective, drives and goals are necessarily organized in a hierarchical fashion, starting with the concrete needs defined by the physical instantiation of the agent, i.e. the nutrients required to maintain the body, to the abstract goals of reaching specific but also abstract goal states, such as for example, having dinner in a specific restaurant or becoming an expert on Byzantine culture. The latter are seen as being bootstrapped on the basis of the former. Below we introduce the layered organization of DAC and discuss how it constitutes a coherent architectural solution to the H4W problem of goal-directed behaviour.

Figure 1.

Figure 1.

The DAC theory of mind and brain (see [10] for a review). Left: highly abstract representation of the DAC architecture. DAC proposes that the brain is organized as a three-layered control structure with tight coupling within and between these layers distinguishing: the soma (SL) and the reactive (RL), adaptive (AL) and contextual (CL) layers. Across these layers, a columnar organization exists that deals with the processing of states of the World or exteroception (left, red), the self or interoception (middle, blue) and action (right, green). See text for further explanation. The reactive layer: the RL comprises dedicated behaviour systems (BS) that combine predefined sensorimotor mappings with drive reduction mechanisms that are predicated on the needs of the body (SL). Right lower panel: each BS follows homeostatic principles supporting the self-essential functions (SEF) of the body (SL). In order to map needs into behaviours, the strength of the essential variables served by the BSs, SEFs, have a specific distribution in task-space called an ‘affordance gradient’. In this example, we consider the (internally represented) ‘attractive force’ of the home position supporting the security SEF or of open space defining the exploration SEF. The values of the respective SEFs are defined by the difference between the sensed value of the affordance gradient (red) and its desired value given the prevailing needs (blue). The regulator of each BS defines the next action as to perform a gradient ascent on the SEF. An integration and action selection process across the different BSs forces a strict winner-take-all decision that defines the specific behaviour emitted. The allostatic controller of the RL regulates the internal homeostatic dynamic of the BSs to set priorities defined by needs and environmental opportunities through the modulation of the affordance gradients, desired values of SEFs and/or the integration process. The adaptive layer: the AL acquires a state space of the agent–environment interaction and shapes action. The learning dynamic of AL is constrained by the SEFs of the RL that define value. The AL crucially contributes to exosensing by allowing the processing of states of distal sensors, e.g. vision and audition, which are not predefined but rather are tuned in somatic time to properties of the interaction with the environment. Acquired sensor and motor states are in turn associated through the valence states signalled by the RL. The contextual layer: the core processes of the CL are divided between a task-model and a self-model. The CL expands the time horizon in which the agent can operate through the use of sequential short-term and long-term memory (STM and LTM) systems respectively. These memory systems operate on integrated sensorimotor representations that are generated by the AL and acquire, retain and express goal-oriented action regulated by the RL. The CL comprises a number of processes (right upper panel): (a) when the discrepancy between predicted and encountered sensory states falls below a STM acquisition threshold, the perceptual predictions (red circle) and motor activity (green rectangle) generated by AL are stored in STM as a, so-called, segment. The STM acquisition threshold is defined by the time-averaged reconstruction error of the perceptual learning system of AL. (b) If a goal state (blue flag) is reached, e.g. reward or punishment, the content of STM is retained in LTM as a sequence conserving its order, goal state and valence marker, e.g. aversive or appetitive, and STM is reset. Every sequence is thus labelled with respect to the specific goal it pertains to and its valence marker. (c) If the outputs generated by the RL and AL to action selection are sub-threshold, the AL perceptual predictions are matched against those stored in LTM. (d) The CL selected action is defined as a weighted sum over the segments of LTM. (e) The contribution of LTM segments to decision-making depends on four factors: perceptual evidence, memory chaining, the distance to the goal state and valence. Working memory (WM) of the CL is defined by the memory dynamics that represents these factors. Active segments that contributed to the selected action are associated with those that were previously active establishing rules for future chaining. The self-model component of the CL monitors task performance and develops (re)descriptions of task dynamics anchored in the self. In this way, the system generates meta-representational knowledge that forms autobiographical memory. This aspect of the DAC CL is not further considered in this paper.

The somatic level (SL) of DAC designates the body itself and defines three fundamental processes: exosensing of states of the environment, endosensing of states of the body or essential variables of survival, defining needs and actuation through control of the skeletal-muscle system. Behaviour is defined as a change in the confirmation and/or position of the SL.

The reactive layer (RL) of DAC supports the basic functionality of the SL and generates control signals that drive and modulate the engagement of higher control layers and their epistemic functions [14] (figure 1, lower right panel). These sensorimotor loops are organized in fundamental and opposing behaviour systems that support basic needs such as fight, flight [15], seek and play [16]. Every behaviour system is homeostatic but importantly their collective properties are in turn regulated by an integrative allostatic loop [17,18]. This allostatic orchestration (see figure 1) is critical to goal-oriented behaviour because both at a behavioural and physiological level, different homeostatic subsystems are in a competitive relationship and priorities and hierarchies must be established on the fly dependent on internal needs, external threats and opportunities. In order to control behaviour, the states of the behaviour systems have a specific neural representation that distributes them across space, called an affordance gradient, such as the ‘attractive force’ of the home position or the ‘repellant force’ of a predator, that thus implicitly encodes action in an egocentric frame of reference [19]. DAC thus sees the reactive agent as operating in a composite behavioural field defined by the dominant behavioural subsystems and their internal representation of affordance gradients. This mechanism links well to the idea of force-field-based control as used in the individual behaviour systems (BS) of RL. This idea has its roots in Gestalt psychology [20] and is a standard approach towards robot navigation [21] but generalizes towards optimal path planning in mobile robots [22] and social interaction in humanoid robots [23].

The adaptive layer (AL) of DAC extends the predefined need-reducing sensorimotor loops of the RL with value-dependent acquired sensor and action states. It allows the agent to escape from the strictly predefined and instantaneous reflexes of RL through learning [24,25]. The AL constructs a state space encoding of both the external (world) and internal (self) a priori unpredictable environment and shapes the amplitude-time course of the predefined RL reflexes. It crucially relies on distal sensors, e.g. vision and audition. The acquired sensor and motor states are in turn associated through the valence states triggered by the RL, following the paradigm of classical conditioning where initially neutral or conditioned stimuli (CS) obtain the ability to trigger actions, or conditioned reponses (CR), by virtue of their contingent presentation with intrinsically motivational stimuli or unconditioned stimuli (US) [12]. In particular, the AL explains the principles underlying the two-phase theory of classical conditioning [26] combining the, so-called, non-specific and specific learning systems into one integrate behavioural architecture. DAC's AL predicts that the former plays a key role in value-dependent stimulus identification in addition to the preparation for action, while the later is dependent on the former and critical for timing rather than ‘consumption’ [27]. The AL allows the agent to overcome the predefined behavioural repertoire of the RL and to engage an a priori unpredictable world, solving the notorious symbol grounding that led to the demise of classical AI [3,28].

The learning dynamics of the AL are defined in terms of minimization of the prediction error between acquired and encountered states of the world relying on local Hebbian learning [24]. This approach has been rephrased in a general formal framework called correlative subspace learning (CSL) where the associations between perceptual states and action are mediated via value representations and perception-value associations are formed on the basis of both perceptual and behavioural prediction [25]. CSL is consistent with ‘predictive brain’ frameworks and ‘free energy minimization’ principles [29,30]. Importantly, DAC demonstrates these principles within an embodied cognitive architecture. Experiments with AL have three important consequences for our understanding of goal-oriented behaviour. First, prediction-error minimization is a multi-scale process that needs to operate in concert with dedicated processes for behaviour control and exploration provided by the RL. Second, prediction-based learning is crucial in real-world behaving systems in order to counteract behavioural feedback, i.e. non-neuronal feedback resulting from both the high spatio-temporal correlation in sensory streams combined with experience-dependent biased sampling of the sensory space [13,24]. Third, the strict distinction between retrospective model-free and prospective model-based decision-making is misleading in the sense that in both cases, the state space on which decision-making operates is defined based on models albeit of varying spatio-temporal complexity [31].

AL extends the need-reducing sensorimotor loops of RL into acquired sense-valence-act triplets that include a priori unknown but now acquired states of the world and the self. This adaptation occurs in a restricted temporal window of relatively immediate interaction, i.e. up to about 1 s, which is the range in which most forms of classical conditioning operate. DAC predicts that this temporal boundary is dependent on the characteristic time constants of cerebellar learning loops [32]. However, in order to escape from the ‘now’ and value-dependent interaction with the world and develop goals in an expanded spatio-temporal window, more advanced memory systems must be engaged: the contextual layer (CL).

The CL of DAC acquires, retains and expresses behavioural plans by combining sequences of AL-defined sensorimotor states with goals in a value-dependent way (figure 1, upper right panel). The CL comprises a dual representational system: one for representations of tasks and the other for representing the self. The former comprises systems for short-term, long-term and working memory (STM, LTM and WM, respectively), while the latter combines a monitoring system of the Task Space with an autobiographical memory. The task space memory systems allow for the formation of sequential representations conditional on the goal achievement of the agent. CL behavioural plans can be recalled through sensory matching and internal chaining among the elements of the retained memory sequences (see figure 1). The dynamic states that this process entails define DAC's WM. The CL organizes LTM along behavioural goals and previous studies have assessed that this together with valence labelling of LTM segments is required, in order to obtain a Bayesian optimal solution to foraging problems [3]. Goals are formed through the integration of sensorimotor states with the termination conditions of behavioural patterns defined through need-reduction and value systems of the RL and AL, respectively. Goals are initially defined in terms of the drives that are guiding the behaviour systems of the RL such as finding a food item, or solving an impasse, i.e. flight, and the valence that they are associated with at the AL. Through learning, goal states can expand to include the sensorimotor states that brought about the change in drive and valence forming a multi-modal construct. Goal states, as termination points of acquired behavioural procedures, together with the behavioural sequence itself exert direct control over how decision-making and action selection is performed.

DAC proposes a four-factor decision-making model that uses: perceptual evidence, memory chaining, the predicted distance to the goal state, or goal fidelity, and the expected value. This four-factor decision-making model of CL predicts that decision-making will display both goal and behavioural procedure fidelity [33,34]. This means that potential actions are weighted both with respect to their distance to the current goal state as measured in the number of events between the memory element and the occurrence of the goal state, and whether they form part of an active behavioural sequence, where active sequences are triggered through perceptual evidence and memory priming processes. This expansion of the decision-making towards goals frees the agent from acting in the restricted temporal window of the AL, its egocentric frame of reference and its dependence on immediate sensory states. Rather, memory allows for the organization of behaviour along allocentric coordinates, prospection and symbolic representations that do not depend on available sensory states. This transition towards allocentric goal-oriented navigation leads to robust robot multi-modal (proximity sensors, chemosensing and vision) foraging including recovery from kidnapping and noise [35] and optimal maze navigation in the presence of distracters [36]. These robot experiments with DAC show that the ability to navigate using allocentric coordinates is closely coupled with the use of goals and goal fidelity. Moreover, the notion of goal fidelity suggests that goals cannot be seen as single scalar values that act as discrete organizers of action (e.g. in the form of a single reward signal), but are defined through the confluence of congruent sensory, motor and value information at the termination points of behavioural sequences, i.e. the sequences formed in LTM. These goal states also exert an implicit influence on decision-making by defining a goal-dependent metric that measures the relevance of specific memory elements to the current task (i.e. goal fidelity). When action depends solely on perceptual evidence and egocentric representations, any variation of these input states (due to occlusions, noise, movement-induced variation, etc.) translates into variability in action selection leading to a nonlinear amplification of behavioural variability and thus the deviation from previously executed and acquired trajectories. This so-called behavioural entropy precludes the agent from reusing acquired egocentrically represented behavioural plans because predicted sensory states will not match encountered ones. To demonstrate how allocentric goal-oriented behaviour enhances fitness by overcoming behavioural entropy, experiments were performed where egocentric action encoding was compared with an allocentrically defined movement vector that points from the current position of the agent to the location of a next landmark, i.e. allocentric goal-based action [37,38]. Using a range of maze learning benchmarks, it was shown how this approach leads to robust maze navigation and learning even when significant noise is applied to the motor output [33]. This study also illustrates the relationship between taxon versus route navigation strategies [39] in a robotic system. The reactive egocentric control system can fulfil the agent's needs when the information from the goal site can be directly detected, there is a direct path between current and goal positions (taxon-based strategy) and approaching it does not conflict with any other behaviour system. However, when a taxon-based strategy fails or relevant landmarks cannot be detected, the contextual control system is required to generate a route using an allocentric navigation method relying on acquired goal states. On the basis of these observations, DAC predicts that AL is focused on real-time egocentric interaction with the world emphasizing interval timing, while CL abstracts its processing towards symbolic events and their order supporting allocentric interaction with a task.

The self-model of the CL includes monitoring and autobiographical memory systems. The former prevents the consolidation of behavioural sequences that are not causally related to goal achievement defining an agency condition for CL memory formation. The later is an episodic memory system that memorizes self-generated goal-oriented behaviours and its context and centres on the self and its agency as opposed to the task and the world in which it is realized.

In summary, the DAC theory introduced proposes that the elements of H4W are processed at each layer of the neuraxis organized along World (What, Where, When), Self (Why) and Action (When, Where, How). Here, ‘Where’ and ‘When’ appear twice to emphasize the distinction between Self and external objects. As we advance along this hierarchy, the agent becomes less dependent on immediate states of the world and relies more on memory increasingly constrained by states of self, from needs to values to goals, in an expanding spatio-temporal horizon. In this hierarchy, the RL need-reduction systems define the value systems of the AL, which in turn constrain the behavioural plans formed by the CL. By combining need reduction and value with the sensorimotor states of the task at hand, goals emerge as CL-based representations of termination points of acquired behavioural procedures that can operate in an allocentric frame of reference freeing the agent from the here-and-now and propelling it towards a goal-oriented future.

2. The neurobiology of drives and goals according to the H4W-DAC taxonomy

The H4W taxonomy and its realization in the DAC architecture suggest a three-level organization of the central nervous system. Below we discuss how each of these three levels of organization is neurally implemented in the mammalian brain, reviewing mainly rodent research (figure 2).

Figure 2.

Figure 2.

Proposed brain architecture representing the neuronal substrate of goal-directed behaviour and its relation to the neurorobotic DAC architecture. The hippocampal formation is proposed to code the organism's world state space (red), the prefrontal cortex (especially its medial and orbitofrontal aspects) to represent task space (i.e. rules, constraints, goals and values of cues and action options, purple) and the striatum (and downstream structures of the basal ganglia) to mediate action selection. In this scheme, the hypothalamus and brain stem contain sensor systems monitoring homeostatic variables and providing information about the motivational needs of the organism that define the pursuit of needs and goals (blue). The arrow from striatum to thalamus represents an indirect projection. The hypothalamic efferents are modelled after those traced for the lateral hypothalamus in relation to feeding behaviour and do not apply to hypothalamic areas in general. For the sake of clarity, the scheme's anatomic connections are by no means complete. For instance, outputs from prefrontal cortex and basal ganglia to the brain stem, or several afferent inputs to amygdala and VTA, have not been included, while several hypothalamic nuclei project directly to ventromedial prefrontal areas. Sensory inputs reach the hippocampus via intermediate stations (parahippocampal areas; not shown) and are supplemented with frontal cortical inputs converging on these intermediate areas. Furthermore, the motor cortices are meant to include premotor, supplementary motor and frontal oculomotor areas (based on [4042]).

(a). Reactive layer: neural mechanisms encoding drives in relation to survival and reproduction

Perhaps the most well-known brain region for controlling motivational drives is the hypothalamus, a phylogenetically ancient, diencephalic structure well connected to sensor and actuator systems in lower CNS centres such as the brain stem, spinal cord and autonomic ganglia. For goal-directed behaviour, both the sensor and effector functions of the hypothalamus are critical. For instance, the preoptic area contains thermosensitive neurons that, at the same time, are involved in the generation and dissipation of body heat [43]. The supraoptic nucleus harbours cells that gauge the osmolality of blood plasma [44], and at the same time this nucleus maintains homeostasis of blood salt levels. Similarly, hypothalamic cell groups are thought to monitor nutrient levels and guard the body's energy balance, while others regulate sexual, maternal and aggressive behaviour as well as sleep (e.g. [4549]). This list of hypothalamic sensor–actuator functions is by no means exhaustive and can be supplemented with numerous brain stem-medulla nuclei that are often positioned even closer to internal sensors and effectors (e.g. the monitoring and regulation of food intake, respiratory and cardiovascular reflexes by the nucleus tractus solitarius, vagal nuclei and connected cell groups). In addition, these functions address not only homeostatic regulation, but also allostasis, referring especially to responses to challenges that require system-wide, dynamic adaptation and predictive regulation in anticipation of upcoming homeostatic disturbances [17,18].

For goal-directed behaviour, a first point to note is that drives and their associated behavioural expression, such as aggression, thirst, hunger, sex, sleep and safety, are determined in basic form at the level of the hypothalamus together with lower-order structures [50], analogous to the functioning of the RL in DAC. A drive arises from the discrepancy between a read-out of a homeostatic parameter (e.g. blood sugar level) and an optimal set point, although for some types of drives the neural basis underlying this comparison is not that clear yet (e.g. for ‘sleep pressure’, e.g. [18,51]). Second, a definition of drive purely in terms of internal homeostasis would be too limited. For instance, organisms will explore novel environments and objects even if they thereby leave their shelter and decrease their own safety level, while no explicit reward is ensured. It is still largely unknown how such a drive for novel information or exploration (or reduction of uncertainty [5255]) would be realized by the brain, although some higher-level control systems have been implicated (e.g. prefrontal cortex and hippocampus; see below). In addition, the question is whether we need to distinguish different levels of drives from ‘basic and innate’ to ‘derived and acquired’ arranged in a hierarchical fashion, with the associated question of how many of such drives must be identified. It is especially with an eye to optimizing the dynamic equilibrium between the multiple competing needs of the organism and the dynamics of its environment that the notion of allostasis has been proposed. Third, our relatively advanced knowledge of homeostatic ‘reflexes’ (i.e. automated sensor–effector reactions to specific disturbances of homeostatic equilibria) has done little to inform us about mechanisms to prioritize particular drives over others. For instance, if an animal is hungry, thirsty and exhausted at the same time, it is less than clear how the brain prioritizes a particular drive as being targeted for satisfaction and deals with conflict resolution. A chosen priority will not only depend on the strength of the drive but also on the context of available options, confidence in available solutions, etc. prompting the organism to estimate how likely a goal (e.g. quenching of thirst) can be achieved given the environment and its affordances (dry or rainy). The DAC architecture solves this challenge relying on allostatic control in the service of need reduction (figure 1).

Returning to the notion of different levels for organizing goal-directed behaviour, we will review evidence implying higher brain structures in the neural implementation of what we have described as its core characteristic: the prospective evaluation of possible state (situation) and action outcomes rather than fixed stimulus–response rules. As mentioned in the Introduction, goal-directedness implies that the organism has implicit or explicit knowledge about its actions being causal in the chain of events leading to the desired goal, i.e. agency. Furthermore, if the value/utility of a goal changes, for instance because of a change in motivational state, the organism should refrain from persistently conducting a fixed stimulus–response sequence or habit, but must adjust its response according to the change in value [7]. Given this definition of goal-directedness and the limitations of hypothalamic-brain stem systems in this respect owing to their reliance on fixed action patterns, there is a need for ‘higher’ systems to be informed about drives expressed by these ‘lower’ reactive levels. This upward projection of the drive state implements a reactive-to-adaptive, flexible type of organization that grounds the affective component of goals and can be realized, among others, through hypothalamic projections to the thalamus, prefrontal cortex, hippocampus and related structures [41].

(b). State representations conducive to goal-directed behaviour: hippocampus

Currently, it is becoming clearer that, when an animal is strongly committed to obtaining a particular goal, information processing in many brain areas is simultaneously affected by inputs that are predictive of, or conducive to, this goal. This effect is more widespread and pervasive than could be captured by the simplification that ‘reward centres’ in the brain are activated. Even structures such as posterior parietal cortex and primary sensory cortex (A1, V1) are deeply affected by associative stimulus–reward and action–reward learning [5662]. Despite this ubiquity, there are good arguments to highlight the roles of hippocampal-prefrontal systems in forming state representations that can be used by action selection systems executing goal-directed behaviour. Put succinctly, when the needs of an agent (‘Why’) have been set at the level of the hypothalamus and brain stem, representations of the state of the world (including the agent's own state) are required to determine where and when this need may be satisfied, and through which particular object (‘What’) within a feasible spatio-temporal range (e.g. an apple to satisfy the need for particular nutrients).

Over the past decades, two classical views of the hippocampal system have been elaborated, both of which are currently in need of updating with respect to goal-directed behaviour. The first view holds that the hippocampus codes an agent's position in space, as inspired by the Cognitive Map theory of Tolman [63] as well as the body of place-cell research initiated by O'Keefe & Dostrovsky [64]. The second notion accounts for hippocampal function as a ‘recorder’ of experience—an organ for the formation of episodic memories that would be transferred, with the passage of time, to neocortical sites where also generalization (semanticization) of memory may take place [65]. The update required on the first view is that a large body of recent research indicates that the hippocampus codes not only for an agent's self-location, but also for specific objects and events—e.g. neutral environmental cues such as odours, and cues that predict reward, as well as the time spent in a situation conducive to goal pursuit [6670]. Moreover, the representation of the task seems to follow a multiplexing of input streams combining sensory, location and action information at both the input and memory stages of hippocampal processing wood [66,7173]. This latter observation is in line with a prediction from the DAC theory, viz. that goal-oriented behaviour is constructed from conjunctive sensorimotor couplets (figure 1). If true, this would mean that modulation of sensory cues should lead to a significant change in the population response of hippocampal neurons [74]. This effect was indeed directly observed in, so-called, rate remapping in environments that were morphed [75] and abolished when sensory cues to the hippocampus are removed [76].

The update on the second notion of the recording of experience is that the hippocampus has turned out not only to record (i.e. encode and store) spatial experiences, but also—and more generally—chains of associated events and sequences of motor actions [7779]. Indeed, the DAC theory predicts that sensorimotor couplets are constructed and combined into sequential representations in STM. The hippocampus seems to display both sensory and motor features. How these chains are formed is unknown, but in the case of motor sequences the information is hypothesized to reach the hippocampus via thalamocortical sensory systems (e.g. somatosensory, proprioceptive and vestibular), where novel information can be associated with retrieved memories to expand these into longer chains. Importantly, the hippocampus appears only to be required for longer, or more complex, sequences as more simple stimulus–response associations (and ensuing habits) can be acquired via the dorsal striatal system [80,81].

In addition to storing information, the hippocampus is able to retrieve previously stored information and to self-generate internal sequences of cell activity that are subsequently used to map novel environments or situations [8284]. Self-initiated retrieval is thought to take place as ‘replay’ mainly occurring during hippocampal local-field potential (LFP) events called sharp-wave ripples [8588], but also during bouts of LFP theta-band activity during running and other ongoing behaviour, resulting in ‘forward sweeps’ of hippocampal place representations ahead of the animal, at choice points in the environment [89]. Thus, during ongoing goal-directed behaviour, as well as ‘off-line’, the hippocampus has multiple modes to (re-)generate and recall information from memory, which can be flexibly used to guide decision-making and/or support consolidation [90,91].

The upshot of these recently emerging insights is to regard the hippocampal system in a broader sense than was hitherto the case, namely as a system for representing the current state of the world including the agent's, incorporating many types of causal and/or non-causal spatio-temporal relationships. The hippocampus is not passively storing episodic memories but rather actively storing information (based on synaptic plasticity) and retrieving information (by way of replay and theta-sequenced firing), where the switch between these two modes can be rapidly and flexibly made, depending on the current needs of the organism, e.g. in planning its behaviour or in consolidating previous experiences. Overall, the strong confluence of sensory thalamocortical information onto the hippocampus—as contrasted with frontal motor information which reaches it more indirectly—emphasizes that this structure is more concerned with the representation of current states of the environment and the organism, and their multisensory derivatives such as place (Where) and time (When), then being confined to the representation of an action or task space per se. However, the inputs to the hippocampus from the grid cells of the medial entorhinal cortex can be interpreted as representing heading direction and thus action, a source of information that in turn strongly dominates memory dynamics [72]. Having said this, the hippocampus engages as well in the representation of goal sites, and these behaviourally significant states are probably encoded with greater density and/or spatial resolution than neutral locations [68,90,92]. These goal sites play an important role in goal-directed choice but according to DAC, they need to be complemented with other information to solve the full H4W problem, and most importantly sequenced rule-based goal-oriented plans for action generated at the level of the CL.

(c). Task representations conducive to goal-directed behaviour: prefrontal cortex

Leaving aspects of prefrontal functions in the cognitive control over memory processes in primates aside, there is accumulating evidence to cast prefrontal functions as controlling goal-directed behaviour (e.g. [93]). Whereas the hippocampus is proposed to engage in world state representations, the prefrontal cortex is more concerned with task- and action-space representations. Neurophysiological studies in rodents and primates indicate that prefrontal neurons can encode task rules that need to be followed to obtain a goal [94], individual actions or chunks of actions leading up to a goal [9597], and goals and goal sites themselves [98,99]. Importantly, orbitofrontal and medial prefrontal-anterior cingulate neurons are sensitive to the motivational value of cues [100104] and actions associated with goal pursuit [105,106]. Lesion studies have confirmed a causal involvement of prefrontal structures in representing goals and task rules, implying orbitofrontal cortex in reversal learning and medial prefrontal cortex in both extra-dimensional shifting [107,108] and learning action–outcome relationships [7,109]. Thus, whereas the hippocampus is proposed to represent objects and events relevant for pursuing a goal in space and time (What, When and Where), the prefrontal cortex appears better equipped to represent a task space, i.e. the set of rules, constraints, goals and goal-predictive values of cues and actions available as options to pursue goals (How) (cf. [110]). In this context, the amygdaloid complex should be included as part of a larger network for affectively driven goal-directed behaviour not only mediating value based Pavlovian response behaviours, but should also be goal-oriented instrumental behaviours which can be invigorated by Pavlovian cues [111,112].

The dominant paradigm to investigate goal-oriented decision-making emphasizes the role of the integration of perceptual evidence in terms of the firing rate given a very limited set of actions [113]. Experiments with the CL of DAC listed above (figure 1) showed that this is a rather restricted perspective because also factors such as memory, value and goals must be considered [57]. DAC theory thus suggests that numerous goals and behavioural procedures can be considered in any given task in a state-dependent fashion. This raises the question of how prefrontal cortex could keep track of this variable set of goal-oriented actions and procedures. A detailed study of the neuronal dynamics of the premotor cortex of the macaque monkey during a countermanding task has shown that the inter trial variability of the neuronal response to the movement cue is directly proportional to the errors the animal has committed (i.e. task memory) and fully predicts performance, i.e. error rate and reaction time rather than firing rate [114]. This suggests that, dependent on task memory or confidence, the neuronal dynamics allows more or less action options to compete for control and that monitoring systems regulate this process by biasing the competition between these options. This raises the question what the neural substrate is that forces a goal relevant decision among the available response options represented in this task space.

(d). Outcome predictions and action selection mechanisms in the basal ganglia

DAC proposes that optimal decision-making depends on the integration across perceptual evidence, memory biases, values and goals [3]. Hence, the question is where in the brain such a comparison and selection could take place. The striatum (i.e. caudate-putamen and nucleus accumbens) is the main recipient of prefrontal output and is organized in different sectors, topographically laid out as a dorsolateral sector (receiving primarily sensorimotor inputs), dorsomedial sector (mainly anterior cingulate and prelimbic inputs), a ventrolateral sector (the ‘core’ of nucleus accumbens; mainly amygdaloid, prelimbic and dorsal hippocampal input) and a ventromedial sector (‘shell’; mainly prelimbic, infralimbic and ventral hippocampal input [115]). An essential organizational feature of the basal ganglia is the grouping of topographical projections in parallel ‘loops’, starting in a particular cortical area and, from there, projecting to specific striatal sectors, external segment of the globus pallidus/pallidum and output structures such as the substantia nigra reticulata [116]. By themselves, these loops do not illuminate a specific mechanism for selecting among available response options. However, striatal principal cells are connected via GABAergic recurrent collaterals, providing a potential mechanism for competitive selection [117,118]. Furthermore, the basal ganglia possesses a funnel-like structure in the sense that the downstream flow of processing in cortico-basal ganglia loops is compressed into lesser and lesser neurons. This structure may provide further competition mechanisms operating at, or in interaction with, the output levels such as substantia nigra reticulata and the internal segment of the globus pallidus [119].

By itself, the presence of GABAergic, inhibitory interactions would suggest an inflexible, learning-insensitive competition mechanism in the striatum. By contrast, recording and pharmacological studies indicate an active role of the basal ganglia in learning goal-directed behaviours. Building on previous models that framed the basal ganglia as an actor–critic architecture for (model-free) reinforcement learning [120,121], we recently argued that especially the ‘Critic’ (goal-predictive) function of the striatum is well supported by the data, whereas its implementation of the ‘Actor’ component is much less clear [9]. This deviation from a classic actor–critic scheme is based on the widespread support for reward-predictive components of cue-, action- and place-related neural coding in multiple striatal sectors (in agreement with a Critic function), whereas clear evidence for a construction in which a ‘Critic’ instructs a separate ‘Actor’ structure within the same basal ganglia is lacking [9]. When surveying the various striatal sectors, it is striking to note that striatal functions in goal prediction (usually described as ‘reward expectancy’ in animal experiments) can be attributed to all sectors, but based on domains of afferent information differing per sector. The dorsomedial striatum, for instance, has been implied in action-outcome learning [122], whereas the ventral striatal core functions in cue-outcome learning, in conjunction with its strong amygdaloid input [9,123]. By contrast, the shell of the nucleus accumbens has been implied in place-outcome learning [124]. In this scheme, the dorsolateral striatum might seem to be the ‘odd one out’ in this company, as it has been implied in habit formation and sensorimotor learning with minor or no dependence on motivational outcome. However, its role can in fact be very well accommodated if the ‘outcome’ is viewed more broadly: outcome can also be constituted by action, so that cue-action (or: stimulus–response) learning is subsumed under an overall basal ganglia architecture for ‘input–outcome’ learning. In conclusion, the anatomical architecture, internal wiring and information resources in afferent structures place the basal ganglia in an eminently suitable position to, first, code state–outcome relationships (where ‘state’ can be stimulus, place or action) and, second, to use this associatively learned information to force an expected outcome-dependent decision among response options represented in task space.

Outcomes (in a broad sense) of situations and actions need to be compared against predictions. The best-known candidate for expressing such as comparison—at least in the domain of reward and appetitive learning—is the reward-prediction-error mechanism that may be implemented by mesodiencephalic dopaminergic neurons projecting to striatum and prefrontal cortex [125], while the habenula has been implied in aversive learning [126]. This type of error is, however, low dimensional in nature and limited to cached value (‘I got less/more than expected’). Error signals incorporating model-based, high-dimensional information (‘I got a banana instead of the apple I expected, but they are worth the same’) may involve cortical systems such as the anterior cingulate and orbitofrontal cortex, but also striatal regions [9,127,128] (see [31] for a more detailed computational analysis of model-based versus model-free reward signals).

It is worth noting that, in a highly adaptive agent capable of planning, world and task space representations will not be static, or merely ‘slowly evolving’ as the agent moves along in search of its goals. The self-generated time-compressed replay and forward-sweeping events found in the hippocampus illustrate how past and future trajectories can be rapidly retrieved and flexibly used for planning (cf. [88,90]). These hippocampal replay events are likely to impact on the ventral striatum, where replay events have been shown to occur in succession to place-cell replay in the hippocampus [85,87]. During hippocampal forward sweeps, ventral striatal neurons code covert reward expectations, while orbitofrontal neurons code expected outcome at alternative sites that had not been selected for visiting [129,130]. Thus, especially the orbitofrontal cortex may not only provide model-based information on cue/object value before choices are made [102,103,131], but also engage in post-decisional evaluation and ‘looking back’ on previous decisions. How hippocampal events are precisely linked to prefrontal processing is less clear yet, but the medial prefrontal cortex does exhibit replay, and its activity coheres with hippocampal theta activity [132,133].

In conclusion, the framework emerging from systems and behavioural neuroscience is that goal-directed behaviour is mediated by a network of highly interconnected brain structures which directly implement H4W: (i) the hypothalamic-brain stem system functions as a key node for signalling homeostatic needs and drives grounding the ‘Why’ of goal-directed action; (ii) the hippocampus encodes episodic state representations configured in space and time, which can be rapidly retrieved online to inform and instruct decision-making systems supporting integrated representations of ‘What’, ‘When’ and ‘Where’; (iii) the prefrontal cortex encodes task space representations, comprising choice options, rules, goals, values of cues and actions relevant to obtaining the goal, and using among others information from the hippocampus and amygdala to shape these representations based on situational relevance and prior experience shaping defining potential candidates for the ‘How’ of action; and (iv) the basal ganglia impose a selection mechanism, including reward/punishment predictions as weighting factors, on the manifold options represented in frontal motor cortical structures further biasing decision-making towards the dominant goal and action defining how a goal can be achieved through a specific action. These systems map to components and processes of the DAC architecture: it initially deals with homeostatic needs using a reactive system but successively acquires new state and task representations as well as a new behavioural repertoire in the adaptive and CLs, which compete for selection. It is the combined contribution of elements at all these levels rather than a monolithic architectural component that realizes goal-directed behaviour.

3. Conclusion

In this article, we identified the central questions of goal-directed choice as ‘Why do I act? What do I need? Where and When can this be obtained, and How do I get it?’ We have analysed this H4W problem from the perspective of a system-level architecture by mapping the H4W objectives onto the DAC theory of mind and brain. Neuroscience and neuroeconomics have often benefited from a close linkage to computational methods but these often abstract from many details of situated action (e.g. model-based reinforcement learning theories [134,135] and several contributions to this special issue). By contrast, we addressed goal-directed choice from the viewpoint of a biomimetic cognitive architecture that also considers the embodied and situated aspects of the choice situation. From this perspective, the brain is a control system and its primary task is to support adaptive action in the real world rather than optimally solving abstract problems detached from perception and action. Subsequently, following the DAC framework, we proposed that the mammalian brain solves the H4W problem by integrating central representations of needs and drives (e.g hypothalamus), valence (e.g. amygdala), world and self-state spaces (e.g. neocortex and hippocampus), task space (e.g. prefrontal cortex) and multi-modal selection (e.g. basal ganglia) on the basis of multiple kinds of outcome predictions. Goal-directed choice thus results from the coherent orchestration of multiple mechanisms within a system-level architecture. That this interpretation is a reasonable approximation of the neuronal substrate of goal-oriented choice is made plausible by the fact that an existence proof of this architectural hypothesis is provided through robot-based experimentation using DAC. The DAC case studies we presented thus also illustrate how biomimetic architectures instantiated in robots can be used to explain neuronal processes and formulate predictions that bridge the gap between real-world robot behaviour and neuronal data [136].

From the DAC perspective, it can be appreciated that goal-directed selection is multilevel; it involves multiple mechanisms that represent information of different qualities (e.g. sensory, memory and different aspects of H4W) at different layers and engage in interactions that have to be coherently orchestrated. The DAC approach can shed light on how more complex goal-directed strategies develop on top of innate reactive control systems, too. At the lowest level, simple representations of valued states (linked to homeostatic variables) might be initially available that trigger stereotyped appetitive or aversive behaviours (e.g. following a sugar gradient in aqueous solution). These behaviours are not goal-directed because they lack key ingredients such as knowledge of causal efficacy of actions and guidance of actions using goal representation. However, they can serve to bootstrap the valuation of states of the external world (e.g. apples have value), the acquisition of goal representations at higher hierarchical levels (e.g. consumption of an apple) and associated goal-directed control strategies (e.g. reaching a specific place and climbing a ladder to obtain the apple). For example, the AL of DAC shows how during learning, Pavlovian mechanisms permit to ‘transfer’ value to novel states, e.g. visual stimuli and spatial representations that can be successively selected as goals. The CL of DAC models how state-action sequences that lead to reinforcement can be stored and increasingly support more complex goal-directed strategies [25,27]. DAC does provide explanations for behaviours such as observed in classical conditioning, navigation and foraging, has made testable and tested predictions on the underlying neuronal substrate and, in addition, has been generalized to the control of robotic systems. These are three key criteria of a scientific theory that should be set as the benchmarks for any theory of mind and brain. With respect to the neuronal substrate of goal-directed choice and its decomposition into the H4W problem, an initial mapping of the DAC taxonomy to the brain would lead to the following system-level decomposition: hypothalamus, central grey and other brain stem structures forming part of the RL self system defining Why; the amygdala as an interface between Why and What at the RL and AL level; neocortex and hippocampus forming key systems of the AL learning machinery establishing What, When and Where, with the caveat that the timing of events is probably relatively coarsely coded in hippocampus [69,70]; detailed timing of When defined through the cerebellum at the level of AL; rule-based task space construction that integrates What, Where and Why at the level of the prefrontal cortex; while the competitive processes of the basal ganglia ultimately orchestrate the How of goal-oriented action. This decomposition confirms the DAC prediction that AL systems are more closely linked to interval based real-time processing, while those of the CL are more dominated by order.

The mapping of H4W to the brain and DAC shows that goals can be seen as emerging from the foundational need systems of the physically instantiated agent. However, through the state space learning systems of the AL and the task learning systems of the CL, these goals become incrementally more abstracted from their homeostatic origins (see also [137,138]). Rather they are defined as models of states of the agent and the world that define the measurable ends of successful behavioural strategies. Hence, what we call ‘goals’ are amalgamations of sensory, affective and action states, stored in different memory systems and defined on the basis of the interaction of the agent with its varying and often conflicting needs with its dynamic environment. This incremental abstraction from needs to goals serves the transition from action in environments with continuously available sensory information that support taxon-based strategies, to tasks that depend on discontinuous and unpredictable environments with intermittent feedback. The latter case can vary from maze navigation to solving complex logical puzzles [34,139]. The predictions of this model are that any representation of sensory-affect-action states (or the DAC triad of world, self, action) that can be part of frontal cortical WM systems can become tagged as a goal state, i.e. a state at which a behavioural sequence terminates. These representations can in turn affect processing at any level of the DAC hierarchy through their ability to drive valence via specific modulatory and feedback circuits. One example would be the ability to drive the inferior olive through projections from frontal areas in this way defining the teaching signals that control plasticity at the level of the cerebellar cortex and thus its goal-based learning capabilities (e.g. [32]). This implies that goal states as defined in these terms are either explicitly tagged in order to play this role or are detected in online processing. Future work has to shed more light on this prediction.

Our analysis of goal-oriented choice, combining a neurobiological perspective with the DAC theory, exemplifies how neurorobotic and experimental methods can work hand-in-hand. As the biologically grounded model constitutes an integrated architectural solution to problems of goal-directed choice and pursuit, it can help studying the systems-level neurobiology of goal-directedness rather than focusing only on its components in isolation. A combined neurophysiological and neurorobotic approach holds the promise to simultaneously tackle the problem of goal-directed choice at multiple levels [140142]: (i) the functional-behavioural level, e.g. how goals are selected and realized in the real world based on their expected value and the cost of achieving them; (ii) the mechanistic-computational level, e.g. how action outcomes are computed based on perception–action systems and how they are evaluated on-the-fly based on limited information and resources; (iii) the physiological level, e.g. how these computations are implemented in neuronal structures, focusing in particular on a system-level architecture formed by hypothalamus, hippocampus, ventral striatum, medial prefrontal cortex and amygdala; and (iv) ecology, e.g. how goal-directed abilities can be learned and adapted in real-time and in a situated environment, thus providing for a convergent multi-scale validation of theories of mind and brain. The alliance between robotics and the empirical research on the brain is beneficial for robotics, too. In developing robots endowed with goal-directed behaviour, it has considerable advantage to base them on the principles identified in biological behaviour, because central nervous systems have evolved as highly successful examples of efficient architectures enabling increasingly more sophisticated perception, cognition and behaviour. The more fundamental methodological question that the approach we sketched addresses is what shape theories of the brain will attain and how system-level questions of brain and behaviour can be effectively pursued. We argue that such answers will take the form of hybrid frameworks that integrate neurobiological research, abstract and biologically detailed computational models realized using biologically grounded real-world artefacts.

Acknowledgements

We thank our reviewers for their constructive feedback and Sytse Wierenga for his help in realizing the graphics.

Endnote

1

A DAC tutorial and source including replications of key studies can be downloaded from http://csnetwork.eu/education/texts-tutorials.

Funding statement

Research funded by the EU FP7 framework under grant agreements no. FP7-ICT-270108 (Goal-Leaders), FP7-ICT-270490 (EFAA), FP7-ICT-612139 (WYSIWYD), ERC-341196 (CDAC) and FP7-ICT-601167 (CSN II).

References

  • 1.Tolman EC. 1932. Purposive behavior in animals and man. New York, NY: Century Co. [Google Scholar]
  • 2.Newell A. 1990. Unified theories of cognition. Cambridge, MA: Harvard University Press. [Google Scholar]
  • 3.Verschure PFMJ, Althaus P. 2003. A real-world rational agent: unifying old and new AI. Cogn. Sci. 27, 561–590. ( 10.1207/s15516709cog2704_1) [DOI] [Google Scholar]
  • 4.Levine J. 1983. Materialism and qualia: the explanatory gap. Pac. Phil. Q. 64, 354–361. [Google Scholar]
  • 5.Chalmers D. 1995. Facing up to the problem of consciousness. J. Conscious. Stud. 2, 200–219. [Google Scholar]
  • 6.Cisek P, Pastor-Bernier A. 2014. On the challenges and mechanisms of embodied decisions. Phil. Trans. R. Soc. B 369, 20130479 ( 10.1098/rstb.2013.0479) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Balleine BW, Dickinson A. 1998. Goal-directed instrumental action: contingency and incentive learning and their cortical substrates. Neuropharmacology 37, 407–419. ( 10.1016/S0028-3908(98)00033-1) [DOI] [PubMed] [Google Scholar]
  • 8.Gläscher J, Daw N, Dayan P, O'Doherty JP. 2010. States versus rewards: dissociable neural prediction error signals underlying model-based and model-free reinforcement learning. Neuron 66, 585–595. ( 10.1016/j.neuron.2010.04.016) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Pennartz C, Ito R, Verschure P, Battaglia F, Robbins T. 2011. The hippocampal–striatal axis in learning, prediction and goal-directed behavior. Trends Neurosci. 34, 548–559. ( 10.1016/j.tins.2011.08.001) [DOI] [PubMed] [Google Scholar]
  • 10.Verschure PFMJ. 2012. Distributed adaptive control: a theory of the mind, brain, body nexus. Biol. Inspired Cogn. Archit. 1, 55–72. ( 10.1016/j.bica.2012.04.005) [DOI] [Google Scholar]
  • 11.Verschure PFMJ. 2013. Formal minds and biological brains II: from the mirage of intelligence to a science and engineering of consciousness. IEEE Expert. 28, 33–36. [Google Scholar]
  • 12.Pavlov IP. 1927. Conditioned reflexes: an investigation of the physiological activity of the cerebral cortex. London, UK: Oxford University Press. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Verschure PFMJ, Voegtlin T, Douglas RJ. 2003. Environmentally mediated synergy between perception and behaviour in mobile robots. Nature 425, 620–624. ( 10.1038/nature02024) [DOI] [PubMed] [Google Scholar]
  • 14.Verschure PFMJ, Krose B, Pfeifer R. 1993. Distributed adaptive control: the self-organization of structured behavior. Robot. Auton. Syst. 9, 181–196. ( 10.1016/0921-8890(92)90054-3) [DOI] [Google Scholar]
  • 15.Gray JA. 1987. The psychology of fear and stress. London, UK: Cambridge University Press. [Google Scholar]
  • 16.Panksepp J, Biven L. 2012. The archaeology of mind: neuroevolutionary origins of human emotions (Norton series on interpersonal neurobiology). New York, NY: WW Norton and Company. [Google Scholar]
  • 17.Sterling P, Eyer J. 1988. Allostasis: a new paradigm to explain arousal pathology. Psychiatry Neurosci. 30, 315–318. [Google Scholar]
  • 18.Sanchez-Fibla M, Bernardet U, Wasserman E, Pelc T, Mintz M, Jackson JC, Lansink C, Pennartz C, Verschure PFJ. 2010. Allostatic control for robot behavior regulation: a comparative rodent-robot study. Adv. Complex Syst. 13, 377–403. ( 10.1142/S0219525910002621) [DOI] [Google Scholar]
  • 19.Sanchez-Fibla M, Duff A, Verschure PFMJ. (eds). 2011. The acquisition of intentionally indexed and object centered affordance gradients: a biomimetic controller and mobile robotics benchmark. In IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS), 25 September 2011, San Francisco, USA IEEE. [Google Scholar]
  • 20.Lewin K. 1936. Principles of topological psychology. New York, NY: McGraw-Hill. [Google Scholar]
  • 21.Khatib O. 1986. Real-time obstacle avoidance for manipulators and mobile robots. Int. J. Robot. Res. 5, 90–98. ( 10.1177/027836498600500106) [DOI] [Google Scholar]
  • 22.Sanchez-Fibla MS, Bernardet U, Verschure PFMJ. (eds). 2010. Allostatic control for robot behaviour regulation: an extension to path planning. Intelligent Robots and Systems (IROS). In IEEE/RSJ Int. Conf. 18–22 October 2010, Taipei, Taiwan IEEE. [Google Scholar]
  • 23.Vouloutsi V, Lallée S, Verschure PMJ. 2013. Modulating behaviors using allostatic control. In Biomimetic and biohybrid systems. Lecture Notes in Computer Science 8064 (eds Lepora N, Mura A, Krapp H, Verschure PMJ, Prescott T.), pp. 287–298. Berlin, Germany: Springer. [Google Scholar]
  • 24.Verschure PFMJ, Pfeifer R. (eds). 1992. Categorization, representations, and the dynamics of system-environment interaction: a case study in autonomous systems. In From animals to animats: Proc. Second Int. Conf. on Simulation of Adaptive behavior Honolulu, Hawaii, 7–11 December 1992 Cambridge, MA: MIT Press. [Google Scholar]
  • 25.Duff A, Verschure PFMJ. 2010. Unifying perceptual and behavioral learning with a correlative subspace learning rule. Neurocomputing 73, 1818–1830. ( 10.1016/j.neucom.2009.11.048) [DOI] [Google Scholar]
  • 26.Miller S, Konorski J. 1928. Sur une forme particuliere des reflexes conditionnels. C.R. Seances Soc. Polonaise Biol. 49, 1155–1157. [Google Scholar]
  • 27.Inderbitzin M, Herreros-Alonso I, Verschure PFMJ. (eds). 2010. An integrated computational model of the two phase theory of classical conditioning. In Neural Networks (IJCNN), The 2010 Int. Joint Conf. 18–23 July 2010, Barcelona, Spain IEEE. [Google Scholar]
  • 28.Searle JR. 1980. Minds brains and programs. Behav. Brain Sci. 3, 417–457. ( 10.1017/S0140525X00005756) [DOI] [Google Scholar]
  • 29.Friston K. 2010. The free-energy principle: a unified brain theory? Nat. Rev. Neurosci. 11, 127–138. ( 10.1038/nrn2787) [DOI] [PubMed] [Google Scholar]
  • 30.Friston K, Schwartenbeck P, FitzGerald T, Moutoussis M, Behrens T, Dolan RJ. 2014. The anatomy of choice: dopamine and decision-making. Phil. Trans. R. Soc. B 369, 20130481 ( 10.1098/rstb.2013.0481) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Daw ND, Dayan P. 2014. The algorithmic anatomy of model-based evaluation. Phil. Trans. R. Soc. B 369, 20130478 ( 10.1098/rstb.2013.0478) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Herreros I, Verschure PF. 2013. Nucleo-olivary inhibition balances the interaction between the reactive and adaptive layers in motor control. Neural Netw. 47, 64–71. ( 10.1016/j.neunet.2013.01.026) [DOI] [PubMed] [Google Scholar]
  • 33.Ringwald M, Verschure PFMJ. (eds). 2007. The fusion of multiple sources of information in the organization of goal-oriented behavior: spatial attention versus integration. In ECMR 2007–3rd European Conf. on Mobile Robots, Freiburg, Germany, 19–21st September 2007 Berlin, Germany: Springer-Verlag. [Google Scholar]
  • 34.Marcos E, Ringwald M, Duff A, Sánchez-Fibla M, Verschure PF. 2013. The hierarchical accumulation of knowledge in the distributed adaptive control architecture. In Computational and robotic models of the hierarchical organization of behavior (eds Baldassarre G, Mirolli M.), pp. 213–234. New York, NY: Springer. [Google Scholar]
  • 35.Mathews Z, Verschure PFMJ, Bermudez i Badia S. (eds). 2010. An insect-based method for learning landmark reliability using expectation reinforcement in dynamic environments. In 2010 IEEE Int. Conf. on Robotics and Automation (ICRA), 3–7 May 2010, Anchorage, Alaska IEEE. [Google Scholar]
  • 36.Marcos E, Duff A, Sanchez-Fibla M, Verschure PFMJ. (eds). 2010. The neuronal substrate underlying order and interval representations in sequential tasks: a biologically based robot study. In The 2010 Int. Joint Conf. on Neural Networks (IJCNN), 18–23 July 2010, Barcelona, Spain. IEEE. [Google Scholar]
  • 37.Mathews Z, et al. (eds). 2009. Insect-like mapless navigation based on head direction cells and contextual learning using chemo-visual sensors. In The 2009 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems IROS, 10 October 2009, St Louis, MI, USA IEEE. [Google Scholar]
  • 38.Marcos E, Sánchez-Fibla M, Verschure PMJ. 2010. The complementary roles of allostatic and contextual control systems in foraging tasks. In From animals to animats 11. Lecture Notes in Computer Science 6226 (eds Doncieux S, Girard B, Guillot A, Hallam J, Meyer J-A, Mouret J-B.), pp. 370–379. Berlin, Germany: Springer. [Google Scholar]
  • 39.Redish AD. 1999. Beyond the cognitive map: from place cells to episodic memory. Cambridge, MA: MIT Press. [Google Scholar]
  • 40.Groenewegen HJ, Uylings H. 2000. The prefrontal cortex and the integration of sensory, limbic and autonomic information. Prog. Brain Res. 126, 3–28. ( 10.1016/S0079-6123(00)26003-2) [DOI] [PubMed] [Google Scholar]
  • 41.Berthoud H-R. 2004. Mind versus metabolism in the control of food intake and energy balance. Physiol. Behav. 81, 781–793. ( 10.1016/j.physbeh.2004.04.034) [DOI] [PubMed] [Google Scholar]
  • 42.Simerly R. 1995. Anatomical substrates of hypothalamic integration. In The rat nervous system (ed. Paxinos G.), pp. 353–376, 2nd edn San Diego, CA: Academic Press. [Google Scholar]
  • 43.Morrison SF, Nakamura K, Madden CJ. 2008. Central control of thermogenesis in mammals. Exp. Physiol. 93, 773–797. ( 10.1113/expphysiol.2007.041848) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Bourque CW. 2008. Central mechanisms of osmosensation and systemic osmoregulation. Nat. Rev. Neurosci. 9, 519–531. ( 10.1038/nrn2400) [DOI] [PubMed] [Google Scholar]
  • 45.Sutcliffe JG, de Lecea L. 2002. The hypocretins: setting the arousal threshold. Nat. Rev. Neurosci. 3, 339–349. ( 10.1038/nrn808) [DOI] [PubMed] [Google Scholar]
  • 46.Blouet C, Schwartz GJ. 2010. Hypothalamic nutrient sensing in the control of energy homeostasis. Behav. Brain Res. 209, 1–12. ( 10.1016/j.bbr.2009.12.024) [DOI] [PubMed] [Google Scholar]
  • 47.Woody EZ, Szechtman H. 2011. Adaptation to potential threat: the evolution, neurobiology, and psychopathology of the security motivation system. Neurosci. Biobehav. Rev. 35, 1019–1033. ( 10.1016/j.neubiorev.2010.08.003) [DOI] [PubMed] [Google Scholar]
  • 48.Flanagan-Cato LM. 2011. Sex differences in the neural circuit that mediates female sexual receptivity. Front. Neuroendocrinol. 32, 124–136. ( 10.1016/j.yfrne.2011.02.008) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Stolzenberg DS, Numan M. 2011. Hypothalamic interaction with the mesolimbic DA system in the control of the maternal and sexual behaviors in rats. Neurosci. Biobehav. Rev. 35, 826–847. ( 10.1016/j.neubiorev.2010.10.003) [DOI] [PubMed] [Google Scholar]
  • 50.Lin D, Boyle MP, Dollar P, Lee H, Lein E, Perona P, Anderson DJ. 2011. Functional identification of an aggression locus in the mouse hypothalamus. Nature 470, 221–226. ( 10.1038/nature09736) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Borbely AA, Achermann PJ. 1992. Concepts and models of sleep regulation: an overview. J. Sleep Res. 1, 63–79. ( 10.1111/j.1365-2869.1992.tb00013.x) [DOI] [PubMed] [Google Scholar]
  • 52.Panksepp J. 1998. Affective neuroscience: the foundations of human and animal emotions. Oxford, UK: Oxford University Press. [Google Scholar]
  • 53.Gray JA, McNaughton N. 2000. The neuropsychology of anxiety: an enquiry into thé functions of thesepto-hippocampal system. Oxford, UK: Oxford University Press. [Google Scholar]
  • 54.Davis KL, Panksepp J. 2011. The brain's emotional foundations of human personality and the affective neuroscience personality scales. Neurosci. Biobehav. Rev. 35, 1946–1958. ( 10.1016/j.neubiorev.2011.04.004) [DOI] [PubMed] [Google Scholar]
  • 55.Winstanley C, et al. 2012. Search, goals and the brain. In Strüngmann Forum volume on cognitive search: evolution, algorithms and the brain (eds Todd PMT, Hills TT, Robbins TW.), pp. 125–156. Cambridge, MT: MIT Press. [Google Scholar]
  • 56.Weinberger NM. 1993. Learning-induced changes of auditory receptive fields. Curr. Opin. Neurobiol. 3, 570–577. ( 10.1016/0959-4388(93)90058-7) [DOI] [PubMed] [Google Scholar]
  • 57.Sugrue LP, Corrado GS, Newsome WT. 2004. Matching behavior and the representation of value in the parietal cortex. Science 304, 1782–1787. ( 10.1126/science.1094765) [DOI] [PubMed] [Google Scholar]
  • 58.Platt ML, Glimcher PW. 1999. Neural correlates of decision variables in parietal cortex. Nature 400, 233–238. ( 10.1038/22268) [DOI] [PubMed] [Google Scholar]
  • 59.Shuler MG, Bear MF. 2006. Reward timing in the primary visual cortex. Science 311, 1606–1609. ( 10.1126/science.1123513) [DOI] [PubMed] [Google Scholar]
  • 60.Vickery TJ, Chun MM, Lee D. 2011. Ubiquity and specificity of reinforcement signals throughout the human brain. Neuron 72, 166–177. ( 10.1016/j.neuron.2011.08.011) [DOI] [PubMed] [Google Scholar]
  • 61.Goltstein PM, Coffey EB, Roelfsema PR, Pennartz CM. 2013. In vivo two-photon Ca2+ imaging reveals selective reward effects on stimulus-specific assemblies in mouse visual cortex. J. Neurosci. 33, 11 540–11 555. ( 10.1523/JNEUROSCI.1341-12.2013) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Kilgard MP, Merzenich MM. 1998. Plasticity of temporal information processing in the primary auditory cortex. Nat. Neurosci. 1, 727–731. ( 10.1038/3729) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Tolman EC. 1948. Cognitive maps in rats and men. Psychol. Rev. 55, 189–208. ( 10.1037/h0061626) [DOI] [PubMed] [Google Scholar]
  • 64.O'Keefe J, Dostrovsky J. 1971. The hippocampus as a spatial map. Preliminary evidence from unit activity in the freely-moving rat. Brain Res. 34, 171–175. ( 10.1016/0006-8993(71)90358-1) [DOI] [PubMed] [Google Scholar]
  • 65.Tulving E. 2002. Episodic memory: from mind to brain. Annu. Rev. Psychol. 53, 1–25. ( 10.1146/annurev.psych.53.100901.135114) [DOI] [PubMed] [Google Scholar]
  • 66.Wood ER, Dudchenko PA, Eichenbaum H. 1999. The global record of memory in hippocampal neuronal activity. Nature 397, 613–616. ( 10.1038/17605) [DOI] [PubMed] [Google Scholar]
  • 67.Leutgeb S, Leutgeb JK, Barnes CA, Moser EI, McNaughton BL, Moser M-B. 2005. Independent codes for spatial and episodic memory in hippocampal neuronal ensembles. Science 309, 619–623. ( 10.1126/science.1114037) [DOI] [PubMed] [Google Scholar]
  • 68.Lansink CS, Jackson JC, Lankelma JV, Ito R, Robbins TW, Everitt BJ, Pennartz CMA. 2012. Reward cues in space: commonalities and differences in neural coding by hippocampal and ventral striatal ensembles. J. Neurosci. 32, 12 444–12 459. ( 10.1523/JNEUROSCI.0593-12.2012) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Kraus BJ, Robinson RJ, II, White JA, Eichenbaum H, Hasselmo ME. 2013. Hippocampal ‘time cells’: time versus path integration. Neuron 78, 1090–1101. ( 10.1016/j.neuron.2013.04.015) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Pastalkova E, Itskov V, Amarasingham A, Buzsáki G. 2008. Internally generated cell assembly sequences in the rat hippocampus. Science 321, 1322–1327. ( 10.1126/science.1159775) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Muller RU, Kubie JL. 1987. The effects of changes in the environment on the spatial firing of hippocampal complex-spike cells. J. Neurosci. 7, 1951–1968. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Rennó-Costa C, Lisman JE, Verschure PFMJ. 2010. The mechanism of rate remapping in the dentate gyrus. Neuron 68, 1051–1058. ( 10.1016/j.neuron.2010.11.024) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Rennó-Costa C, Lisman JE, Verschure PFMJ. 2014. A signature of attractor dynamics in the CA3 region of the hippocampus. PLoS Comput. Biol. 10, e1003641 ( 10.1371/journal.pcbi.1003641) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Lisman JE. 2007. Role of the dual entorhinal inputs to hippocampus: a hypothesis based on cue/action (non-self/self) couplets. Prog. Brain Res. 163, 615–625. ( 10.1016/S0079-6123(07)63033-7) [DOI] [PubMed] [Google Scholar]
  • 75.Leutgeb JK, Leutgeb S, Moser M-B, Moser EI. 2007. Pattern separation in the dentate gyrus and CA3 of the hippocampus. Science 315, 961–966. ( 10.1126/science.1135801) [DOI] [PubMed] [Google Scholar]
  • 76.Lu L, Leutgeb JK, Tsao A, Henriksen EJ, Leutgeb S, Barnes CA, Witter MP, Moser M-B, Moser EI. 2013. Impaired hippocampal rate coding after lesions of the lateral entorhinal cortex. Nat. Neurosci. 16, 1085–1093. ( 10.1038/nn.3462) [DOI] [PubMed] [Google Scholar]
  • 77.Rondi-Reig L, Petit GH, Tobin C, Tonegawa S, Mariani J, Berthoz A. 2006. Impaired sequential egocentric and allocentric memories in forebrain-specific–NMDA receptor knock-out mice during a new task dissociating strategies of navigation. J. Neurosci. 26, 4071–4081. ( 10.1523/JNEUROSCI.3408-05.2006) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Cabral HO, Vinck M, Fouquet C, Pennartz C, Rondi-Reig L, Battaglia FP. 2014. Oscillatory dynamics and place field maps reflect hippocampal ensemble processing of sequence and place memory under NMDA receptor control. Neuron 81, 402–415. ( 10.1016/j.neuron.2013.11.010) [DOI] [PubMed] [Google Scholar]
  • 79.Eichenbaum H, Dudchenko P, Wood E, Shapiro M, Tanila H. 1999. The hippocampus, memory, and place cells: is it spatial memory or a memory space? Neuron 23, 209–216. ( 10.1016/S0896-6273(00)80773-4) [DOI] [PubMed] [Google Scholar]
  • 80.Packard MG, McGaugh JL. 1996. Inactivation of hippocampus or caudate nucleus with lidocaine differentially affects expression of place and response learning. Neurobiol. Learn. Mem. 65, 65–72. ( 10.1006/nlme.1996.0007) [DOI] [PubMed] [Google Scholar]
  • 81.Yin BJK. 2006. The role of the basal ganglia in habit formation. Nat. Rev. Neurosci. 7, 464–476. ( 10.1038/nrn1919) [DOI] [PubMed] [Google Scholar]
  • 82.Wilson MA, McNaughton BL. 1994. Reactivation of hippocampal ensemble memories during sleep. Science 265, 676–679. ( 10.1126/science.8036517) [DOI] [PubMed] [Google Scholar]
  • 83.Carr MF, Jadhav SP, Frank LM. 2011. Hippocampal replay in the awake state: a potential substrate for memory consolidation and retrieval. Nat. Neurosci. 14, 147–153. ( 10.1038/nn.2732) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Dragoi G, Tonegawa S. 2010. Preplay of future place cell sequences by hippocampal cellular assemblies. Nature 469, 397–401. ( 10.1038/nature09633) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Pennartz CMA, Uylings HBM, Barnes CA, McNaughton BL. 2002. Memory reactivation and consolidation during sleep: from cellular mechanisms to human performance. In Progress in brain research, vol. 138: plasticity in the adult brain: from genes to neurotherapy (eds Hofman MA, Boer GJ, Holtmaat AJGD, Van Someren EJW, Verhaagen J, Swaab DF.), pp. 143–166. Amsterdam, The Netherlands: Elsevier. [DOI] [PubMed] [Google Scholar]
  • 86.Kudrimoti HS, Barnes CA, McNaughton BL. 1999. Reactivation of hippocampal cell assemblies: effects of behavioral state, experience, and EEG dynamics. J. Neurosci. 19, 4090–4101. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Lansink CS, Goltstein PM, Lankelma JV, McNaughton BL, Pennartz CM. 2009. Hippocampus leads ventral striatum in replay of place-reward information. PLoS Biol. 7, e1000173 ( 10.1371/journal.pbio.1000173) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Jadhav SP, Kemere C, German PW, Frank LM. 2012. Awake hippocampal sharp-wave ripples support spatial memory. Science 336, 1454–1458. ( 10.1126/science.1217230) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Johnson A, Redish AD. 2007. Neural ensembles in CA3 transiently encode paths forward of the animal at a decision point. J. Neurosci. 27, 12 176–12 189. ( 10.1523/JNEUROSCI.3761-07.2007) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Pfeiffer BE, Foster DJ. 2013. Hippocampal place-cell sequences depict future paths to remembered goals. Nature 497, 74–79. ( 10.1038/nature12112) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91.Pezzulo G, van der Meer MAA, Lansink CS, Pennartz CMA. In press Internally generated sequences in learning and executing goal-directed behavior. Trends Cogn. Sci. [DOI] [PubMed] [Google Scholar]
  • 92.Hollup SA, Hollup SA, Molden S, Molden S, Donnett JG, Donnett JG, Moser M-B, Moser EI. 2001. Place fields of rat hippocampal pyramidal cells and spatial learning in the watermaze. Eur. J. Neurosci. 13, 1197–1208. ( 10.1046/j.0953-816x.2001.01487.x) [DOI] [PubMed] [Google Scholar]
  • 93.Passingham RE, Wise SP. 2012. The neurobiology of the prefrontal cortex: anatomy, evolution, and the origin of insight. Oxford, UK: Oxford University Press. [Google Scholar]
  • 94.Wallis JD, Anderson KC, Miller EK. 2001. Single neurons in prefrontal cortex encode abstract rules. Nature 411, 953–956. ( 10.1038/35082081) [DOI] [PubMed] [Google Scholar]
  • 95.Mulder AB, Nordquist RE, Pennartz CMA. 2003. Learning-related changes in response patterns of prefrontal neurons during instrumental conditioning. Behav. Brain Res. 146, 77–88. ( 10.1016/j.bbr.2003.09.016) [DOI] [PubMed] [Google Scholar]
  • 96.Averbeck B, Sohn J, Lee D. 2006. Activity in prefrontal cortex during dynamic selection of action sequences. Nat. Neurosci. 9, 276–282. ( 10.1038/nn1634) [DOI] [PubMed] [Google Scholar]
  • 97.Histed MH, Miller EK. 2006. Microstimulation of frontal cortex can reorder a remembered spatial sequence. PLoS Biol. 4, e134 ( 10.1371/journal.pbio.0040134) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98.Hok V, Save E, Lenck-Santini P, Poucet B. 2005. Coding for spatial goals in the prelimbic/infralimbic area of the rat frontal cortex. Proc. Natl Acad. Sci. USA 102, 4602–4607. ( 10.1073/pnas.0407332102) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99.Tsujimoto S, Genovesio A, Wise SP. 2008. Transient neuronal correlations underlying goal selection and maintenance in prefrontal cortex. Cereb. Cortex 18, 2748–2761. ( 10.1093/cercor/bhn033) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 100.Watanabe M. 1992. Frontal units of the monkey coding the associative significance of visual and auditory stimuli. Exp. Brain Res. 89, 233–247. ( 10.1007/BF00228241) [DOI] [PubMed] [Google Scholar]
  • 101.Tremblay L, Schultz W. 1999. Relative reward preference in primate orbitofrontal cortex. Nature 398, 704–708. ( 10.1038/19525) [DOI] [PubMed] [Google Scholar]
  • 102.Schoenbaum G, Setlow B, Saddoris MP, Gallagher M. 2003. Encoding predicted outcome and acquired value in orbitofrontal cortex during cue sampling depends upon input from basolateral amygdala. Neuron 39, 855–867. ( 10.1016/S0896-6273(03)00474-4) [DOI] [PubMed] [Google Scholar]
  • 103.van Wingerden M, Vinck M, Tijms V, Ferreira IR, Jonker AJ, Pennartz C. 2012. NMDA receptors control cue-outcome selectivity and plasticity of orbitofrontal firing patterns during associative stimulus-reward learning. Neuron 76, 813–825. ( 10.1016/j.neuron.2012.09.039) [DOI] [PubMed] [Google Scholar]
  • 104.Padoa-Schioppa C, Assad JA. 2006. Neurons in the orbitofrontal cortex encode economic value. Nature 441, 223–226. ( 10.1038/nature04676) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 105.Hosokawa T, Kennerley SW, Sloan J, Wallis JD. 2013. Single-neuron mechanisms underlying cost-benefit analysis in frontal cortex. J. Neurosci. 33, 17 385–17 397. ( 10.1523/JNEUROSCI.2221-13.2013) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 106.Kennerley SW, Dahmubed AF, Lara AH, Wallis JD. 2009. Neurons in the frontal lobe encode the value of multiple decision variables. J. Cogn. Neurosci. 21, 1162–1178. ( 10.1162/jocn.2009.21100) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 107.Dias R, Robbins T, Roberts A. 1996. Dissociation in prefrontal cortex of affective and attentional shifts. Nature 380, 69–72. ( 10.1038/380069a0) [DOI] [PubMed] [Google Scholar]
  • 108.Birrell JM, Brown VJ. 2000. Medial frontal cortex mediates perceptual attentional set shifting in the rat. J. Neurosci. 20, 4320–4324. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109.Balleine BW KA, Dickinson A. 2003. The effect of lesions of the basolateral amygdala on instrumental conditioning. J. Neurosci. 23, 666–675. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 110.Kouneiher F, Charron S, Koechlin E. 2009. Motivation and cognitive control in the human prefrontal cortex. Nat. Neurosci. 12, 939–945. ( 10.1038/nn.2321) [DOI] [PubMed] [Google Scholar]
  • 111.Corbit LH, Leung BK, Balleine BW. 2013. The role of the amygdala-striatal pathway in the acquisition and performance of goal-directed instrumental actions. J. Neurosci. 33, 17 682–17 690. ( 10.1523/JNEUROSCI.3271-13.2013) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 112.Rhodes SE, Murray EA. 2013. Differential effects of amygdala, orbital prefrontal cortex, and prelimbic cortex lesions on goal-directed behavior. J. Neurosci. 33, 3380–3389. ( 10.1523/JNEUROSCI.4374-12.2013) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 113.Shadlen MN, Kiani R. 2013. Decision making as a window on cognition. Neuron 80, 791–806. ( 10.1016/j.neuron.2013.10.047) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 114.Marcos E, Pani P, Brunamonti E, Deco G, Verschure P. 2013. Neural variability in premotor cortex is modulated by trial history and predicts behavioral performance. Neuron 78, 1–7. ( 10.1016/j.neuron.2013.02.006) [DOI] [PubMed] [Google Scholar]
  • 115.Voorn P, Vanderschuren LJ, Groenewegen HJ, Robbins TW, Pennartz C. 2004. Putting a spin on the dorsal–ventral divide of the striatum. Trends Neurosci. 27, 468–474. ( 10.1016/j.tins.2004.06.006) [DOI] [PubMed] [Google Scholar]
  • 116.Alexander GE, Crutcher MD, DeLong MR. 1991. Basal ganglia-thalamocortical circuits: parallel substrates for motor, oculomotor,'prefrontal’ and ‘limbic’ functions. Progress Brain Res. 85, 119–146. ( 10.1016/S0079-6123(08)62678-3) [DOI] [PubMed] [Google Scholar]
  • 117.Tunstall MJ, Oorschot DE, Kean A, Wickens JR. 2002. Inhibitory interactions between spiny projection neurons in the rat striatum. J. Neurophysiol. 88, 1263–1269. [DOI] [PubMed] [Google Scholar]
  • 118.Taverna S, Van Dongen YC, Groenewegen HJ, Pennartz CM. 2004. Direct physiological evidence for synaptic connectivity between medium-sized spiny neurons in rat nucleus accumbens in situ. J. Neurophysiol. 91, 1111–1121. ( 10.1152/jn.00892.2003) [DOI] [PubMed] [Google Scholar]
  • 119.Redgrave P, Prescott T, Gurney KN. 1999. The basal ganglia: a vertebrate solution to the selection problem? Neuroscience 89, 1009–1023. ( 10.1016/S0306-4522(98)00319-4) [DOI] [PubMed] [Google Scholar]
  • 120.Barto AG. 1995. Adaptive critics and the basal ganglia. In Models of information processing in the basal ganglia (eds James CH, Davis JL, Beiser DG.), pp. 215–232. Cambridge, MA: MIT Press. [Google Scholar]
  • 121.Schultz W. 1998. Predictive reward signal of dopamine neurons. J. Neurophysiol. 80, 1–27. [DOI] [PubMed] [Google Scholar]
  • 122.Yin HH, Ostlund SB, Knowlton BJ, Balleine BW. 2005. The role of the dorsomedial striatum in instrumental conditioning. Eur. J. Neurosci. 22, 513–523. ( 10.1111/j.1460-9568.2005.04218.x) [DOI] [PubMed] [Google Scholar]
  • 123.Di Ciano P, Everitt BJ. 2004. Direct interactions between the basolateral amygdala and nucleus accumbens core underlie cocaine-seeking behavior by rats. J. Neurosci. 24, 7167–7173. ( 10.1523/JNEUROSCI.1581-04.2004) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 124.Ito R, Robbins TW, Pennartz CM, Everitt BJ. 2008. Functional interaction between the hippocampus and nucleus accumbens shell is necessary for the acquisition of appetitive spatial context conditioning. J. Neurosci. 28, 6950–6959. ( 10.1523/JNEUROSCI.1615-08.2008) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 125.Schultz W. 2007. Multiple dopamine functions at different time courses. Annu. Rev. Neurosci. 30, 259–288. ( 10.1146/annurev.neuro.28.061604.135722) [DOI] [PubMed] [Google Scholar]
  • 126.Hikosaka O. 2010. The habenula: from stress evasion to value-based decision-making. Nat. Rev. Neurosci. 11, 503–513. ( 10.1038/nrn2866) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 127.McDannald MA, Takahashi YK, Lopatina N, Pietras BW, Jones JL, Schoenbaum G. 2012. Model-based learning and the contribution of the orbitofrontal cortex to the model-free world. Eur. J. Neurosci. 35, 991–996. ( 10.1111/j.1460-9568.2011.07982.x) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 128.Jones JL, Esber GR, McDannald MA, Gruber AJ, Hernandez A, Mirenzi A, Schoenbaum G. 2012. Orbitofrontal cortex supports behavior and learning using inferred but not cached values. Science 338, 953–956. ( 10.1126/science.1227489) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 129.Van der Meer MA, Redish AD. 2009. Covert expectation-of-reward in rat ventral striatum at decision points. Front. Integr. Neurosci. 3, 1 ( 10.3389/neuro.07.001.2009) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 130.Steiner AP, Redish AD. 2012. The road not taken: neural correlates of decision making in orbitofrontal cortex. Front. Neurosci. 6, 131 ( 10.3389/fnins.2012.00131) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 131.Balleine BW, Leung BK, Ostlund SB. 2011. The orbitofrontal cortex, predicted value, and choice. Ann. NY Acad. Sci. 1239, 43–50. ( 10.1111/j.1749-6632.2011.06270.x) [DOI] [PubMed] [Google Scholar]
  • 132.Jones MW, Wilson MA. 2005. Theta rhythms coordinate hippocampal–prefrontal interactions in a spatial memory task. PLoS Biol. 3, e402 ( 10.1371/journal.pbio.0030402) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 133.Euston DR, Tatsuno M, McNaughton BL. 2007. Fast-forward playback of recent memory sequences in prefrontal cortex during sleep. Science 318, 1147–1150. ( 10.1126/science.1148979) [DOI] [PubMed] [Google Scholar]
  • 134.Daw ND, Niv Y, Dayan P. 2005. Uncertainty-based competition between prefrontal and dorsolateral striatal systems for behavioral control. Nat. Neurosci. 8, 1704–1711. ( 10.1038/nn1560) [DOI] [PubMed] [Google Scholar]
  • 135.Kalenscher T, Pennartz C. 2008. Is a bird in the hand worth two in the future? The neuroeconomics of intertemporal decision-making. Prog. Neurobiol. 84, 284–315. ( 10.1016/j.pneurobio.2007.11.004) [DOI] [PubMed] [Google Scholar]
  • 136.Edelman GM. 2007. Learning in and from brain-based devices. Science 318, 1103–1105. ( 10.1126/science.1148677) [DOI] [PubMed] [Google Scholar]
  • 137.Pezzulo G, Castelfranchi C. 2009. Thinking as the control of imagination: a conceptual framework for goal-directed systems. Psychol. Res. 73, 559–577. ( 10.1007/s00426-009-0237-z) [DOI] [PubMed] [Google Scholar]
  • 138.Pezzulo G, Castelfranchi C. 2007. The symbol detachment problem. Cogn. Process. 8, 115–131. ( 10.1007/s10339-007-0164-0) [DOI] [PubMed] [Google Scholar]
  • 139.Duff A, Sanchez-Fibla M, Verschure PFMJ. 2011. A biologically based model for the integration of sensory-motor contingencies in rules and plans: a prefrontal cortex based extension of the distributed adaptive control architecture. Brain Res. Bull. 85, 289–304. ( 10.1016/j.brainresbull.2010.11.008) [DOI] [PubMed] [Google Scholar]
  • 140.Verschure PFMJ. 2012. The distributed adaptive control architecture of the mind, brain, body nexus. Biol. Inspired Cogn. Archit. 1, 55–72. [Google Scholar]
  • 141.Pezzulo G, Rigoli F, Chersi F. 2013. The mixed instrumental controller: using value of Information to combine habitual choice and mental simulation. Front. Psychol. 4, 92 ( 10.3389/fpsyg.2013.00092) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 142.Pezzulo G, Barsalou LW, Cangelosi A, Fischer MH, McRae K, Spivey MJ. 2011. The mechanics of embodiment: a dialog on embodiment and computational modeling. Front. Psychol. 2, 5 ( 10.3389/fpsyg.2011.00005) [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Philosophical Transactions of the Royal Society B: Biological Sciences are provided here courtesy of The Royal Society

RESOURCES