Abstract
Recent advances in artificial intelligence (AI) and neuroscience are impressive. In AI, this includes the development of computer programs that can beat a grandmaster at GO or outperform human radiologists at cancer detection. A great deal of these technological developments are directly related to progress in artificial neural networks—initially inspired by our knowledge about how the brain carries out computation. In parallel, neuroscience has also experienced significant advances in understanding the brain. For example, in the field of spatial navigation, knowledge about the mechanisms and brain regions involved in neural computations of cognitive maps—an internal representation of space—recently received the Nobel Prize in medicine. Much of the recent progress in neuroscience has partly been due to the development of technology used to record from very large populations of neurons in multiple regions of the brain with exquisite temporal and spatial resolution in behaving animals. With the advent of the vast quantities of data that these techniques allow us to collect there has been an increased interest in the intersection between AI and neuroscience, many of these intersections involve using AI as a novel tool to explore and analyze these large data sets. However, given the common initial motivation point—to understand the brain—these disciplines could be more strongly linked. Currently much of this potential synergy is not being realized. We propose that spatial navigation is an excellent area in which these two disciplines can converge to help advance what we know about the brain. In this review, we first summarize progress in the neuroscience of spatial navigation and reinforcement learning. We then turn our attention to discuss how spatial navigation has been modeled using descriptive, mechanistic, and normative approaches and the use of AI in such models. Next, we discuss how AI can advance neuroscience, how neuroscience can advance AI, and the limitations of these approaches. We finally conclude by highlighting promising lines of research in which spatial navigation can be the point of intersection between neuroscience and AI and how this can contribute to the advancement of the understanding of intelligent behavior.
Keywords: neuroscience, artificial intelligence, spatial navigation, deep learning, reinforcement learning, memory, learning
Introduction
Artificial General Intelligence (AGI), understood as the capability to produce intelligent behavior at the level of humans, has been a matter of debate. Even at the operational level, a definition that we can use to classify the behavior of any agent as intelligent or not, lacks consensus. Some define intelligence as the whole coordination of brain, body, and environment (Pfeifer and Scheier, 1999). Others also require the existence of a task to define intelligence (Almássy et al., 1998). From this perspective, in the absence of the elements of this definition of intelligence, adaptive intelligent behavior does not exist (Chiel and Beer, 1997). Another perspective is that intelligence can exist without actuators and even without an environment. The common denominator across perspectives and fields is that intelligence requires a brain. If we accept that premise, in order to understand intelligence, natural, or artificial, we must study the brain. The goal of neuroscience is precisely that—to understand how the brain works. If we assume that intelligent behavior can be understood by studying how it emerges, it is reasonable to attempt to learn from a working example: biological brains.
Historically, Artificial Intelligence (AI) researchers followed this approach. In fact, many of the initial ideas of numerous state-of-the-art algorithms in AI were derived from psychology and neuroscience (Hassabis et al., 2017). For example, Artificial Neural Networks (ANNs) were initially proposed in the 1940's, inspired by the organization and learning mechanisms observed in the brain (McCulloch and Pitts, 1943; Hebb, 1949). Up until a few years ago, ANNs were mainly used by academics. However, the recent success of Deep Learning (DL) in real-world problems has led to growing interest in industry that has fueled an unprecedented growth of artificial neural networks research (Sinz et al., 2019). In particular, DL and Reinforcement Learning (RL) have received a great deal of attention, not only from the scientific community but also from the general public due to the diverse sectors on which they are being applied such as health care, finance, and technology. DL is an area of machine learning in which ANNs with multiple layers are used to extract high-level features from their inputs (i.e., “deep”). RL on the other hand, is a methodology inspired by conditioning experiments in psychology in which agents learn to maximize rewards and minimize punishment by interacting in their environments (Sutton and Barto, 2018).
Despite the undeniable success of machine learning applications to diverse problems and progress in this field with regard to improving the algorithms and technology to train ANNs—which sometimes has lead to misleading remarks in the media about the real state of the field – there are criticisms about the real advancement in the AI field. One of the main criticisms is the lack of generalization and the amount of training examples that DL algorithms require for learning to solve even simple and structured tasks. In contrast, biological systems can learn complex tasks quickly and extract semantic knowledge from a relatively small number of instances. From this point of view, AI can greatly benefit from applying general principles that real brains employ to solve complex tasks.
In parallel to the advances in AI, the field of neuroscience has experienced tremendous progress in recent years due to the technological advances that allow high density recordings of brain activity with unprecedented spatiotemporal resolution from multiple parts of the brain simultaneously (Steinmetz et al., 2018). Some of these recent advances in the neuroscience of spatial navigation led to the Nobel Prize in Medicine being awarded to John O'Keefe and Edvard and May-Brit Moser (Colgin, 2019). Despite this notable progress in neuroscience, there is a concern that with big data, the field needs to improve the analytical tools to help us make sense of the complex and vast quantities of data that we are now capable of recording (Jonas and Kording, 2017; Vogt, 2018; Vu et al., 2018). Besides analytical tools to make sense of the data, there is a need for tools that help generate new hypotheses based on these richer and larger datasets.
With progress in both Neuroscience and AI, there is a recent renewed interest to conduct research bridging these two fields so that they may benefit from each other (Hassabis et al., 2017; Jonas and Kording, 2017; Richards et al., 2019). One obvious and already successful interaction between AI and Neuroscience is to use machine learning (ML), an area of AI that applies computer science and statistical techniques in data analysis, to study complex and large datasets in Neuroscience (Vogt, 2018; Vu et al., 2018). For example, ML methods are used to analyze motion characteristics of behaving animals to predict cognitive function (Ryait et al., 2019), and have been used in studies of sensory processing to determine optimal stimuli for representations in primary visual cortex (Walker et al., 2019). Besides using ML as an analytical tool, there are attempts to go further and use artificial neural networks as a model to understand brain function (Musall et al., 2019; Richards et al., 2019).
A potential convergent point in which neuroscience and AI could be combined to learn more about the brain, and advance both fields, is spatial navigation. From a neuroscience perspective, being able to actively explore the world might have been one of the key factors that provided organisms evolutionary advantages that triggered the development of cognitive process such as prediction, attention, learning and memory (Swanson, 2003). For example, it has been hypothesized that the memory processing mechanisms involving the entorhinal cortex and the hippocampus evolved from the mechanisms that compute the relationships of spatial landmarks and the tracking of movements of the body in the world (Buzsáki and Moser, 2013). Even more, these computations enabling neural representations of the animal location carried out in the parietal and entorhinal cortices have been proposed as a general mechanism implemented across the neocortex to represent spatial relationships between objects and as a general mechanism for many conceptual “spaces” (Constantinescu et al., 2016; Behrens et al., 2018; Hawkins et al., 2019). Therefore, by understanding how spatial navigation is carried out in biological systems, we can learn about the underlying cognitive processes that are also important components of intelligent behavior which may further advance AI (Bellmund et al., 2018).
In this paper we propose spatial navigation as a common ground for neuroscience and AI to converge and exchange ideas and expand our knowledge of the brain and, ultimately, complex intelligent behavior. In the following sections we first go into detail about why spatial navigation could be used to learn about the brain and advance AI. Then we review the neurobiology of the rodent spatial navigation system, highlighting the structures that form the main concepts of what we know about space representations in the brain: head-direction, place, grid, and border cells. After, we review the models used to study these structures and the processes involved in spatial navigation. Finally, we highlight the limitations of the proposed approach and conclude by providing future directions in which a closer interaction between the fields could improve our understanding of the brain and ultimately of intelligent behavior.
The Neuroscience of Spatial Navigation
In order to describe how neuroscience and AI are well-suited to benefit from one another, we begin by summarizing what is known about the neuroscience of spatial navigation. Here we briefly describe some of the key findings about the neural correlates of spatial navigation and the computational bases of these neural substrates. In addition, we also provide a summary of the neuroscience of reinforcement learning which is a key ingredient in the development of newer AI approaches to understand how spatial navigation tasks might be solved by biological systems. This section aims to denote the limits of what is known about the neurobiology of spatial navigation but also provides important directions and constrains that models should take into consideration.
First, it is important to note that accurate navigation involves several different strategies to reach a goal location: one can follow a sensory cue that marks a goal location, one can follow a determined sequence of actions (a route), or one can determine which way to proceed by following an internal representation of space (map). Depending on the strategy, multiple cognitive processes (or combinations) are required and therefore involve coordination across several brain regions (see Figure 1). In addition, spatial navigation involves several cognitive processes that are crucial for a broad range of intelligent behavior. For example, spatial navigation is strongly linked to memory and learning, planning, attention, and decision making, among others (O'Keefe and Nadel, 1978; Gallistel, 1990; McNaughton et al., 1991; Skaggs and McNaughton, 1996).
Second, spatial navigation has been proposed to follow two different complementary learning strategies that reflect the processes that are computed in the hippocampus and the striatum (Chersi and Burgess, 2015). Briefly, when the striatum is involved, local and incremental reinforcement learning rules facilitate subsequent learning of spatial navigation tasks based on egocentric information (Figure 1A). In this type of navigation, loops between the cortex and the basal ganglia are proposed to support stimulus-response associations and procedural memory, which are linked to route or cue-based navigation. In contrast, when the hippocampus is involved, faster one-shot associative learning rules are applied to solve spatial navigation. Recent studies in humans link these mechanisms for decision making, in which model-free choice guides route-based navigation and model-based choice directs map-based navigation (Anggraini et al., 2018).
Spatial Navigation and the Neural Basis of the Cognitive Map
The theory of the “cognitive map” proposes that the brain creates a representation (or model) of the environmental space that is used to navigate (Tolman, 1948; O'Keefe and Nadel, 1978; Gallistel, 1990). Experimental support for this theory is derived from studies that require rodents (and humans) to solve navigational tasks where the goal location is not visible from an animals current location (Knierim and Hamilton, 2011). Navigation to precise “hidden” locations, or place navigation, can be performed by referencing distant landmarks (or allocentric frame of reference), or by referencing one's body orientation in relation to cues and executing a sequence of actions to the goal (an egocentric frame of reference). Some theoretical work suggests that allocentric and egocentric frames of reference can operate sequentially such that information is decoded to determine a subject's egocentric orientation in the environment and vice versa (McNaughton et al., 1995; Byrne and Becker, 2007; Burgess, 2008; Clark et al., 2018). For example, an allocentric to egocentric transformation may allow a subject to select an action (turn left) at a specific intersection (a particular allocentric location and orientation) in a city. In addition, animals can localize their position and produce trajectories to goal locations by using self-motion cues, e.g., vestibular, proprioceptive, optic flow often referred to as path integration or dead reckoning (Gallistel, 1990; McNaughton et al., 1991; Whishaw et al., 2001).
The neurobiological basis of map-like spatial representations is thought to involve a network of spatially selective neurons in the mammalian nervous system. These include populations of cells that code for spatial location such as place cells (O'Keefe and Nadel, 1978), grid cells (Hafting et al., 2005; Bonnevie et al., 2013), border cells (Solstad et al., 2008), landmark or object vector cells (Deshmukh and Knierim, 2013; Wilber et al., 2014; Høydal et al., 2019), cells that code for head direction (Taube et al., 1990), cells that code for an animals egocentric orientation with respect to environmental features (Wilber et al., 2014; Hinman et al., 2019; LaChance et al., 2019; Alexander et al., 2020), position along a route (Nitz, 2006), and angular and linear locomotor speed (McNaughton et al., 1994; Sharp et al., 2001; Wilber et al., 2014, 2017; Kropff et al., 2015; Munn et al., 2020). These spatial cell types have been identified in a neural circuit that includes the hippocampal formation and several limbic-thalamic and limbic-cortical regions (see Figures 1A,B). While hippocampal circuitry has been linked with allocentric spatial processing, subcortical regions such as a basal ganglia-cortical circuit are thought to contribute to some forms of egocentric action-based navigation. The latter circuit has also been associated with stimulus-response learning, procedural memory and reward prediction. We briefly describe these spatial cell types in greater detail below to provide relevant biological restrictions that can be used in the development of models to study spatial navigation that can inform neuroscience.
Place Cells and Grid Cells
The hippocampus contains neurons that discharge in specific environmental locations (Figure 2A; O'Keefe and Dostrovsky, 1971) such that populations of these cells encode the present position much like a GPS (O'Keefe and Nadel, 1978). Grid cells have been identified in parahippocampal cortex (medial entorhinal cortex, presubiculum, parasubiculum) and differ from place cells in that they fire in multiple locations forming a hexagonal grid pattern (Figure 2B; Hafting et al., 2005). The location by which place/grid cells form their firing fields are modulated by self-motion stimuli or path integration (McNaughton et al., 2006), and are also modulated by landmarks such as local or distant environmental cues or its overall shape (O'Keefe and Burgess, 1996; Yoder et al., 2011; Krupic et al., 2015).
Border, Landmark or Object Vector Cells
While place and grid cells can be modulated by the shape of the environment, recent work has shown that the parahippocampal cortex (medial entorhinal cortex, presubiculum, parasubiculum) also contains neurons that respond specifically to boundary stimuli (Figure 2C; Solstad et al., 2008; Boccara et al., 2010). In the subiculum, these “border” cells can also discharge at specific distances relative to a boundary (Lever et al., 2009). In addition, cells have been identified in both the hippocampus and medial entorhinal cortex that discharge at a specific direction and distance in relation to a cue (Figure 2D). These place responses have been described as landmark or object vector cell activity (McNaughton et al., 1995; Deshmukh and Knierim, 2013; Wilber et al., 2014; Høydal et al., 2019).
Head Direction Cells
Several interconnected limbic and parahippocampal regions contain populations of neurons termed head direction (HD) cells (Cullen and Taube, 2017; Peyrache et al., 2019; Angelaki and Laurens, 2020; Munn and Giocomo, 2020). HD cells are neurons that fire maximally when an animal points its head in a specific direction (Figure 2E) and a small population of HD cells can accurately track the animals HD (Peyrache et al., 2015; Xu Z. et al., 2019). Research has identified this signal in the anterior thalamic nuclei, retrosplenial, parietal, and parahippocampal (entorhinal, postsubiculum, and parasubiculum) cortices (Taube, 2007). Similar to location specific firing in the hippocampus and parahippocampal cortex (place, grid, border, and object vector cells), the preferred direction of HD cells can be controlled by self-motion cues (angular path integration) and environmental cues (reviewed in Taube, 2007).
Egocentric, Action, and Route Modulated Cells
Recent work has established that a circuit including the parietal and retrosplenial cortex and neighboring hippocampal and subcortical regions play a central role in processing an egocentric coordinate system (Clark et al., 2018; Wang et al., 2018, 2020; Hinman et al., 2019; LaChance et al., 2019). For instance, neural populations in the parietal and retrosplenial cortex fire in response to an animal's egocentric actions or posture (McNaughton et al., 1994; Whitlock et al., 2012; Wilber et al., 2017; Mimica et al., 2018), egocentric orientation relative to a landmark or environmental boundary (Wilber et al., 2014; Alexander et al., 2020), and location along a complex route (Nitz, 2006). The parietal cortex has also been linked to allocentric information processing with some parietal neurons exhibiting allocentric HD correlates, and others modulated by the conjunction of egocentric and HD correlates (Chen et al., 1994; Wilber et al., 2014). Importantly, these conjunctive cell populations and other cells encoding primarily in action centered coordinates anticipate upcoming actions, for example, anticipating a left or right turn (Whitlock et al., 2012; Wilber et al., 2014). In addition, recent work has shown that retrosplenial neurons exhibit HD, position, and spike in the relation to the animals distance between path segments, as well as a conjunctive combination of these firing characteristics (Alexander and Nitz, 2015; Mao et al., 2017, 2018). Thus, the parietal and retrosplenial cortex may be part of a circuit that interfaces between allocentric and egocentric frames of reference (Pennartz et al., 2011; Stoianov et al., 2018). Therefore, these computations performed in the parietal and retrosplenial cortex might be crucial for understanding how transformations between self-centered experiences is related to map-like representations of space.
In summary, the mammalian nervous system encodes a map-like representation of space. This map-like representation is built from cells that code for: spatial location such as place, head direction, the position in the environment with repeating geometric firing patterns—grid cells, borders, the direction and distance of landmarks or objects, an animals egocentric orientation with respect to environmental features, position along a route, and specific combinations of angular and linear locomotor speed. The cells that encode space in allocentric or map-like coordinates are generally found in the hippocampal formation and several limbic-thalamic and limbic-cortical regions. While the cells that encode egocentric or body-centered coordinates are generally found in subcortical regions such as a basal ganglia and posterior cortical regions such as the parietal cortex. The computations carried out by the cells and circuits involved in spatial navigation presented here are thought to be similar to the substrates for cognitive processes such as memory, semantic knowledge extraction, prediction, and decision making which in turn are building blocks of intelligent behavior.
Spatial Navigation and Reinforcement Learning
Another important cognitive process strongly linked to spatial navigation and important in intelligent behavior is learning. The brain structures that are involved in spatial navigation and memory formation are also involved in learning (Bellmund et al., 2018). During spatial navigation, learning can occur first as a trial-and-error process that links memory and reward or punishment signals. However, later with experience, a cognitive map is formed and can be used to infer useful spatial information (Buzsáki and Moser, 2013). This adaptive process exploits previous experience to improve the outcome of future choices using different strategies that are implemented in different areas of the brain, including the hippocampus (Johnson and Venditto, 2015). Covering all aspects of learning underlying spatial navigation is beyond the scope of this review where we will focus an aspects that is amenable to AI approaches, RL. We briefly summarize the area of reinforcement learning and the brain structures that are involved in the process of sequential decision making that are crucial to navigate.
In RL there are two main approaches to implement learning, model-free and model-based (and also hybrid approaches). In model-free learning, there is no representation of the world. Instead, learning happens based on more immediate sensory-actions associations. In contrast, in model-based learning there is an internal representation of the environment (Lee et al., 2012). These two approaches nicely overlap with the egocentric and allocentric frames of reference for spatial navigation and have been proposed to work together (Khamassi and Humphries, 2012). Model-free learning is proposed to be implemented by cortical-basal ganglia loops that participate in sensory-actions associations used to update the value function (Figure 1B). In spatial navigation, analogous computations are thought to be egocentric (or route-based) in which no cognitive map is used to reach a goal location. Instead, in model-based learning the animal (agent) uses an internal representation of the environment to update the value function. This representation corresponds to the cognitive map in spatial navigation in which the position of the animal in the environment is updated. In this case, these processes are thought to be implemented by the interaction between the hippocampus and the ventral striatum (Pennartz et al., 2011). In AI and neuroscience, research in RL and rodent spatial navigation has already proved to be a successful approach to understand how these two concepts are closely related and the interaction between these fields can help to inform one another. For example (Stoianov et al., 2018), demonstrated how a RL model can replicate results in rodent experiments in which contextual cues are manipulated to explore the behavioral and brain constrains in goal directed navigation tasks.
Although more research is needed to clarify the details about the neuroscience of the interaction between the navigation and learning systems, there is increasing progress in this area. For example, it is known that the basis of route-based navigation involves brain structures which encode sensory-action associations such as the striatum. In particular, the sensory information reaches the dorsal striatum from the corresponding cortical areas. The dorsal striatum sends projections to the substantia nigra reticulata, which in turn receives inputs from the substantia niagra compacta and the ventral tegmental area (VTA) (Figure 1B). The reward signals are thought to originate in the VTA. Therefore, it has been hypothesized that, since the ventral striatum receives direct input from the SNc/VTA and the hippocampus, the associations between place and reward signals are performed in the latter structure (Chersi and Burgess, 2015).
In summary, different structures interact in spatial navigation and learning depending on the strategy used. Although more research in this area is needed, this section provides a brief review of the neuroscience underlying this ability within the context of RL. This knowledge can inform and guide some of the parameters used in artificial agents solving spatial navigation tasks. This informed development of AI systems will in turn provide opportunities to validate the result against rodent experiments and therefore, to generate hypothesis that can inform neuroscience.
Models for Spatial Navigation and Their Contribution to the Understanding of the Brain
Major advances in our understanding of how the brain is involved in spatial navigation has been achieved in part, due to modeling work. Depending on what the goal of the model is, it can be classified as descriptive, mechanistic, or normative (Dayan and Abbott, 2001). Descriptive models of spatial navigation have the goal of characterizing what the system does, usually reproducing experimental data (Sutherland and Hamilton, 2004). Mechanistic models of the spatial navigation system provide an explanation of how spatial navigation is solved using processes and mechanisms. In this modeling approach, explicit implementations and assumptions are derived from observations and hypotheses from experimental work. Here we will use the term “hypothesis driven models” to encapsulate both descriptive and mechanistic models. Finally, in normative models (or end-to-end models in AI) of spatial navigation the goal is to understand why the brain might solve spatial navigation in a particular way, so the assumptions are less explicit. In this section, we describe the modeling work that encompasses all these approaches that have been developed to understand how animals navigate in space. In addition, we describe modeling work on reinforcement learning which has been important for the development of end-to-end AI approaches that tackle spatial navigation tasks.
Hypothesis-Driven Models of Head-Direction, Place and Grid Cells for Spatial Navigation
When navigating using path integration, it is necessary for the brain to encode the spatial location and update this information with the direction and the speed of motion. Hippocampal place cells are thought to provide this critical information. The modeling complexity of the activity of place cells largely varies depending on the goal of the study. For example, place cells have been modeled as an arrangement of radial basis functions (e.g., Gaussian) each centered at its corresponding place field and collectively the set of place fields cover the environment (Bonnevie et al., 2013; Cazin et al., 2020). In these cases, the position of the agent has been modeled as the actual position in the environment (Arleo and Gerstner, 2000). In other models for which the goal is to study the spatial representations, the current position and distance from the centers of the place field is derived from sensory and idiothetic information (Banino et al., 2018; Cueva and Wei, 2018).
In turn, HD cells have been modeled as a recurrent network that maintains a representation of the current orientation and is modulated by self-motion signals (Figures 2E, 3A). The architecture of this network has been proposed as a ring attractor network (Skaggs et al., 1994; McNaughton et al., 2006; Clark and Taube, 2012; Knierim and Zhang, 2012). Briefly, these models involve HD cells conceptually arranged in an anatomically connected ring with each HD cell occupying a position corresponding to their preferred firing direction. Thus, adjacent HD cells on the “ring” share similar, but slightly offset, preferred firing directions (though not necessarily physically adjacent positions in the brain). The anatomical ring is organized such that functionally adjacent HD cells share strong excitatory connections while cells that occupy different directions (e.g., are 180 degrees apart) share weaker excitatory connectivity (or are inhibited). A consequence of this framework is a sustained hill of excitation centered on the animal's current HD. In most models, external inputs from environmental cues and angular head velocity derived from idiothetic self-motion cues (angular path integration) move the activity hill around the ring (Taube, 2007). Thus, the activity hill is organized to move corresponding to the animal's current HD.
Besides these models of head orientation, there is an extensive body of modeling work to understand how place is represented in the brain. Modeling work has followed two approaches to study the organization of grid cells (Figure 3B). One in which a recurrent network is optimized with hand-chosen parameters to reproduce the hexagonal pattern of activation observed in electrophysiological recordings (McNaughton et al., 2006; Giocomo et al., 2011; Knierim and Zhang, 2012; Navratilova et al., 2012). This approach has been shown to work in simulated conditions (Samu et al., 2009). In these simulations allocentric information derived from a model of grid cells during path integration can correct the accumulated error generated by a noisy representation of speed and direction. In a more recent model (Bush et al., 2015) modeled grid cells to support vector navigation and provide provide a framework for testing experimental predictions. As the neuroscience of spatial navigation uncovers more about how space is represented and manipulated in the brain, models reflect this progress as well. For example, with more studies about grid cells, models that aim to understand how place cells and grid cells interact have been very important to understand the restrictions in the circuitry between the enthorinal cortex and the hippocampus (Solstad et al., 2006).
Hypothesis Driven Models of Spatial Reference Frame Coordination for Navigation
Spatial navigation systems, in mammals at least, are highly robust and adaptable to different levels of sensory information and environmental conditions. For example, mammals are capable of navigating in darkness using internal representations of space or using sensory cues and are capable of rapidly updating these representations when distant cues and landmarks are available (Rosenzweig et al., 2003). These complex navigational strategies rely on the ability to change the frames of reference to use spatial information (Figure 3C). In this section we review the modeling work developed to understand how the brain transforms spatial information between different frames of reference. This is important not only to understand the brain but also because these transformations can be important to extract knowledge by deriving semantic knowledge form episodic memories (Figure 3D; Buzsáki and Moser, 2013; Wang et al., 2020).
Allocentric-Egocentric Reference Frame Transformation
Computational descriptive models propose that cell populations within the anterior thalamic nuclei, parietal cortex, and retrosplenial cortex operate as a network that transforms spatial information from an egocentric (e.g., body centered) to allocentric (i.e., map-like) frame of reference and vice versa (reviewed in Clark et al., 2018). For example, the view of a landmark, which is body-centered in nature, can be transformed into an allocentric, map-like, frame of reference. The details of these reference frame transformations vary slightly between models but are similar regarding the neurobiological subcomponents. For example, in one variant of this framework, McNaughton et al. (1995) model an egocentric-to-allocentric transformation using a linear mapping across a three-layered neural circuit. The initial computations involve a layer composed of two neural populations: an allocentric HD cell signal, which is generated within a subcortical circuit including anterior thalamic-to-cortical projections, and an egocentric cell signal by parietal cortex neurons which are modulated by the animals' egocentric heading relative to a landmark. These two cell types converge post-synaptically on a second layer of parietal cortex cells that encode the conjunction of HD and egocentric signals. The biological validity of these conjunctive cells is supported by recent work (Wilber et al., 2014). Conjunctive cells are then associated with a third layer encoding the bearing of a landmark (allocentric direction), which when combined with information regarding the relative distance from the landmark, produces a signal reflecting a vector relative to the landmark. In sum, this model demonstrates how sensory information can be used to support spatial localization by transforming egocentric information into a location in the environment. In other words, transform body centered encoding of a landmark into map-like landmark representations (e.g., a cell that fires in a specific map-like location relative to a landmark independent of which direction the animal is facing; Figure 2D). Recent studies have identified neural correlates resembling exactly this model output, or landmark vector cell responses, within the hippocampus and entorhinal cortex (Deshmukh and Knierim, 2013; Wilber et al., 2014; Høydal et al., 2019). In addition, the opposite transformation in which allocentric location is decoded to determine egocentric orientation could be performed using a similar 3-layer network. For instance, if a subject is required to select the appropriate action (turn left) for a specific intersection in a city (a specific place and direction) given the current goal (e.g., to go to the bank vs. coffee).
There are mechanistic modeling studies in which spatial representations found in the mammalian navigation system are used to study how different frames of reference can be used to navigate when different sensory information is available. For example (Byrne and Becker, 2007), implemented a model that shows how egocentric and allocentric frames of references can be built and how transformation from one to another can be carried out. Moreover, in a related study, it is shown how these transformations relate to short and long-term memory (Krichmar and Edelman, 2005). More recently (Oess et al., 2017), showed how the hippocampus, the parietal cortex and retrosplenial cortices could interact to solve spatial navigation tasks using an egocentric, an allocentric or route-centric frames of references. These simulations can help to understand how the transformation of egocentric and allocentric frames of reference can be employed by the brain when using different navigation strategies. Another example in which modeling aspects of the rodent spatial navigation system has helped to understand the integration of self-motion and visual information to represent the localization in space is by using an attractor-based network model (Campbell et al., 2018). In this work they use this model to understand how optic-flow, locomotion and landmark cues produce activity patterns in the medial entorhinal cortex to represent spatial position during navigational tasks.
An ANN Model Approach to Solve Spatial Navigation
In a traditional approach to modeling brain function, such as spatial navigation, the model parameters are specified by the experimenter and optimized to reproduce experimental data. More recently, with the advancement in ANNs, there are more AI end-to-end (normative) approaches to model spatial navigation in which the parameters that determine the representations and how they are exploited are not specified explicitly. Instead, ANNs are trained to solve spatial navigation tasks and the representations and parameters employed by the network are restricted to match biological constraints. In this section we recapitulate modeling work that has the goal of advancing the understanding of the rodent spatial navigation system using both AI and neuroscience approaches.
The spatial representations that ANNs use to solve spatial navigation resemble different properties reported in rodent experiments. For example Kanitscheider and Fiete (2017), trained a RNN to perform a series of spatial navigation tasks that require self-localization and mapping, a very well-studied and useful property in robotic navigation. In this work, the authors show that the representations that are exploited by the trained network resemble characteristics of the biological spatial navigation system such as place cells that remap between environments (thought to represent the neural substrate of unique cognitive maps for different locations). Similarly Cueva and Wei (2018), showed how an agent using a recurrent neural network (RNN) can solve a spatial navigation task to study the spatial representations used by such network. In the trained RNN, they found a grid-cell like representation of space in which a hexagonal periodic pattern of activity was used to keep track of the location of the agent in the environment. The spatial representation exploited by this network did not combine the sensory raw input and motion signals as in other models (Samu et al., 2009). Instead, this representation emerged because of the integration of the speed and direction signals given to the network and the metabolic restriction implemented in the cost function used to train the network. One limitation of these RNN approaches is that the learning mechanism used to train the agent to solve the task were not biologically plausible.
Recently, there are studies in which applying biologically relevant restrictions to the ANNs led to understanding how these processes occur in the brain. For example Whittington et al. (2018), propose a model inspired by the hippocampal-entorhinal cortex system in which grid and place cell like representations emerge when an ANNs is trained to solve navigation in a 2D environment. These representations constructed by the ANN using the end-to-end approach, generalizes from sensory exposures from different environments. Moreover Sorscher et al. (2019), analyzed the conditions in which grid cell-like representations emerge from models optimized for spatial navigation. In this work, the authors were able to ascertain which constraints favor the hexagonal activation pattern of grid cell like representation emerged in three different network architectures.
In summary, this end-to-end approach in which ANNs are used to model brains in embodied agents that learn to navigate in space using relevant biological restrictions provides a promising tool to study the representations of space that might resemble those used in nature and further our understanding of how such spatial representations may “emerge.”
Reinforcement Learning and Spatial Navigation
An important related element in the mammalian spatial navigation system is learning. One way to model how embodied agents learn to navigate is using RL. In this section we summarize how RL modeling has been successfully integrated in AI approaches to understand spatial navigation.
Recently, DNNs have been combined with RL (termed as Deep RL) in spatial navigation tasks. In this framework, spatial representations are learned by interacting with the environment instead of provided by the experimenter. In this version of RL, deep NN are used to approximate value and transition functions using environmental (sensory) information (instead of a look up table or another function approximation method) (Botvinick et al., 2019). Due to considering multiple layers (i.e., deep), deep RL leverages this organization to learn spatial representations that generalize well and can be transferred to different tasks (Mirowski et al., 2017; Banino et al., 2018; Botvinick et al., 2019). For instance Banino et al. (2018), used deep learning in simulated agents to study how space representations can be used to facilitate flexible navigation strategies that closely resemble experimental data from rodents. In this work, the authors trained the deep network to perform path integration using trajectories from real rodents. One of the intermediate representations that the simulated agents used to keep track its location when doing path integration was grid cell-like activity patterns. Moreover, the authors investigated whether this representation could be exploited to perform goal directed navigation. This time, the authors included a RL module which learned to associate values to specific locations in the environment. Actions that brought the agent closer to the goal were associated with higher value. Using this strategy, these agents solved the goal directed navigation tasks using shortcuts when possible, compared to agents only using place and head direction cells to navigate (Figure 3E). This type of work in which similar representations to the ones found in real brains are used to solve navigation tasks is important because they provide opportunities to learn more about how similar processes might happen in the brain.
Besides using ANNs and RL to solve spatial navigation tasks, important concepts, and mechanisms found in neuroscience experiments have been used to improve algorithms in AI. For example, an important mechanism that links what we know about learning and memory and spatial navigation are the theories about how memories are consolidated in the brain. In one theory about memory, hippocampal replay plays a crucial role in forming an index or memory trace that binds together experience components in the neocortex for long-term storage and knowledge extraction during sleep (Frankland and Bontempi, 2005). There are numerous examples of experimental evidence that replays occurs in multiple areas of the brain (Skaggs and McNaughton, 1996; Kudrimoti et al., 1999; Euston et al., 2007; Bermudez Contreras et al., 2013; Wilber et al., 2017). This replay mechanism has been proposed to reduce the number of “iterations” required to explore an environment in RL (Cazé et al., 2018; Momennejad et al., 2018; Cazin et al., 2019).
Finally, despite the contribution of more abstract models of spatial navigation, there are models in which AI takes a more ethological and embodied approach to study spatial navigation. From this perspective, cognition is not only a product of isolated computations occurring in the brain but instead emerge from the interaction between the body and the environment (Noe and O'Regan, 2001; Thelen and Smith, 2007; Bonner and Epstein, 2017). Therefore, from this perspective, instead of solely looking for the responsible brain structures involved in spatial navigation and their neural codes, the study of the restrictions imposed by the environment and anatomy of the organism, might help to better understand how the internal representations are constructed and how they are manipulated to navigate in space (Krichmar and Edelman, 2005; Evans et al., 2016; Brette, 2019; Santoro et al., 2019).
We have summarized modeling work that has the goal of improving our understanding of the mechanisms involved in spatial navigation and how they are implemented by the brain. The range of such work varies from descriptive and mechanistic models in which the goal is to reproduce experimental data using explicit hypotheses about brain organization, to more recent approaches that rely less on explicit experimenter definitions and use ANNs as a model of the brain. In addition, we have summarized the neurobiology of RL and how RL has been implemented to solve spatial navigation tasks. Regardless of the level of abstraction and the questions they aim to answer, this work expands our knowledge of the brain by providing predictions, generating new hypotheses and demonstrating how the cognitive processes necessary for complex behavior might rise from spatial navigation.
Discussion
Here we reviewed the neuroscience and modeling work of spatial navigation. We proposed that by understanding how spatial navigation is solved by the brain, we could provide useful insights to alleviate some current problems for AI. Conversely, a way to expand our current understanding of the neuroscience of spatial navigation is to use AI and machine learning techniques to aid in analyzing some large data sets. This has historically been demonstrated by testing theories of how the brain performs spatial navigation using descriptive and mechanistic models of the hippocampal formation. More recently, the development of end-to-end or normative approaches using ANNs to solve spatial navigation problems while following relevant biological restrictions can further our understanding of how and why representations might emerge and be manipulated by the brain. In this section we outline how considering spatial navigation as the intersection point between neuroscience and AI research can provide a valuable opportunity to advance both fields and we review the limitations of the approaches presented in this paper.
Neuroscience Contributions to Advance AI
As previously mentioned, there are strong examples of contributions from neuroscience and psychology to the advancement of AI, such as the inspiration for connectionism and of ANNs (Rumelhart et al., 1988), the hierarchical organization of the mammalian visual processing in the cortex for the development of deep learning (Schmidhuber, 2014), the successful application of attentional mechanisms to active computer vision (Bermudez Contreras et al., 2008) or training ANNs (Graves, 2013; Sutskever et al., 2014), and the impressive development of RL systems that can beat world-class players at highly-complex games (Botvinick et al., 2019). Despite these contributions that have propelled the recent impressive progress and applications of DL and RL, there are important limitations in these areas which can benefit from the knowledge generated in neuroscience. First, a limitation of the current state of DL is that, despite being used in impressive applications, these networks suffer from poor generalization or fail to extract semantic knowledge from the large training data sets. For example, successful applications of classic DL are highly optimized non-linear classifier systems that require many training examples to fine tune a large number of parameters rather than systems that extract knowledge by building robust semantic understanding of the inputs. In contrast, the analogous biological networks show a great deal of generalization during learning. The way that the brain performs spatial navigation might provide valuable insights into how to solve this limitation in current AI methods. First, memory consolidation is hypothesized to use a mechanism in which relevant episodes (time and space dependent) are formed into memories from which semantic knowledge (context independent) is extracted. These mechanisms, possibly implemented in the same network, might be similarly employed for navigation in the formation of cognitive maps from repeated exposure to self-centered exploration episodes (Figure 3C; Lever et al., 2002; Buzsáki and Moser, 2013).
Another related limitation is the number of training examples that DL requires to learn. This limitation is contrasted with the biological counterparts in which learning happens very rapidly in most cases. This might be due to the fact that brains are not completely randomly connected at birth such that we have to learn everything from scratch. Instead, there are developmental processes that determine pre-wired networks and mechanisms that bootstrap innate behaviors (Zador, 2019). Thus, neuroscience may be able to inform AI so that models can combine learning with evolutionary and developmental approaches in which plasticity and circuit refinement build upon pre-wired brain networks. A similar limitation in RL arises in complex environments where agents require a large number of exposures to the environment in order to improve policies (which is the way that determines how the agent interact with its environment). A neuroscience-inspired mechanism to reduce the number of required exposures for learning that is also implemented by structures involved in spatial navigation, is to use previous experiences to select possible actions for new situations. This approach, known as learning to learn (episodic meta RL), has already been successfully applied to solve spatial navigation tasks (Chalmers et al., 2016, 2017; Duan et al., 2017; Finn et al., 2017; Botvinick et al., 2019). In addition, a well-studied process in neuroscience is memory consolidation in which the replay of previous experiences helps to extract semantic knowledge from episodic instances. There is also evidence that not only previous experiences but also unexplored possibilities are used to evaluate possible outcomes in navigation tasks in rodents (Dragoi and Tonegawa, 2011). A similar approach has been applied in ANNs to solve spatial navigation in simulated agents and robots (Cazin et al., 2019, 2020). Another important interaction between AI and Neuroscience in spatial navigation has been the idea that the hippocampus is not a spatial cognitive map but instead, a prediction map (Evans and Burgess, 2019). From this perspective, hippocampal activity encodes the animal's future locations which are restricted by the environment and their value (rewards) (Stachenfeld et al., 2017; Brunec and Momennejad, 2019).
An additional aspect in which neuroscience could provide ideas to advance AI and in particular ANNs is by incorporating what is known about brain architecture. At the moment, most of the deep learning approaches use a limited repertoire of what is known about how brain cells compute information. For example, the richness of neuronal types, network topologies or the additional dynamics provided by neuromodulators, the combination of local and long-range synaptic connections might improve the capabilities or reduce the limitations of AI approaches (Ullman, 2019). One important aspect of the representations derived from ANNs is their robustness. The generalization achieved by these systems can be enhanced using regularization techniques during training such as dropout (Srivastava et al., 2014; Hawkins and Ahmad, 2016). Introducing variability or “noise” in the training data or the computing units (Destexhe and Contreras, 2006; Guo et al., 2018; Wu et al., 2019) (arguably modeling their biological counterparts) can shape the properties of their representations (Faisal et al., 2008). In spatial navigation for example, this variability is useful for favoring the emergence of robust representations that resemble the spatial representations found in the medial temporal lobe (Banino et al., 2018). The robustness to environment variability, noisy sensors, and actuators of these emergent spatial representations has been proposed to be crucial in the efficiency of biological systems to navigate (Vickerstaff and Cheung, 2010).
From the AI perspective, one of the main criticisms of the current development of DL and AI in trying to understand the brain is that, recently, the main focus of such developments has been to exploit the computational power of deep architectures for technological advancement. This emphasis produced powerful classification devices that are poor at generalizing and at extracting semantic knowledge. However, such processes have been proposed to advance our understanding of intelligent behavior (Hassabis et al., 2017; Zador, 2019). Analogously, artificial autonomous navigation is an active area of AI research for engineering driverless vehicles (Lipson and Kurman, 2016). Even though one of the most popular algorithms in autonomous vehicles has a version based on certain aspects of the neuroscience of the navigation system in rodents (Milford et al., 2010; Ball et al., 2013; Xu L. et al., 2019), this particular approach has not been designed to advance what we know about the brain, suggesting a potentially unrealized opportunity for synnergy between the neuroscience of spatial navigation and AI (Dudek and Jenkin, 2002; Zafar and Mohanta, 2018).
AI Contributions in Neuroscience Research
The contribution of AI and machine learning in neuroscience is 2-fold. At a more straightforward level, machine learning provides an analysis tool to make sense of brain activity and behavior in animal models in neuroscience. For example, DNNs have also been applied to analyze animal behavior to predict motor impairments in a mouse model of stroke. In this work, the trained network was able to predict the volume of brain tissue affected by stroke (Ryait et al., 2019). Similarly, DNNs have been used for pose estimation of animal videos (Mathis et al., 2018). In addition, and more related to spatial navigation, DNNs have also been applied to decode sensory and behavioral information such as animal position and orientation from hippocampal activity (Frey et al., 2019; Xu Z. et al., 2019). In a less direct way, AI has advanced neuroscience by providing a model of the brain. For example, DNNs have been used to reproduce brain activity in the visual system to learn about the organization of this network in primates (Walker et al., 2019) and mice (Cadena et al., 2019). Analogously and as previously mentioned, AI approaches have been used as a model of the brain to understand how spatial representations emerge and under what conditions (Banino et al., 2018; Cueva and Wei, 2018; Sorscher et al., 2019).
There are limitations and criticisms for these contributions in neuroscience. One common criticism for machine learning approaches in neuroscience is the employment of DL as a model of the brain. This criticism comes from two perspectives. One is the complexity, because by employing such models it is difficult to analyze them and to fully explain their behavior, particularly when using end-to-end approaches. However, there is progress in the development of approaches to understand the computations that complex deep ANNs carry out to produce their outputs. For example Cohen et al. (2020), propose a layer-by-layer approach in which the features represented at different layers of a deep network trained to classify images can be understood as a hierarchical architecture that gradually builds the concepts exploited by the network. A complementary approach proposed by Tishby and Zaslavsky (2015) and Alemi et al. (2019) consists of analyzing what information is passed onto the next layer, a process called “information bottleneck.” For a review of these approaches see Montavon et al. (2018). Despite this limitation, this approach might still provide controlled, reproducible experimental sand-boxes to improve our current analytical tools (that can be applied to real brain data) or to generate and test new hypotheses (Jonas and Kording, 2017). Another concern is that, in some cases, ANNs employ non-biologically plausible algorithms (Zador, 2019). For example, in some cases, popular and successful algorithms such as backpropagation are used to train such ANNs, which do not emulate the way in which the brain “learns” (Richards et al., 2019). This criticism raises the possibility that even if we can train ANNs that perform spatial navigation, this is not a guarantee that the brain might solve the task in a similar way (Burak and Fiete, 2009; Kanitscheider and Fiete, 2017). It is in these cases where a reciprocal interaction with neuroscience research can provide inspiration to propose new biologically relevant learning algorithms. For example, there are attempts to implement biologically plausible algorithms (including variants of backpropagation) to train deep ANNs (Roelfsema and Holtmaat, 2018; Sacramento et al., 2018; Pozzi et al., 2019). Therefore, imposing biologically plausible restrictions to the artificial environment, body and brain, might allow one to directly compare the solutions found by AI to their biological counterparts to determine whether these solutions might inform us about how the brain performs spatial navigation (Sinz et al., 2019).
Another criticism of the machine learning approach and in particular of using DNNs to understand brain function and ultimately, intelligent behavior, is that this approach disregards the “refinement phylogenetic” of biological organisms (Cisek, 2019). One proposal to address this is to start with the minimal cognitive functions that gave rise to navigation in simple organisms (Yamauchi and Beer, 1996; Beer and Williams, 2015). The feedback from neuroscience can provide useful insights for the advancement of such approaches (Webb and Wystrach, 2016; Graham and Philippides, 2017). Although, the neurobiological basis of spatial navigation and its relationship to learning and memory in simpler organisms like insects are not as well-understood as in rodents.
Overall, despite the great advances in the understanding of the neurobiology of the spatial navigation system in rodents, there are important open questions in neuroscience that can benefit from AI approaches. In particular, end-to-end approaches to solve navigation tasks can help in the advancement of the neuroscience of spatial navigation because the potential solutions are not restricted to the current knowledge of the experimenter. This is a very important point in the generation of new hypotheses about how the brain might solve a complex task. For example, the manipulation of spatial representations is difficult to study with current approaches in neuroscience (Kanitscheider and Fiete, 2017). This is due to the difficulty of experimental preparations and lack of tools to analyze such complex data. AI can help neuroscience research in both regards. On the one hand, a successfully trained ANN that solves a navigation task provides the opportunity to repeat and manipulate environmental conditions (e.g., sensory inputs) and parameters (e.g., network topology) to gain insights into possibly interesting avenues to follow in rodent experiments. On the other hand, it provides an opportunity to develop analytical tools to understand the complex mechanisms employed in these models and apply them to real data.
Finally, we want to clarify that the classification of the models presented here is not necessarily exhaustive, mutually exclusive or discrete. There are end-to-end models that have different levels of assumptions or use a hypothesis driven approach to different extents. Similarly, in theoretical neuroscience normative models might not be considered completely equivalent to end-to-end models in AI. However, the purpose of the classification presented in this paper was to highlight the differences of end-to-end AI approaches that can further our understanding of brain function and provide a common language that can bridge the communication gap between the neuroscience of spatial navigation and AI.
Conclusions and Future Directions
In this paper we propose spatial navigation as a common ground where research in neuroscience and AI can converge to expand our understanding of the brain. We suggest that by understanding how the brain carries out the cognitive processes to solve a complex task such as spatial navigation, we will be in a better place to understand how intelligent behavior might arise. There are several reasons why we propose this. First, spatial navigation is a complex task that involves areas and cognitive processes in the brain that are crucial for intelligent complex behavior. Second, by using spatial navigation as a problem to be solved by artificial systems that follow biologically relevant restrictions, we can use this as a “sandbox” to improve our analytical tools. Our hope is that this paper provides a starting point for researchers of both fields to further the understanding of spatial navigation as a path to advance the study of how biological brains produce intelligent behavior.
To conclude, by building models and agents that solve spatial navigation tasks following the restrictions imposed by the interactions of the body and environment found in biological systems, we argue that we can not only learn more about the brain but also how the processes involved in complex intelligent behavior might rise. Moreover, by producing comparable solutions that can be validated against experimental results in neuroscience, we might advance the development of ANNs and overcome current limitations. Promising research avenues can be drawn from the approaches and studies presented here. For example, further investigation of learning of spatial representation at multiple scales in time and levels of abstraction, the role of memory in these processes, knowledge extraction, learning to learn, and understanding how the brain performs coordinate transformation between body-centered and map-like representations. These are crucial cognitive components of intelligence which can have a great impact in neuroscience and AI.
Despite much room for future growth, the early joint efforts between AI and neuroscience to understand the neural substrates for spatial navigation is gaining traction and is becoming an exciting and promising approach. We propose that this synergetic interaction will help to understand how cognitive processes give rise to intelligent behavior but also will contribute to the development of better tools to do so.
Author Contributions
EB-C conceived and presented the original idea. All authors discussed the contents and contributed to the final manuscript.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
We thank Paul Graham and Andrew Philippides for useful comments in early versions of this paper and Victoria Roy for assistance with figures. We thank Jenna Bailey for useful suggestions about the organization of the text. BC was supported by a National Institute of Health grant AA024983 and an Alzheimer's Association grant AARG-17-531572. AW was supported through the National Institute of Health grant AG049090 and the Florida Department of Health grant 20A09.
References
- Alemi A. A., Fischer I., Dillon J. V., Murphy K. (2019). Deep variational information bottleneck, in 5th International Conference on Learning Representations ICLR 2017e Track Proc (Toulon: ), 1–19. [Google Scholar]
- Alexander A. S., Carstensen L. C., Hinman J. R., Raudies F., William Chapman G., Hasselmo M. E. (2020). Egocentric boundary vector tuning of the retrosplenial cortex. Sci. Adv. 6:eaaz2322. 10.1126/sciadv.aaz2322 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alexander A. S., Nitz D. A. (2015). Retrosplenial cortex maps the conjunction of internal and external spaces. Nat. Neurosci. 18, 1143–1151. 10.1038/nn.4058 [DOI] [PubMed] [Google Scholar]
- Almássy N., Edelman G. M., Sporns O. (1998). Behavioral constraints in the development of neuronal properties: a cortical model embedded in a real-world device. Cereb. Cortex 8, 346–361. 10.1093/cercor/8.4.346 [DOI] [PubMed] [Google Scholar]
- Angelaki D. E., Laurens J. (2020). The head direction cell network: attractor dynamics, integration within the navigation system, and three-dimensional properties. Curr. Opin. Neurobiol. 60, 136–144. 10.1016/j.conb.2019.12.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Anggraini D., Glasauer S., Wunderlich K. (2018). Neural signatures of reinforcement learning correlate with strategy adoption during spatial navigation. Sci. Rep. 8:10110. 10.1038/s41598-018-28241-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- Arleo A., Gerstner W. (2000). Spatial cognition and neuro-mimetic navigation: a model of hippocampal place cell activity. Biol. Cybern. 83, 287–299. 10.1007/s004220000171 [DOI] [PubMed] [Google Scholar]
- Ball D., Heath S., Wiles J., Wyeth G., Corke P., Milford M. (2013). OpenRatSLAM: an open source brain-based SLAM system. Auton. Robots 34, 149–176. 10.1007/s10514-012-9317-9 [DOI] [Google Scholar]
- Banino A., Barry C., Uria B., Blundell C., Lillicrap T., Mirowski P., et al. (2018). Vector-based navigation using grid-like representations in artificial agents. Nature 26, 429–433. 10.1038/s41586-018-0102-6 [DOI] [PubMed] [Google Scholar]
- Beer R. D., Williams P. L. (2015). Information processing and dynamics in minimally cognitive agents. Cogn. Sci. 39, 1–38. 10.1111/cogs.12142 [DOI] [PubMed] [Google Scholar]
- Behrens T. E. J., Muller T. H., Whittington J. C. R., Mark S., Baram A. B., Stachenfeld K. L., et al. (2018). What is a cognitive map? Organizing knowledge for flexible behavior. Neuron 100, 490–509. 10.1016/j.neuron.2018.10.002 [DOI] [PubMed] [Google Scholar]
- Bellmund J. L. S., Gärdenfors P., Moser E. I., Doeller C. F. (2018). Navigating cognition: spatial codes for human thinking. Science 362:eaat6766. 10.1126/science.aat6766 [DOI] [PubMed] [Google Scholar]
- Bermudez Contreras E., Buxton H., Spier E. (2008). Attention can improve a simple model for visual object recognition. Image Vis. Comput. 26, 776–787. 10.1016/j.imavis.2007.08.014 [DOI] [Google Scholar]
- Bermudez Contreras E. J., Schjetnan A. G. P., Muhammad A., Bartho P., McNaughton B. L., Kolb B., et al. (2013). Formation and reverberation of sequential neural activity patterns evoked by sensory stimulation are enhanced during cortical desynchronization. Neuron 79, 555–566. 10.1016/j.neuron.2013.06.013 [DOI] [PubMed] [Google Scholar]
- Boccara C. N., Sargolini F., Thoresen V. H., Solstad T., Witter M. P., Moser E. I., et al. (2010). Grid cells in pre-and parasubiculum. Nat. Neurosci. 13, 987–994. 10.1038/nn.2602 [DOI] [PubMed] [Google Scholar]
- Bonner M. F., Epstein R. A. (2017). Coding of navigational affordances in the human visual system. Proc. Natl. Acad. Sci. U.S.A. 114, 4793–4798. 10.1073/pnas.1618228114 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bonnevie T., Dunn B., Fyhn M., Hafting T., Derdikman D., Kubie J. L., et al. (2013). Grid cells require excitatory drive from the hippocampus. Nat. Neurosci. 16, 309–317. 10.1038/nn.3311 [DOI] [PubMed] [Google Scholar]
- Botvinick M., Ritter S., Wang J. X., Kurth-Nelson Z., Blundell C., Hassabis D. (2019). Reinforcement learning, fast and slow. Trends Cogn. Sci. 23, 408–422. 10.1016/j.tics.2019.02.006 [DOI] [PubMed] [Google Scholar]
- Brette R. (2019). Is coding a relevant metaphor for the brain. Behav. Brain Sci. 42:e215. 10.1017/S0140525X19001997 [DOI] [PubMed] [Google Scholar]
- Brunec I. K., Momennejad I. (2019). Predictive representations in hippocampal and prefrontal hierarchies. bioRxiv [Preprint]. 1–42. 10.1101/786434 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Burak Y., Fiete I. R. (2009). Accurate path integration in continuous attractor network models of grid cells. PLoS Comput. Biol. 5:e1000291. 10.1371/journal.pcbi.1000291 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Burgess N. (2008). Spatial cognition and the brain. Ann. N.Y. Acad. Sci. 1124, 77–97. 10.1196/annals.1440.002 [DOI] [PubMed] [Google Scholar]
- Bush D., Barry C., Manson D., Burgess N. (2015). Using grid cells for navigation. Neuron 87, 507–520. 10.1016/j.neuron.2015.07.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Buzsáki G., Moser E. I. (2013). Memory, navigation and theta rhythm in the hippocampal-entorhinal system. Nat. Neurosci. 16, 130–138. 10.1038/nn.3304 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Byrne P., Becker S. (2007). Remembering the past and imagining the future : a neural model of spatial memory and imagery. Psychol. Rev. 114, 340–375. 10.1037/0033-295X.114.2.340 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cadena S. A., Sinz F. H., Muhammad T., Froudarakis E., Cobos E., Walker E. Y., et al. (2019). How well do deep neural networks trained on object recognition characterize the mouse visual system? in NeurIPS Workshop Neuro-AI (Vancouver, BC: ). [Google Scholar]
- Campbell M. G., Ocko S. A., Mallory C. S., Low I. I. C., Ganguli S., Giocomo L. M. (2018). Principles governing the integration of landmark and self-motion cues in entorhinal cortical codes for navigation. Nat. Neurosci. 21, 1096–1106. 10.1038/s41593-018-0189-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cazé R., Khamassi M., Aubin L., Girard B. (2018). Hippocampal replays under the scrutiny of reinforcement learning models. J. Neurophysiol. 120, 2877–2896. 10.1152/jn.00145.2018 [DOI] [PubMed] [Google Scholar]
- Cazin N., Llofriu Alonso M., Scleidorovich Chiodi P., Pelc T., Harland B., Weitzenfeld A., et al. (2019). Reservoir computing model of prefrontal cortex creates novel combinations of previous navigation sequences from hippocampal place-cell replay with spatial reward propagation. PLoS Comput. Biol. 15:e1006624. 10.1371/journal.pcbi.1006624 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cazin N., Scleidorovich P., Weitzenfeld A., Dominey P. F. (2020). Real-time sensory-motor integration of hippocampal place cell replay and prefrontal sequence learning in simulated and physical rat robots for novel path optimization. Biol. Cybern. 114, 249–268. 10.1007/s00422-020-00820-2 [DOI] [PubMed] [Google Scholar]
- Chalmers E., Bermudez Contreras E., Robertson B., Luczak A., Gruber A. (2016). Context-switching and adaptation: brain-inspired mechanisms for handling environmental changes. Proc. Int. Jt. Conf. Neural. Netw. 2016, 3522–3529. 10.1109/IJCNN.2016.7727651 [DOI] [Google Scholar]
- Chalmers E., Bermudez-Contreras E., Robertson B., Luczak A., Gruber A. (2017). Learning to predict consequences as a method of knowledge transfer in reinforcement learning. IEEE Trans Neural Netw. Learn Syst. 29, 1–12. 10.1109/TNNLS.2017.2690910 [DOI] [PubMed] [Google Scholar]
- Chen L. L., Lin L. H., Green E. J., Barnes C. A., McNaughton B. L. (1994). Head-direction cells in the rat posterior cortex - anatomical distribution and behavioral modulation. Exp. Brain Res. 101, 8–23. 10.1007/BF00243212 [DOI] [PubMed] [Google Scholar]
- Chersi F., Burgess N. (2015). The cognitive architecture of spatial navigation: hippocampal and striatal contributions. Neuron 88, 64–77. 10.1016/j.neuron.2015.09.021 [DOI] [PubMed] [Google Scholar]
- Chiel H. J., Beer R. D. (1997). The brain has a body: adaptive behavior emerges from interactions of nervous system, body and environment. Trends Neurosci. 20, 553–557. 10.1016/S0166-2236(97)01149-1 [DOI] [PubMed] [Google Scholar]
- Cisek P. (2019). Resynthesizing behavior through phylogenetic refinement. Atten. Percep. Psychophys. 81, 2265–2287. 10.3758/s13414-019-01760-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Clark B. J., Simmons C. M., Berkowitz L. E., Wilber A. A. (2018). The retrosplenial-parietal network and reference frame coordination for spatial navigation. Behav. Neurosci. 132, 416–429. 10.1037/bne0000260 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Clark B. J., Taube J. S. (2012). Vestibular and attractor network basis of the head direction cell signal in subcortical circuits. Front Neural Circuits. 6:7. 10.3389/fncir.2012.00007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cohen U., Chung S. Y., Lee D. D., Sompolinsky H. (2020). Separability and geometry of object manifolds in deep neural networks. Nat. Commun. 11, 1–13. 10.1038/s41467-020-14578-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Colgin L. L. (2019). Five decades of hippocampal place cells and EEG rhythms in behaving rats. J. Neurosci. 40:0741-19. 10.1523/JNEUROSCI.0741-19.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Constantinescu A. O., O'Reilly J. X., Behrens T. E. J. (2016). Organizing conceptual knowledge in humans with a gridlike code. Science 352, 1464–1468. 10.1126/science.aaf0941 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cueva C. J., Wei X.-X. (2018). Emergence of grid-like representations by training recurrent neural networks to perform spatial localization, in International Conference on Learning Representations (ICLR), (Vancouver, BC: ), 1–19. [Google Scholar]
- Cullen K. E., Taube J. S. (2017). Our sense of direction: progress, controversies and challenges. Nat. Neurosci. 20, 1465–1473. 10.1038/nn.4658 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dayan P., Abbott L. (2001). Theoretical Neuroscience. Computational and Mathematical Modeling of Neural Systems. Cambridge, MA: MIT Press. [Google Scholar]
- Deshmukh S. S., Knierim J. J. (2013). Influence of local objects on hippocampal representations: landmark vectors and memory. Hippocampus 23, 253–267. 10.1002/hipo.22101 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Destexhe A., Contreras D. (2006). Neuronal computations with stochastic network states. Science 85, 85–90. 10.1126/science.1127241 [DOI] [PubMed] [Google Scholar]
- Dragoi G., Tonegawa S. (2011). Preplay of future place cell sequences by hippocampal cellular assemblies. Nature 469, 397–401. 10.1038/nature09633 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Duan Y., Schulman J., Chen X., Bartlett P. L., Sutskever I., Abbeel P., et al. (2017). RL2: Fast reinforcement learning via slow reinforcement learning, in Fifth International Conference on Learning Representations (ICLR), 1–18. [Google Scholar]
- Dudek G., Jenkin M. (2002). Computational Principles of Mobile Robotics. Cambridge, UK: Cambridge University Press. [Google Scholar]
- Euston D. R., Tatsuno M., McNaughton B. L. (2007). Fast-forward playback of recent memory sequences in prefrontal cortex during sleep. Science 318, 1147–1150. 10.1126/science.1148979 [DOI] [PubMed] [Google Scholar]
- Evans T., Bicanski A., Bush D., Burgess N. (2016). How environment and self-motion combine in neural representations of space. J. Physiol. 594, 6535–6546. 10.1113/JP270666 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Evans T., Burgess N. (2019). Coordinated hippocampal-entorhinal replay as structural inference, in Advances in Neural Information Processing Systems (NeurIPS) (Vancouver, BC: ). [Google Scholar]
- Faisal A. A., Selen L. P. J., Wolpert D. M. (2008). Noise in the nervous system. Nat. Rev. Neurosci. 9, 292–303. 10.1038/nrn2258 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Finn C., Abbeel P., Levine S. (2017). Model-agnostic meta-learning for fast adaptation of deep networks. 34th International Conference on Machine Learning (ICML) 3, 1856–1868. [Google Scholar]
- Frankland P. W., Bontempi B. (2005). The organization of recent and remote memories. Nat. Rev. Neurosci. 6, 119–130. 10.1038/nrn1607 [DOI] [PubMed] [Google Scholar]
- Frey M., Tanni S., Perrodin C., Leary A. O., Nau M., Kelly J., et al. (2019). Deep insight : a general framework for interpretting wide-band neural activity. bioRxiv [Preprint]. 1–23. 10.1101/871848 [DOI] [Google Scholar]
- Gallistel C. R. (1990). The Organization of Learning. Cambridge: Bradform Books; MIT Press. [Google Scholar]
- Giocomo L. M., Moser M. B., Moser E. I. (2011). Computational models of grid cells. Neuron. 71, 589–603. 10.1016/j.neuron.2011.07.023 [DOI] [PubMed] [Google Scholar]
- Graham P., Philippides A. (2017). Vision for navigation: what can we learn from ants? Arthropod. Struct. Dev. 46, 718–722. 10.1016/j.asd.2017.07.001 [DOI] [PubMed] [Google Scholar]
- Graves A. (2013). Generating sequences with recurrent neural networks. arXiv [Preprint]. arXiv:1308.0850. [Google Scholar]
- Guo D., Perc M., Liu T., Yao D. (2018). Functional importance of noise in neuronal information processing. Europhys. Lett. 124, 1–7. 10.1209/0295-5075/124/50001 [DOI] [Google Scholar]
- Hafting T., Fyhn M., Molden S., Moser M. B., Moser E. I. (2005). Microstructure of a spatial map in the entorhinal cortex. Nature 436, 801–806. 10.1038/nature03721 [DOI] [PubMed] [Google Scholar]
- Harvey R. E., Berkowitz L. E., Hamilton D. A., Clark B. J. (2019). The effects of developmental alcohol exposure on the neurobiology of spatial processing. Neurosci. Biobehav. Rev. 107, 775–794. 10.1016/j.neubiorev.2019.09.018 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hassabis D., Kumaran D., Summerfield C., Botvinick M. (2017). Neuroscience-inspired artificial intelligence. Neuron 95, 245–258. 10.1016/j.neuron.2017.06.011 [DOI] [PubMed] [Google Scholar]
- Hawkins J., Ahmad S. (2016). Why neurons have thousands of synapses, a theory of sequence memory in neocortex. Front. Neural Circ. 10:23. 10.3389/fncir.2016.00023 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hawkins J., Lewis M., Klukas M., Purdy S., Ahmad S. (2019). A framework for intelligence and cortical function based on grid cells in the neocortex. Front. Neural Circ. 12:121. 10.3389/fncir.2018.00121 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hebb D. O. (1949). The Organization of Behavior. New York, NY: Wiley. [Google Scholar]
- Hinman J. R., Chapman G. W., Hasselmo M. E. (2019). Neuronal representation of environmental boundaries in egocentric coordinates. Nat. Commun. 10:10722. 10.1038/s41467-019-10722-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- Høydal Ø. A., Skytøen E. R., Andersson S. O., Moser M. B., Moser E. I. (2019). Object-vector coding in the medial entorhinal cortex. Nature 568, 400–404. 10.1038/s41586-019-1077-7 [DOI] [PubMed] [Google Scholar]
- Johnson A., Venditto S. (2015). Reinforcement learning and hippocampal dynamics, in Analysis and Modeling of Coordinated Multi-neuronal Activity, ed M. Tatsuno (New York, NY: Springer; ), 299–312. [Google Scholar]
- Jonas E., Kording K. P. (2017). Could a neuroscientist understand a microprocessor? PLoS Comput. Biol. 13, 1–24. 10.1371/journal.pcbi.1005268 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kanitscheider I., Fiete I. (2017). Training recurrent networks to generate hypotheses about how the brain solves hard navigation problems, in Advances in Neural Information Processing Systems (NIPS), (Long Beach, CA: ), 4530–4539. [Google Scholar]
- Khamassi M., Humphries M. D. (2012). Integrating cortico-limbic-basal ganglia architectures for learning model-based and model-free navigation strategies. Front. Behav. Neurosci. 6:79. 10.3389/fnbeh.2012.00079 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Knierim J. J., Hamilton D. A. (2011). Framing spatial cognition: Neural representations of proximal and distal frames of reference and their roles in navigation. Physiol Rev. 91, 1245–1279. 10.1152/physrev.00021.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Knierim J. J., Zhang K. (2012). Attractor dynamics of spatially correlated neural activity in the limbic system. Annu. Rev. Neurosci. 35, 267–285. 10.1146/annurev-neuro-062111-150351 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Krichmar J. L., Edelman G. M. (2005). Brain-based devices for the study of nervous systems and the development of intelligent machines. Artif. Life 11, 63–77. 10.1162/1064546053278946 [DOI] [PubMed] [Google Scholar]
- Kropff E., Carmichael J. E., Moser M. B., Moser E. I. (2015). Speed cells in the medial entorhinal cortex. Nature 523, 419–424. 10.1038/nature14622 [DOI] [PubMed] [Google Scholar]
- Krupic J., Bauza M., Burton S., Barry C., O'Keefe J. (2015). Grid cell symmetry is shaped by environmental geometry. Nature 518, 232–235. 10.1038/nature14153 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kudrimoti H. S., Barnes C. A., McNaughton B. L. (1999). Reactivation of hippocampal cell assemblies: effects of behavioral state, experience, and EEG dynamics. J. Neurosci. 19, 4090–4101. 10.1523/JNEUROSCI.19-10-04090.1999 [DOI] [PMC free article] [PubMed] [Google Scholar]
- LaChance P. A., Todd T. P., Taube J. S. (2019). A sense of space in postrhinal cortex. Science 365:eaax4192. 10.1126/science.aax4192 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lee D., Seo H., Jung M. W. (2012). Neural basis of reinforcement learning and decision making. Annu. Rev. Neurosci. 35, 287–308. 10.1146/annurev-neuro-062111-150512 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lever C., Burton S., Jeewajee A., O'Keefe J., Burgess N. (2009). Boundary vector cells in the subiculum of the hippocampal formation. J. Neurosci. 29, 9771–9777. 10.1523/JNEUROSCI.1319-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lever C., Wills T., Cacucci F., Burgess N., Keefe J. O. (2002). Long-term plasticity in hippocampal place-cell representation of environmental geometry. Nature 416, 236–238. 10.1038/416090a [DOI] [PubMed] [Google Scholar]
- Lipson H., Kurman M. (2016). Driverless: Intelligent Cars and the Road Ahead. Cambridge, MA: MIT Press. [Google Scholar]
- Mao D., Kandler S., McNaughton B. L., Bonin V. (2017). Sparse orthogonal population representation of spatial context in the retrosplenial cortex. Nat. Commun. 8:243. 10.1038/s41467-017-00180-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mao D., Neumann A. R., Sun J., Bonin V., Mohajerani M. H., McNaughton B. L. (2018). Hippocampus-dependent emergence of spatial sequence coding in retrosplenial cortex. Proc. Natl. Acad. Sci. U.S.A. 115, 8015–8018. 10.1073/pnas.1803224115 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mathis A., Mamidanna P., Cury K. M., Abe T., Murthy V. N., Mathis M. W., et al. (2018). DeepLabCut: markerless pose estimation of user-defined body parts with deep learning. Nat. Neurosci. 21, 1281–1289. 10.1038/s41593-018-0209-y [DOI] [PubMed] [Google Scholar]
- McCulloch W. S., Pitts W. H. (1943). A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 5, 115–133. 10.1007/BF02478259 [DOI] [PubMed] [Google Scholar]
- McNaughton B. L., Battaglia F. P., Jensen O., Moser E. I., Moser M.-B. (2006). Path integration and the neural basis of the cognitive map. Nat. Rev. Neurosci. 7, 663–678. 10.1038/nrn1932 [DOI] [PubMed] [Google Scholar]
- McNaughton B. L., Knierim J. J., Wilson M. A. (1995). Vector encoding and the vestibular foundations of spatial cognition: neurophysiological and computational mechanisms, in The Cognitive Neurosciences, ed M. Gazzaniga (Cambridge: MIT Press; ), 585–595. [Google Scholar]
- McNaughton B. L., Mizumori S. J. Y., Barnes C. A., Leonard B. J., Marquis M., Green E. J. (1994). Cortical representation of motion during unrestrained spatial navigation in the rat. Cereb. Cortex 4, 27–39. 10.1093/cercor/4.1.27 [DOI] [PubMed] [Google Scholar]
- McNaughton L., Chen L. L., Markus J. (1991). Dead reckoning, landmark learning, and the sense of direction: a neurophysiological and computational hypothesis. J. Cogn. Neurosci. 3, 190–202. 10.1162/jocn.1991.3.2.190 [DOI] [PubMed] [Google Scholar]
- Milford M. J., Wiles J., Wyeth G. F. (2010). Solving navigational uncertainty using grid cells on robots. PLoS Comput. Biol. 6:e1000995. 10.1371/journal.pcbi.1000995 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mimica B., Dunn B. A., Tombaz T., Bojja V. P. T. N. C. S., Whitlock J. R. (2018). Efficient cortical coding of 3D posture in freely behaving rats. Science 589, 584–589. Available online at: https://science.sciencemag.org/content/362/6414/584 [DOI] [PubMed] [Google Scholar]
- Mirowski P., Pascanu R., Viola F., Soyer H., Ballard A. J., Banino A., et al. (2017). Learning to navigate in complex environments, in International Conference on Learning Representations (ICLR). (Toulon: ). [Google Scholar]
- Momennejad I., Otto A. R., Daw N. D., Norman K. A. (2018). Offline replay supports planning in human reinforcement learning. Elife 7, 1–25. 10.7554/eLife.32548 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Montavon G., Samek W., Müller K. R. (2018). Methods for interpreting and understanding deep neural networks. Digit. Signal Process A Rev. J. 73, 1–15. 10.1016/j.dsp.2017.10.011 [DOI] [Google Scholar]
- Munn R. G., Giocomo L. M. (2020). Multiple head direction signals within entorhinal cortex: origin and function. Curr. Opin. Neurobiol. 64, 32–40. 10.1016/j.conb.2020.01.015 [DOI] [PubMed] [Google Scholar]
- Munn R. G. K., Mallory C. S., Hardcastle K., Chetkovich D. M., Giocomo L. M. (2020). Entorhinal velocity signals reflect environmental geometry. Nat. Neurosci. 23, 239–251. 10.1038/s41593-019-0562-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Musall S., Urai A., Sussillo D., Churchland A. (2019). Harnessing behavioral diversity to understand circuits for cognition. Curr. Opin. Neurobiol. 58, 229–238. 10.1016/j.conb.2019.09.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Navratilova Z., Giocomo L. M., Fellous J. M., Hasselmo M. E., McNaughton B. L. (2012). Phase precession and variable spatial scaling in a periodic attractor map model of medial entorhinal grid cells with realistic after-spike dynamics. Hippocampus. 22, 772–789. 10.1002/hipo.20939 [DOI] [PubMed] [Google Scholar]
- Nitz D. A. (2006). Tracking route progression in the posterior parietal cortex. Neuron 49, 747–756. 10.1016/j.neuron.2006.01.037 [DOI] [PubMed] [Google Scholar]
- Noe A., O'Regan K. (2001). What it is like to see: a sensorimotor theory of perceptual experience. Synthese 129, 79–103. 10.1023/A:1012695023768 [DOI] [Google Scholar]
- Oess T., Krichmar J. L., Röhrbein F. (2017). A computational model for spatial navigation based on reference frames in the hippocampus, retrosplenial cortex, and posterior parietal cortex. Front. Neurorobot. 11:4. 10.3389/fnbot.2017.00004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- O'Keefe J., Burgess N. (1996). Geometric determinants of the neurons. Nature 381, 425–428. 10.1038/381425a0 [DOI] [PubMed] [Google Scholar]
- O'Keefe J., Dostrovsky J. (1971). The hippocampus as a spatial map. Preliminary evidence from unit activity in the freely-moving rat. Brain Res. 34, 171–175. 10.1016/0006-8993(71)90358-1 [DOI] [PubMed] [Google Scholar]
- O'Keefe J., Nadel L. (1978). The Hippocampus as a Cognitive Map. New York, NY: Oxford University Press. [Google Scholar]
- Pennartz C. M. A., Ito R., Verschure P. F. M. J., Battaglia F. P., Robbins T. W. (2011). The hippocampal-striatal axis in learning, prediction and goal-directed behavior. Trends Neurosci. 34, 548–559. 10.1016/j.tins.2011.08.001 [DOI] [PubMed] [Google Scholar]
- Peyrache A., Duszkiewicz A. J., Viejo G., Angeles-Duran S. (2019). Thalamocortical processing of the head-direction sense. Prog Neurobiol. 183:101693. 10.1016/j.pneurobio.2019.101693 [DOI] [PubMed] [Google Scholar]
- Peyrache A., Lacroix M. M., Petersen P. C., Buzsáki G. (2015). Internally organized mechanisms of the head direction sense. Nat. Neurosci. 18, 569–575. 10.1038/nn.3968 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pfeifer R., Scheier C. (1999). Understanding Intelligence. London: MIT Press. [Google Scholar]
- Pozzi I., Bohté S., Roelfsema P. (2019). A biologically plausible learning rule for deep learning in the brain. arXiv [Preprint]. arXiv:1811.01768. [Google Scholar]
- Richards B. A., Lillicrap T. P., Beaudoin P., Bengio Y., Bogacz R., Christensen A., et al. (2019). A deep learning framework for neuroscience. Nat. Neurosci. 22, 1761–1770. 10.1038/s41593-019-0520-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Roelfsema P. R., Holtmaat A. (2018). Control of synaptic plasticity in deep cortical networks. Nat. Rev. Neurosci. 19, 166–180. 10.1038/nrn.2018.6 [DOI] [PubMed] [Google Scholar]
- Rosenzweig E. S., Redish A. D., McNaughton B. L., Barnes C. A. (2003). Hippocampal map realignment and spatial learning. Nat. Neurosci. 6, 609–615. 10.1038/nn1053 [DOI] [PubMed] [Google Scholar]
- Rumelhart D. E., McClelland J. L., Research Group P. D. P. (1988). Parallel Distributed Processing, Vol. 1. Cambridge, MA: Bradford Books; MIT Press. [Google Scholar]
- Ryait H., Bermudez-Contreras E., Faraji J., Mirza Agha B., Gomez-Palacio Schjetnan A. J., et al. (2019). Data-driven analyses of motor impairments in animal models and neurological disorders. PLoS Biol. 17:e3000516. 10.1371/journal.pbio.3000516 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sacramento J., Bengio Y., Costa R. P., Senn W. (2018). Dendritic cortical microcircuits approximate the backpropagation algorithm, in: Advances in Neural Information Processing Systems 31. 8721–8732. Available online at: http://papers.nips.cc/paper/8089-dendritic-cortical-microcircuits-approximate-the-backpropagation-algorithm.pdf
- Samu D., Eros P., Ujfalussy B., Kiss T. (2009). Robust path integration in the entorhinal grid cell system with hippocampal feed-back. Biol. Cybern. 101, 19–34. 10.1007/s00422-009-0311-z [DOI] [PubMed] [Google Scholar]
- Santoro A., Hill F., Barrett D., Raposo D., Botvinick M., Lillicrap T. (2019). Is coding a relevant metaphor for building AI? A commentary on is coding a relevant metaphor for the brain?, by romain brette. arXiv. 1904:1–4. 10.1017/S0140525X19001365 [DOI] [PubMed] [Google Scholar]
- Schmidhuber J. (2014). Deep learning in neural networks: an overview. Neural Netw. 61, 85–117. 10.1016/j.neunet.2014.09.003 [DOI] [PubMed] [Google Scholar]
- Schultheiss N. W., Redish A. D. (2015). The compass within. Nat. Neurosci. 18, 482–483. 10.1038/nn.3977 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sharp P. E., Tinkelman A., Cho J. (2001). Angular velocity and head direction signals recorded from the dorsal tegmental nucleus of gudden in the rat: Implications for path integration in the head direction cell circuit. Behav. Neurosci. 115, 571–588. 10.1037/0735-7044.115.3.571 [DOI] [PubMed] [Google Scholar]
- Sinz F. H., Pitkow X., Reimer J., Bethge M., Tolias A. S. (2019). Engineering a less artificial intelligence. Neuron 103, 967–979. 10.1016/j.neuron.2019.08.034 [DOI] [PubMed] [Google Scholar]
- Skaggs W. E., Knierim J. J., Kudrimoti H. S., McNaughton B. L. (1994). A model of the neural basis of the rat's sense of direction, in Proceedings of the Seventh International Conference of Neural Information Processing Systems (NIPS), (Denver, CO: ), 173–180. [PubMed] [Google Scholar]
- Skaggs W. E., McNaughton B. L. (1996). Replay of neuronal firing sequences in rat hippocampus during sleep following spatial experience. Science 271, 1870–1873. 10.1126/science.271.5257.1870 [DOI] [PubMed] [Google Scholar]
- Solstad T., Boccara C. N., Kropff E., Moser M., Moser E. I. (2008). Representation of geometric borders in the enthorinal cortex. Science 322, 1865–1868. 10.1126/science.1166466 [DOI] [PubMed] [Google Scholar]
- Solstad T., Moser E. I., Einevoll G. T. (2006). From grid cells to place cells: a matematical model. Hippocampus 16, 1026–1031. 10.1002/hipo.20244 [DOI] [PubMed] [Google Scholar]
- Sorscher B., Mel G. C., Ganguli S., Ocko S. A. (2019). A unified theory for the origin of grid cells through the lens of pattern formation, in Advances in Neural Information Processing Systems (NeurIPS), (Vancouver, BC: ), 1–11. [Google Scholar]
- Srivastava N., Hinton G., Krizhevsky A., Sutskever I., Salakhutdinov R. (2014). Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15, 1929–1958. 10.5555/2627435.2670313 [DOI] [Google Scholar]
- Stachenfeld K. L., Botvinick M. M., Gershman S. J. (2017). The hippocampus as a predictive map. Nat. Neurosci. 20, 1643–1653. 10.1038/nn.4650 [DOI] [PubMed] [Google Scholar]
- Steinmetz N. A., Koch C., Harris K. D., Carandini M. (2018). Challenges and opportunities for large-scale electrophysiology with Neuropixels probes. Curr. Opin. Neurobiol. 50, 92–100. 10.1016/j.conb.2018.01.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stoianov I. P., Pennartz C. M. A., Lansink C. S., Pezzulo G. (2018). Model-based spatial navigation in the hippocampus-ventral striatum circuit: a computational analysis. PLoS Comput. Biol. 14, 1–28. 10.1371/journal.pcbi.1006316 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sutherland R. J., Hamilton D. A. (2004). Rodent spatial navigation: At the crossroads of cognition and movement. Neurosci. Biobehav. Rev. 28, 687–697. 10.1016/j.neubiorev.2004.09.012 [DOI] [PubMed] [Google Scholar]
- Sutskever I., Vinyals O., Le Q. V. (2014). Sequence to sequence learning with neural networks, in Advances in Neural Information Processing Systems (NIPS), 4, 3104–3112. Available online at: http://papers.nips.cc/paper/5346-sequence-to-sequence-learning-with-neural-networks.pdf [Google Scholar]
- Sutton R. S., Barto A. G. (2018). Reinforcement Learning. An Introduction, 2nd Edn. Cambridge: MIT Press. [Google Scholar]
- Swanson L. W. (2003). Brain Architecture. New York, NY: Oxford University Press. [Google Scholar]
- Taube J., Muller R., Ranck J., Jr. (1990). Head-direction cells recorded from the postsubiculum in freely moving rats. I. Description and quantitative analysis. J. Neurosci. 10, 420–435. 10.1523/JNEUROSCI.10-02-00420.1990 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Taube J. S. (2007). The head direction signal: origins and sensory-motor integration. Annu. Rev. Neurosci. 30, 181–207. 10.1146/annurev.neuro.29.051605.112854 [DOI] [PubMed] [Google Scholar]
- Thelen E., Smith L. B. (2007). Dynamic systems theories. Handb. Child. Psychol. 258–312. 10.1002/9780470147658.chpsy0106 [DOI] [Google Scholar]
- Tishby N., Zaslavsky N. (2015). Deep learning and the information bottleneck principle in 2015 IEEE Information Theory Workshop ITW 2015 (Jeju Island: ). [Google Scholar]
- Tolman E. C. (1948). Cognitive maps in rats and men. Psychol. Rev. 55, 189–208. 10.1037/h0061626 [DOI] [PubMed] [Google Scholar]
- Ullman S. (2019). Using neuroscience to develop artificial intelligence. Science 363, 692–693. 10.1126/science.aau6595 [DOI] [PubMed] [Google Scholar]
- Vickerstaff R. J., Cheung A. (2010). Which coordinate system for modelling path integration? J. Theor. Biol. 263, 242–261. 10.1016/j.jtbi.2009.11.021 [DOI] [PubMed] [Google Scholar]
- Vogt N. (2018). Machine learning in neuroscience. Nat. Methods 15:33 10.1038/nmeth.4549 [DOI] [Google Scholar]
- Vu M.-A. T., Adali T., Ba D., Buzsáki G., Carlson D., Heller K., et al. (2018). A shared vision for machine learning in neuroscience. J. Neurosci. 38, 1601–1607. 10.1523/JNEUROSCI.0508-17.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Walker E. Y., Sinz F. H., Cobos E., Muhammad T., Froudarakis E., Fahey P. G., et al. (2019). Inception loops discover what excites neurons most using deep predictive models. Nat. Neurosci. 22, 2060–2065. 10.1038/s41593-019-0517-x [DOI] [PubMed] [Google Scholar]
- Wang C., Chen X., Knierim J. J. (2020). Egocentric and allocentric representations of space in the rodent brain. Curr. Opin. Neurobiol. 60, 12–20. 10.1016/j.conb.2019.11.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang C., Chen X., Lee H., Deshmukh S. S., Yoganarasimha D., Savelli F., et al. (2018). Egocentric coding of external items in the lateral entorhinal cortex. Science 362, 945–949. 10.1126/science.aau4940 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Webb B., Wystrach A. (2016). Neural mechanisms of insect navigation. Curr. Opin. Insect. Sci. 15, 27–39. 10.1016/j.cois.2016.02.011 [DOI] [PubMed] [Google Scholar]
- Whishaw I. Q., Hines D. J., Wallace D. G. (2001). Dead reckoning (path integration) requires the hippocampal formation: evidence from spontaneous exploration and spatial learning tasks in light (allothetic) and dark (idiothetic) tests. Behav. Brain Res. 127, 49–69. 10.1016/S0166-43280100359-X [DOI] [PubMed] [Google Scholar]
- Whitlock J. R., Pfuhl G., Dagslott N., Moser M. B., Moser E. I. (2012). Functional split between parietal and entorhinal cortices in the rat. Neuron 73, 789–802. 10.1016/j.neuron.2011.12.028 [DOI] [PubMed] [Google Scholar]
- Whittington J. C. R., Muller T. H., Barry C., Mark S., Behrens T. E. J. (2018). Generalisation of structural knowledge in the hippocampal-entorhinal system, in Proceedings of the International Conference of Neural Information Processing Systems (NeurIPS), 8484–8495. [Google Scholar]
- Wilber A. A., Clark B. J., Forster T. C., Tatsuno M., McNaughton B. L. (2014). Interaction of egocentric and world-centered reference frames in the rat posterior parietal cortex. J. Neurosci. 34, 5431–5446. 10.1523/JNEUROSCI.0511-14.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wilber A. A., Skelin I., Wu W., McNaughton B. L. (2017). Laminar organization of encoding and memory reactivation in the parietal cortex. Neuron 95, 1406–1419.e5. 10.1016/j.neuron.2017.08.033 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wu S., Zhang Y., Cui Y., Li H., Wang J., Guo L., et al. (2019). Heterogeneity of synaptic input connectivity regulates spike-based neuronal avalanches. Neural Netw. 110, 91–103. 10.1016/j.neunet.2018.10.017 [DOI] [PubMed] [Google Scholar]
- Xu L., Feng C., Kamat V. R., Menassa C. C. (2019). An occupancy grid mapping enhanced visual SLAM for real-time locating applications in indoor GPS-denied environments. Autom. Constr. 104, 230–245. 10.1016/j.autcon.2019.04.011 [DOI] [Google Scholar]
- Xu Z., Wu W., Winter S. S., Mehlman M. L., Butler W. N., Simmons C. M., et al. (2019). A comparison of neural decoding methods and population coding across thalamo-cortical head direction cells. Front. Neural Circ. 13:75. 10.3389/fncir.2019.00075 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yamauchi B., Beer R. (1996). Spatial learning for navigation in dynamic environments. IEEE Trans. Syst. Man Cybern. Part B Cybern. 26, 496–505. 10.1109/3477.499799 [DOI] [PubMed] [Google Scholar]
- Yoder R. M., Clark B. J., Taube J. S. (2011). Origins of landmark encoding in the brain. Trends Neurosci. 34, 561–571. 10.1016/j.tins.2011.08.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zador A. M. (2019). A critique of pure learning: what artificial neural networks can learn from animal brains. Nat. Commun. 10:3770. 10.1038/s41467-019-11786-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zafar M. N., Mohanta J. C. (2018). Methodology for path planning and optimization of mobile robots: a review. Procedia Comput. Sci. 133, 141–152. 10.1016/j.procs.2018.07.018 [DOI] [Google Scholar]