Abstract
Environmental information is required to stabilize estimates of head direction (HD) based on angular path integration. However, it is unclear how this happens in real-world (visually complex) environments. We present a computational model of how visual feedback can stabilize HD information in environments that contain multiple cues of varying stability and directional specificity. We show how combinations of feature-specific visual inputs can generate a stable unimodal landmark bearing signal, even in the presence of multiple cues and ambiguous directional specificity. This signal is associated with the retrosplenial HD signal (inherited from thalamic HD cells) and conveys feedback to the subcortical HD circuitry. The model predicts neurons with a unimodal encoding of the egocentric orientation of the array of landmarks, rather than any one particular landmark. The relationship between these abstract landmark bearing neurons and head direction cells is reminiscent of the relationship between place cells and grid cells. Their unimodal encoding is formed from visual inputs via a modified version of Oja’s Subspace Algorithm. The rule allows the landmark bearing signal to disconnect from directionally unstable or ephemeral cues, incorporate newly added stable cues, support orientation across many different environments (high memory capacity), and is consistent with recent empirical findings on bidirectional HD firing reported in the retrosplenial cortex. Our account of visual feedback for HD stabilization provides a novel perspective on neural mechanisms of spatial navigation within richer sensory environments, and makes experimentally testable predictions.
Author summary
Animals need a ‘compass’ to maintain a sense of orientation relative to their environment in order to support spatial navigation and memory. This ‘compass’ has been identified with head direction cells, found in various species. These cells fire at a high rate when an animals’ head is facing in a specific allocentric direction regardless of the current location. Their firing needs to be stabilized by sensory inputs to prevent drift. Since the discovery of head direction cells most experiments and computational models have employed visual input in the form of a single polarizing cue. It is unknown how the global sense of direction can be maintained in more complex environments with multiple cues of differing stability, salience, and directional specificity. Our model suggests that (contrary to models with simple cues) a new type of neuron (abstract landmark bearing cells) endows the head direction system with powerful abilities, including unimodal landmark encoding at a sensory level despite partially conflicting cues, robustness against unreliable and ephemeral cues, and a high encoding capacity across environments. The model is consistent with numerous empirical findings, and provides a novel perspective on the neural mechanisms of spatial navigation in more realistic, cue-rich settings across multiple environments.
Introduction
The global sense of direction is believed to be supported by head direction (HD) cells. First discovered in rodents, a HD neuron only fires at a high and steady rate when an animal’s head is facing in a specific allocentric direction regardless of the current location [1,2]. HD signals have been found in various brain areas in several organisms [3,4], and data consistent with signatures of HD coding have been found in humans [5–8].
HD cells are found along the diencephalic-cingulate-parahippocampal Papez’ circuit [9], including the retrosplenial cortex (RSC) [10], dorsal tegmental nucleus of Gudden [11], lateral mammillary nucleus [12,13], anterodorsal thalamic nucleus [14], dorsal presubiculum [1], and medial entorhinal cortex [15]. HD signals were also found in other cortical regions, including the medial parietal cortex in humans [5,16], as well as medial prestriate cortex (or visual association cortex) [17] and deep layers of the primary visual cortex (V1) [18].
HD cells are controlled by both external landmarks and self-motion information [19–21]. The self-motion information includes angular path integration in the vestibular system, which generates unimodal HD signals even before eye-opening [3,22,23]. Path integration via vestibular inputs can maintain HD in darkness, yet is subject to accumulated path integration error resulting in drift in the absence of sensory reset [24].
Anatomical considerations suggest that distinct sub-regions in RSC, namely dysgranular RSC (dRSC) and granular RSC (gRSC), map onto different aspects of information processing, in which dRSC shares more reciprocal connections with visual regions while gRSC is more densely connected with subcortical regions of the HD system [25,26]. HD cells with both directional and direction-by-location firing characteristics have been reported in RSC [10], which has been proposed to be involved in learning external landmark information and location-specific feedback [27], transmitting cortical visual information to subcortical HD circuits [17]. Consistent with this, lesions to dRSC do not disrupt the generation of HD signals but destabilize HD signals across time [3].
More recently, Jacob et al. [28] report so-called bidirectional (BD) HD cells in dRSC. These cells exhibit a bimodal activity profile, firing for two directions across two connected environments that contain identical visual landmarks in opposite directions (180° mirrored environments). See Table 1 for all abbreviations used in this paper. HD signals in gRSC and dorsal presubiculum remain unimodal across such conflicting (at the sensory level) environments [28]. This study corroborates the notion that dRSC acts as an integrator of self-motion and visual information, thus providing integrated signals for correcting accumulated path integration error [3,22,23].
Table 1. The ascending sorted acronyms.
Acronyms | Descriptions |
---|---|
aLB | Abstract landmark-bearing |
BC | Between-compartment |
BD | Bidirectional |
dRSC | Dysgranular retrosplenial cortex |
gRSC | Granular retrosplenial cortex |
HD | Head direction / head-directional |
IoU | Intersection over Union |
mOSA | Modified Oja’s Subspace Algorithm |
OSA | Oja’s Subspace Algorithm |
RSC | Retrosplenial cortex |
V1 | Primary visual cortex |
Vis | Visual (or V1) |
WC | Within-compartment |
Acronyms that only occur in figures are not shown in this table, as they are explained in all related captions.
However, several crucial questions regarding the integration of visual information with HD signals remain unanswered. How does the HD system extract useful information from complex visual inputs to enable HD cells to maintain a unimodal representation across the whole environment? Almost all HD modelling studies simply assume that an isolated Gaussian activity packet, represented in visual areas, codes for a single polarizing cue in the environment, e.g. a stereotypical white cue card on the wall of a recording arena. However, although useful for stabilizing HD, it is unclear how such a unimodal Gaussian representation is derived from more complex environments with multiple cues of differing stability, salience, and directional specificity. Considering that real animals likely visit multiple such complex environments also raises the question of capacity: If visual feedback connections allow HD to be reset upon re-entry to a familiar environment, what might the capacity of the system be, and how does the HD system deal with unreliable or ephemeral visual information in rich real-world environments? Here the term ‘capacity’ refers to the number of distinct environments that can be stored and recalled (i.e. re-instating the appropriate HD) in the future. The present model investigates these questions in order to identify new properties of the HD-visual circuitry, including retrosplenial cortex (RSC).
Computational models have only recently begun to explore the integration of visual landmark information with HD circuits in detail [27], though mostly restricted to simple cue representations and a direct V1-RSC connection (e.g. [29–32]). Here we propose a computational model of visual landmark processing in cortical regions, capable of extracting a unimodal orientation signal under panoramic visual landmarks for supporting the HD system in subcortical regions across multiple complex environments. This is achieved via cells downstream of primary visual areas that integrate visual information via a modified version of Oja’s learning rule with negative feedback control on input layers of visual cells [33–35]. The model can maintain unimodal landmark encoding at a sensory level despite conflicting visual information; is robust against unreliable and ephemeral cues; and exhibits a high encoding capacity across multiple environments. We also suggest an experimentally practical scheme to test for novel model components, and discuss the utility of such a unimodal encoding for HD retrieval across multiple environments. The resulting account is coherent with numerous empirical findings, and provides a novel perspective on the neural mechanisms of spatial navigation in more realistic, cue-rich settings (with cues of varying directional specificity) across multiple environments.
Materials and methods
Model overview
Fig 1A depicts a schematic of the model, containing two pathways, processing visual landmarks (purple) and HD self-motion cue (orange) respectively. The two pathways meet in dRSC (pink) where the integrated visual/self-motion signal is generated and projected back to the HD attractor in order to correct for drift. The subcortical pathway (orange) transmits and maintains global HD. For simplicity, we employ a single ring attractor to generate HD information regulated by HD angular velocity from the vestibular system [36]. Other models have explored more biologically plausible architectures with spiking neural networks (e.g. [27,31,32,37,38]). For the purpose of modelling, the attractor is assumed to substitute not only the entire generative circuitry for HD signals (i.e. dorsal tegmental nucleus of Gudden and lateral mammillary nucleus), but also anterodorsal thalamic nucleus (being the principal source of HD information entering gRSC), and presubiculum being the main target for feedback from dRSC to the attractor (i.e. pink arrow from dRSC to the attractor).
For the cortical pathway (purple), we employ a new algorithm (see below) to learn a sparse, unimodal representation of a set of visual landmarks in complex scenery containing multiple cues/features with differing directional specificity. Fig 1B provides an illustration of feature-specific visual receptive fields of the visual inputs in the model. Red, blue, and green colors stand for different sensory attributes/features (e.g. colors, texture, contrast etc.) encoded by corresponding arrays of neurons with egocentric directional receptive fields (the top row). Conversely, cues with the same ‘color’ stand for indistinguishable sensory features encoded by the same array of receptive fields. For instance, the firing rates of visual cells for the ‘red’ feature exhibit a bimodal distribution of activity when two undistinguishable ‘red’ cues are located in opposing directions. That is, solely on the basis of the ‘red’ cue the agent would not be able to distinguish two orientations 180° apart. Visual cells for the ‘green’ feature are silent in the example due to the absence of such a feature in the current scene.
Downstream of visual neurons the learning algorithm (see below) generates a representation of what we term ‘abstract landmark-bearing’ (aLB) cells. Receiving afferent projections from visual model components, these cells sparsely encode an abstract directional representation, i.e. one that does not correspond to the bearing of any one particular cue but rather a global orientation of the entire egocentric scenery (i.e. a conjunction of cues; Fig 1D). For simplicity we treat all cues as distal so that parallax effects are absent. However, the model is in principle compatible with the proposed role of RSC in location-specific feedback by expanding the dRSC population to include place modulated HD cells [27].
aLB cells anchor more complex arrays of sensory inputs, to HD signals, as compared to single isolated cues. Their downstream targets are dRSC cells, suggested to integrate landmark information with global HD signals received from gRSC (which receives global HD signals from the ring attractor). The model leaves open where aLB cells reside, e.g. in retrosplenial cortex or elsewhere (see Discussion). Key model mechanisms are summarized below. More detailed information is available in S1-S4 Appendices. All hyperparameters chosen for simulations described in the results are given in S1 Table.
Neural dynamics
We employ a firing-rate based neuron model with firing rates normalized to range from 0 to 1. The firing rate f(t) is derived from the activation level a(t) of a single neuron via a modified hyperbolic tangent function. Unless specified otherwise,
(1) |
if a(t)≥α, and f(a(t)) = 0 otherwise. Here α and β are parameters of the activation function. The activation level vector a(t) (or a) of the output layer is given by standard neural rate decay dynamics,
(2) |
where τ is the membrane potential time constant controlling the speed of natural decay of neural activation, gj is a gain constant controlling the amplitude of input from the jth input layer, nI is the total number of input layers (or brain areas), f(j) is the firing-rate vector of the jth input layer, and W(j) is the weight matrix of the connection from the jth input layer to the output layer. All elements in W(j) are assumed to be non-negative (except the inhibitory self-connection in the HD attractor) [36]. Therefore, gj>0 indicates excitatory projections while gj<0 indicates inhibitory projections.
Non-plastic connections in the model include one-to-one feed-forward projections (e.g. between the attractor and gRSC) and imply the same number of neurons N both in the input and output layer. Inhibitory connections can be of two types: global inhibition to all neurons within a layer
(3) |
and lateral inhibition originating from one neuron and equivalently affecting all other neurons in the same layer (e.g. as in Fig 1D, bottom, among aLB cells), i.e.
(4) |
where I(N) is an N×N identity matrix and J(N) is an N×N all-ones matrix. These connections are kept stable at all times, i.e. they are not subject to learning. See S1 Appendix for details and S2 Table for a summary of all neural connections used in this paper.
The HD attractor ring
For simplicity we employ a single HD ring attractor (see S2 Appendix) [36], updated by angular velocity and consisting of NHD = 360 HD neurons with equally spaced preferred HDs. Symmetric connections within the attractor (not subject to learning) sustain a stable Gaussian firing pattern without external inputs. During head turning (Fig 1C), asymmetric self-connections within the attractor, scaled by the magnitude of angular velocity, update the firing pattern of HD cells. That is, path integration is performed.
The subcortical pathway
The HD attractor provides unimodal HD signals to gRSC, which are in turn projected further to dRSC for the integration with aLB signals (purple pathway in Fig 1A). Similar to the HD attractor, both gRSC and dRSC consist of neurons with different preferred directions, inherited from the upstream HD attractor, and each contains the same number neurons as the HD attractor.
For the connections among these structures (see S2 Table for a summary), the HD attractor and gRSC are connected by 1-to-1 connections, containing the same classic (unimodal) HD cells. gRSC and dRSC are fully connected (with initially random connectivity), subject to Hebbian plasticity (S1 Appendix; also in Fig 1A, marked as ‘Hebbian’). This allows the global HD signal to be associated with environment-specific aLB signals conveying visual information for HD calibration. Without plasticity in this connection, dRSC could only receive unimodal HD signals from gRSC, contradicting the existence of bimodal firing patterns (BD cells) observed in dRSC during darkness (when sensory inputs and hence aLB signals would be absent) after the rat experienced two environments with conflicting visual cues [28]. We also assume global self-inhibitory connections (Eq 3) among dRSC and gRSC in order to restrict the retrosplenial activation from spilling over.
Feature-specific visual signals
Visual inputs from a specific scenery are separately provided through different feature-based channels (see Fig 1B for an illustration). That is, each channel selectively codes visual information based on its own preferred feature. The term ‘feature’ stands for perceptual attributes (e.g. colors, edges, contrast, luminance, etc.), possibly mediated via known anatomical projections from visual areas 17 and 18 to RSC [39,40]. Each feature-specific channel contains neurons with their own egocentric tuning directions θ. The egocentric visual space of the agent ranges from -180° to 180° (‘ahead’ being 0°; Fig 1B), and changes with the direction of the agent.
We compose our sceneries of three basic types of visual signals. The first type is a scaled von Mises distribution with a single peak (a circular Gaussian):
(5) |
where κ is the encoding precision and φ is the preferred direction with maximum firing rate fmax. The second type (e.g. the ‘blue’ cue in Fig 1B) extends fvI by enlarging the range of high firing rate, yielding a unimodal signal with less directional specificity, calculated as the cumulative distribution function of a von Mises distribution (i.e. summing multiple offset circular Gaussians with encoding precision κ1 and a range of parameters φ). The third type (e.g. ‘red’ cues in Fig 1B with encoding precision κ1) consists of two von Mises distributions with different peaks, yielding a multimodal firing pattern (e.g. the ‘red’ cue in Fig 1B). See S3 Appendix for details and Fig 2A for exact firing curves.
A key premise of the model is that visual signals for every feature (plus low amplitude background noise), if not zero for all directions, are scaled to the same mean, yielding fluctuation of firing rates about the overall mean stimulus intensity. This property is inspired by neural adaptation in sensory systems [41], hypothesized to occur upstream of the brain areas that comprise the model, and allows for a distinction of salient vs non-salient cues and associated novelty detection (see Results). To avoid parallax effects [27] and focus on properties of cue rich environments, all cues are treated as distal cues (i.e. egocentric bearing and HD are separated only by a constant, arbitrary offset).
Abstract landmark bearing and the mOSA learning algorithm
We propose that abstract landmark bearing (aLB) cells sparsely encode the conjunction (i.e. the co-occurrence in space and time, or a ‘local view’) of all feature-specific visual inputs, driven by afferent projections from early visual areas. That is, a single aLB cell encodes the egocentric conjunction of all cues at a given orientation rather than encoding single cues separately. A population of aLB cells that is active in a given environment encodes all possible local views with a finite angular resolution, given by the total number of aLB cells active in that environment.
The firing of aLB cells is determined by their inputs from visual cells, via connections whose weights are adjusted using a version of Oja’s subspace algorithm (OSA) [34,35]. The original OSA can extract principal subspaces as components from time-series information within a neural network architecture, compared to standard Hebbian learning an additional feedback term (for weight decrements) is present [34,42] (see below for details). Here the OSA acts between visual cells and alB cells. Compared to the original OSA we introduce two modifications: 1. Lateral inhibition among aLB cells (Eq 6); and 2. Enforcement of non-negative weights from visual cells (Eq 7). We first describe each modification and then outline the rationale behind each and how the interplay of those changes yields the desired properties of aLB cells.
1. Lateral inhibition: aLB cells are subject to lateral inhibition among each other, allowing for winner-takes-all dynamics [43,44], similar to sensory pattern discrimination [45,46]. The decay dynamics of aLB cells can be written as
(6) |
Here, WaLB is the matrix of constant lateral inhibition weights among aLB cells (see Eq 4), the subscript ‘V2aLB’ refers to the parameters for the connection from visual areas to aLB cells, and ‘Visual’ refers to the parameters for the related visual layer. For instance, gaLB is the gain factor controlling the amplitude of lateral inhibition weights among aLB cells, is the connection weight matrix between aLB cells and visual cells within the jth visual feature, and is the firing-rate vector of visual cells for the jth visual feature channel (red, green, blue in Fig 1). NF is the total number of visual features. All parameter values are summarized in S1 Table.
The original OSA does not assume lateral inhibition among postsynaptic neurons [34]. The main effect of lateral inhibition is to enforce winner-take-all dynamics among aLB cells and a yields sparse encoding of a given egocentric view of landmarks that is consistent with a small range of HDs (cf. Fig 1D). The sparse recruitment of different sets of aLB cells in different environments ensures high capacity (see Results).
2. Non-negative weights in the Oja’s Subspace Algorithm: The synaptic weights in Eq 6 are updated via the modified version of Oja’s Subspace Algorithm (mOSA) shown in Eq 7. The original OSA ensures global convergence without explicit weight normalization [34,35,42]. However, the original algorithm does not stipulate the sign of the synaptic weights. To ensure positive feedforward connectivity, [∙]+ maps negative elements of the weight matrix to zero, since biological synapses do not change sign.
(7) |
Here ηV2aLB refers to the learning rate for all visual features. Note that the convergence of weights during learning is not affected by the [∙]+ operator (see [47] for a theoretical interpretation of non-negative OSA).
The feedback term in Eq 7 contains the transpose of the incoming weights multiplied by the aLB firing rates, corresponding to the expected feature-specific visual pattern that should yield the current aLB firing pattern. That is, it drives the plastic connections to be updated in ratio with the difference between real and expected visual inputs (i.e. controls the learning rate of both the feedback and feedforward term and effectively introduces a threshold for potentiation, further enforcing sparseness (in addition to the initial effect of lateral inhibition). The feedback term also enables the disconnection (depression of weights) of activated aLB cells from inactive visual cells (when an expected feature is not present, i.e. when in Eq 7). This renders the aLB population highly flexible, allowing it to update its encoding of the visual scene when changes to the visual world are introduced. Moreover, the feedback term supports the desired effect of the lateral inhibition, as it allows the winner-take-all dynamics among aLB cells to unfold before multiple aLB cells develop strong weights for the same local view. This is particularly useful if an ambiguous cue (such as the red feature in Fig 1D) provides residual input to an aLB cell that already codes for a different local view.
In summary, lateral inhibition and the feedback term in mOSA support sparseness and responsiveness to changes in visual inputs in the aLB population, while the restriction to non-negative weights respects the fact that synapses should not change sign. Note that mOSA (like the original OSA) is a non-local algorithm, but could be equivalently realized by a local algorithm in terms of extra interneurons encoding feedback signals (e.g. [48]).
Note that Eq 7 can be rewritten in the following form:
(7B) |
Introducing the prefactor ζ (>1) in front of the negative feedback term can be compensated for by scaling down the learning rate ηV2aLB and scaling up the gain factor gV2aLB in Eq 6 to yield the same outcome for all simulations reported in the Results section. However, the interpretation (and predictions) for what happens at the level of biological implementation changes. The prefactor ζ suggests that the weight decay is modulated by a feedback connection from aLB cells to visual areas. Conversely, down-regulating the scaling factor gV2aLB (compared to Eq 7B) potentially corresponds to a neuro-modulatory effect on the connection converging onto aLB cells from visual areas.
It should be stressed that other learning algorithms fail to yield appropriate feedback connection for HD in the complex environments considered here (including ambiguous cues). That is, the use of complex environments with multiple cues (some of which are multimodal or ephemeral, e.g. the red cue in Fig 1B) allows for model properties that are not evident in simpler setups. In other words, the failure of other learning algorithms to produce feedback signals appropriate to reset HD only becomes apparent by considering environments with cue configurations more akin to real-world environments (e.g. containing repeating or highly similar visual patterns). We compared mOSA to results obtained with other neural learning algorithms acting between visual and aLB layers. Neither Hebbian learning with lateral inhibition among aLB cells, nor Hebbian covariance learning [49,50], Intrator’s BCM algorithm [51], or the original Oja Subspace Algorithm (i.e. OSA with no self-inhibition) [34] could generate unimodal aLB representations (and thus appropriate HD feedback) in the environment depicted in Fig 1B (see S4 Appendix for details and S1 Fig for simulation results of these alternative algorithms).
The cortical pathway and dRSC
Receiving egocentric visual input from visual layers, aLB cells encode conjunctions of egocentric bearings on multiple features/cues. This signal is integrated with HD in dRSC, anchoring the global HD signals and egocentric visual signals to dRSC via Hebbian learning between aLB cells and dRSC as well as between gRSC and dRSC (see S2 Table for a summary of all neural connections used here), i.e.
(8) |
where the term ‘Input’ refers to either aLB or gRSC, with W and η refers to the corresponding connection matrix and learning rate. During the test phase, we set η = 0 to fix synaptic connections. Weights subject to Hebbian learning are normalised to ensure their stability with this learning rule. More specifically,
(9) |
where ‖∙‖2 is the two-norm for calculating the total connection, wmax is the maximum total connection strength, and wl is the lth row vector (i.e., synaptic connections targeting the lth output neuron) of the weight matrix W.
To stabilize the subcortical HD attractor against drift, a projection with both 1-to-1 excitation and global inhibition from dRSC transmits the integrated signal back to the HD attractor (S2 Appendix). Choosing dRSC as the locus of integration is coherent with experimental results, showing that landmark-dominant HD signals (i.e. BD cells) were only discovered in the dRSC, instead of in the subcortical pathway (i.e. not in gRSC, anterodorsal thalamic nucleus, or dorsal presubiculum) [28].
Visualization of HD and aLB representations
The representation of internal HD signals is naturally defined as the firing-rate weighted mean of preferred directions of HD neurons, i.e. the population vector, both in the attractor and gRSC. More specifically,
(10) |
as fHD is the firing rate vector of the HD attractor, θ = {θl = −180+(l−1)Δθ|1≤l≤NHD} is the set of preferred directions, and ∙ is the dot product. gRSC cells share the same representation.
To visualize the global representation of activated aLB cells, their firing rates during the testing phase are binned according to the real HD where they show maximum firing rates above a threshold εaLB = 0.5. Binning according to true HD reveals unimodality or multimodality of aLB firing patterns, even though aLB cells, unlike HD cell, are not presumed to have a certain preferred direction in advance.
We use the Jaccard similarity coefficient [52], or ‘Intersection over Union’ (IoU), to measure the similarity of neural firing patterns (over aLB cells) between different testing environments. The IoU is defined as the proportion of cells that is highly active in both environments (the overlap between two set of active cells) relative to the number of cells that is highly active in only one of the two environments being compared (i.e., the union of cells exclusive to one environments).
(11) |
where is the set of activated aLB cells for the ith tested environment and is the firing rate of the lth aLB cell.
Simulations
Every simulation consists of the learning phase and testing phase. Learning with neural plasticity turned on is based on a 20-minute HD trajectory from rat recordings, with an average angular velocity of 72.88 deg/s (Fig 1C and S1 Dataset). In the testing phase, all synaptic changes are frozen and a slow uniform rotation with 60 deg/s for 60 seconds is simulated, starting from the final HD of the learning phase. This way each HD is sampled 10 times and a balanced estimate of the learned representation across all directions is sampled for visualization. All simulations are conducted via MATLAB R2017a.
Results
We first examine the capability of aLB cells to form a landmark representation from unstable cues within an environment and the capacity of aLB cells to learn distinct sets of landmark representations across multiple environments. We then show how visual/sensory feedback may be conveyed from aLB cells to HD cells via RSC, and how the model is coherent with recent empirical findings. Finally, we consider model-specific predictions.
aLB cells with unimodal encoding in complex sceneries
We first show how aLB cells learn a sparse and unimodal representation over a complex scenery. Visual projections to RSC arise in early visual areas [26], and we do not posit upstream curation of landmark inputs. Instead, the learning algorithm must generate an egocentric sensory representation suitable to convey feedback to HD cells. We set up two visual landmarks in a single environment (Fig 2A, same as Fig 1B): an unambiguous, broad ‘blue’ cue located due East with low encoding precision, and conflicting, narrow ‘red’ cues located due North and due South. The simulation is based on the HD trajectory shown in Fig 1C. The synaptic weights (Fig 2B) between visual areas and aLB cells quickly converge (Fig 2C), and the unimodal representation of aLB cells stably emerges at an early stage of learning (S2 Fig). Moreover, nearly all aLB cells are active for only one egocentric bearing (Fig 2D, right). aLB also exhibit high encoding precision for each preferred direction (Fig 2D). This aspect is extracted from visual signals with high encoding precision and narrow firing range (‘red’ cues in Fig 2A), and shows that aLB cells integrate useful properties from multiple cues (i.e. high encoding precision from the two narrow cues and unidirectionality from the isolated wide cue). Overall, these results suggest that the algorithm successfully achieves a suitable unimodal aLB cell representation when exposed to scenery with multiple and ambiguous features.
Robustness against unreliable information
aLB cells can maintain unimodal encoding of a complex scene in the presence of unreliable visual information without any extra top-down control (e.g. visual attention). Unreliable sensory inputs in our simulations include landmarks moving with a constant speed, cues that vanish and reappear somewhere else (i.e. teleported cues), or ephemeral cues that only temporarily exist.
First, for the case of moving cues, we set up the same single environment as Fig 2A but with the unambiguous ‘blue’ cue moving along the cylinder at an observable speed of 90 deg/s anticlockwise (Fig 3A, left). No unimodal aLB representation develops in this case when tested in the same environment with the moving ‘blue’ cue (Fig 3A, middle). However, freezing the developing weights and stabilizing the moving cue yields an approximately unimodal aLB representation (Fig 3A, right), suggesting that the learning algorithm tries to maintain a set of weights coherent with the most recently experienced scene containing both features.
Second, for the case of teleporting cues, we set up the same environment but with the unambiguous landmark periodically teleporting to a random position on the cylinder (every 10 seconds; Fig 3B, left). Similar to the case of moving cues, we find an approximately unimodal aLB representation once the developing weights are frozen and the teleporting landmark is stabilized, without additional learning once the landmark stops teleporting (Fig 3B, right). Together these results suggest that the mOSA algorithm can continuously update the weights and unlearn weights from unstable cues (locations in the field of view occupied prior to the most recent teleportation event) and extract weights from stable cues. Thus the model exhibits the preference for encoding stable cues, consistent with increased fMRI signals in RSC for more stable landmarks [53].
Note, the fact that the same set of active aLB cells is maintained throughout the above simulations depends on the properties of the visual inputs. Given the relatively low encoding precision of the ‘blue’ cue (and hence lower maximum firing rate given its width, see Materials and Methods) when the unstable ‘blue’ cue moves, the remaining stable ‘red’ cue can enforce the persistence of the same set of aLB cells through all cue-configurations without recruiting new aLB cells. This same set of cells simply develops incoming weights from visual cells coding for the updated position of the blue cue, while unlearning (see the learning rule of mOSA, Eq 7) weights from visual cells coding for the previous position of the blue cue. That is, when the blue cue moves, the drive via the initial random connections from the ‘blue’ visual cells covering the new position is insufficient to override the currently active set of aLB cells receiving input from the stable ‘red’ visual neurons. Naturally, adding more stable cues will make the current set of active aLB cells even more resistant to change (as might be expected in cue-rich natural environments). On the other hand, adding a novel cue with saliency comparable or higher than existing cues will lead to the recruitment of a new set of aLB cells (see the next subsection).
Incorporating novel cues
Here we consider the edge case where a newly added cue is of similar salience as the previously present array of cues. We set up an environment with the stable, conflicting ‘red’ cues and the unambiguous, broad ‘blue’ cue, similar to previous simulations. However, a novel and unambiguous ‘green’ cue (with encoding precision κ2) appears due West after 400 seconds for a similar duration (Fig 3C). In this way, the agent is first confronted with the ‘red-blue’ scenery for 400 seconds, then the ‘red-blue-green’ scenery for another 400 seconds, and then again with the ‘red-blue’ scenery for the next 400 seconds.
After 1200 seconds, we tested the model on both the ‘red-blue’ scenery (Fig 3D, top) and the ‘red-blue-green’ scenery (Fig 3D, middle). Unimodal representations of abstract landmark bearing are successfully expressed in both cases, however with different sets of active aLB cells (with little to no overlap, S4 Fig). Contrary to the previous section, the inputs from the ‘red-blue’ scenery cannot enforce the persistence of the same aLB cells, because once the high-saliency green cue appears, it creates a strong drive to random aLB cells via its initial random connections, silencing other aLB cells via lateral inhibition in the aLB layer. This leads to the recruitment of a new set of aLB cells (S3 and S4 Figs). This may be viewed as the model treating the new ‘red-blue-green’ scenery as a novel environment, and this novelty detection is dependent on the cues being normalized to equal total firing.
Moreover, aLB cells also express a unimodal representation when tested on the ‘green’ cue alone (with the ‘red’ and ‘blue’ cues removed, Fig 3D, bottom), which has not been explicitly learned on its own before. This shows that the aLB representation is robust to cue removal on one hand, and that the model can incorporate a novel visual feature (even in the presence of already learned cues) into its landmark bearing representation on the other. Hence blocking by previously learned features is not an issue for the model, which is similar to place cells in the hippocampus [54].
Also note, the ability to incorporate a newly added cue is distinct from the question of whether or not a new set of aLB cells is recruited. If the newly added cue is of low saliency (i.e. of low directional specificity, with low peak firing rates among feature-specific visual cells), no new aLB cells are recruited (compared to the previously active ensemble) but the cue nonetheless contributes to driving previously active aLB cells following continued learning.
High capacity of scene-based encoding
We suggest that the present model approximates the learning of HD feedback signals from sensory inputs in environments that are richer than stereotypical simulation environments with a single Gaussian cue, and hence more indicative of the mechanisms at play in rich real-world scenes. If this is the case, the model should be able to cope with a large number of environments concurrently. That is, the learning rule should ensure a large capacity for distinct environments, using a limited number of neurons overall, and recruit new sets of aLB cells in new environments to provide unique feedback connections to HD cells without interference from or forgetting of previously learned feedback for a different environment.
To test the capacity of the model, we set up 10 environments that the agent experiences sequentially (Fig 4A). All environments share the same stable unambiguous narrow ‘red’ cue and stable unambiguous broad ‘blue’ cue. The distinguishing cue between all 10 environments is an additional unambiguous narrow ‘green’ cue, occupying a different position in each environment (starting due West in the first environment) and shifted by 36° anticlockwise in each successive environment. During the learning phase, the agent stays in each environment for 120 seconds before moving to the next environment. In these simulations continuity between the environments is assumed. That is, the HD at the last moment in environment 1 is the starting HD in environment 2, as if the agent went from one environment into an adjacent one.
To quantify the similarity (or lack thereof) among recruited sets of aLB cells across different environments, we test the model (in each environment) with snapshots of the learned synaptic connections between visual cells and aLB cells. Testing the model generates a set of active aLB cells for each synaptic snapshot (Figs 4B and S5). Similarity between the active sets of aLB cells across environments is quantified via IoU similarity maps [52].
The first comparison is based on snapshots of synaptic weights taken after the first exposure to each environment (‘intermediate vs. intermediate’ in Fig 4C), i.e. we quantify the similarity between the recruited aLB population with one set of weights (e.g. immediately after learning environment I) and the recruited population of aLB cells with another set of weights (e.g. immediately after learning environment II).
The second comparison is based on the sets of recruited aLB cells with synaptic weights taken after learning has ended (‘final vs. final’ in Fig 4C), i.e. once all 10 environments have been experienced. In both cases, the similarity between sets of recruited aLB cells is maximal (i.e. 1) when tested for the same environment and drops to near-zero when tested for all distinct environments, even with the smallest separation of the green cue between ‘adjacent’ (in time/cue separation) environments, suggesting aLB cells providing distinguishable signals for multiple environments even when these environments share some cues.
To quantify the similarity among recruited sets of aLB cells for the same environment across time, the third comparison is based on the sets of recruited aLB cells with synaptic weights taken after the first exposure to each environment, vs. the sets of recruited aLB cells with synaptic weights taken after the entire learning period (‘intermediate vs. final’ in Fig 4C). For all 10 environments tested, we find the high similarity between ‘intermediate’ and ‘final’ sets of recruited aLB cells within a given environment, even when other, distinct environments, have been learned in the meantime. This suggests the set of aLB cells recruited upon first exposure is preserved over time. Apart from the capacity to maintain distinct sets of visual-to-aLB feedback connections Fig 4 also demonstrates the ability of the system to perform pattern separation, given the similarity of the environments. S9 Fig shows similar results for a more varied set of cues in each environment. The capacity also scales with the number of aLB cells (S9C Fig).
Together the above results suggest a high capacity to maintain distinct feedback connections for a large number of environments. Since aLB cells convey visual feedback to HD cells (via RSC; see next sections), these distinct feedback connections (from distinct sets of aLB cells) can ensure HD remapping between distant environments and continuity between adjacent environments [55]. The effective remapping (in 1D) of aLB cells is reminiscent of place cell remapping (in 2D) [56,57]. That is, when an animal is re-introduced into a familiar environment, the sparse subset of aLB cells specific to that environment will be recruited to convey feedback, similar to the associative memory recall shown in hippocampus [58], and HD cells will adopt the preferred firing direction they exhibited previously in any given environment.
Multimodal dRSC signals across mirrored environments
To investigate how the processed landmark information from aLB cells can act on HD cells, and to check for consistency with experimental data [28], the sensory aLB cells are now connected with dRSC, which receives HD information from the HD attractor (via gRSC) and in turn sends feedback to the attractor to compensate for natural drift (see Fig 1A for the full network).
Constraints on the systems-level function of RSC in HD processing come from recent experiments reported by Jacob et al. [28]. In these experiments, rats could freely explore two environments with an identical simple visual cue card located in both two environments but in opposite directions (e.g. West and East, respectively) with distinct odor cues in each environment. After exploration across these two simple environments, two types of bidirectional (BD) HD cells were reported. Some cells fired for two HDs, one per environment. These cells were termed between-compartment BD cells (BC-BD). In addition, within-compartment BD cells (WC-BD) showed bimodal HD firing patterns when restricted in a single environment. Both BD cell types (BC and WC) were solely found in dRSC.
To test that the aLB model of landmark processing is compatible with these findings, we set up two connected environments similar to the experimental design from Jacob et al. [28] (also see [30] for related modelling work). Both environments were additionally enriched with multiple cues constituting complex sceneries (Figs 5A and S6A). The second environment has the same scenery but is rotated by 180°. In our simulations, the agent alternates between Env. I and Env. II on a short time scale (relative to the whole learning time) of 60 seconds (Fig 5A), mimicking the free access to both environments experienced by the animals in Jacob et al.’s experiment [28]. Similar to earlier simulations, the learning phase is followed by testing phases where no more learning occurs.
After learning, neurons in dRSC show bimodal firing patterns during the testing phase when the agent alternates between both environments (Fig 5C, top). Coherent with experimental findings on WC-BD cells, dRSC cells in the model show a bimodal activity profile firing for the duplicate (arrays of) visual landmarks in opposite directions. Moreover, bimodal firing patterns are preserved when the agent is restricted to one of the environments for testing (Figs 5C, bottom, and S6B), as the learned connections between gRSC and dRSC preserve bimodal firing patterns when restricted in a single environment (Fig 5B). I.e. with HD coming into gRSC, the agent is exposed to two landmark signals across the two environments (one being the 180° rotated version of the other), leading to the same HD being associated with two landmark bearings. In addition, similar multimodal firing profiles of dRSC cells are also found in the simulation with three conflicting environments (S5 Appendix and S7 Fig), replicating a similar simulation result from [30]. These results suggest that the simulated dRSC cells in our model are consistent with the WC-BD cells reported by Jacob et al. [28], integrating abstract visual information with internal HD signals.
Jacob et al. [28] also used distinct odor cues in the two connected environments. To simulate olfactory cues, ‘odors’ were treated as constant ‘visual’ signals (spanning 360°), distinct from the red, blue, and green input channels in Fig 5. Each environment is accompanied by a unimodal ‘green’ visual cue in opposite directions (S6C Fig). After alternating exploration, dRSC cells still show bimodal firing patterns when tested in any single environment (S6D Fig), i.e. bimodal cells are still of the WC-BD type. This suggests that BC-BD cells are unlikely to be solely explained by different odor cues, as modelled here. Neither are they likely to stand for aLB cells themselves with observable unimodal firing patterns, since BC-BD cells preserve bimodal firing patterns during darkness and are thus not solely of sensory origin [28], unlike aLB cells, which are driven by visual signals alone in the present model. The bimodal (WC-BD) firing patterns of many dRSC cells are also preserved when tested in darkness given unimodal odor cues in each environment (S6E Fig), in accordance with such experimental findings.
To test whether BC-BD cells could be generated by other means, we explore heterogeneous learning rates. For half of dRSC cells (180 out of 360), the learning rate on incoming connections is reduced to near zero (see S1 Table for value settings). Under these conditions, the model generates both WC-BD and BC-BD cells simultaneously (S6F Fig). This suggests that experimentally reported BC-BD cells [28] may correspond to dRSC cells that have not (yet) fully learnt bimodal firing in a single environment (S6F Fig, middle) due to variations in the learning rates across all dRSC cells (also see Discussion section).
Stabilizing global HD signals
Compared to the present model, the majority of previous studies presumed a direct projection from visual cue representations to the HD attractor (e.g. [36]). Only recently the role of RSC has been considered in the stabilization of HD signals (e.g. [27,30]). Since we suggest a unimodal representation of complex sensory environments via aLB cells as opposed to direct projections from cue representations, we test whether aLB signals can indeed stabilize HD signals to correct accumulated path integration error.
In contrast to the natural drift without sensory inputs (Fig 6A), Fig 6B shows subcortical HD signals from the simulations in the previous section (cf. Fig 5). Across the whole learning phase, the HD representation as well as the HD firing field does not incur drift (Fig 6B, compared to red dashed lines indicating no change to the expected HD change). Feedback is reliably conveyed by the learned connections, so that the accumulated path integration error in global HD signals is corrected via the feedback connection from dRSC to the attractor (pink connection in Fig 1A). That is, the visual information that is conveyed via aLB signals and integrated with HD signals in dRSC can correct the accumulated path integration error in subcortical HD cells.
Note that bimodal dRSC signals (i.e. WC-BD cells) do not disturb the unimodality of the subcortical HD representation. For example, if a WC-BD dRSC cell fires for North and South, while the simulated agent is facing North, the HD attractor would get North-specific feedback. The same bimodal dRSC cell fires when the HD is South. However, South-specific feedback from that dRSC cell would be compensated by corresponding feedback from another dRSC cell directed at the now active HD cells. Thus HD is not affected adversely.
aLB cells vs. a direct visual-RSC connection
The core feature of aLB cells is their unimodal encoding of an array of multimodal sensory signals. These cells code for the conjunction of multiple cues for a given egocentric view. To test whether the subcortical HD system indeed benefits from such ‘scenery-based encoding’, we compare the model to an alternative model without aLB cells that employs a direct projection from visual areas to dRSC subject to classic Hebbian learning [27,30]. Crucially we employ a complex environment of the sort used above for both comparisons, since the benefits of aLB cells only become apparent in such environments.
The agent starts out exploring the same environment as in Fig 2A. After 10 minutes, the scenery is rotated by 120° anticlockwise (Fig 7A). If the array of landmarks is not treated as a coherent whole (i.e. a scene), the ‘red’ cue on its own can be perceived as having either rotated 60° clockwise, or 120° anticlockwise consistent with the ‘blue’ cue rotation. If such a conflicting ‘red’ cue has higher saliency, its contribution to feedback can dominate if a direct connection from visual areas to dRSC is employed in lieu of aLB cells (Fig 7C, top). For currently active HD cells receiving such conflicting feedback, preferred HDs 60° clockwise lie closer on the HD attractor ring than preferred HDs 120° anticlockwise. Thus, the HD signal is more likely to shift towards the clockwise direction (Fig 7C and 7D, right). That is, if individual cues in an array possess high saliency and ambiguous directional specificity, they may lead to this kind of erroneous feedback. With aLB signals, on the other hand, HD cells can correctly follow the 120° anticlockwise rotation (Fig 7B and 7D, left) as the whole scene (all cues) is encoded as a conjunction of cues.
Introducing complex cue arrays suggests Hebbian learning acting on a direct visual-to-dRSC connection cannot guarantee HD firing field rotation following environmental changes in complex environments. Together with the added computational power bestowed by aLB cells (robustness, the ability to disconnect themselves from unstable cues, resistance to blocking, and high capacity), these results suggest that a neural representation akin to aLB cells may serve as an intermediate processing layer for the scene-based sensory information within RSC or upstream of it.
To further test the utility of aLB cells, we compare the two models (one with aLB cells, one with a Hebbian visual to RSC connection) under the same multi-environmental case as in Fig 4 (capacity simulations). Each model is simulated in all environments sequentially, 2 minutes per environment. The distinguishing ‘green’ cue shifts by 36° anticlockwise in each successive environment.
Since distinct sets of aLB cells convey scene-based landmark bearing from multiple scenes to dRSC (with high capacity; Fig 4), different sets of aLB cells are recruited among these environments to provide unique feedback to dRSC without interference (Figs 4C and S5). For visualization purposes we test for environment-specific feedback by choosing HD between environments to be continuous. That is, the initial HD in each environment is different, whilst the HD at the last moment in Env. I is the same as the initial HD in Env. II. As a consequence, the preferred directions of HD cells (and gRSC cells) should not change across all environments if each environment provides the correct feedback signals. This is the case with aLB cells (Fig 8A–8C, left). That is, aLB cells have anchored new scenes to the current HD such that the correct HD is retrieved on each previously learned scene, with minimal heading errors (-2.66° ± 0.03°).
However, this is not the case with a direct visual-to-dRSC Hebbian connection (cf. Fig 7), in which the preferred direction of dRSC cells (Fig 8A, right), as well as the HD cells (Fig 8B, right), would shift to follow the single changing cue (note that only 1 out of 3 cues shifts). In the absence of the conjunctive encoding provided by aLB cells, connections from the salient and changing ‘green’ cue learned in one environment will perturb HD in the next (cf. Fig 4). That is, HD cells would change their preferred directions between connected environments, despite the continuity of externally measured head direction (Fig 8B and 8C, right). This gives rise to large HD retrieval errors (151.91° ± 0.04°). Thus, we suggest that aLB cells (providing a scene-based, conjunctive encoding) can enable HD retrieval with near zero HD errors across multiple sceneries that are partially overlapping at the sensory level, contrary to a direct V1-dRSC connection.
Discussion
Head direction models have traditionally focused on the nature of the HD attractor circuitry (e.g. [36]). Visual feedback that anchored the attractor against drift is commonly represented as a simple Gaussian function, approximating classical HD recordings with a single cue card in the environment. Only recently, models have begun to investigate the mechanisms by which visual feedback for HD is conveyed in any detail [27]. Here we have proposed that abstract landmark-bearing (aLB) cells represent the conjunction of all available sensory cues. Produced by a modified version of the Oja’s Subspace Algorithm [34], aLB cells can cope with more complex environments, with individual cues potentially being multimodal (ambiguous with regard to directional feedback), or of low directional specificity. We have shown that landmark processing by this algorithm can extract unimodality and directional specificity from distinct cues where neither cue by itself combines both directional precision and unimodality. The learning is robust with regards to unstable cues, can incorporate novel cues, endows the model with a high capacity for distinct environments, and is compatible with the most recent experimental data on retrosplenial HD coding.
Abstract landmark bearing (aLB) cells
The strongest prediction of the model is the presence of a landmark signal that is purely sensory and unimodal, and that is not necessarily referencing a specific landmark, but rather the bearing of an array of cues, potentially including multimodal cues and cues with low directional specificity. This signal is encoded by cells we termed abstract landmark bearing (aLB) cells. Here, these cells encode scene specific bearings derived from visual signals, though it is conceivable that other sensory modalities contribute to their firing. aLB cells could contribute to panoramic memory supported by RSC [59]. Crucially, the firing of aLB cell constitutes a signal wholly dependent on sensory inputs and independent of vestibular inputs, unlike HD cells. This property could be tested in experiments that dissociate visual and vestibular signals, e.g. with sub-threshold vestibular rotations of an animal relative to its environment [60]. In addition, being of sensory origin aLB cells should cease firing in darkness, unless their firing is supported by sensory modalities other than vision. However, given the winner-take-all dynamics at play in the aLB layer, these cells may have a narrower tuning width compared to retrosplenial HD cells.
The present model is agnostic as to the exact location of these cells. Anatomically, dRSC holds strong reciprocal connections with visual areas [61] and parietal cortex [62], which are believed to code for the egocentric visual scenery prior to RSC [26]. In addition, besides RSC [10], HD-specific signals were also found in human medial parietal cortex [16], prestriate cortex [17], and deep layers of V1 [18] in rats. RSC has been implicated in sensory integration [63]. However, the classical subdivision of RSC into dysgranular and granular regions also hides a more fine-grained anatomical parcellation [64]. Therefore, aLB cells could be located in any of these cortical regions. It also remains a possibility that both direct and indirect pathways (with only the latter containing aLB cells) map visual inputs onto RSC concurrently.
Lateral inhibition and feedback transmission
A clear role of aLB cells is that they represent a compressed representation of the entire visual scene, which is realized through lateral inhibition and the modified Oja’s Subspace Algorithm (mOSA). Neither the lateral inhibition alone (paired with classical Hebbian learning) nor the feedback alone (i.e. original OSA without lateral inhibition among aLB cells) yields a unimodal representation if the simulated agent is confronted with a complex scenery that contains conflicting cues. This suggests both lateral inhibition and a reasonably strong feedback term in the learning rule are necessary.
The feedback-term in mOSA implements long-term depression on the synaptic weights from individual visual cue representations (input channels) to aLB cells. This allows the algorithm to unlearn connections that may interfere with encoding, e.g. pruning initial random connections and disconnecting aLB cells from visual cells that represented an ephemeral cue. This suggests that pruning visual connections based on ‘expected inputs’ (conveyed by the feedback term) is necessary for the development of an aLB-like representation. Alternative algorithms, such as in Hebbian covariance learning [49,50] or Intrator’s BCM [51] did not yield a unimodal representations of landmark-bearing when faced with a complex scenery containing conflicting cues, even when combined with lateral self-inhibition (S4 Appendix and S1A Fig).
At the computational level, the problem of HD feedback includes computing the egocentric landmark signals best suited to stabilize a drifting HD attractor. At the algorithmic level, we have proposed a solution in the form of a modified OSA. The present learning algorithm effectively depresses weights from expected visual inputs that are not consistent with the current scene, via the OSA feedback term for weight normalization. How could this feedback term be implemented? Separate regimes of synaptic depression and potentiation regimes, as well as the forward and backward transmission of information, could be orchestrated by the theta rhythm, similar to the proposed role of theta in the encoding and decoding processes in spatial memory [65–68]. That is, theta modulation of HD cells outside the generative circuitry of the attractor (e.g. in the retrosplenial cortex medial entorhinal cortex and parasubiculum [69]) could orchestrate feed-forward (at one phase of theta) vs feedback passes (at the opposite phase of theta) from visual and parietal to hippocampal regions via RSC and back, in order to associate visual inputs with the head direction system. Thus the present model may provide the first stepping stone towards a functional explanation of theta modulation in some HD cells [69], a property that is currently lacking theoretical understanding.
Note that mOSA with lateral inhibition and feedback transmission is a non-local algorithm, and could be equivalently realized by a local neural algorithm in terms of extra interneurons encoding feedback signals (e.g. [48]). Such a neural-network architecture could potentially be akin to the hippocampal CA3 back-projection onto dentate gyrus for pattern separation and storage [70,71]. Alternatively, the sparse coding in dentate gyrus may also be compatible with learning rules other than the mOSA in conjunction with autoassociative networks [71] or attractor neural networks for conflicting cue disambiguation [72].
Lateral inhibition leads to fairly narrow aLB tuning curves. We note that less sparse and more broadly tuned aLB activity could result if subsets of aLB cells do not share inhibitory connections with each other. Thus they could fire together while suppressing other aLB cells, e.g. when the strict lateral inhibition in Eq 4 is changed to other forms for Gaussian-like output firing patterns (e.g. Eq 8 in S2 Appendix). However, such broadening of aLB tuning curves would likely come at the cost of a decline in storage capacity [73–75].
Coping with environmental changes, and ‘remapping’
We have shown that aLB cells exhibit steady unimodal encoding in the presence of unstable visual features (including moving cues and teleporting cues) without any extra top-down control. If the ephemeral cue is not more salient than the remaining (stable) array of cues, the mOSA algorithm can disconnect aLB cells from such ephemeral landmarks via long-term depression (see above discussion of the feedback-term in the learning rule) and tries to maintain synaptic connections coherent with the most recently experienced scenery. This effect offers a neural-level explanation for why stable cues tend to leave stronger fMRI signatures in RSC [53].
A novel cue can be incorporated by aLB cells, showing that blocking by previously learned features is not an issue. Absence of blocking of directional cues has been reported among HD cells [2] and place cells [54], which is likely to be explained by the absence of blocking in their upstream directional inputs (possibly from aLB cells). Moreover, even if the novel feature has been learned as part of a complex scenery, it can convey feedback (via aLB cells) on its own.
Note that, when the novel cue is more salient than the remaining array of cues, a novel set of aLB cells is recruited. This raises the question of whether or not the set of active aLB cells could provide a signal to distinguish environments at the sensory level. If only minor changes to the cue configuration are to be expected within a given environment, and more radical changes between environments, then aLB cells may help an agent distinguish between these. However, other mechanisms, such as place cell remapping [56] can also convey that information. Similarly, an agent can reasonably expect that a change of environment is accompanied by substantial path integration inputs and should not rely on sensory discrimination alone. Finally, to put the effect of cue saliency in the model into perspective, we note that within a single real-world environment with a multitude of stable features, it is very unlikely that a single added cue will overpower an already active aLB representation to recruit new cells. Hence this should only happen in novel environments, but could in principle be tested by substituting environmental (distal) cues with the cues from another familiar environment, expecting aLB remapping and a shift in the HD cell representation. Completely novel cues on the other hand, would likely only lead to the recruitment of a new set of aLB cells, and visual feedback being associated with the currently active HD cells (no HD shift), since the novel cues are not associated with strong feedback connections onto HD cells yet.
We note that the relationship between aLB cells and HD cells in the model is somewhat analogous to that suggested between place cells and grid cells. Place cells have been shown to develop links to multiple cues in a cue-rich environment [58], and their firing can subsequently be supported by a subset of cues. aLB cells are similarly associated with multiple cues (Fig 3) and remap (i.e. to a different set of active aLB cells; S3C, S4, and S5 Figs) after substantial changes to the cue array (cf. place cell remapping [56]). HD cells, on the other hand, are ‘re-used’ throughout all environments (like grid cells), except for a coherent offset (angular for HD cells, spatial for grid cells, see [76]). HD cells implement angular path integration, whilst grid cells implement translational path integration. The sensory and path integration estimates are then combined [21] by associating aLB cells and HD cells analogous to the association of place and grid cells. That is, in the 2-dimensional spatial domain, grid cells have been proposed to update (or even generate) place cell responses [77,78], while place cells may also contribute to grid cells firing [47,79,80] and provide location feedback for grid cells (performing inherently noisy path integration) when an animal is far from environmental boundaries.
Coding capacity
To enable spatial navigation with a large spatial scale, the navigational system with internal heading representations should be able to cope with numerous rich environments concurrently. We have shown how aLB cells and the mOSA algorithm may help accomplish a large capacity for distinct environments by allowing the model to develop distinct sets of feedback connections (activating distinct sets of aLB cells) to provide environment-specific feedback. That is, upon re-entry into familiar environments (populated by a complex array of sensory cues) with a given heading, the present model will retrieve the correct, environment-specific HD in each environment. Capacity simulations cover both highly similar environments (demonstrating pattern separation) and more varied environments. Moreover, the capacity appears to scale, at a minimum, linearly (and possibly supra-linearly) with the number of aLB cells, with only 1080 aLB cells allowing for the correct recall of up to 16 out of 20 environments (S9C Fig). This suggests that distinct sets of feedback connections for hundreds of environments may be maintained by an aLB population with typical cortical population numbers (e.g. compared to O(105) neurons estimated to reside in the subiculum of rodents [81]), likely outstripping any real-world need for a rodent in its lifetime. These considerations strongly suggest that the present model can, in principle, cope with a large number of distinct environments.
Compatibility with bidirectional (BD) cells
Jacob and co-workers [28] reported bidirectional (BD) HD cells in RSC. These cells exhibit a bimodal activity profile, firing in response to identical visual landmarks in mirrored, connected environments. BD cells were solely found in dRSC [28]. These data provide further evidence for dRSC as an integrator of HD signals and visual landmark information. dRSC cells in the present model correspond to within-compartment BD (WC-BD) cells that preserve the bimodal activity profile when restricted in a single environment after exploration (similar to [30]).
Concurrently reported between-compartment BD (BC-BD) cells, seemingly solely fire for the bearing of mirrored and otherwise equal external landmarks across two rooms [28]. According to our simulations, these cells also cannot be solely explained by different, diffuse odor cues. Given that BC-BD cells keep firing in darkness [28], they are not likely to be of solely sensory origin (unlike aLB cells). Instead, BC-BD cells could occur due to inconsistent learning rates among dRSC cells that give rise to within-environment unimodality (cf. S6F Fig). However, since some BC-BD cells were reported to be spatially-modulated [28], these BC-BD cells may be derived from inhibitory control on WC-BD cells from hippocampal place cells (not part of the present model), providing a contextual signal to distinguish environments. Such inhibitory control on RSC via place cells has been proposed for parallax correction [27], providing a functional account of inhibitory connections from the hippocampus to RSC [82,83].
In the present model we have focused exclusively on landmark coding and the properties of complex environments. However, we note that making aLB cells or dRSC cells subject to spatial modulation should in principle make the present model compatible with properties of our previous model [27]. Both these models try to suggest normative explanations of HD coding in RSC: addressing the necessary computational mechanisms to cope with complex environments and location-specific feedback (e.g. for parallax correction), respectively.
Model predictions
Beyond the replication of recent experiments, we also suggest an experimentally practical scheme to test for the presence of an aLB-like representation (Fig 7). Short of directly recording these cells (appearing as possibly narrowly tuned), HD-like cells whose responses may vanish in darkness, unless supported by other sensory modalities, such as tactile, olfactory, or acoustic cues), the conjunctive encoding of the cue array has strong implications for how conflicting cues within an array of cues act on the HD attractor. E.g. two identical cues separated by 180° can appear as two different rotations when considered in isolation (60° clockwise or 120° anticlockwise, red cues in Fig 7). Only by taking the relative angle between the bimodal red and the unimodal blue cue into account can those rotation angles be distinguished. We have shown that contrary to the traditional view on the HD system (with a direct V1-RSC connection), HD cells driven by aLB cells (which treat the entire array of cues as one entity) could successfully follow such environmental rotations. Therefore, in complex environments similar to the ones used here, animals should be subject to the same shortcomings seen in the present model when aLB cells are removed.
With regard to bidirectional HD coding in RSC [28], our simulations suggest that BD-WC cells could be the product of variable learning rates across the dRSC population. Acetylcholine has been suggested to modulate learning in dRSC [82,84,85], and differential learning rates could thus be related to variations in bound neuromodulator.
We note that after removing aLB cells (i.e. treating all cues/input channels individually; Figs 7 and 8), the cortical pathway of our model resembles the snapshot model of visual feedback suggested e.g. for bees [86]. In fact, in various insects [87–90], including bees utilizing polarized skylight cues [89,91,92], the compass system (akin to HD system in rats) that provides the sense of direction with ring attractor dynamics and angular velocity integration is found in the central complex [36,93–95]. In the fly brain, such compass system in the central brain shows unimodal heading representation of the scenery with multiple salient cues [96] as well as the flexibility to adapt to multiple sensory sceneries [96,97]. Although the fast all-to-all inhibition that results in winner-takes-all dynamics on these compass neurons would induce directional ambiguation with multiple salient cues [98], this could be avoided with additional aLB cells (Figs 1 and 2). Thus, our model predicts that, if the snapshot model correctly accounts for insect data, these animals should exhibit difficulties in orienting in environments similar to the ones employed here. Similarly, the high encoding capacity shown in our model (c.f. Fig 4) could be absent.
Another prediction is that the firing of aLB cell constitutes a signal wholly dependent on sensory inputs and independent of vestibular inputs, unlike HD cells. This property could be tested in experiments that dissociate visual and vestibular signals. Finally, we have proposed that the capabilities of the model may be dependent on the theta modulation of some HD cells associating visual inputs with the HD system (see above). aLB cells also show remapping-like behavior between distinct environments, suggesting they may be 1-dimensional analogues of place cells [36,57], in stark contrast to classical HD cells.
Conclusions
We have proposed a theoretical model of environmental stabilization of the HD system supported by abstract landmark bearing (aLB) cells. aLB cells form a compressed, sparse representation of the sensory environment. Based on a novel learning algorithm this model supports a large capacity for diverse environments, and extracts useful information across complex arrays of cues to achieve a unimodal representation of landmark bearing abstracted from any individual cue. RSC then integrates this landmark information with the HD attractor and provides feedback to stabilize HD signals. The model is robust against unstable cues, and able to incorporate new stable cues (i.e. it is resistant to blocking) without any top-down curation of landmark information. This theoretical framework integrates numerous empirical findings of various types of HD-modulated firing in RSC, and predicts a new type of cell, as well as the outcome of behavioral manipulations on the HD attractor in the presence of conflicting cues.
The present model suggests a new perspective on the neural mechanisms of spatial navigation in rich sensory environments more akin to real-world scenarios. However, like the hippocampal formation, including entorhinal cortex [5,99–101], RSC has been implicated in non-spatial coding as well [102]. With aLB cells being the product of a modified Oja’s Subspace Algorithm (which is related to PCA), an instructive avenue for future research may be to consider how non-spatial data could be processed by RSC and relayed to the medial temporal lobe, while preserving the capacity of RSC to act on the HD system. The results may suggest more general coding principles for RSC that apply to spatial and non-spatial coding alike.
Supporting information
Acknowledgments
The authors thank Daniel Manson for providing real head direction trajectory data (S1 Dataset).
Data Availability
Animal behavioural data are within the Supporting Information files (S1 Dataset). Simulation results data are available on Kaggle: www.kaggle.com/chronowanderer/aLBcells-for-HDsystem-simulation-results Relevant computational modelling codes (in MATLAB 2017a) are available on GitHub: www.github.com/Chronowanderer/aLBcells-for-HDsystem and www.github.com/Bicanski/aLBcells-for-HDsystem.
Funding Statement
AB and NB acknowledge funding by the European Research Council (erc.europa.eu) No. 694779, the Wellcome Trust (wellcome.org) No. 202805/Z/16/Z, and Horizon 2020 (ec.europa.eu/programmes/horizon2020) No. 785907. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Taube JS, Muller RU, Ranck JBJ. Head-direction cells recorded from the postsubiculum in freely moving rats. I. Description and quantitative analysis. J Neurosci. 1990; 10(2):420–35. doi: 10.1523/JNEUROSCI.10-02-00420.1990 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Goodridge JP, Dudchenko PA, Worboys KA, Golob EJ, Taube JS. Cue control and head direction cells. Behav Neurosci. 1998; 112(4):749–61. doi: 10.1037//0735-7044.112.4.749 [DOI] [PubMed] [Google Scholar]
- 3.Taube JS. The Head Direction Signal: Origins and Sensory-Motor Integration. Annu Rev Neurosci. 2007; 30(1):181–207. doi: 10.1146/annurev.neuro.29.051605.112854 [DOI] [PubMed] [Google Scholar]
- 4.Welberg L. The place of head-direction cells. Nat Rev Neurosci. 2014; 15(3):136–136. doi: 10.1038/nrn3693 [DOI] [Google Scholar]
- 5.Doeller CF, Barry C, Burgess N. Evidence for grid cells in a human memory network. Nature. 2010; 463(7281):657–61. doi: 10.1038/nature08704 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Jacobs J, Kahana MJ, Ekstrom AD, Mollison M V, Fried I. A sense of direction in human entorhinal cortex. Proc Natl Acad Sci U S A. 2010; 107(14):6487–92. doi: 10.1073/pnas.0911213107 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Shine JP, Valdés-Herrera JP, Hegarty M, Wolbers T. The Human Retrosplenial Cortex and Thalamus Code Head Direction in a Global Reference Frame. J Neurosci. 2016; 36(24):6371–81. doi: 10.1523/JNEUROSCI.1268-15.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Kim M, Maguire EA. Encoding of 3D head direction information in the human brain. Hippocampus. 2019; 29(7):619–29. doi: 10.1002/hipo.23060 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Dillingham CM, Frizzati A, Nelson AJD, Vann SD. How do mammillary body inputs contribute to anterior thalamic function? Neurosci Biobehav Rev. 2015; 54:108–19. doi: 10.1016/j.neubiorev.2014.07.025 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Cho J, Sharp PE. Head direction, place, and movement correlates for cells in the rat retrosplenial cortex. Behav Neurosci. 2001; 115(1):3–25. doi: 10.1037/0735-7044.115.1.3 [DOI] [PubMed] [Google Scholar]
- 11.Sharp PE, Tinkelman A, Cho J. Angular velocity and head direction signals recorded from the dorsal tegmental nucleus of gudden in the rat: Implications for path integration in the head direction cell circuit. Behav Neurosci. 2001; 115(3):571–88. doi: 10.1037/0735-7044.115.3.571 [DOI] [PubMed] [Google Scholar]
- 12.Blair HT, Cho J, Sharp PE. Role of the Lateral Mammillary Nucleus in the Rat Head Direction Circuit: A Combined Single Unit Recording and Lesion Study. Neuron. 1998; 21(6):1387–97. doi: 10.1016/s0896-6273(00)80657-1 [DOI] [PubMed] [Google Scholar]
- 13.Stackman RW, Taube JS. Firing Properties of Rat Lateral Mammillary Single Units: Head Direction, Head Pitch, and Angular Head Velocity. J Neurosci. 1998; 18(21):9020–37. doi: 10.1523/JNEUROSCI.18-21-09020.1998 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Taube JS. Head direction cells recorded in the anterior thalamic nuclei of freely moving rats. J Neurosci. 1995; 15(1):70–86. doi: 10.1523/JNEUROSCI.15-01-00070.1995 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Sargolini F, Fyhn M, Hafting T, McNaughton BL, Witter MP, Moser M-B, et al. Conjunctive Representation of Position, Direction, and Velocity in Entorhinal Cortex. Science (80-). 2006; 312(5774):758–62. doi: 10.1126/science.1125572 [DOI] [PubMed] [Google Scholar]
- 16.Baumann O, Mattingley JB. Medial parietal cortex encodes perceived heading direction in humans. J Neurosci. 2010; 30(39):12897–901. doi: 10.1523/JNEUROSCI.3077-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Chen LL, Lin L-H, Green EJ, Barnes CA, McNaughton BL. Head-direction cells in the rat posterior cortex. Exp Brain Res. 1994; 101(1):8–23. doi: 10.1007/BF00243212 [DOI] [PubMed] [Google Scholar]
- 18.Keshavarzi S, Margrie TW, Vélez-Fort M, Cossell L, Strom M, Bracey EF, et al. A Circuit for Integration of Head- and Visual-Motion Signals in Layer 6 of Mouse Primary Visual Cortex. Neuron. 2018; 98(1):179–191.e6. doi: 10.1016/j.neuron.2018.02.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Taube JS, Burton HL. Head direction cell activity monitored in a novel environment and during a cue conflict situation. J Neurophysiol. 1995; 74(5):1953–71. doi: 10.1152/jn.1995.74.5.1953 [DOI] [PubMed] [Google Scholar]
- 20.Yoder RM, Clark BJ, Brown JE, Lamia M V., Valerio S, Shinder ME, et al. Both visual and idiothetic cues contribute to head direction cell stability during navigation along complex routes. J Neurophysiol. 2011; 105(6):2989–3001. doi: 10.1152/jn.01041.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Evans T, Bicanski A, Bush D, Burgess N. How environment and self-motion combine in neural representations of space. J Physiol. 2016; 594(22):6535–46. doi: 10.1113/JP270666 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Yoder RM, Taube JS. The vestibular contribution to the head direction signal and navigation. Front Integr Neurosci. 2014; 8:32. doi: 10.3389/fnint.2014.00032 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Tan HM, Bassett JP, O’Keefe J, Cacucci F, Wills TJ. The development of the head direction system before eye opening in the rat. Curr Biol. 2015; 25(4):479–83. doi: 10.1016/j.cub.2014.12.030 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Mizumori SJ, Williams JD. Directionally selective mnemonic properties of neurons in the lateral dorsal nucleus of the thalamus of rats. J Neurosci. 1993; 13(9):4015–28. doi: 10.1523/JNEUROSCI.13-09-04015.1993 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Shibata H. Organization of projections of rat retrosplenial cortex to the anterior thalamic nuclei. Eur J Neurosci. 1998; 10:3210–9. doi: 10.1046/j.1460-9568.1998.00328.x [DOI] [PubMed] [Google Scholar]
- 26.Vann SD, Aggleton JP, Maguire EA. What does the retrosplenial cortex do? Nat Rev Neurosci. 2009; 10(11):792–802. doi: 10.1038/nrn2733 [DOI] [PubMed] [Google Scholar]
- 27.Bicanski A, Burgess N. Environmental Anchoring of Head Direction in a Computational Model of Retrosplenial Cortex. J Neurosci. 2016; 36(46):11601–18. doi: 10.1523/JNEUROSCI.0516-16.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Jacob PY, Casali G, Spieser L, Page H, Overington D, Jeffery K. An independent, landmark-dominated head-direction signal in dysgranular retrosplenial cortex. Nat Neurosci. 2017; 20(2):173–5. doi: 10.1038/nn.4465 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Page HJI, Walters DM, Knight R, Piette CE, Jeffery KJ, Stringer SM. A theoretical account of cue averaging in the rodent head direction system. Philos Trans R Soc B Biol Sci. 2013; 369(1635):20130283. doi: 10.1098/rstb.2013.0283 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Page HJI, Jeffery KJ. Landmark-Based Updating of the Head Direction System by Retrosplenial Cortex: A Computational Model. Front Cell Neurosci. 2018; 12(July):1–17. doi: 10.3389/fncel.2018.00191 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Skaggs WE, Knierim JJ, Kudrimoti HS, McNaughton BL. A model of the neural basis of the rat’s sense of direction. Adv Neural Inf Process Syst. 1995; 7:173–80. [PubMed] [Google Scholar]
- 32.Redish AD, Elga AN, Touretzky DS. A coupled attractor model of the rodent Head Direction system. Netw Comput Neural Syst. 1996; 7(4):671–85. doi: 10.1088/0954-898X_7_4_004 [DOI] [Google Scholar]
- 33.Oja E. Simplified neuron model as a principal component analyzer. J Math Biol. 1982; 15(3):267–73. doi: 10.1007/BF00275687 [DOI] [PubMed] [Google Scholar]
- 34.Oja E. NEURAL NETWORKS, PRINCIPAL COMPONENTS, AND SUBSPACES. Int J Neural Syst. 1989; 01(01):61–8. doi: 10.1142/S0129065789000475 [DOI] [Google Scholar]
- 35.Oja E. Principal components, minor components, and linear neural networks. Neural Networks. 1992; 5(6):927–35. doi: 10.1016/S0893-6080(05)80089-9 [DOI] [Google Scholar]
- 36.Zhang K. Representation of Spatial Orientation by the Intrinsic Dynamics of the Head-Direction Cell Ensemble: A Theory. J Neurosci. 1996; 16(6):2112–26. doi: 10.1523/JNEUROSCI.16-06-02112.1996 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Song P, Wang X-J. Angular Path Integration by Moving “Hill of Activity”: A Spiking Neuron Model without Recurrent Excitation of the Head-Direction System. J Neurosci. 2005; 25(4):1002–14. doi: 10.1523/JNEUROSCI.4172-04.2005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Stratton P, Wyeth G, Wiles J. Calibration of the head direction network: a role for symmetric angular head velocity cells. J Comput Neurosci. 2010; 28(3):527–38. doi: 10.1007/s10827-010-0234-7 [DOI] [PubMed] [Google Scholar]
- 39.van Groen T, Wyss JM. Connections of the Retrosplenial Dysgranular Cortex in the Rat. J Comp Neurol. 1992; 315:200–16. doi: 10.1002/cne.903150207 [DOI] [PubMed] [Google Scholar]
- 40.van Groen T, Wyss JM. Connections of the retrosplenial granular b cortex in the rat. J Comp Neurol. 2003; 463(3):249–63. doi: 10.1002/cne.10757 [DOI] [PubMed] [Google Scholar]
- 41.Harth E, Pertile G. The role of inhibition and adaptation in sensory information processing. Kybernetik. 1972; 10(1):32–7. doi: 10.1007/BF00288781 [DOI] [PubMed] [Google Scholar]
- 42.Chen T, Hua Y, Yan W. Global convergence of Oja’s subspace algorithm for principal component extraction. IEEE Trans Neural Networks. 1998; 9(1):58–67. doi: 10.1109/72.655030 [DOI] [PubMed] [Google Scholar]
- 43.Maass W. On the computational power of winner-take-all. Neural Comput. 2000; 12(11):2519–35. doi: 10.1162/089976600300014827 [DOI] [PubMed] [Google Scholar]
- 44.Binas J, Rutishauser U, Indiveri G, Pfeiffer M. Learning and stabilization of winner-take-all dynamics through interacting excitatory and inhibitory plasticity. Front Comput Neurosci. 2014; 8:68. doi: 10.3389/fncom.2014.00068 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Douglas RJ, Martin KAC. NEURONAL CIRCUITS OF THE NEOCORTEX. Annu Rev Neurosci. 2004; 27(1):419–51. doi: 10.1146/annurev.neuro.27.070203.144152 [DOI] [PubMed] [Google Scholar]
- 46.Chen Y. Mechanisms of Winner-Take-All and Group Selection in Neuronal Spiking Networks. Front Comput Neurosci. 2017; 11:20. doi: 10.3389/fncom.2017.00020 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Dordek Y, Soudry D, Meir R, Derdikman D. Extracting grid cell characteristics from place cell inputs using non-negative principal component analysis. Elife. 2016; 5:e10094. doi: 10.7554/eLife.10094 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Fyfe C. PCA Properties of Interneurons. In: ICANN ‘93. London: Springer London; 1993. pp. 183–8. doi: 10.1007/978-1-4471-2063-6_41 [DOI] [Google Scholar]
- 49.Sejnowski TJ. Storing covariance with nonlinearly interacting neurons. J Math Biol. 1977; 4(4):303–21. doi: 10.1007/BF00275079 [DOI] [PubMed] [Google Scholar]
- 50.Young DL, Poon C-S. Hebbian Covariance Learning. In Springer, Boston, MA; 1998. pp. 73–83. doi: 10.1007/978-1-4757-9077-1_14 [DOI] [PubMed] [Google Scholar]
- 51.Intrator N, Cooper LN. Objective function formulation of the BCM theory of visual cortical plasticity: Statistical connections, stability conditions. Neural Networks. 1992; 5(1):3–17. doi: 10.1016/S0893-6080(05)80003-6 [DOI] [Google Scholar]
- 52.Jaccard P. THE DISTRIBUTION OF THE FLORA IN THE ALPINE ZONE.1. New Phytol. 1912; 11(2):37–50. doi: 10.1111/j.1469-8137.1912.tb05611.x [DOI] [Google Scholar]
- 53.Auger SD, Maguire EA. Assessing the mechanism of response in the retrosplenial cortex of good and poor navigators. Cortex. 2013; 49(10):2904–13. doi: 10.1016/j.cortex.2013.08.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Barry J, Muller R. Updating the hippocampal representation of space: Place cell firing fields are controlled by a novel spatial stimulus. Hippocampus. 2011; 21(5):481–94. doi: 10.1002/hipo.20764 [DOI] [PubMed] [Google Scholar]
- 55.Dudchenko PA, Zinyuk LE. The formation of cognitive maps of adjacent environments: Evidence from the head direction cell system. Behav Neurosci. 2005; 119(6):1511–23. doi: 10.1037/0735-7044.119.6.1511 [DOI] [PubMed] [Google Scholar]
- 56.Muller RU, Kubie JL. The effects of changes in the environment on the spatial firing of hippocampal complex-spike cells. J Neurosci. 1987; 7(7):1951–68. doi: 10.1523/JNEUROSCI.07-07-01951.1987 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Jeffery KJ. Place cells, grid cells, attractors, and remapping. Neural Plast. 2011; 2011(1–2):182602. doi: 10.1155/2011/182602 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Nakazawa K, Quirk MC, Chitwood RA, Watanabe M, Yeckel MF, Sun LD, et al. Requirement for hippocampal CA3 NMDA receptors in associative memory recall. Science (80-). 2002; 297(5579):211–8. doi: 10.1126/science.1071795 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Robertson CE, Hermann KL, Mynick A, Kravitz DJ, Kanwisher N. Neural Representations Integrate the Current Field of View with the Remembered 360° Panorama in Scene-Selective Cortex. Curr Biol. 2016; 26(18):2463–8. doi: 10.1016/j.cub.2016.07.002 [DOI] [PubMed] [Google Scholar]
- 60.Jeffery KJ, Donnett JG, Burgess N, O’Keefe JM. Directional control of hippocampal place fields. Exp Brain Res. 1997; 117(1):131–42. doi: 10.1007/s002210050206 [DOI] [PubMed] [Google Scholar]
- 61.Vogt BA, Miller MW. Cortical connections between rat cingulate cortex and visual, motor, and postsubicular cortices. J Comp Neurol. 1983; 216(2):192–210. doi: 10.1002/cne.902160207 [DOI] [PubMed] [Google Scholar]
- 62.Kobayashi Y, Amaral DG. Macaque monkey retrosplenial cortex: II. Cortical afferents. J Comp Neurol. 2003; 466(1):48–79. doi: 10.1002/cne.10883 [DOI] [PubMed] [Google Scholar]
- 63.Nelson AJD, Hindley EL, Vann SD, Aggleton JP. When is the rat retrosplenial cortex required for stimulus integration? Behav Neurosci. 2018; 132(5):366–77. doi: 10.1037/bne0000267 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Shibata H, Honda Y, Sasaki H, Naito J. Organization of intrinsic connections of the retrosplenial cortex in the rat. Anat Sci Int. 2009; 84(4):280–92. doi: 10.1007/s12565-009-0035-0 [DOI] [PubMed] [Google Scholar]
- 65.Hasselmo ME, Bodelón C, Wyble BP. A proposed function for hippocampal theta rhythm: Separate phases of encoding and retrieval enhance reversal of prior learning. Neural Comput. 2002; 14(4):793–817. doi: 10.1162/089976602317318965 [DOI] [PubMed] [Google Scholar]
- 66.Hasselmo ME. What is the function of hippocampal theta rhythm?—Linking bahavioral data to phasic properties of field potential and unit recording data. Hippocampus. 2005; 15(7):936–49. doi: 10.1002/hipo.20116 [DOI] [PubMed] [Google Scholar]
- 67.Byrne P, Becker S, Burgess N. Remembering the past and imagining the future: a neural model of spatial memory and imagery. Psychol Rev. 2007; 114(2):340–75. doi: 10.1037/0033-295X.114.2.340 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Bicanski A, Burgess N. A Neural Level Model of Spatial Memory and Imagery. Elife. 2018; 7:e33752. doi: 10.7554/eLife.33752 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Kornienko O, Latuske P, Bassler M, Kohler L, Allen K. Non-rhythmic head-direction cells in the parahippocampal region are not constrained by attractor network dynamics. Elife. 2018; 7:1–25. doi: 10.7554/eLife.35949 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Treves A, Rolls ET. Computational analysis of the role of the hippocampus in memory. Hippocampus. 1994; 4(3):374–91. doi: 10.1002/hipo.450040319 [DOI] [PubMed] [Google Scholar]
- 71.Myers CE, Scharfman HE. Pattern separation in the dentate gyrus: A role for the CA3 backprojection. Hippocampus. 2011; 21(11):1190–215. doi: 10.1002/hipo.20828 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Hedrick K, Zhang K. Analysis of an Attractor Neural Network’s Response to Conflicting External Inputs. J Math Neurosci. 2018; 8(1):6. doi: 10.1186/s13408-018-0061-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.Marr D. Simple memory: a theory for archicortex. Philos Trans R Soc Lond B Biol Sci. 1971; 262(841):23–81. doi: 10.1098/rstb.1971.0078 [DOI] [PubMed] [Google Scholar]
- 74.Treves A, Rolls ET. Computational constraints suggest the need for two distinct input systems to the hippocampal CA3 network. Hippocampus. 1992; 2(2):189–99. doi: 10.1002/hipo.450020209 [DOI] [PubMed] [Google Scholar]
- 75.O’Reilly RC, McClelland JL. Hippocampal conjunctive encoding, storage, and recall: Avoiding a trade-off. Hippocampus. 1994; 4(6):661–82. doi: 10.1002/hipo.450040605 [DOI] [PubMed] [Google Scholar]
- 76.Fyhn M, Hafting T, Treves A, Moser MB, Moser EI. Hippocampal remapping and grid realignment in entorhinal cortex. Nature. 2007; 446(7132):190–4. doi: 10.1038/nature05601 [DOI] [PubMed] [Google Scholar]
- 77.Solstad T, Moser EI, Einevoll GT. From grid cells to place cells: A mathematical model. Hippocampus. 2006; 16(12):1026–31. doi: 10.1002/hipo.20244 [DOI] [PubMed] [Google Scholar]
- 78.Fuhs MC, Touretzky DS. A spin glass model of path integration in rat medial entorhinal cortex. J Neurosci. 2006; 26(16):4266–76. doi: 10.1523/JNEUROSCI.4353-05.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Kropff E, Treves A. The emergence of grid cells: Intelligent design or just adaptation? Hippocampus. 2008; 18(12):1256–69. doi: 10.1002/hipo.20520 [DOI] [PubMed] [Google Scholar]
- 80.Si B, Treves A. A model for the differentiation between grid and conjunctive units in medial entorhinal cortex. Hippocampus. 2013; 23(12):1410–24. doi: 10.1002/hipo.22194 [DOI] [PubMed] [Google Scholar]
- 81.Taube JS, Bassett JP. Persistent Neural Activity in Head Direction Cells. Cereb Cortex. 2003; 13(11):1162–72. doi: 10.1093/cercor/bhg102 [DOI] [PubMed] [Google Scholar]
- 82.Miyashita T, Rockland KS. GABAergic projections from the hippocampus to the retrosplenial cortex in the rat. Eur J Neurosci. 2007; 26(5):1193–204. doi: 10.1111/j.1460-9568.2007.05745.x [DOI] [PubMed] [Google Scholar]
- 83.Yamawaki N, Li X, Lambot L, Ren LY, Radulovic J, Shepherd GMG. Long-range inhibitory intersection of a retrosplenial thalamocortical circuit by apical tuft-targeting CA1 neurons. Nat Neurosci. 2019; 22(4):618–26. doi: 10.1038/s41593-019-0355-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 84.Hasselmo ME. The role of acetylcholine in learning and memory. Curr Opin Neurobiol. 2006; 16(6):710–5. doi: 10.1016/j.conb.2006.09.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Anzalone S, Roland J, Vogt B, Savage L. Acetylcholine efflux from retrosplenial areas and hippocampal sectors during maze exploration. Behav Brain Res. 2009; 201(2):272–8. doi: 10.1016/j.bbr.2009.02.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Cartwright BA, Collett TS. How honey bees use landmarks to guide their return to a food source. Nature. 1982; 295(5850):560–4. doi: 10.1038/295560a0 [DOI] [Google Scholar]
- 87.Pisokas I, Heinze S, Webb B. The head direction circuit of two insect species. Elife. 2020; 9:1–49. doi: 10.7554/eLife.53985 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88.Pfeiffer K, Homberg U. Organization and Functional Roles of the Central Complex in the Insect Brain. Annu Rev Entomol. 2014; 59(1):165–84. doi: 10.1146/annurev-ento-011613-162031 [DOI] [PubMed] [Google Scholar]
- 89.Homberg U. In search of the sky compass in the insect brain. Naturwissenschaften. 2004; 91(5):199–208. doi: 10.1007/s00114-004-0525-9 [DOI] [PubMed] [Google Scholar]
- 90.Homberg U, Heinze S, Pfeiffer K, Kinoshita M, el Jundi B. Central neural coding of sky polarization in insects. Philos Trans R Soc B Biol Sci. 2011; 366(1565):680–7. doi: 10.1098/rstb.2010.0199 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91.Srinivasan M V. Honeybees as a model for the study of visually guided flight, navigation, and biologically inspired robotics. Physiol Rev. 2011; 91(2):413–60. doi: 10.1152/physrev.00005.2010 [DOI] [PubMed] [Google Scholar]
- 92.Evangelista C, Kraft P, Dacke M, Labhart T, Srinivasan M V. Honeybee navigation: critically examining the role of the polarization compass. Philos Trans R Soc B Biol Sci. 2014; 369(1636):20130037. doi: 10.1098/rstb.2013.0037 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 93.Stone T, Webb B, Adden A, Weddig N Ben, Honkanen A, Templin R, et al. An Anatomically Constrained Model for Path Integration in the Bee Brain. Curr Biol. 2017; 27(20):3069–3085.e11. doi: 10.1016/j.cub.2017.08.052 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 94.Kim SS, Rouault H, Druckmann S, Jayaraman V. Ring attractor dynamics in the Drosophila central brain. Science (80-). 2017; 356(6340):849–53. doi: 10.1126/science.aal4835 [DOI] [PubMed] [Google Scholar]
- 95.Turner-Evans D, Wegener S, Rouault H, Franconville R, Wolff T, Seelig JD, et al. Angular velocity integration in a fly heading circuit. Elife. 2017; 6. doi: 10.7554/eLife.23496 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.Kim SS, Hermundstad AM, Romani S, Abbott LF, Jayaraman V. Generation of stable heading representations in diverse visual scenes. Nature. 2019; 576(7785):126–31. doi: 10.1038/s41586-019-1767-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Fisher YE, Lu J, D’Alessandro I, Wilson RI. Sensorimotor experience remaps visual input to a heading-direction network. Nature. 2019; 576(7785):121–5. doi: 10.1038/s41586-019-1772-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98.Hulse BK, Haberkern H, Franconville R, Turner-Evans DB, Takemura S, Wolff T, et al. A connectome of the Drosophila central complex reveals network motifs suitable for flexible navigation and context-dependent action selection. bioRxiv. 2020;2020.12.08.413955. doi: 10.1101/2020.12.08.413955 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Constantinescu AO, O’Reilly JX, Behrens TEJ. Organizing conceptual knowledge in humans with a gridlike code. Science (80-). 2016; 352(6292):1464–8. doi: 10.1126/science.aaf0941 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 100.Bottini R, Doeller CF. Knowledge Across Reference Frames: Cognitive Maps and Image Spaces. Trends Cogn Sci. 2020; 24(8):606–19. doi: 10.1016/j.tics.2020.05.008 [DOI] [PubMed] [Google Scholar]
- 101.Kaplan R, Friston KJ. Entorhinal transformations in abstract frames of reference. Rushworth MFS, editor. PLOS Biol. 2019; 17(5):e3000230. doi: 10.1371/journal.pbio.3000230 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102.Todd TP, Fournier DI, Bucci DJ. Retrosplenial cortex and its role in cue-specific learning and memory. Neurosci Biobehav Rev. 2019; 107:713–28. doi: 10.1016/j.neubiorev.2019.04.016 [DOI] [PMC free article] [PubMed] [Google Scholar]