Skip to main content

This is a preprint.

It has not yet been peer reviewed by a journal.

The National Library of Medicine is running a pilot to include preprints that result from research funded by NIH in PMC and PubMed.

bioRxiv logoLink to bioRxiv
[Preprint]. 2024 Feb 4:2023.12.27.573490. Originally published 2023 Dec 28. [Version 2] doi: 10.1101/2023.12.27.573490

Hippocampal sequences span experience relative to rewards

Marielena Sosa 1,*, Mark H Plitt 1,2, Lisa M Giocomo 1,*
PMCID: PMC10793396  PMID: 38234842

Abstract

Hippocampal place cells fire in sequences that span spatial environments and non-spatial modalities, suggesting that hippocampal activity can anchor to the most behaviorally salient aspects of experience. As reward is a highly salient event, we hypothesized that sequences of hippocampal activity can anchor to rewards. To test this, we performed two-photon imaging of hippocampal CA1 neurons as mice navigated virtual environments with changing hidden reward locations. When the reward moved, the firing fields of a subpopulation of cells moved to the same relative position with respect to reward, constructing a sequence of reward-relative cells that spanned the entire task structure. The density of these reward-relative sequences increased with task experience as additional neurons were recruited to the reward-relative population. Conversely, a largely separate subpopulation maintained a spatially-based place code. These findings thus reveal separate hippocampal ensembles can flexibly encode multiple behaviorally salient reference frames, reflecting the structure of the experience.

INTRODUCTION

Memories of positive experiences are essential for reinforcing rewarding behaviors. These memories must also have the capacity to update when knowledge of reward changes (i.e. a water source moves to a new location)1,2. How does the brain both amplify memories of events surrounding a reward and maintain a stable representation of the external world? The hippocampus provides a potential neural circuit for this process. Hippocampal place cells fire in one or a few spatially restricted locations, defined as their place fields3. In different environments, place cells “remap”, such that the preferred location or firing rate of their place field changes412. Together, the population of place cells creates reliable sequences of activity as an animal traverses an environment, resulting in a unique neural code for each environment and a unique sequence of spiking for each traversal9,1321. While these patterns support spatial navigation2226, recent evidence has revealed that hippocampal neurons also show reliable tuning relative to a variety of non-spatial modalities and create sequences of activity for these modalities17,27,28. For instance, sequential hippocampal firing has been observed across time2935, in relation to progressions of auditory36, olfactory3742, or multisensory stimuli4345, and during accumulation of evidence for decision-making tasks46. Together, these findings suggest that the hippocampus can be understood as generating sequential activity13,17,27,29,47,48 to encode the progression of events as they unfold in a given experience27,28,4951.

Multiple lines of evidence demonstrate that the hippocampus prioritizes coding for aspects of experience that are particularly salient or that can affect the animal’s behavior27,3346,5262. The presence of food or water reward is one such highly salient event that is consistently prioritized, as demonstrated by prior work finding that place cells tend to cluster near (i.e. “over-represent”) reward locations1,2,54,55,6366. In addition, a small subpopulation of hippocampal cells are active precisely at rewarded locations54,64, even when the reward is moved to a new location54. Moreover, optogenetic activation of cells with place fields near reward drives the animal to engage in reward-seeking actions24, suggesting a causal role for hippocampal activity in reward-related behaviors. Yet it remains unclear whether the hippocampus encodes complete sequences of events around rewards separately from the influence of other sensory stimuli. Moreover, it is incompletely understood how different aspects of experience are represented simultaneously.

We therefore hypothesized that reward may anchor the sequential activity of the hippocampus in a subpopulation of pyramidal cells. Remapping dynamics supporting this hypothesis have been reported at locations very close to reward sites1,2,54,55,6366, but in order to support memory of events leading up to and following rewarding experiences, hippocampal activity encoding events at distances further away from reward must also be able to update when reward conditions change. Yet predictable remapping that spans the entire environment in response to changing reward locations has not been demonstrated. We reasoned that previously reported reward-specific cells54 may comprise a subset of a larger population encoding an entire sequence of events relative to reward. This hypothesis predicts that moving the reward within a constant environment should induce remapping even at locations far from reward, in a manner that preserves sequential firing relative to each reward location. In parallel, we predicted that a subset of hippocampal neurons should preserve their firing relative to the spatial environment, reflecting an ability of the hippocampus to flexibly anchor to both the spatial environment and the experience of finding reward as two salient reference frames45,59,60,6773 that are required to solve the task at hand.

To investigate whether the hippocampus simultaneously encodes sequential experience relative to reward and the spatial environment, we used 2-photon (2P) calcium imaging74,75 to monitor large populations of CA1 neurons in head-fixed mice learning a virtual reality (VR) task. The use of VR provided tight control of the sensory stimuli and constrained the animal’s trajectory, allowing us to more readily observe remapping of neurons in relation to distant rewards, a phenomenon potentially less visible in freely moving scenarios. Further, the task included multiple updates to hidden reward zones across two environments, allowing us to dissociate spatially-driven and reward-driven remapping. We found a subpopulation of neurons that, when rewards moved, remapped to the same relative position with respect to reward and thus constructed a sequence of activity relative to the reward location. With increasing experience, more neurons were recruited to these “reward-relative sequences”. These results suggest that the hippocampus learned a generalized representation of the task anchored to reward, while also maintaining a spatial map in largely separate neural ensembles.

RESULTS

Monitoring neural activity in mice during a virtual reality navigation and reward learning task

We performed 2P calcium imaging to monitor CA1 neurons expressing GCaMP7f in head-fixed mice learning a VR linear navigation task (Fig. 1AC). To observe putative remapping related to rewards, and examine how this remapping might develop with experience, we designed a task with multiple changing hidden reward locations. In the task, animals were required to run down a 450 cm virtual linear environment (Env 1) with a single hidden, 50-cm reward zone. Sucrose water reward was delivered operantly for licking within the hidden reward zone (Fig. 1DE). Imaging began on day 1 of task acquisition. On day 3, the reward zone was moved within a “switch” session to a new location on the track. The reward switch was signaled by automatic reward delivery on the first 10 trials after the switch only if the mouse failed to lick in the new zone. On day 5, a third reward zone was introduced, followed by a return to the original zone on day 7. On day 8, the reward switch coincided with the introduction of a novel environment (Env 2) that evokes spatially-driven global remapping7. The sequence of switches was subsequently reversed in Env 2 (Fig. 1E). On intervening non-switch (“stay”) days, the last reward location from the previous day was maintained. Reward switch sequences were counterbalanced across animals (n=7 “switch” mice, Fig. S1AD). The three possible reward locations were positioned equidistant from each of the “tower” landmarks along the track, thus controlling for proximity of each reward to visual landmarks8,55,7680. A separate “fixed-condition” group of mice (n=3) experienced only one reward location in a single environment (Env 1) (Fig. S1D, F), to control for the effect of extended experience alone compared to learning about the reward switches. To keep animals engaged and prevent the movement of the reward zone from being a completely novel experience, on ~15% of trials the reward was randomly omitted for all animals. At the end of the track, the mouse passed through a brief gray “teleport” tunnel (~50 cm + temporal jitter up to 10 s; see Methods) before starting the next lap in the same direction. Time in the teleport tunnel was randomly jittered so that exact time of track entry could not be anticipated.

Figure 1. 2P imaging of CA1 neurons in a VR task with changing hidden reward locations.

Figure 1.

A) Top-down view (left) and side view (right) of head-fixed VR set-up. 5% sucrose water reward is automatically delivered when the mouse licks a capacitive lick port.

B) Coronal histology showing imaging cannula implantation site and calcium indicator GCaMP7f expression under the pan-neuronal synapsin promoter (AAV1-Syn-jGCaMP7f, green) in dorsal CA1 (DAPI, blue).

C) Example field of view (mean image) from the same mouse in (B) (mouse m12) with identified CA1 neurons in shaded colors (n=1780 cells).

D) Task timeline. On “stay” days, the reward remains at the same location throughout the session; on “switch” days, it moves to a new location after the first 30 trials (typically 80 trials/session).

E) Side views of virtual linear tracks showing an example reward switch sequence. Shaded regions indicate hidden reward zones, for illustration only. On intervening days, the last reward location from the previous day is maintained.

F) Licking behavior for an example mouse (m14) on the first day and first 3 switches. Top row: Smoothed lick rasters. Rewarded trials in black; omission trials in magenta. Shaded regions indicate the active reward zone. Bottom row: Mean ± s.e.m. lick rate for trials at each reward location (blue=A, purple=B, red=C).

G) Licking behavior for a second example mouse (m3) that experienced a different starting reward zone and switch sequence from the mouse in (F). See Fig. S1D for the full set of sequence allocations per mouse.

H) Behavior across animals performing reward switches (n=7 mice). An anticipatory lick ratio of 1 indicates licking exclusively in the 50 cm before a reward zone (“anticipatory zone”, see Fig. S1E); 0 (horizontal dashed line) indicates chance licking across the whole track outside of the reward zone. Thin lines represent individual mice, with lines colored according to the reward zone active for each mouse on a given set of trials (blue=A, purple=B, red=C); thick black and teal lines show mean ± s.e.m. across mice for pre- and post-switch trials, respectively. See also Figure S1.

We observed that mice developed an anticipatory ramp of licking that, on average, peaked at the beginning of the reward zone (Fig. 1FG, Fig. S1AC). This is the earliest location where operant licking triggered reward delivery, demonstrating successful task acquisition. After a reward switch, mice entered an exploratory period of licking but then adapted their licking to anticipate each new zone, typically within a session (Fig. 1FG, Fig. S1AC). We quantified this improvement in licking precision across the switch as the ratio of anticipatory lick rate 50 cm before the currently rewarded zone compared to everywhere else outside the zone (Fig. S1E). We found that all mice were able to improve and retain their licking precision for the new reward zone (Fig. 1H, n=7 mice), demonstrating accurate spatial learning and memory.

Moving the reward induces remapping that spans a constant environment

We focused on the switch days to analyze single-cell remapping patterns. First, we identified place cells active either before or after the reward switch (see Methods; means ± standard deviation [s.d.] across 7 mice and 7 switch sessions: 408 ± 216 place cells out of 868 ± 356 cells imaged; 48.6% ± 16.7% place cells active before or after the switch). After a reward switch within an environment, a subset of place cells maintained a stable field at the same location on the track (“track-relative” cells, 20.2% ± 8.8%, mean ± s.d. of all place cells, averaged across all 6 switches within environment, 7 mice; Fig. 2A, Fig. S2A). However, many place cells remapped, despite the constant visual stimuli and visual landmarks (Fig. 2A, Fig. S2AD). First, we observed subsets of cells with place fields that disappeared (“disappearing”, 11.5% ± 7.8%) or appeared (“appearing”, 8.3% ± 5.2%) after the reward location switched (Fig. 2A, Fig. S2A). Other cells precisely followed the reward location (“remap near reward”, 4.7% ± 3.9%, Fig. 2A, Fig. S2AD), consistently firing within ±50 cm of the beginning of the reward zone, similar to previous reports54. Notably, a subset of place cells with fields distant from reward (>50 cm) also remapped after the reward switch (“remap far from reward”, 15.6% ± 3.7%; Fig. 2A, Fig. S2A, right). A remaining 39.9% ± 19.3% of place cells did not show sufficiently stereotyped remapping patterns to be classified by these criteria. At the population level, we observed that a reward switch within a constant environment induced more remapping than the spontaneous instability seen with a fixed reward and fixed environment8185, but less remapping than the introduction of a novel environment with the reward switch511 (Fig. S2BD).

Figure 2. A subpopulation of CA1 cells remaps relative to reward.

Figure 2.

A) Five place cells from an example mouse (m4) with fields that are either track-relative, disappearing, appearing, following the reward location closely (“remap near reward”, with spatial firing peaks ≤ 50 cm from both reward zone starts), or “remap far from reward” (peaks >50 cm from at least one reward zone start; see Methods). Data are from an example switch day in Env 2 (day 14). Left of each cell: Spatial activity per trial shown as smoothed deconvolved calcium events normalized to the mean activity for each cell within session. White lines indicate beginnings of reward zones. Right of each cell: Trial-by-trial correlation matrix for that cell.

B) Example place cells that remap to the same relative distance from reward (orange cell ID labels), compared to cells that do not (black cell ID labels). Top row: Mean-normalized spatial activity plotted on the original track axis. Bottom row: Post-switch trials (after the horizontal blue dashed line) are circularly shifted to align the reward locations, to illustrate alignment of fields at a similar relative distance from reward both within (left, examples from day 14) and across environments (right, day 8).

C) Illustration of method to quantify remapping relative to reward. Left: A scatter of peak activity positions before vs. after the reward switch in linear track coordinates reveals track-relative place cells along the diagonal (dashed line; shown for example animal m12 on day 14, switching from reward zone “B” before to “A” after). An off-diagonal band at the intersection of the starts of reward zones suggests cells remapped relative to both rewards. Middle left: We center this reward-relative band by converting the linear coordinates to periodic and rotating the data to align the starts of reward zones at 0. Middle right: Track-relative place cells are removed to isolate remapping cells. Right: The data are then compared to a random remapping shuffle (one shuffle iteration shown in red). Together, these shuffles (n=1000) produce the distribution marked in red lines in (E) and (G). All points are jittered slightly for visualization (see Methods) and are translucent; darker shades indicate overlapping points.

D) Peak activity positions relative to reward for all animals on the first switch (day 3), for remapping cells with significant spatial information both before and after the switch (n=677 cells, 7 mice). Orange shaded region indicates a range of ≤50 cm (~0.698 radians) between relative peaks, corresponding to points included in the histogram in (E, bottom).

E) The fraction of cells remapping to a consistent reward-relative position (≤ 50 cm between relative peaks) on the first switch day is higher than expected by chance. Top: histogram of the circular distance between relative peaks after minus before the switch (i.e. orthogonal to the unity line in D), compared to the mean and upper 95% of the shuffle distribution (solid and dotted red lines, respectively). Bottom: Distribution of mean firing position relative to reward for cells with ≤ 50 cm distance between relative peaks, corresponding to orange shaded area along the unity line in (D) (n=261 cells, 7 mice, non-zero mean = 0.358 radians, 95% confidence interval [lower, upper] = [0.196, 0.519], circular mean test). Black dotted line marks the mean, blue shading marks the extent of the reward zone.

F) Same as (D) but for the last switch (day 14) (n=707 cells, 7 mice).

G) Same as (E) but for the last switch (day 14). Bottom: n=398 cells with ≤ 50 cm distance between relative peaks (non-zero mean = 0.206 radians, 95% confidence interval [lower, upper] = [0.103, 0.309], circular mean test).

H) Fraction of cells above chance that remap relative to reward grows linearly with task experience (number of switches). Each dot shown is the combined fraction across n=7 mice; mean ± confidence interval of linear regression is shown in gray.

See also Figure S2.

An above-chance fraction of reward-relative remapping cells

We next investigated whether a subpopulation of cells encoded experience anchored to reward. To test this hypothesis, we attempted to identify neurons with fields far from the start of the reward zone (>50 cm) that shifted their place fields to match the shift in reward location. To visualize this form of remapping, we circularly shifted the spatial activity of cells on trials following the switch to align the reward locations (Fig. 2B). This approach revealed that a subset of cells remapped to a similar relative distance from the reward, both within and across environments (“reward-relative” cells; see below and Methods for full criteria) (Fig. 2B). This field alignment relative to reward was observed even for fields that, from the perspective of the linear track position, remapped from the latter half of the track to the beginning, or vice versa (e.g. Fig. 2B cells m4.109, m14.482). Such alignment occurred despite the variable length of the teleport zone, suggesting that these cells do not simply track distance run from the last reward. Consistent with this interpretation, the distance run in the teleport zone did not predict the cells’ spatial firing variability on the following lap in the vast majority (~92%) of reward-relative cells (Fig. S2EF). Instead, reward-relative cells seemed to maintain their preferred firing position with respect to reward according to where the animal was in the progression of the overall task, which spans across the teleport and the trial start but is informed by these boundaries. Altogether, these observations suggest that a reward-relative subpopulation maps a periodic axis of experience from reward to reward, consistent with the cyclical manner in which the animal repeats the task structure. At the same time, other cells seemed to remap randomly, as circular shifting did not align their fields (“non-reward-relative remapping”, 12.3% ± 3.1% of all place cells; Fig. 2B, black cell ID labels).

To assess whether reward-relative remapping occurred at greater than chance levels, we plotted the peak spatial firing of all cells that maintained significant spatial information before versus after the switch. This revealed the track-relative place cells along the diagonal, as they maintained their peak firing positions pre- to post-switch. In addition, this analysis revealed an off-diagonal band of cells at the intersection of the reward locations, which extended linearly away from the cluster of reward cells that has been previously described54 (Fig. 2C, far left panel; Fig. S2B, middle). As noted above, we reasoned that the animals may experience the task as periodic instead of a sequence of discrete linear episodes. We therefore transformed the linear track coordinates to the phase of a periodic variable (i.e. 0 to 450 cm becomes -π to π radians), allowing us to shift the data to align the reward zones at zero and isolate the band of cells with peaks at this intersection (Fig. 2C, middle-left panel). Cells which fall along the unity line in this analysis thus putatively remap to the same relative distance from reward. We then excluded the track-relative cells to focus on the remapping cells across animals (Fig. 2C, middle-right panel). We measured the circular distance of each cell’s peak firing position relative to the reward zone pre- or post-switch (Fig. 2D) and calculated the difference between each of these relative peaks. A difference between relative peak positions close to zero indicates reward-relative remapping. We compared the distribution of differences to a “random-remapping” shuffle (Fig. 2C, far right panel) and found that the fraction of place cells that exhibited reward-relative remapping across animals exceeded chance on the first switch day (Fig. 2E, top). Further, the fraction of cells that exceeded the chance level increased by the last switch day (Fig. 2FG, Fig. S2G) and significantly increased linearly across task experience with each switch (Fig. 2H). A linear mixed effects model accounting for mouse identity revealed a similar, significant increase in reward-relative remapping over experience (Fig. S2H). Across days and animals, the fraction of cells exceeding chance spanned a mean range of ~0.65 radians around the unity line, or nearly 50 cm, suggesting there is some variability in the precision of reward-relative remapping (Fig. 2E, G, Fig. S2G). This distance thus provided a candidate region to quantitatively identify reward-relative remapping cells. We next analyzed the distribution of reward-relative peak firing positions within this candidate zone (i.e. position along the unity line in Fig. 2D) and observed that the mean of the distribution was significantly greater than zero (Fig. 2E, bottom). This indicates that more reward-relative place fields are in locations following reward delivery (as indicated by maximal licking, Fig. 1FG, Fig. S1AC) than in locations preceding the reward. This post-reward shift of the distribution of cells was consistent across days (Fig. 2G, bottom).

Next, to investigate the degree to which cells active in close proximity to the reward contribute to the above-chance levels of reward-relative remapping, we excluded cells with peaks within ± 50 cm of the start of both reward zones. Note that this excludes both the reward and anticipatory zones. With this exclusion, we still found above-chance reward-relative remapping (Fig. S2IL). In addition, the mean of the remaining reward-relative firing positions continued to follow the reward zone (Fig. S2J, L, bottom). However, there was not a significant increase in remapping at these distances across days (Fig. S2M), suggesting there is more growth in the population of reward-relative cells at closer proximities to reward. Nevertheless, this analysis confirmed that reward-relative remapping is not limited to neurons with close proximity to reward.

Finally, we implemented two main criteria to identify a robust subpopulation of reward-relative cells for further analysis. First, we identified candidate place cells with reward-relative spatial peaks that were within 50 cm of each other before versus after a reward switch (i.e. within the orange candidate zone highlighted in Fig. 2D, F). This criterion included cells with peaks both very close to and potentially very far from the reward. Second, to reduce the influence of noise in spatial peak detection and take into account the shape of the firing fields, we performed a cross-correlation of the cells’ activity aligned to the reward zone pre- vs. post- switch. We required the offset at the maximum of this cross-correlation to be ≤50 cm from zero lag and exceed chance levels relative to a shuffle for each cell (Fig. S2NP). This approach provides a metric for how stable each cell’s mean firing position is relative to reward. We confirmed that these criteria identified a subpopulation of cells whose fields were maximally correlated in periodic coordinates relative to reward, in contrast to the track-relative cells which were maximally correlated relative to the original linear track coordinates (Fig. S2OP). From here on, we refer to cells that passed both of these criteria as “reward-relative” (RR) (15.4% ± 5.3% of place cells averaged over all switch days; 12.4% ± 3.3% on the first switch, 18.9% ± 6.5% on the last switch).

Reward-relative sequences are preserved when the reward moves

Next, to ask whether the reward-relative cells constructed sequences of activity anchored to reward, we sorted the reward-relative cells within each animal by their peak firing position on trials before the reward switch, using split-halves cross-validation (see Methods). We applied this sort to the post-switch trials and found that the reward-relative subpopulation indeed constructed sequences that strongly preserved their firing order relative to reward across the switch within an environment (Fig. 3AB, Fig. S3A). Sequence preservation was quantified as a circular-circular correlation coefficient between the peak firing positions in the pre- and post-switch sequences, which was much higher than the shuffle of cell identities for nearly all animals and days (48 out of 49 sessions positively correlated, with rho >95% of the shuffle, two-sided permutation test) (Fig. 3C). These sequences spanned the task structure from reward to reward, such that cells firing at the furthest distances from reward would occasionally “wrap around” from the beginning to the end of the linear track or vice versa (e.g. Fig. 3A, mouse 4 and Fig. S3A). When we introduced the novel environment on day 8, despite apparent global remapping in the total population of place cells (Fig. S3B), reward-relative sequences were clearly preserved (Fig. S3C, top). The track-relative place cells likewise constructed robust sequences (48 out of 49 sessions positively correlated, with rho >95% of the shuffle, two-sided permutation test) both within (Fig. 3DF) and even across environments (Fig. S3C, middle) in a minority of place cells (20.2% ± 8.8% of place cells within vs. 8.4% ± 3.6% across, mean ± s.d. across 7 mice). The sequence stability of both the reward-relative and track-relative subpopulations contrasted starkly with the absence of sequence preservation for most of the non-reward-relative remapping cells (40 out of 49 sessions did not exceed the shuffle; Fig. 3GI; Fig. S3C, bottom) and for the disappearing and appearing cells (Fig. S3DE), as expected.

Figure 3. Preserved sequences relative to reward and space.

Figure 3.

A) Preserved firing order of reward-relative sequences across the 1st switch (day 3) (2 example mice, top and bottom rows; m4 is the same mouse shown in Fig. 2A, m12 had the maximum number of cells). Reward-relative remapping cells within each animal are sorted by their cross-validated peak activity before the switch, and this sort is applied to activity after the switch. Data are plotted as the spatially-binned, unsmoothed dF/F normalized to each cell’s mean within session, from which peak firing positions are identified in linear track coordinates that have been converted to radians (see Fig. 2C). White dotted lines indicate starts and ends of each reward zone. Each sequence’s circular-circular correlation coefficient (rho), p-value relative to shuffle, and number of cells is shown above, where p < 0.001 indicates that the observed rho exceeded all 1000 shuffles. Color scale is applied in (B), (D-E), (G-H).

B) Reward-relative sequences on the last switch (day 14), for the same mice shown in (A).

C) Circular-circular correlation coefficients for reward-relative sequences before vs. after the reward switch, for each mouse (n=7), with the upper 95th percentile of 1000 shuffles of cell IDs marked as “x”. Closed circles indicate coefficients with p < 0.05 compared to shuffle using a two-sided permutation test; open circles indicate p ≥ 0.05. D) Sequences of track-relative place cells on the 1st switch day for the same mice as in (A-B).

E) Sequences of track-relative place cells on the last switch day for the same mice as in (A-B,D).

F) Circular-circular correlation coefficients for track-relative place cell sequences (n=7 mice), colored as in (C).

G-H) Sequences of non-reward-relative remapping cells with significant spatial information before and after the switch, for the same mice as above and on the 1st (G) and last (H) switch day. Note that a cross-validated sort of the “before” trials produces a sequence, but in general this sequence is not robustly maintained across the switch. I) Circular-circular correlation coefficients for non-reward-relative remapping cells (n=7 mice), colored as in (C).

J) Distribution of peak firing positions for the reward-relative sequences after the switch on each switch day, shown as a fraction of total place cells within each animal, averaged across animals (n=7 mice). Standard errors are omitted for visualization clarity. Horizontal dotted lines indicate the expected uniform distribution for each day (see Methods); vertical gray dashed lines indicate the reward zone start and surrounding ±50 cm.

K-L) Quantification of changes in reward-relative sequence shape across animals and days, using linear mixed effects models with switch index as a fixed effect and mice as random effects (n=7 mice). β is the coefficient for switch index, shown with corresponding p-value (Wald test). Each point is the fraction of place cells within that mouse (colored as in C) in the reward-relative sequences, specifically within ±50 cm (K) or >50 cm (L) from the start of the reward zone.

M) Mean lick rate across post-switch trials and across animals (10 cm bins), colored by switch as in (J).

N-O) Quantification of change in lick rate across days within ±50 cm of the reward zone start (N) or >50 cm (O), using a linear mixed effects model as in (K-L).

P) Mean running speed across post-switch trials and across animals (2 cm bins), colored by switch as in (J).

Q-R) Quantification of change in speed across days, as in (K-L) and (N-O).

S) Distribution of peak firing positions for the track-relative sequences after the switch on each switch day, shown on the track coordinates converted to radians (not reward-relative). Arrows indicate the start position of each reward zone (“A”, “B”, “C”). Horizontal dotted lines indicate the expected uniform distribution for each day.

T-U) Quantification of changes in track-relative sequence shape across animals and days as in (K-L), here within ≤50 cm of either end of the track (T) or >50 cm from the track ends (U). See also Figure S3.

We next quantified the mean sequence shape for each subpopulation, either relative to reward or to the VR track coordinates. The reward-relative sequences strongly over-represented the reward zone1,2,54,55,6366 and were most concentrated just following the start of the reward zone (Fig. 3J). Further, this density near the reward zone grew across task experience (Fig. 3JK). The fraction of place cells at further positions (>50 cm) from the start of the reward zone showed a modest increase over days, but to a much lesser degree than positions near the reward (Fig. 3L). In parallel, mice increased their average post-switch lick rate prior to the reward zone (Fig. 3MO) and their running speed away from the reward zone over days (Fig. 3PR). These changes suggest that reward-relative sequence strength increases as behavioral performance becomes more robust. In addition, on any given day, the precision of the animal’s licking correlated with the precision of the reward-relative sequence (i.e. the narrowness of the cell distribution around reward, measured as the circular variance) (Fig. S3FI). In contrast to the reward-relative sequences, the track-relative place cell sequences tended to over-represent the ends of the track, as opposed to the reward zones (Fig. 3S). In addition, the track-relative sequences showed a subtle but significant drop in density at positions away from the ends of the track across days, consistent with selective stabilization of place fields near key landmarks8,55,7680,86 (Fig. 3SU). These differences in sequence shape between the reward-relative and track-relative cells suggest that each of these subpopulations anchors to the most salient point in the “space” they encode. For the reward-relative sequences, this is the reward zone, and for the track-relative sequences, this is likely the start and end of the track where the animal exits and enters the teleport zone. By comparison, the disappearing, appearing, and non-reward-relative remapping cells encoded a more uniform representation of the track (Fig. S3JK), with only subtle over-representation of the track ends and an observable decrease across the track for the disappearing cells across days (Fig. S3JK).

Non-track-relative place cells are increasingly recruited to the reward-relative population over days

The growth in reward-relative sequence density across task experience raised three possibilities: (1) reward-relative cells may become more consistent at remapping to precise reward-relative positions with extended learning, (2) more cells might be recruited from other populations to the reward-relative population, or (3) a combination of both changes occurs. To investigate these possibilities and understand how individual neurons are dynamically reallocated to represent updated reward information within and across environments, we followed the activity of the same neurons across days by matching the aligned regions of interest (ROIs) for each cell. Specifically, we followed single neurons across pairs of switch days and analyzed their spatial firing patterns from one switch to the next (Fig. 4). Using linear mixed effects models to account for variance across animals, we found that there was an increase in the fraction of cells that remained reward-relative on consecutive reward switches, indicating that reward-relative cells are more likely to maintain their reward-relative coding across experience (Fig. 4AB). In addition, an increasing proportion of non-track-relative cells (appearing, disappearing, and non-reward-relative remapping) were recruited into the reward-relative population with task experience (Fig. 4CD). Non-place cells and track-relative cells likewise were recruited into the reward-relative population, though this occurred at a constant rate of turnover that did not increase over days (Fig. 4EH). In addition, the track-relative cell population did not recruit more cells from any other population (Fig. 4IK) and did not show an increased likelihood of remaining track-relative across experience (Fig. 4L), in contrast to the reward-relative population. These results suggest that the track-relative place cells remain a more independent population from the reward-relative cells, and that cells exhibiting high flexibility (such as appearing/disappearing) are more likely to be recruited into the reward-relative representation with extended experience in this task.

Figure 4. Tracking cells over days reveals increased recruitment into the reward-relative population.

Figure 4.

A) Example cells from 2 separate mice that were tracked from switch 6 to 7 (days 12–14) that were identified as reward-relative (RR) on both switch 6 and switch 7, and which maintained their firing position relative to reward. Top of each cell: spatial firing over trials and linear track position on each day (deconvolved activity normalized to the mean of each cell). White vertical lines indicate the beginnings of the reward zones. Bottom of each cell: Mean local field of view (FOV) around the ROI of the cell on both days (blue shading highlights ROI, indicated by the yellow arrow).

B) Fraction of tracked RR cells on each switch day that remain RR on subsequent switch days increases over task experience. All β coefficients and p-values in (B, D, F, H, and I-L) are from linear mixed effects models with switch day pair as a fixed effect and mice as random effects (n=7 mice in each plot, key in H). Coefficients correspond to the fractional increase in recruitment per switch day pair, from the model fit (gray line).

C) Same as (A) but for 2 example cells that were identified as appearing (top) or non-RR remapping (bottom) on switch 6 but were converted to RR on switch 7.

D) Fraction of “non-track-relative” place cells (combined appearing, disappearing, and non-RR remapping) that convert into RR cells on subsequent switch days increases over task experience.

E) Same as (A) but for 2 example non-place cells on switch 6 (did not have significant spatial information in either trial set) that were converted to RR place cells on day 7.

F) Fraction of non-place cells converting to RR cells on subsequent switch days shows a modest increase over task experience.

G) Same as (A) but for 2 example track-relative (TR) cells on switch 6 that were converted to RR on switch 7.

H) Fraction of TR cells converting to RR cells does not increase over task experience.

I) Fraction of RR cells converting to TR cells does not change significantly over task experience.

J) Fraction of non-TR cells (combined appearing, disappearing, and non-RR remapping) converting to TR does not change significantly.

K) Fraction of non-place cells converting to TR does not change significantly.

L) Fraction of TR cells remaining TR does not change significantly.

See also Figure S4.

The track-relative population itself showed a steady rate of turnover over days (Fig. 4L), consistent with representational drift8185. Indeed, when we analyzed place cell sequences identified on each reference day and followed two days later, we observed drift even in the fixed-condition animals (n=3 mice) that experienced a single environment and reward location for all 14 days (Fig. S4A, F). Drift was prevalent as well in the reward-relative and track-relative sequences tracked across consecutive switch days (Fig. S4BE, GJ), at a slightly higher degree than the fixed-condition animals (Fig. S4F), consistent with our observation that moving a reward induces more remapping than not moving a reward (Fig. 2A, Fig. S2AD). In a subset of animals and day pairs, cells that remained reward-relative or track-relative tended to maintain their firing order across days. However, sequence preservation across days was generally lower (i.e. closer to the shuffle) (Fig. S4C, E, H, J) than within-day sequences (Fig. 3C, F; Fig. S3A, C). These results indicate that rather than being dedicated cell classes, reward-relative and track-relative ensembles are both subject to the same plasticity across days of experience as the overall hippocampal population8185. However, the increasing recruitment of cells into the reward-relative population, despite this population having different membership on different days, suggests an increased allocation of resources to the reward-relative representation. We conclude that as animals become more familiar with the task demands—specifically the requirement to estimate a hidden reward location that can be dissociated from spatial stimuli—the hippocampus builds an increasingly dense representation of experience relative to reward, while preserving a fixed representation of the spatial environment in a largely separate population of cells on any given day.

Encoding of reward proximity versus movement covariates in reward-relative cells

The animals’ approach to and departure from the reward zone in this task is associated with stereotyped running and licking behaviors that comprise an important aspect of the animal’s experience surrounding reward. For instance, these behaviors must be timed with an accurate estimate of distance from reward to inform behavioral adaptations after the reward switch. We therefore took two complementary approaches, detailed below, to disentangle which features of the reward-relative experience are encoded in the reward-relative subpopulation. These approaches further allowed us to control for the potential influence of movement covariates on the neural activity.

First, we tested the hypothesis that reward-relative firing following the reward could be locked to the running speed profile of the animal, as opposed to the location where the animal expected to receive reward. For this, we leveraged the ~15% omission trials to compare the activity of reward-relative neurons on rewarded versus unrewarded trials. To consider the activity of reward-relative cells under conditions where running speed was generally comparable between rewarded and omission trials, we restricted our analysis to: reward-relative cells with peaks following the start of the reward zone, sessions in which the reward zone began at location “A” (80 cm) or “B” (200 cm), and trials before a reward switch. To then align the running speed profiles across rewarded and omission trials, we fit a time warping model87 on the running speed data and applied this time warped transformation to the neural data (Fig. S5AD, see Methods). We then computed a “reward versus omission index” (RO index) to quantify the difference in each cell’s firing rate following rewards versus omissions (Fig. S5E).

Reward-relative cells exhibited heterogeneity in their firing rates following rewards versus omissions (Fig. 5AC). A subset of cells fired at the same position relative to reward regardless of the animal’s running speed (Fig. 5A, top and middle row of cells, Fig. 5B, middle, Fig. 5C, top and middle), with firing rates often differing between rewarded and omission trials (e.g. Fig. 5A, top, Fig. 5C, top and middle). Another subset of cells fired relative to a particular phase of the speed profile (Fig. 5A, bottom, Fig. 5B, top and bottom, Fig. 5C, bottom). However, at the population level, the distribution of RO indices for reward-relative cells was skewed toward reward-preferring (Fig. 5D). This reward-preference was also clearly observable at the single-cell and population levels in trials after the reward switch (Fig. S5FI), although running speed is also more variable on these trials (Fig. S5F). Finally, combining RO indices across days (Fig. S5J) revealed that, while there is heterogeneity among individual cells, there is a significant preference for the reward-relative population to fire more on rewarded rather than omission trials (Fig. 5E, Fig. S5K; note results were similar when all sessions were included, Fig. S5L).

Figure 5. Diversity in reward-relative cell coding on rewarded vs. omission trials, with a population preference for reward.

Figure 5.

A) Raw and time warp model-aligned speed (first row) and neural activity for 3 reward-relative example cells (second to fourth rows) on the first switch day, in mouse 14. Throughout Fig. 5, analysis is performed only on trial sets before the reward switch which have at least 3 omission trials. In (A-C), the left columns show raw (untransformed) data, with the start of the reward zone indicated by the gray dashed line. The right columns show data transformed by the time warp model fit to the trial-by-trial speed in that session (see Methods, Fig. S5). All included cells have peak firing following the start of the reward zone. First row: mean ± s.e.m. speed profiles for rewarded (black) and omission (magenta) trials. Note precise alignment of the transformed speed profiles on the right. Second to fourth rows: mean ± s.e.m. deconvolved activity for rewarded and omission trials, normalized to each cell’s mean within the session, showing an example reward-relative cell in each row. The difference in model-transformed activity between rewarded and omission trials is captured as a “reward vs. omission” index (bold text, top right of each panel; see Methods). The top row highlights neurons with much higher activity on rewarded trials. A positive index indicates higher activity on rewarded trials, a negative index indicates higher activity on omission trials.

B) Raw and time warp model-aligned speed and neural activity for 3 reward-relative example cells on the last switch day, in mouse 14. Note cell m14.474 is also shown in Fig. 2B.

C) Raw and time warp model-aligned speed and neural activity for 3 reward-relative example cells on the last switch day, in mouse 12. Note that speed is still aligned well with the reward zone at position “B” (at 220 cm) for m12 vs. position “A” (at 80 cm) for m14. Note cell m12.379 is also shown in Fig. 2B.

D) Reward vs. omission index for the population of reward-relative cells that have spatial firing peaks following the start of the reward zone, in the session shown in (A) (left), (B) (middle), and (C) (right). Each dot is a cell. *p<0.05, ***p<0.001, one-sample Wilcoxon signed-rank test against a 0 median. Switch 1, m14: W = 183, p = 0.018, n = 36 cells; Switch 7, m14: W = 721, p = 2.69e-5, n = 79 cells; Switch 7, m12: W = 2358, p = 3.70e-4, n = 122 cells.

E) Reward vs. omission index across all switch days where the reward zone was at “A” or “B”, for the population of reward-relative cells that have spatial firing peaks following the start of the reward zone within each mouse (n = 7 mice). Each dot is a cell, colors correspond to individual mice. Significance level is set at p<0.007 with Bonferroni correction. **p<0.007, ****p<0.0001. m3: median = 0.14, W = 1.50e3, p = 4.66e-7, n = 3 days, 114 cells; m4: median = 0.20, W = 1.03e3, p = 5.93e-10, n = 3 days, 112 cells; m7: median = 0.09, W = 4.44e3, p = 2.08e-3, n = 4 days, 157 cells; m11: median = 0.23, W = 194, p = 5.32e-5, n = 4 days, 48 cells; m12: median = 0.09, W = 3.41e4, p = 4.40e-8, n = 5 days, 441 cells; m13: median = 0.09, W = 523, p = 0.13, n = 3 days, 52 cells; m14: median = 0.11, W = 1.39e4, p = 3.74e-8, n = 4 days, 296 cells. See also Figure S5.

Second, we implemented a generalized linear model (GLM)88 to dissociate the contribution of task versus movement variables to the moment-to-moment deconvolved calcium activity of each cell. For task variables, we included linear track position, reward-relative position, and whether the animal was rewarded on each trial, convolved with the linear track position. For movement variables, we included running speed, acceleration, and licking (Fig. S6A, see Methods for details). We trained the GLM using 5-fold cross-validation and tested its prediction on the activity of each neuron in held out trials (Fig. 6AB, Fig. S6BC). Only well-fit neurons (fraction deviance explained > 0.15)88 were included in the analysis (35% of place cells across 7 mice and 7 switch days; Fig. S6BC). This included 61% (2377/3901) of track-relative cells, 37% (1157/3104) of reward-relative cells (Fig. 6AB), and 43% (1091/2518) of non-RR remapping cells.

Figure 6. A generalized linear model reveals reward-relative position as a leading predictor of neural activity in the reward-relative population.

Figure 6.

A) Example spatial firing maps of two reward-relative cells that were fit well (fraction deviance explained > 0.5) by the GLM, on the first switch day (top) and last switch day (bottom). White lines mark the starts of reward zones. Note that cell m14.77 (bottom) is the same cell as shown in Fig. 5B, bottom row.

B) Deconvolved calcium activity (black) and GLM prediction (red) of held-out test data for the two example cells shown in (A). Fraction Poisson deviance explained (FDE) of the test data is reported as a measure of model performance.

C) Relative contribution (“fraction explained deviance”, see Methods), or encoding strength, for each variable included in the GLM binned by track position, for the two example cells shown in (A). Task variables are “pos.”: linear track position; “rew.-rel. pos.”: reward-relative position; “rewarded”: whether the animal was rewarded as a function of position (i.e. a binary that stays high from the reward delivery time to the end of the trial). Movement variables are speed, acceleration, and licking (see Fig. S6A for implementation). Relative contribution is calculated from an ablation procedure where each variable is individually removed from the full model (see Methods). Gray dashed lines mark the starts of the two reward zones in each session, also indicated in (A). Boxes correspond to color coding in (E). Note that reward-relative position provides the maximum relative contribution for both example cells, including m14.77 even though it fired equivalently on rewarded and omission trials (see Fig. 5B).

D) Distributions of relative contribution of each variable in the track-relative, reward-relative, and non-reward-relative remapping subpopulations across animals. Boxes indicate the interquartile range, whiskers extend from the 2.5th to 97.5th percentile, horizontal line indicates the median, and notches indicate the confidence interval around the median computed from 10,000 bootstrap iterations. Colored dots mark the medians of each individual mouse.

E) Distributions of top predictor variables (maximum relative contribution) for individual cells within each subpopulation, shown as % of cells in the subpopulation; same n as in (D).

See also Figure S6.

After testing the full model, we performed an ablation procedure to measure the relative contribution of each predictor variable (see Methods). This ablation procedure revealed clear peaks in the contribution of reward-relative position to the activity of many reward-relative cells as a function of position along the track, recapitulating their firing fields (Fig. 6C). We then quantified the relative contribution of each variable within the track-relative, reward-relative, and non-RR remapping subpopulations that were well-fit by the full model (Fig. 6D). We found that linear track position was the strongest predictor for the track-relative population (Fig. 6D) and the top predictor for 81% of included track-relative cells (Fig. 6E). This result provided confirmation of our classification of these cells as stable place cells that remain locked to the spatial environment within a session. The non-RR remapping cells likewise were best predicted by linear track position followed by reward-relative position (Fig. 6E), consistent with their recruitment to the reward-relative population over experience (Fig. 4). In contrast, for the reward-relative population, reward-relative position provided the highest relative contribution (Fig. 6D) and the top predictor for 28.8% of reward-relative cells (Fig. 6E). Reward-relative position was followed by speed (top predictor for 23.4% of reward-relative cells), linear track position (17.8%), and the receipt of reward (13.3%), with minimal contributions of acceleration and licking. Within the reward-relative cells that exhibited reward-relative position or speed as their top predictors, we observed a mixed contribution of the other variables as well (Fig. S6DE). These results suggest that the reward-relative population supplies a heterogeneous code for multiple aspects of the reward-related experience. While movement covariates are challenging to fully disentangle from the animal’s progression through the task, the GLM, combined with our analyses of rewarded versus omission trials, provides evidence that reward and reward-relative position are strong predictors of activity in the reward-relative cell population.

DISCUSSION

Here, we reveal that the hippocampus encodes an animal’s physical location and its position relative to rewards through parallel sequences of neural activity. We employed 2-photon calcium imaging to observe large groups of CA1 neurons in the hippocampus as mice navigated virtual linear tracks with changing hidden reward locations. This paradigm allowed us to identify reward-relative sequences that are dissociable from the spatial visual stimuli, as they instead remain anchored to the movable reward location. This suggests that the brain employs a parallel coding system to simultaneously track an animal’s physical and reward-relative positions, pointing to a mechanism for how the brain could monitor an animal’s experience relative to multiple frames of reference45,59,60,6773,89. Further, we found that the reward-relative sequences, but not the “track-relative” sequences locked to the spatial environment, show a notable increase in the number of neurons involved as the animal gains more experience with the task. These results suggest that the hippocampus builds a strengthening representation of generalized events surrounding reward as animals become more familiar with the task structure. Moreover, this work raises the possibility that the hippocampal code for space, provided by a subpopulation of place cells, can update the progression of reward-relative sequences when the reward location changes. In this way, hippocampal ensembles encoding different aspects of experience may interact to support accurate memory while amplifying behaviorally salient features such as reward.

We validated prior work that found a subpopulation of hippocampal neurons precisely encodes reward locations54. Here, we found that this subset of reward-relative neurons, which fire very close to rewards, comprised a central component of a more extensive reward-relative sequence. Our finding of reward-relative coding at farther distances from reward is reminiscent of hippocampal cells that encode distances and directions to goals in freely flying bats90. At the same time, we extended these previous findings in two key ways: (1) we found that the reward-relative code spans the entire environment, involving cells firing at distances up to hundreds of centimeters away (i.e. as far as from one reward to the next), and (2) additional neurons are recruited into the reward-relative sequences over time, demonstrating a dynamic and evolving coding strategy. Specifically, the cells firing closest to reward were most likely to increase in number across days (Fig. 2DH, Fig. S2I-M, Fig. 3JL), suggesting that they may be the most detectable in other paradigms. Importantly, we also observed a noticeable turnover in the neurons that form the reward-relative population from day to day, indicating a high degree of reorganization in terms of which neurons participate. Despite this daily fluctuation in individual neurons, the overall sequential structure appears to be maintained within the population.

This latter finding of day to day fluctuation in the reward-relative population is consistent with recent studies which have highlighted a phenomenon known as representational drift8185. In the hippocampus8185 and in other cortical areas9193, this drift is characterized by changing spatial tuning and stability of single neurons as a function of extended experience82,83. Despite these single neuron changes, the population often continues to accurately code for the spatial environment81 or task91. This flexibility is thought to increase the storage capacity of neuronal networks9497. Related changes over time have also been proposed to support the decoding of temporal separation between unique events50,92,94,98, which could be particularly important for separating memories of events surrounding different rewarding experiences. We observed population drift in both the reward-relative and track-relative sequence membership. This frequent reorganization suggests that these are not dedicated subpopulations; neurons in the reward-relative population occasionally switched coding modes—though this flexibility may vary by anatomical location within CA154,99,100. At the same time, we found that more resources (i.e. neurons) are allocated to the reward-relative representation at the population level with extended experience. This finding indicates an increasingly robust network code that can be reliably anchored to reward within a day despite day-to-day drift. This stability amidst flux offers insight into the brain’s ability to balance dynamic coding with the preservation of population-level representations91,92,97.

In the hippocampus, spatial firing can also be highly flexible at short timescales even within a fixed environment. This flexibility is evident in the variability of place cell firing even within a field57,85 and especially when an animal takes different routes101106 or targets a different goal89,90,107109 through the same physical position. Our work is consistent with such trajectory specific coding and raises the possibility that the reward-relative population may split even further by specific destinations or reward sites when there are multiple possible reward locations. Indeed, in cases where multiple rewards are present at once, hippocampal activity may multiplex spatial and reward-relative information to dissociate different goal locations64,89,110. This discrimination is enhanced over learning110, consistent with the manner in which hippocampal maps are likewise dependent on learned experience4,6,7,12,52,53,111.

Furthermore, place cells undergo significant remapping in response to changes in the animal’s motivational state, such as during task engagement61,62,70 or disengagement112, or upon removal of a reward113. Importantly, in our task the animal must remain engaged when the reward zone updates in order to accurately recall and seek out the new reward location. Consistent with this enhanced attention58,59,61,62, we see predictable structure in how the neurons remap surrounding reward (i.e. they maintain their firing order relative to reward), in contrast to the more random reorganization observed with task disengagement112 or the absence of reward113. However, an important consideration for all of these short timescale firing differences, and for representational drift, is the contribution of behavioral variability to neural activity changes114,115. For example, small changes in path shape116 and the movement dynamics involved in path integration117121 may influence aspects of hippocampal activity. Our results from a generalized linear model indicate that such movement dynamics influence but do not dominate the activity in the reward-relative population, as reward-relative position was a leading predictor of activity. Further work will be required to disentangle which aspects of reward-seeking behavior are encoded in the hippocampus.

The origin of the input that informs reward-relative neurons remains an open question. One possibility is that subpopulations of neurons in CA1 receive stronger neuromodulatory inputs than others1, increasing their plasticity in response to reward-related information122,123. CA1 receives reward-related dopaminergic input from both the locus coeruleus (LC)124,125 and the ventral tegmental area (VTA)113,126,127 that could potentially drive plasticity to recruit neurons into reward-relative sequences. Both of these inputs have recently been shown to restructure hippocampal population activity, with both LC and VTA input contributing to the hippocampal reward over-representation113,125,126, and VTA input additionally stabilizing place fields113 in a manner consistent with reward expectation113,128. Alternatively, it is possible that upstream cortical regions, such as the prefrontal cortex129, contribute to shaping these reward-relative sequences. Supporting this idea, the prefrontal cortex has been observed to exhibit comparable neural activity sequences that generalize across multiple reward-directed trajectories130134 to encode the animal’s progression towards a goal135,136. Another potential source of input is the medial entorhinal cortex, which exhibits changes in both firing rate and the location where cells are active in response to hidden reward locations137,138. In particular, layer III of the medial entorhinal cortex is thought to contribute significantly to the over-representation of rewards by the hippocampus63. Additionally, the lateral entorhinal cortex, which also projects to the hippocampus, exhibits a strong over-representation of positions and cues just preceding and following rewards120,139141 and provides critical input for learning about updated reward locations139,140.

These insights point to a complex network of brain regions potentially interacting to inform reward-relative sequences in the hippocampus. In turn, the hippocampal-entorhinal network may send information about distance from expected rewards back to the prefrontal cortex and other regions where similar activity has been observed1, such as the nucleus accumbens142147. Such goal-related hippocampal outputs to prefrontal cortex148 and nucleus accumbens149 have recently been reported, specifically in the deep sublayer of CA1148, where goal-related remapping is generally more prevalent148,150. Over learning, reward-relative sequences generated between the entorhinal cortices and hippocampus may become linked with codes for goal-related action sequences in the prefrontal cortex and striatum, helping the brain to generalize learned information surrounding reward to similar experiences1. Understanding these interactions in future work will provide new insight regarding how the brain integrates spatial and reward information, which is crucial for navigation and decision-making processes.

The significance of reward-relative sequences in the brain may lie in their connection to hidden state estimation7,43,151155, a concept closely related to generalization28,156. Hidden state estimation refers to the process by which an animal or agent infers an unobservable state, such as the probability of receiving a reward under a certain set of conditions, in order to choose an optimal behavioral strategy151. For example, in our study involving mice, reward-relative sequences could be instrumental in helping the animals discern different reward states, such as whether they are in a state of the task associated with reward A, B, or C. By building a code during learning that allows the hippocampus to accurately infer such states, this inference process can generalize to similar future scenarios in which the animal must make predictions about which states lead to reward151,152. Interestingly, however, the density of the reward-relative sequences we observed was highest following the beginning of the zone where animals were most likely to get reward, rather than preceding it. This finding contrasts with computational simulations of predictive codes in the hippocampus that are skewed preceding goals157. Instead, the shape of the reward-relative sequences may be more consistent with anchoring to the animal’s estimate of the reward zone, with decreasing density around the zone as a function of distance which mirrors the organization seen around other salient landmarks86,158.

Our work is consistent with previous proposals that suggest a primary function of the hippocampus is to generate sequential activity at a baseline state17,27,29,47,48. These sequences could then act as a substrate that can become anchored to particular salient reference points16,17,27,47, allowing maximal flexibility to learn new information as it becomes relevant and layer it onto existing representations159161. This role for hippocampal sequences extends to computational studies28,159,162 and human research163167, where such coding schemes have been linked to the organization of knowledge and the chronological order of events in episodic memory27,49,163166,168170. The ability of the hippocampus to parallelize coding schemes for different features of experience may help both build complete memories and generalize knowledge about specific features without interfering with the original memory. As observed in our study, the hippocampus may then amplify the representations of some of these features over others according to task demands. A similar effect has been observed in humans, in that memory recall is particularly strong for information that either precedes or follows a reward171,172. This suggests that rewards can significantly enhance the strength and precision of memory storage and retrieval, especially for events closely associated with these rewards. Thus, reward-relative sequences may play a crucial role in how we form and recall our experiences, with implications stretching beyond spatial navigation.

METHODS

Subjects

All procedures were approved by the Institutional Animal Care and Use Committee at Stanford University School of Medicine. Male and female (n = 5 male, 5 female) C57BL/6J mice were acquired from Jackson Laboratory. Mice were housed in a transparent cage (Innovive) in groups of five same-sex littermates prior to surgery, with access to an in-cage running wheel for at least 4 weeks. After surgical implantation, mice were housed in groups of 1–3 same-sex littermates, with all mice per cage implanted. All mice were kept on a 12-h light–dark schedule, with experiments conducted during the light phase. Mice were between ~2.5 to 4.5 months at the time of surgery (weighing 18–31 g). Before surgery, animals had ad libitum access to food and water, and ad libitum access to food throughout the experiment. Mice were excluded from the study if they failed to perform the behavioral pre-training task described below.

Surgery: calcium indicator virus injections and imaging window implants

We adapted previously established procedures for 2-photon imaging of CA1 pyramidal cells7,173. The imaging cannula was constructed using a ~1.3-mm long stainless-steel cannula (3 mm outer diameter, McMaster) affixed to a circular cover glass (Warner Instruments, number 0 cover glass, 3 mm in diameter; Norland Optics, number 81 adhesive). Any excess glass extending beyond the cannula edge was removed with a diamond-tipped file. During the imaging cannula implantation procedure, animals were anesthetized through intraperitoneal injection of ketamine (~85 mg/kg) and xylazine (~8.5 mg/kg). Anesthesia was maintained during the procedure with inhaled 0.5–1.5% isoflurane and oxygen at a flow rate of 0.8–1 L/min using a standard isoflurane vaporizer. Prior to surgery, animals received a subcutaneous administration of ~2 mg/kg dexamethasone and 5–10 mg/kg Rimadyl (to reduce inflammation and promote analgesia, respectively). First, the viral injection site targeting the left dorsal CA1 was determined by stereotaxis (AP −1.94 mm, ML −1.10 to −1.30 mm) and an initial hole was drilled only deep enough to expose the dura. An automated Hamilton syringe microinjector (World Precisions Instruments) was used to lower the syringe (35-gauge needle) to the target depth at the CA1 pyramidal layer (DV −1.33 to −1.37 mm) and inject 500 nL adenovirus (AAV) at 50 nL/min, to express the genetically encoded calcium indicator GCaMP under the pan-neuronal synapsin promoter (AAV1-Syn-jGCaMP7f-WPRE, AddGene, viral prep 104488-AAV1, titer 2×1012). The needle was left in place for 10 min to allow for virus diffusion.

After retracting the needle, a 3 mm diameter circular craniotomy was then performed over the left hippocampus using a robotic surgery drill for precision (Neurostar). The craniotomy was centered at AP −1.95 mm, ML −1.8 to −2.1 mm (avoiding the midline suture). During drilling, the skull was kept moist by applying cold sterile artificial cerebrospinal fluid (ACSF; sterilized using a vacuum filter). The dura was then delicately removed using a bent 30-gauge needle. To access CA1, the cortex overlying hippocampus was carefully aspirated using a blunt aspiration needle, with continuous irrigation of ice-cold, sterile ACSF. Aspiration was stopped when the fibers of the external capsule were clearly visible, leaving the external capsule intact. Following hemostasis, the imaging cannula was gently lowered into the craniotomy until the cover glass lightly contacted the fibers of the external capsule. To optimize an imaging plane tangential to the CA1 pyramidal layer while minimizing structural distortion, the cannula was positioned at an approximate 10° roll angle relative to horizontal. The cannula was affixed in place using cyanoacrylate adhesive. A thin layer of adhesive was also applied to the exposed skull surface, which was pre-scored with a number 11 scalpel before the craniotomy to provide increased surface area for adhesive binding. A headplate featuring a left-offset 7-mm diameter beveled window and lateral screw holes for attachment to the imaging rig was positioned over the imaging cannula at a matching 10° angle. The headplate was cemented to the skull using Metabond dental acrylic dyed black with India ink or black acrylic powder.

Upon completion of the procedure, animals were administered 1 mL of saline and 10 mg/kg of Baytril, then placed on a warming blanket for recovery. Monitoring continued for the next several days, and additional Rimadyl and Baytril were administered if signs of discomfort or infection appeared. A minimum recovery period of 10 days was required before initiation of water restriction, head-fixation, and virtual reality training.

Histology

After the conclusion of all experiments, mice were deeply anesthetized and administered an overdose of Euthasol, then perfused transcardially with PBS followed by 4% paraformaldehyde (PFA) in 0.1M PB. Brains were removed and post-fixed in PFA for 24 hours, followed by incubation in 30% sucrose in PBS for >4 days. 50 μm coronal sections were cut on a cryostat, mounted on gelatin-coated slides, and coverslipped with DAPI mounting medium (Vectashield). Histological images were taken on a Zeiss widefield fluorescence microscope.

Virtual Reality (VR) Design

All VR tasks were designed and operated using custom code written for the Unity game engine (https://unity.com). Virtual environments were displayed on three 24-inch LCD monitors surrounding the mouse at 90° angles relative to each other. The VR behavior system included a rotating fixed axis cylinder to serve as the animal’s treadmill and a rotary encoder (Yumo) to read axis rotations to record the animal’s running. A capacitive lick port, consisting of a feeding tube (Kent Scientific) wired to a capacitive sensor, detected licks and delivered sucrose water reward via a gravity solenoid valve (Cole Palmer). Two separate Arduino Uno microcontrollers operated the rotary encoder and lick detection system. Unity was controlled on a separate computer from the calcium imaging acquisition computer. Behavioral data were sampled at 50 Hz, matching the VR frame rate. Both the start of the VR task as well as each 50 Hz frame were synchronized with the ~15.5 Hz sampled imaging data via Unity-generated TTL pulses from an Arduino to the imaging computer.

Behavioral training and VR tasks

Handling and pre-training

After ~1 week of recovery, the mice were handled for 2–3 days for 10 minutes each day, then acclimated to head-fixation (at least 10 days after surgery) on the cylindrical treadmill in a stepwise fashion for at minimum 15-min sessions for 2–3 days (increasing to 30 min-1 hour on the second to third day). Mice were then acclimated to the lick port by 1–2 days of pre-training to lick for water rewards delivered at 2-second intervals (i.e. a minimum of 2 s between rewards if mice were actively licking, otherwise no reward was delivered). For this pre-training and all subsequent behavior, sucrose water reward (5% sucrose w/v) was delivered when the mouse licked the capacitive lick port. The water delivery system was calibrated to release ~4 μL of liquid per drop. To motivate behavior, mice were water restricted to 85% or higher of their baseline body weight. Mice were weighed daily to monitor health and underwent hydration assessments (via skin tenting). The total volume of water consumed during the training and behavioral sessions was measured, and after the experiment each day, supplemental water was supplied up to a total of ~0.045 mL/gm per day (typically 0.8–1 mL per day, adjusted to maintain body weight). After the training and water consumption, the animals were returned to their home cage each day.

Once acclimated to the lick port, mice were pre-trained on a “running” task on a 350 cm virtual linear track with random black and white checkerboard walls and a white floor to collect a cued reward. The reward location was marked by two large gray towers, initially positioned at 50 cm down the track. If the mouse licked within 25 cm of the towers, it received a liquid reward; otherwise, an automatic reward was given at the towers. After dispensing the reward, the towers advanced (disappeared and quickly reappeared further down the track). Covering the distance to the current reward in under 20 s increased the inter-reward distance by 10 cm, but if the mouse took longer than 30 s, the distance decreased by 10 cm, with a maximum reward location of 300 cm. Upon consistent completion of 300 cm distance in under 20 s, the automatic reward was turned off, requiring the mice to lick to receive reward. At the end of the track, the mice passed through a black “curtain” and into a gray teleport zone (50 cm long) that was equiluminant with the VR environment before re-entering the track from the beginning. Once mice were reliably licking and completing 200 laps of the pre-training track within a 30–40 min period, they were advanced to the main task and imaging (mean ± s.d.: 6.1 ± 1.8 days of running pre-training across 10 mice).

Hidden reward zone “switch” VR task

The main “switch” task reported in this study involved two virtual environments highly similar to those previously used to study hippocampal remapping7, each with visually distinct features from the pre-training environment. Both environments consisted of a 450 cm linear track, with two colored towers and two patterned towers along the walls. Environment 1 (Env 1) had diagonal low-frequency black and white gratings on the walls, a gray floor and dark gray sky, and began with two green towers. Environment 2 (Env 2) had higher frequency gratings on the wall, a gray floor, a very light gray sky, and began with two blue towers. The reward zone was a “hidden” (not explicitly marked) 50-cm span at one of 3 possible locations along the track, each equidistantly spaced between the towers but with different proximity to the start or end of the track: zone A: 80–130 cm, zone B: 200–250 cm, zone C: 320–370 cm. Only one reward zone was ever active at a time. On the first 10 trials of any new condition, including the first day and each switch subsequently described, the reward was automatically delivered at the end of the zone if the mouse had not yet licked within the zone to signal reward availability. Otherwise, reward was delivered at the first location within the zone where the mouse licked. After these 10 trials, reward was operantly delivered for licking within the zone. Reward was randomly omitted on approximately 15% of trials (i.e. if a random number generator exceeded 0.85). Each lap terminated in a black curtain and gray teleportation “tunnel” to return the mouse to the beginning of the track. Time in the teleport tunnel was randomly jittered between 1 and 5s (5–10s on trials following reward omissions or trials where the mouse did not lick in the reward zone), followed by 50 cm of running during which the beginning of the track was visible to provide smooth optic flow.

Each mouse encountered a different starting reward zone and sequence of reward zone switches, counterbalanced across mice (n = 7 mice). Mice were allowed to learn an initial reward zone (e.g. A as in Fig. 1E) for days 1–2 of the task. On day 3 (Switch 1), the zone was moved to one of the two other possible locations on the track after 30 trials (e.g. B). For the first 10 trials at the new location, automatic delivery was activated at stated above, but empirically, we observed that the mice often started licking at the new location before that 10 trials had elapsed (Fig. S1AC). The new reward zone was maintained on day 4. On day 5 (Switch 2), the zone was moved to the third possible reward zone (e.g. C), maintained on day 6, and moved back to the original location on day 7 (e.g. A). Each switch occurred after 30 trials. On day 8, the reward zone switch coincided with a switch into Env 2, where the sequence of zone switches was then reversed on the same day-to-day schedule for a total of 14 days (Fig. 1, Fig. S1).

An additional “fixed-condition” cohort (n=3 mice) experienced only Env 1 and one fixed reward location throughout the 14 days (Fig. S1D,F).

We targeted 80–100 trials per session with simultaneous two-photon imaging (described below), with 1 imaging session per day (mean ± s.d: 80.9 ± 8.8 trials across 10 mice). The session was terminated early if the mouse ceased licking for an extended period of time or ceased running consistently, and/or if the imaging time exceeded 50 minutes. Prior to the imaging session for this task, mice were provided 30 “warm-up” trials using the task and reward zone from the previous day to re-acclimate them to the VR setup. On day 1 of imaging, this warm-up session used the pre-training environment from prior days. On days 2 onward, the warm-up session was whichever environment and reward zone was active at the end of the previous day. Following the completion of the daily imaging session, mice were given another ~100 training trials without imaging on the last reward zone seen in the imaging session until they acquired their daily water allotment.

Two-photon (2P) imaging

We used a resonant-galvo scanning 2P microscope from Neurolabware to image the calcium activity of CA1 neurons. Excitation was achieved using 920-nm light from a tunable output femtosecond pulsed laser (Coherent Chameleon Discovery). Laser power modulation was accomplished with a Pockels cell (Conoptics). The typical excitation power, measured at the front face of the objective (Leica 25X, 1.0NA, 2.6 mm working distance), ranged from 13–40 mW for mice m12, m13, m14; 15–68 mW for mice m2, m3, m4, m6, m7; and 50–100 mW for m10 and m11. The hidden reward zone task was imaged starting from day 1 for all mice except m11, for whom imaging started on day 3 due to lower viral expression. Continuous imaging under constant laser power persisted through each trial of the imaging session. For most animals and sessions, to minimize photodamage and photobleaching, the Pockels cell was used to reduce laser power to minimal levels between trials (during the teleport zone). For mice m11, m12, m13, and m14, laser power was maintained and imaging continued throughout the teleport zone on days 1, 3, 7, 8, and 14 of the task. Photons were collected using Hamamatsu gated GAsP photomultiplier tubes (part H11706–401 for Neurolabware microscopes). Imaging data were acquired via Scanbox software (https://scanbox.org/), operated through MATLAB (Mathworks). The imaging field of view (FOV) was collected at 1.0 magnification with unidirectional scanning at ~15.5 Hz, resulting in a ~0.64 × 0.64 mm FOV (512 × 716 pixels). The same FOV was acquired each session by aligning to a reference image from previous days prior to the start of data acquisition, with the aid of the ‘searchref’ plugin in Scanbox. This allowed us to track single cells across days (see Calcium data processing).

Calcium data processing

The Suite2P software package174 (v0.10.3) was used to perform xy motion correction (rigid and non-rigid) and identify putative cell regions of interest (ROIs). Manual curation was performed to eliminate ROIs containing multiple somas or dendrites, lacking visually obvious transients, suspected of overexpressing the calcium indicator, or exhibiting high and continuous fluorescence fluctuation typical of putative interneurons. This approach yielded between 155 and 1982 putative pyramidal neurons per session, due to variation in imaging window implant quality and viral expression. Custom code was used to identify cells that were successfully tracked across imaging sessions using the intersection over union (IOU) of their ROI pixels. The threshold for ROI matching was chosen algorithmically for each dataset such that the IOU for the best match for an ROI pair was always greater than the IOU for any second best match.

To compute the ΔF/F (dF/F) for each ROI, baseline fluorescence was calculated within each trial independently using a maximin procedure with a 20-s sliding window (modification of default suite2p procedure). Limitation to individual trials both accounts for potential photobleaching over the session and avoids the teleport periods for sessions where the laser power was reduced (see Two-photon imaging). dF/F was then calculated for each cell as the fluorescence minus the baseline, divided by the absolute value of the baseline. dF/F was smoothed with a 2-sample (~0.129 s) s.d. Gaussian kernel. The activity rate was then extracted by deconvolving dF/F with a canonical calcium kernel using the OASIS algorithm175 as used in suite2p. It is important to note that this deconvolution is not interpreted as a spike rate but rather as a method to eliminate the asymmetric smoothing effect on the calcium signal induced by the indicator kinetics. Additional putative interneurons were detected for exclusion from further analysis via a Pearson correlation >0.5 between their dF/F timeseries and the animal’s running speed, excluding 0.24 ± 0.41% of cells (mean ± standard deviation across mice and days).

Statistics

To avoid assumptions about data distributions, we used nonparametric tests or permutation tests in most cases, except where a Shapiro–Wilk test for normality confirmed that parametric tests were reasonable. Specific tests are described in further detail in specific Methods sections. In the main text, percentages of cells classified by remapping type are reported as mean ± standard deviation out of all place cells identified on the specified days (n = 7 mice for the “switch” task). Place cell criteria are defined below in Place cell and peak firing identification. Averaged data in figures are shown as mean ± s.e.m. unless otherwise indicated. For the distributions of sequence positions over days in Fig. 3 and Fig. S3, s.e.m. shading is omitted for clarity, but the variance for these distributions is taken into account in linear mixed effects models to quantify these plots (see Linear mixed effects models).

All analyses were performed in Python 3.8.5. Linear mixed effects models were performed using the statsmodels package (https://www.statsmodels.org/stable/index.html). Pairwise repeated measures (paired) t-tests were run using the pingouin package (https://pingouin-stats.org/) with a Holm step-down Bonferroni correction for multiple comparisons. Linear regressions and Wilcoxon signed-rank tests were performed using the SciPy (https://scipy.org/) statistics module, with confidence intervals computed with the Uncertainties package (https://pypi.org/project/uncertainties/). Circular statistics were performed using Astropy (https://docs.astropy.org/en/stable/stats/circ.html), pycircstat176 (https://github.com/circstat/pycircstat), and circular-circular correlation code originally written to analyze hippocampal theta phase precession177 (https://github.com/CINPLA/phase-precession). The time warp models and GLM were implemented from publicly available repositories (time warp: https://github.com/ahwillia/affinewarp; GLM: https://github.com/HarveyLab/GLM_Tensorflow_2).

The number of mice to include was determined by having coverage of every possible reward sequence permutation (6 sequences) with at least one mouse. In addition, our mouse sample sizes were similar to those reported in previous publications7,150.

Quantification of licking behavior

The capacitive lick sensor allowed us to detect single licks. A very small number of trials where the sensor had become corroded were removed from subsequent licking analysis by setting these values to NaN (~0.8% of all imaged trials, n=64 removed trials out of 7736 trials across the 7 switch mice). These trials were detected by >30% of the imaging frame samples in the trial containing ≥3 Unity frames where licks were detected, as this would have produced a sustained lick rate ≥20 Hz. Remaining lick counts were converted to a binary vector at the imaging frame rate and spatially binned at the same resolution as neural activity (10 cm bins), then divided by the time occupancy in each bin to yield a lick rate. We quantified the licking precision of the animals over blocks of 10 trials using an anticipatory lick ratio inspired by the ramp of anticipatory licking we observed preceding the reward zone (Fig. 1FG, Fig. S1AC). The ratio was computed as:

lickratio=lickin-lickoutlickin+lickout

Where lickin is the mean lick rate in a 50 cm “anticipatory” zone before the start of the reward zone, and lickout is the mean rate outside of this zone and the reward zone. The reward zone itself is thus excluded to exclude consummatory licks (Fig. S1E). A ratio of 1 indicates perfect licking only in the anticipatory zone, a ratio of −1 indicates licking only outside of this zone, and a ratio of 0 indicates chance licking everywhere excluding the reward zone.

Place cell and peak firing identification

For all neural spatial activity analyses we excluded activity when the animal was moving at <2 cm/s. To get a trial-by-trial activity rate matrix for each cell, we binned the 450 cm linear track into 45 bins of 10 cm each. For each cell, we took the mean of the activity samples (dF/F or deconvolved calcium activity) on each trial within each position bin (note that taking the mean activity over time samples is equivalent to normalizing by the occupancy within a trial), producing a matrix of trials x position bins.

We defined place cells as cells with significant spatial information178 (SI) in either trial set 1 (pre-switch) or trial set 2 (post-switch) in a session. On days without a switch, we used the 1st and 2nd half of trials. For the spatial information computation and for plotting spatial firing over trials, we used the deconvolved calcium activity as this reduces the asymmetry of the calcium signal (smoothed only for plotting with a 10 cm s.d. Gaussian kernel). SI was calculated using a formula implemented for calcium imaging179 as:

SI=i=1Npififlog2fif

where pi is the occupancy probability in position bin i for the whole session, fi is the trial-averaged activity per position bin i, and f is the mean activity rate over the whole session, computed as the sum of fi*pi over all N bins. To get pi, we calculated the occupancy (number of imaging samples) in each bin on each trial and divided this by the total number of samples in each trial to get an occupancy probability per position bin per trial. We then summed the occupancy probabilities across trials and divided by the total per session to get an occupancy probability per position bin per session. To get the spatial “tuning curve” over the session, we averaged the activity in each bin across trials. To determine the significance of the SI scores, we created a null distribution by circularly permuting the position data relative to the timeseries of each cell, by a random amount of at least 1 s and a maximum amount of the length of the trial, independently on each trial. SI was calculated from the trial-averaged activity of each shuffle, and this shuffle procedure was repeated 100 times per cell. A cell’s true SI was considered significant if it exceeded 95% of the SI scores from all shuffles within animal (i.e. shuffled scores were pooled across cells within animal to produce this threshold, which is more stringent than comparing to the shuffle of each individual cell180).

To identify spatial peaks, we used the unsmoothed spatially binned dF/F, as this signal is the closest to the raw data. Place cell firing peaks or “spatial peaks” refer to the position bin of the maximum spatially-binned activity, usually averaged across trials within a set. There was no restriction to field boundaries, thus allowing cells to have multiple fields. To plot single cell place fields and cell sequences, neural activity was normalized to the mean of each cell within a session.

Trial-by-trial similarity matrices

Correlation matrices for single cells were computed using the spatially binned deconvolved activity on each trial, smoothed with a 2-bin (20 cm) s.d. Gaussian. This resulted in a matrix Ai for each cell i of j trials × M position bins. Each trial was z-scored across the position axis, and the trial-by-trial correlation matrix C was computed as:

C=1M-1AiAiT

Cell category definitions for remapping subtypes

We defined the remapping types shown in Fig. 2 and Fig. S2 as follows: track-relative: significant SI before and after the switch, with spatial peaks (position bin of maximum trial-averaged dF/F) ≤50 cm from each other before vs. after; disappearing: significant SI before but not after the switch, with a mean spatially-binned dF/F after that is less than the 50th percentile of the per-trial mean dF/F before; appearing: significant SI after but not before, with a mean spatially-binned dF/F after that is greater than 1 s.d. above the mean in trials before; remap near reward: significant SI before and after, with spatial peaks ≤50 cm from the starts of both reward zones; remap far from reward: significant SI before and after, not “track-relative” (i.e. peaks >50 cm apart before vs. after) and not “near reward” cells (i.e. peaks >50 cm from the start of at least one reward zone).

Reward-relative remapping

Quantification of reward-relative remapping compared to chance

To perform statistics on putative reward-relative remapping (Fig. 2, Fig. S2), we restricted the set of included place cells to require significant spatial information in the 1st and 2nd trial set in each session (pre- and post- switch), to produce a stringent set of spatial peaks. For all scatter plots of spatial peaks before vs. after a reward switch, points are jittered by a random amount between −π /100 and +π/100 for visualization only. As outlined in Fig. 2, we converted the track axis (i.e. the original position timeseries of the animal) to circular coordinates, setting the beginning of the track to −π and the end of the track (beginning of the teleport zone) to π. Spatial peaks for each cell were re-computed using dF/F binned into 45 bins on this axis (bin size 2π/45, corresponding to ~10 cm). Putative track-relative cells were removed by excluding cells with spatial peaks before vs. after the switch that fell within 0.698 radians (~50 cm) of each other. The coordinates of the remaining data were then circularly rotated to align the start of each reward zone at 0, such that spatial peaks could be measured relative to the start of each reward zone (as the signed circular distance). We measured the distance between relative peaks as the “after” peak minus the “before” peak, circularly wrapped, creating a distribution that can be thought of as orthogonal to the unity line shown in the scatters in Fig. 2D, F. We compared this distribution (shown in Fig. 2E, G) to a “random-remapping” shuffle. This shuffle was generated by maintaining the pre-switch position of each cell’s peak and circularly permuting the cell’s post-switch firing 1000 times by a random offset of 0 to 44 bins, creating a set of shuffled post-switch positions for each cell. To define a candidate range of reward-relative remapping variability (i.e. a range around the unity line in Fig. 2D, F), we computed the circular difference between the maximum and minimum bin of the distribution that exceeded the upper 95% of the shuffle, divided by 2, and averaged across switches (Fig. S2G). This produced a mean range of ~0.656 radians (46.9 cm) around 0, which is captured by a 50 cm threshold given our 10 cm bin size. This threshold was subsequently used to identify candidate reward-relative remapping cells (see below).

Criteria to define reward-relative cells

Reward-relative cells could encompass “remap near reward” and “remap far from reward” cells as defined in Cell category definitions for remapping subtypes. However, we observed variable dynamics over trials with which the spatial activity of these cells remapped after a reward switch (Fig. 2AB, Fig. S2A), which could affect spatial information scores. We therefore relaxed the SI criterion for reward-relative cells such that they were required to have significant SI either before or after the switch but not necessarily both, to allow for these dynamics. We began by finding place cells which had peak spatial firing within ≤50 cm (0.698 radians) of each other when these peaks were calculated using circular bins relative to reward zone starts. As the detection of peak firing can be noisy, we further took into account the shape of the field by performing a cross-correlation between each place cell’s trial-averaged spatial activity before vs. after, using the activity matrices (trials x position bins, unsmoothed dF/F) where reward zone starts are circularly aligned between trial sets 1 and 2. This procedure allowed us to test the hypothesis that reward-relative cells should have a peak cross-correlation close to zero on this reward-centered axis. We calculated a shuffle for the cross-correlation by circularly permuting the activity in trial set 2 by a random offset between 1 and 45 bins on each trial 500 times, and required that the real cross-correlogram had a peak exceeding the 95% confidence interval (upper 97.5% threshold) of the shuffle, and that the offset of this peak was within 5 bins (~50 cm) of zero (see Fig. S2NP). We validated that the population of cells passing this criterion had high Pearson correlation coefficients between their circularly aligned activity maps in set 1 vs. set 2 (whereas, by contrast “track-relative” cells had high correlation coefficients and a cross-correlation offset near zero only in the original linear track coordinates; Fig. S2OP). Place cells that passed this final circular cross-correlation criterion were therefore defined as reward-relative.

Relationship between distance run in the teleport zone and trial-to-trial variability

We measured the distance run in the teleport zone at the end of each trial by integrating the animal’s speed between the entry to the teleport zone (i.e. first frame at which the temporal jitter started, see Behavioral training and VR tasks) and the last frame before the re-entry to the track (i.e. start of the next trial). We then computed the trial-wise spatial peak error for each reward-relative cell as phase difference between the cell’s spatial peak location in circular coordinates on each trial and its mean spatial peak location within either trial set 1 (before the reward switch) or trial set 2 (after the switch). Circular error was then converted back to centimeters. We next performed a Spearman correlation between distance run during the teleport and the error on the following trial; population correlation coefficients are reported in Fig. S2F. Results were similar using linear root-squared error rather than circular coordinates (data not shown).

Sequence detection and quantification

The sequential firing of neural subpopulations was detected using a split-halves cross-validation procedure7. For these analyses, we used the unsmoothed, spatially binned dF/F in circular track coordinates (not aligned to reward, though aligning to reward makes no difference for the circular sequence order). To sort the activity of neurons, we used the activity averaged over the odd trials before the switch to find and sort the peak firing positions. This sort was validated by plotting the trial-averaged activity of the even trials before the switch (e.g. left-hand columns of Fig. 3A). The same sort was then applied to the trials after the switch. The sequence positions before vs. after were then taken as the peak firing positions of the trial-averaged “even-before” trials and the trial-averaged “after” trials. Preservation of sequence order was quantified as the circular-circular correlation coefficient176,177,181 between the sequence positions before vs. after. Though a p-value can be directly computed for this correlation coefficient, its significance was further validated through a permutation procedure. We randomly permuted the cell identities of the neurons after the switch and re-computed the correlation coefficient 1000 times to get a null distribution. The observed correlation coefficient was considered significant if it was outside 95% of the null distribution using a two-tailed test. The p-value was calculated as:

p=n|coefshuf||coefobs|+1Nperm+1

where n|coefshuf||coefobs| is the number of shuffled coefficients with absolute values ≥ the absolute value of the observed coefficient, and Nperm is the total number of shuffles.

For sequences tracked across days, we again used the cross-validated sort from trial set 1 (before the switch, or the 1st half of trials if the reward did not move) on the reference day (the first day in each day pair) and applied this sort to trial set 2 (after the switch, or 2nd half) on the reference day as well as activity in trial sets 1 and 2 on the target day (the second day in the pair). To measure drift between days, we computed the across-day circular-circular correlation coefficient between the sequence positions on reference day trial set 2 vs. target day trial set 1.

To quantify the density of sequences, we divided the number of cells in the sequence with peaks in each position bin by the total number of place cells for that animal. These density curves were smoothed only for visualization in Fig. 3J, S and Fig. S3J with a 1-bin (2π/45) s.d. Gaussian and shown as the mean across animals per each switch day. To visualize the expected uniform distribution across animals, we calculated this as the mean fraction of cells in each sequence out of the place cells in each animal on a given switch day, divided by the number of position bins.

Linear mixed effects models

To quantify neural activity changes across task experience, we used linear mixed effects models to take into account variance across animals. Linear mixed effects models were implemented using the mixedlm method of the statsmodels.formula.api package (https://www.statsmodels.org/stable/mixed_linear.html). Reported p-values are from Wald tests supplied by mixedlm. The fixed effect was either the switch index or day-pair index as a continuous variable, except for Fig. S5J where additional fixed effects were considered (see figure legend). Random effects were mouse identity. Random intercepts for each mouse were allowed; we also confirmed that including random slopes did not affect the results. When the dependent variable was reported in fractions of cells, we applied a logit transform to the fractions when a large proportion of the values were near zero. In most cases, however, we chose to display results in the original fractions for interpretability. In these cases, we confirmed that performing a logit transform on the fractions did not qualitatively change the results and only modestly changed the model fit.

Analysis of rewarded vs. omission trials

All following analyses were restricted to trial sets (i.e. before or after the reward switch) that had at least 3 omission trials within the set.

Time warp modeling

We first fit 5 different time warp model types—shift, linear, piecewise 1 knot, piecewise 2 knots, and piecewise 3 knots87—on the matrix of speed profiles within a trial set (T trials × M linear position bins × 1, expanded in the third dimension for compatibility with the time warp algorithm). These models apply warping functions of increasing nonlinear complexity to stretch and compress the data on each trial for maximal alignment87. We included both rewarded and omission trials in the fitting procedure to find the best alignment across them. Model fit was assessed using the mean squared error (MSE) between the time warped speed profile of each trial and the mean across time warped trials (Fig. S5A). The piecewise 3 knots model most often produced the lowest MSE, or best fit, across sessions (Fig. S5C). We therefore re-fit all sessions with piecewise 3 knots to ensure that model selection could not influence the reward vs. omission metric described below. We then applied the model transform to the neural data matrix for the same set of trials (T trials × M linear position bins × N neurons; excluding neural data at movement speeds <2cm/s, similar to all other neural analyses). For neural analysis in Fig. 5, we then focused on trial sets before the switch where the reward zone was at position “A” or “B”. This is because the mean rewarded vs. omission speed profiles were more similar on these trials than after the switch (Fig. S5F) or when the reward was at “C” (as the animals tended to lick for longer on omission trials and there is less room at the end of the track for running speed to reach its maximum from position “C”; see Fig. S1 and Fig. S5H for examples). Restricting sessions to “A” and “B” reward zones produced similar results to restricting sessions to those best fit by the time warp model (data not shown). Complementary analyses for trials after the switch and for all reward zones are reported in Fig. S5.

Reward vs. omission index

We compared neural activity on rewarded vs. omission trials for reward-relative cells with spatial peaks between the start of the reward zone and the end of the track (under the hypothesis that these cells could receive information about whether reward was received or not on that trial). Using the time warp model-transformed neural activity averaged across rewarded or omission trials within the trial set of interest, we calculated a “reward vs. omission index” (RO index) for each of these reward-relative cells as:

ROindex=iwi--mi-iwi-+mi-

Where wi- is the mean activity in position bin i averaged across rewarded trials, and mi- is the mean activity in position bin i averaged across omission trials. RO indices of 1 therefore indicate an exclusive firing preference for rewarded trials, −1 indicates an exclusive preference for omission trials, and 0 indicates equal firing between rewards and omissions. We confirmed that the RO index calculated from the model-transformed activity was highly correlated with the RO index calculated from the original activity (Fig. S5E). We then quantified the median RO index for each animal and day. Using a linear mixed effects model (Fig. SJ), we found that there was no significant change across days, as a result of the model fit, or as a result of the original MSE between rewarded and omission trials. We therefore combined neurons across days for presentation in Fig. 5E.

Generalized linear model (GLM)

We implemented a Poisson GLM developed previously88 to predict the deconvolved calcium event time series of individual neurons from a set of task and movement variables. All behavioral and neural time series were sampled at ~15.5 Hz, the imaging frame rate.

Design matrix

Our design matrix is schematized in Fig. S6A and included 6 predictor variables split into 3 task variables and 3 movement variables.

Task variables were linear track position (from 0 to 450 cm), reward-relative position (from −π to π, centered around the start of each reward zone at 0 radians), and “rewarded”, a binary that goes from 0 to 1 when the reward is delivered on each trial and stays at 1 until the end of the trial (rewarded stays at 0 on omission trials). Linear track position and reward-relative position were separately expanded into 45 cosine basis functions (or “bumps”) each, one for each 10 cm or ~0.14 radian bin used in other analyses. The rewarded variable was multiple with the linear position basis functions to yield the interaction between position and whether the animal had been rewarded on a given trial.

Movement variables were running speed and acceleration (smoothed with a 5-sample s.d. Gaussian) and smoothed lick count (binary licks per frame smoothed with a 2-sample s.d. Gaussian to approximate an instantaneous rate). Each movement variable was separately quantile-transformed to encode the distribution of movement dynamics similarly across animals. The quantiles were expanded with B-spline basis functions using the SplineTranformer method of Python’s sklearn package, with 3 polynomial degrees and 5 knots, resulting in 7 bases (5+3–1) per movement variable. Spline choices were made for consistency with previous work88.

All predictors were concatenated into a design matrix with 156 total features: 45 position bases + 45 reward-relative position bases + 45 rewarded-×-position bases + 7 speed bases + 7 acceleration bases + 7 licking bases. The features were independently z-scored across samples before model fitting on the “response matrix” of deconvolved activity of all cells (samples × neurons).

Model fitting and testing

Trial identities were used to group data for training and testing. Individual sessions (including trials both before and after the reward switch) were split into 85% training trials and 15% testing trials, using the GroupShuffleSplit method of Python’s sklearn package with a consistent random seed. Specifically, trials were allocated such that the training trials included 85% of rewarded trials before the switch, 85% of rewarded trials after the switch, and 85% of all omission trials (likewise the testing set received the remaining trials from each group). The training trials were further divided for 5-fold cross validation during the fitting procedure. An optimal model was selected according to the deviance explained on the 5-fold cross-validation data. We then tested this model on the held-out test data to assess model performance as the fraction deviance explained (FDE):

FDE=1-devmodeldevnull

where devmodel is the Poisson deviance of the full model and devnull is the Poisson deviance of a null model that predicts the mean of the neural activity across time samples.

On average, the model’s FDE for all place cells was 0.10 ± 0.19 (mean ± s.d.) (Fig. S6BC), 0.32 ± 0.13 for track-relative cells, 0.29 ± 0.11 for reward-relative cells (examples in Fig. 6AB), and 0.29 ± 0.11 for non-reward-relative remapping cells. For analysis of the relative contribution of individual variables, we only included cells with FDE > 0.15 in accordance with previous procedures88.

Relative contribution of individual variables via model ablation

To assess how much each variable contributed to the model’s ability to predict a given cell’s activity, we performed a model ablation procedure. In brief, after fitting the full model, we zeroed the coefficients for each variable and compared the performance of the ablated versus the full model on the cross-validation data. We calculated the reduction in model fit (deviance) of the ablated model relative to the full model, normalized by the full model’s deviance relative to the null model. This is known as “fraction explained deviance”88, which we term “relative contribution” to distinguish it from fraction deviance explained. Relative contribution was thus computed as:

Relativecontribution=devablated-devfulldevnull-devfull

Relative contribution was binned by linear track position to visualize the contribution at each position along the track, then averaged across position bins for quantifying the contribution at the population level (Fig. 6D) and identifying the top predictor per cell (variable with the maximum mean relative contribution) (Fig. 6E).

Supplementary Material

Supplement 1

Acknowledgements

We thank Charlotte Herber, John Wen, and Jon Rueckemann for comments on the manuscript; Can Dong for assistance with animal training; Adriana Diaz for histology assistance and lab infrastructure support; Alex Gonzalez, Tucker Fisher, and Eric Denovellis for guidance on model implementation and evaluation; and Shih-Yi Tseng for feedback on the GLM. We thank members of the Giocomo laboratory and the Simons Collaboration for the Global Brain Remapping postdoc group for extensive feedback and discussions, in particular P. Dylan Rich, E. Mika Diamanti, and Eric Denovellis for suggesting edits to draft figures. We thank William Newsome and Scott Linderman for frequent feedback on this work, Charline Tessereau and Feng Xuan for insightful discussions, Simón Carrillo-Segura and André Fenton for generously sharing code and valuable insights, and Jeff Gauthier for input on calcium signal processing and discussions about reward-related remapping. This work was supported by a Helen Hay Whitney Foundation fellowship (MS), National Institute of Health Grants 1R01MH126904-01A1 (LMG), R01MH130452 (LMG), BRAIN Initiative U19NS118284 (LMG), P50 DA042012 (LMG), The Vallee Foundation (LMG), The James S. McDonnell Foundation (LMG), The Simons Foundation 542987SPI (LMG), and the Champalimaud Vision Award to William Newsome, who we thank for sharing his support.

Footnotes

Declarations of Interests: The authors declare no competing interests.

Data and Code Availability: Upon publication, data will be made available on a publicly accessible data sharing site (e.g. figshare, DANDI, Mendeley). Custom scripts for analyzing the data will be available on GitHub.

REFERENCES

  • 1.Sosa M., and Giocomo L.M. (2021). Navigating for reward. Nat Rev Neurosci 22, 472–487. 10.1038/s41583-021-00479-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Nyberg N., Duvelle É., Barry C., and Spiers H.J. (2022). Spatial goal coding in the hippocampal formation. Neuron 110, 394–422. 10.1016/j.neuron.2021.12.012. [DOI] [PubMed] [Google Scholar]
  • 3.O’Keefe J., and Dostrovsky J. (1971). The hippocampus as a spatial map. Preliminary evidence from unit activity in the freely-moving rat. Brain Res. 34, 171–175. [DOI] [PubMed] [Google Scholar]
  • 4.Frank L.M., Stanley G.B., and Brown E.N. (2004). Hippocampal plasticity across multiple days of exposure to novel environments. Journal of Neuroscience 24, 7681–7689. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Colgin L.L., Moser E.I., and Moser M.B. (2008). Understanding memory through hippocampal remapping. Trends in neurosciences 31, 469–477. 10.1016/j.tins.2008.06.008. [DOI] [PubMed] [Google Scholar]
  • 6.Bostock E., Muller R.U., and Kubie J.L. (1991). Experience-dependent modifications of hippocampal place cell firing. Hippocampus. 1, 193–205. [DOI] [PubMed] [Google Scholar]
  • 7.Plitt M.H., and Giocomo L.M. (2021). Experience-dependent contextual codes in the hippocampus. Nat Neurosci 24, 705–714. 10.1038/s41593-021-00816-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Knierim J.J., Kudrimoti H.S., and McNaughton B.L. (1998). Interactions between idiothetic cues and external landmarks in the control of place cells and head direction cells. J.Neurophysiol. 80, 425–446. [DOI] [PubMed] [Google Scholar]
  • 9.Leutgeb S., Leutgeb J.K., Treves A., Moser M.B., and Moser E.I. (2004). Distinct ensemble codes in hippocampal areas CA3 and CA1. Science 305, 1295–1298. [DOI] [PubMed] [Google Scholar]
  • 10.O’Keefe J., and Conway D.H. (1978). Hippocampal place units in the freely moving rat: why they fire where they fire. Exp Brain Res 31, 573–590. [DOI] [PubMed] [Google Scholar]
  • 11.Leutgeb J.K., Leutgeb S., Treves A., Meyer R., Barnes C.A., McNaughton B.L., Moser M.B., and Moser E.I. (2005). Progressive transformation of hippocampal neuronal representations in “morphed” environments. Neuron. %20;48, 345–358. [DOI] [PubMed] [Google Scholar]
  • 12.Cohen J.D., Bolstad M., and Lee A.K. (2017). Experience-dependent shaping of hippocampal CA1 intracellular activity in novel and familiar environments. Elife 6, e23040. 10.7554/eLife.23040. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Buzsáki G. (2010). Neural syntax: cell assemblies, synapsembles, and readers. Neuron 68, 362–385. 10.1016/j.neuron.2010.09.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.O’Keefe J., and Nadel L. (1978). The hippocampus as a cognitive map (Oxford University Press; ). [Google Scholar]
  • 15.Schiller D., Eichenbaum H., Buffalo E.A., Davachi L., Foster D.J., Leutgeb S., and Ranganath C. (2015). Memory and Space: Towards an Understanding of the Cognitive Map. J Neurosci 35, 13904–13911. 10.1523/jneurosci.2618-15.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Levy E.R.J., Carrillo-Segura S., Park E.H., Redman W.T., Hurtado J.R., Chung S., and Fenton A.A. (2023). A manifold neural population code for space in hippocampal coactivity dynamics independent of place fields. Cell Rep 42, 113142. 10.1016/j.celrep.2023.113142. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Buzsaki G., and Tingley D. (2018). Space and Time: The Hippocampus as a Sequence Generator. Trends Cogn Sci 22, 853–869. 10.1016/j.tics.2018.07.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Brown E.N., Frank L.M., Tang D., Quirk M.C., and Wilson M.A. (1998). A statistical paradigm for neural spike train decoding applied to position prediction from ensemble firing patterns of rat hippocampal place cells. J Neurosci 18, 7411–7425. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Redish A.D., Rosenzweig E.S., Bohanick J.D., McNaughton B.L., and Barnes C.A. (2000). Dynamics of hippocampal ensemble activity realignment: time versus space. Journal of Neuroscience 20, 9298–9309. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Wilson M.A., and Mcnaughton B.L. (1993). Dynamics of the hippocampal ensemble code for space. Science 261, 1055–1058. [DOI] [PubMed] [Google Scholar]
  • 21.Skaggs W.E., McNaughton B.L., Wilson M.A., and Barnes C.A. (1996). Theta phase precession in hippocampal neuronal populations and the compression of temporal sequences. Hippocampus 6, 149–172. [DOI] [PubMed] [Google Scholar]
  • 22.Knierim J.J. (2015). From the GPS to HM: Place cells, grid cells, and memory. Hippocampus 25, 719–725. 10.1002/hipo.22453. [DOI] [PubMed] [Google Scholar]
  • 23.Packard M.G., and McGaugh J.L. (1996). Inactivation of hippocampus or caudate nucleus with lidocaine differentially affects expression of place and response learning. Neurobiol.Learn.Mem. 65, 65–72. [DOI] [PubMed] [Google Scholar]
  • 24.Robinson N.T.M., Descamps L.A.L., Russell L.E., Buchholz M.O., Bicknell B.A., Antonov G.K., Lau J.Y.N., Nutbrown R., Schmidt-Hieber C., and Hausser M. (2020). Targeted Activation of Hippocampal Place Cells Drives Memory-Guided Spatial Behavior. Cell 183, 2041–2042. 10.1016/j.cell.2020.12.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Eichenbaum H. (2017). The role of the hippocampus in navigation is memory. J Neurophysiol 117, 1785–1796. 10.1152/jn.00005.2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Liu C., Todorova R., Tang W., Oliva A., and Fernandez-Ruiz A. (2023). Associative and predictive hippocampal codes support memory-guided behaviors. Science 382, eadi8237. 10.1126/science.adi8237. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Rueckemann J.W., Sosa M., Giocomo L.M., and Buffalo E.A. (2021). The grid code for ordered experience. Nat Rev Neurosci 22, 637–649. 10.1038/s41583-021-00499-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Whittington J.C.R., Muller T.H., Mark S., Chen G., Barry C., Burgess N., and Behrens T.E.J. (2020). The Tolman-Eichenbaum Machine: Unifying Space and Relational Memory through Generalization in the Hippocampal Formation. Cell 183, 1249–1263.e23. 10.1016/j.cell.2020.10.024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Pastalkova E., Itskov V., Amarasingham A., and Buzsaki G. (2008). Internally generated cell assembly sequences in the rat hippocampus. Science. 321, 1322–1327. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Eichenbaum H. (2014). Time cells in the hippocampus: a new dimension for mapping memories. Nature reviews. Neuroscience 15, 732–744. 10.1038/nrn3827. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Shimbo A., Izawa E.I., and Fujisawa S. (2021). Scalable representation of time in the hippocampus. Sci Adv 7. 10.1126/sciadv.abd7013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Kraus B.J., Brandon M.P., Robinson R.J. 2nd, Connerney M.A., Hasselmo M.E., and Eichenbaum H. (2015). During running in place, grid cells integrate elapsed time and distance run. Neuron 88, 578–589. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.MacDonald C.J., Lepage K.Q., Eden U.T., and Eichenbaum H. (2011). Hippocampal “time cells” bridge the gap in memory for discontiguous events. Neuron 71, 737–749. 10.1016/j.neuron.2011.07.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Taxidis J., Pnevmatikakis E.A., Dorian C.C., Mylavarapu A.L., Arora J.S., Samadian K.D., Hoffberg E.A., and Golshani P. (2020). Differential Emergence and Stability of Sensory and Temporal Representations in Context-Specific Hippocampal Sequences. Neuron 108, 984–998 e9. 10.1016/j.neuron.2020.08.028. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Abramson S., Kraus B.J., White J.A., Hasselmo M.E., Derdikman D., and Morris G. (2023). Flexible coding of time or distance in hippocampal cells. Elife 12, e83930. 10.7554/eLife.83930. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Aronov D., Nevers R., and Tank D.W. (2017). Mapping of a non-spatial dimension by the hippocampal-entorhinal circuit. Nature 543, 719–722. 10.1038/nature21692. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Shahbaba B., Li L., Agostinelli F., Saraf M., Cooper K.W., Haghverdian D., Elias G.A., Baldi P., and Fortin N.J. (2022). Hippocampal ensembles represent sequential relationships among an extended sequence of nonspatial events. Nat Commun 13, 787. 10.1038/s41467-022-28057-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Radvansky B.A., and Dombeck D.A. (2018). An olfactory virtual reality system for mice. Nat Commun 9, 839. 10.1038/s41467-018-03262-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Ginther M.R., Walsh D.F., and Ramus S.J. (2011). Hippocampal neurons encode different episodes in an overlapping sequence of odors task. J Neurosci 31, 2706–2711. 10.1523/JNEUROSCI.3413-10.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Allen T.A., Salz D.M., McKenzie S., and Fortin N.J. (2016). Nonspatial Sequence Coding in CA1 Neurons. J Neurosci 36, 1547–1563. 10.1523/JNEUROSCI.2874-15.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Fischler-Ruiz W., Clark D.G., Joshi N.R., Devi-Chou V., Kitch L., Schnitzer M., Abbott L.F., and Axel R. (2021). Olfactory landmarks and path integration converge to form a cognitive spatial map. Neuron 109, 4036–4049.e5. 10.1016/j.neuron.2021.09.055. [DOI] [PubMed] [Google Scholar]
  • 42.Zhang S., and Manahan-Vaughan D. (2015). Spatial olfactory learning contributes to place field formation in the hippocampus. Cerebral Cortex 25, 423–432. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Barron H.C., Reeve H.M., Koolschijn R.S., Perestenko P.V., Shpektor A., Nili H., Rothaermel R., Campo-Urriza N., O’Reilly J.X., Bannerman D.M., et al. (2020). Neuronal Computation Underlying Inferential Reasoning in Humans and Mice. Cell 183, 228–243 e21. 10.1016/j.cell.2020.08.035. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Terada S., Sakurai Y., Nakahara H., and Fujisawa S. (2017). Temporal and Rate Coding for Discrete Event Sequences in the Hippocampus. Neuron 94, 1248–1262 e4. 10.1016/j.neuron.2017.05.024. [DOI] [PubMed] [Google Scholar]
  • 45.Radvansky B.A., Oh J.Y., Climer J.R., and Dombeck D.A. (2021). Behavior determines the hippocampal spatial mapping of a multisensory environment. Cell Rep 36, 109444. 10.1016/j.celrep.2021.109444. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Nieh E.H., Schottdorf M., Freeman N.W., Low R.J., Lewallen S., Koay S.A., Pinto L., Gauthier J.L., Brody C.D., and Tank D.W. (2021). Geometry of abstract learned knowledge in the hippocampus. Nature 595, 80–84. 10.1038/s41586-021-03652-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Dragoi G. (2020). Cell assemblies, sequences and temporal coding in the hippocampus. Curr Opin Neurobiol 64, 111–118. 10.1016/j.conb.2020.03.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Villette V., Malvache A., Tressard T., Dupuy N., and Cossart R. (2015). Internally Recurring Hippocampal Sequences as a Population Template of Spatiotemporal Information. Neuron 88, 357–366. 10.1016/j.neuron.2015.09.052. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Eichenbaum H. (2017). On the Integration of Space, Time, and Memory. Neuron 95, 1007–1018. 10.1016/j.neuron.2017.06.036. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Manns J.R., Howard M.W., and Eichenbaum H. (2007). Gradual changes in hippocampal activity support remembering the order of events. Neuron. 56, 530–540. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Fortin N.J., Agster K.L., and Eichenbaum H.B. (2002). Critical role of the hippocampus in memory for sequences of events. Nature Neuroscience 5, US http://www. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Moore J.J., Cushman J.D., Acharya L., Popeney B., and Mehta M.R. (2021). Linking hippocampal multiplexed tuning, Hebbian plasticity and navigation. Nature 599, 442–448. 10.1038/s41586-021-03989-z. [DOI] [PubMed] [Google Scholar]
  • 53.Zhao X., Hsu C.-L., and Spruston N. (2022). Rapid synaptic plasticity contributes to a learned conjunctive code of position and choice-related information in the hippocampus. Neuron 110, 96–108.e4. 10.1016/j.neuron.2021.10.003. [DOI] [PubMed] [Google Scholar]
  • 54.Gauthier J.L., and Tank D.W. (2018). A Dedicated Population for Reward Coding in the Hippocampus. Neuron 99, 179–193.e7. 10.1016/j.neuron.2018.06.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Sato M., Mizuta K., Islam T., Kawano M., Sekine Y., Takekawa T., Gomez-Dominguez D., Schmidt A., Wolf F., Kim K., et al. (2020). Distinct Mechanisms of Over-Representation of Landmarks and Rewards in the Hippocampus. Cell Rep 32, 107864. 10.1016/j.celrep.2020.107864. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Zemla R., Moore J.J., Hopkins M.D., and Basu J. (2022). Task-selective place cells show behaviorally driven dynamics during learning and stability during memory recall. Cell Rep 41, 111700. 10.1016/j.celrep.2022.111700. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Johnson A., Fenton A.A., Kentros C., and Redish A.D. (2009). Looking for cognition in the structure within the noise. Trends Cogn Sci. 13, 55–64. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Muzzio I.A., Levita L., Kulkarni J., Monaco J., Kentros C., Stead M., Abbott L.F., and Kandel E.R. (2009). Attention enhances the retrieval and stability of visuospatial and olfactory representations in the dorsal hippocampus. PLoS Biol 7, e1000140. 10.1371/journal.pbio.1000140. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Muzzio I.A., Kentros C., and Kandel E. (2009). What is remembered? Role of attention on the encoding and retrieval of hippocampal representations. J Physiol 587, 2837–2854. 10.1113/jphysiol.2009.172445. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Levita L., and Muzzio I.A. (2010). Role of the hippocampus in goal-oriented tasks requiring retrieval of spatial versus non-spatial information. Neurobiol Learn Mem 93, 581–588. 10.1016/j.nlm.2010.02.006. [DOI] [PubMed] [Google Scholar]
  • 61.Fenton A.A., Lytton W.W., Barry J.M., Lenck-Santini P.P., Zinyuk L.E., Kubik S., Bures J., Poucet B., Muller R.U., and Olypher A.V. (2010). Attention-like modulation of hippocampus place cell discharge. Journal of Neuroscience 30, 4613–4625. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Kentros C.G., Agnihotri N.T., Streater S., Hawkins R.D., and Kandel E.R. (2004). Increased attention to spatial context increases both place field stability and spatial memory. Neuron. 42, 283–295. [DOI] [PubMed] [Google Scholar]
  • 63.Grienberger C., and Magee J.C. (2022). Entorhinal cortex directs learning-related changes in CA1 representations. Nature 611, 554–562. 10.1038/s41586-022-05378-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Lee J.S., Briguglio J.J., Cohen J.D., Romani S., and Lee A.K. (2020). The Statistical Structure of the Hippocampal Code for Space as a Function of Time, Context, and Value. Cell 183, 620–635 e22. 10.1016/j.cell.2020.09.024. [DOI] [PubMed] [Google Scholar]
  • 65.Dupret D., O’Neill J., Pleydell-Bouverie B., and Csicsvari J. (2010). The reorganization and reactiation of hippocampal maps predict spatial memory performance. Nat Neurosci 13, 995–1002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Hollup S.A., Molden S., Donnett J.G., Moser M.B., and Moser E.I. (2001). Accumulation of hippocampal place fields at the goal location in an annular watermaze task. J Neurosci 21, 1635–1644. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Gothard K.M., Skaggs W.E., Moore K.M., and McNaughton B.L. (1996). Binding of hippocampal CA1 neural activity to multiple reference frames in a landmark-based navigation task. Journal of Neuroscience 16, 823–835. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Zinyuk L., Kubik S., Kaminsky Y., Fenton A.A., and Bures J. (2000). Understanding hippocampal activity by using purposeful behavior: place navigation induces place cell discharge in both task-relevant and task-irrelevant spatial reference frames. Proc.Natl.Acad.Sci.U.S.A 97, 3771–3776. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Park E.H., Keeley S., Savin C., Ranck J.B., and Fenton A.A. (2019). How the Internally Organized Direction Sense Is Used to Navigate. Neuron 101, 285–293.e5. 10.1016/j.neuron.2018.11.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Markus E.J., Qin Y.L., Leonard B., Skaggs W.E., McNaughton B.L., and Barnes C.A. (1995). Interactions between location and task affect the spatial and directional firing of hippocampal neurons. Journal of Neuroscience 15, 7079–7094. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Nitz D. (2009). Parietal cortex, navigation, and the construction of arbitrary reference frames for spatial information. Neurobiol Learn Mem 91, 179–185. 10.1016/j.nlm.2008.08.007. [DOI] [PubMed] [Google Scholar]
  • 72.McKenzie S., Frank A.J., Kinsky N.R., Porter B., Riviere P.D., and Eichenbaum H. (2014). Hippocampal representation of related and opposing memories develop within distinct, hierarchically organized neural schemas. Neuron 83, 202–215. 10.1016/j.neuron.2014.05.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Bures J., Fenton A.A., Kaminsky Y., and Zinyuk L. (1997). Place cells and place navigation. Proc Natl Acad Sci U S A 94, 343−-50. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Helmchen F., and Denk W. (2005). Deep tissue two-photon microscopy. Nat Methods 2, 932–940. 10.1038/nmeth818. [DOI] [PubMed] [Google Scholar]
  • 75.Grienberger C., Giovannucci A., Zeiger W., and Portera-Cailliau C. (2022). Two-photon calcium imaging of neuronal activity. Nat Rev Methods Primers 2, 1–23. 10.1038/s43586-022-00147-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Knierim J.J. (2002). Dynamic interactions between local surface cues, distal landmarks, and intrinsic circuitry in hippocampal place cells. Journal of Neuroscience 22, 6254–6264. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Bourboulou R., Marti G., Michon F.-X., El Feghaly E., Nouguier M., Robbe D., Koenig J., and Epsztein J. (2019). Dynamic control of hippocampal spatial coding resolution by local visual cues. Elife 8, e44487. 10.7554/eLife.44487. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Geiller T., Fattahi M., Choi J.S., and Royer S. (2017). Place cells are more strongly tied to landmarks in deep than in superficial CA1. Nat Commun 8, doi 10.1038/ncomms14531. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Mizuta K., Nakai J., Hayashi Y., and Sato M. (2021). Multiple coordinated cellular dynamics mediate CA1 map plasticity. Hippocampus 31, 235–243. 10.1002/hipo.23300. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Youngstrom I.A., and Strowbridge B.W. (2012). Visual landmarks facilitate rodent spatial navigation in virtual reality environments. Learn Mem 19, 84–90. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Keinath A.T., Mosser C.-A., and Brandon M.P. (2022). The representation of context in mouse hippocampus is preserved despite neural drift. Nat Commun 13, 2415. 10.1038/s41467-022-30198-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.Khatib D., Ratzon A., Sellevoll M., Barak O., Morris G., and Derdikman D. (2023). Active experience, not time, determines within-day representational drift in dorsal CA1. Neuron 111, 2348–2356.e5. 10.1016/j.neuron.2023.05.014. [DOI] [PubMed] [Google Scholar]
  • 83.Geva N., Deitch D., Rubin A., and Ziv Y. (2023). Time and experience differentially affect distinct aspects of hippocampal representational drift. Neuron 111, 2357–2366.e5. 10.1016/j.neuron.2023.05.005. [DOI] [PubMed] [Google Scholar]
  • 84.Ziv Y., Burns L.D., Cocker E.D., Hamel E.O., Ghosh K.K., Kitch L.J., El Gamal A., and Schnitzer M.J. (2013). Long-term dynamics of CA1 hippocampal place codes. Nature neuroscience 16, 264–266. 10.1038/nn.3329. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Fenton A.A., and Muller R.U. (1998). Place cell discharge is extremely variable during individual passes of the rat through the firing field. Proc. Natl. Acad. Sci. USA 95, 3182–3187. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86.Nagelhus A., Andersson S.O., Cogno S.G., Moser E.I., and Moser M.-B. (2023). Object-centered population coding in CA1 of the hippocampus. Neuron 111, 2091–2104.e14. 10.1016/j.neuron.2023.04.008. [DOI] [PubMed] [Google Scholar]
  • 87.Williams A.H., Poole B., Maheswaranathan N., Dhawale A.K., Fisher T., Wilson C.D., Brann D.H., Trautmann E.M., Ryu S., Shusterman R., et al. (2020). Discovering Precise Temporal Patterns in Large-Scale Neural Recordings through Robust and Interpretable Time Warping. Neuron 105, 246–259.e8. 10.1016/j.neuron.2019.10.020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Tseng S.-Y., Chettih S.N., Arlt C., Barroso-Luque R., and Harvey C.D. (2022). Shared and specialized coding across posterior cortical areas for dynamic navigation decisions. Neuron 110, 2484–2502.e16. 10.1016/j.neuron.2022.05.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Aoki Y., Igata H., Ikegaya Y., and Sasaki T. (2019). The Integration of Goal-Directed Signals onto Spatial Maps of Hippocampal Place Cells. Cell Rep 27, 1516–1527 e5. 10.1016/j.celrep.2019.04.002. [DOI] [PubMed] [Google Scholar]
  • 90.Sarel A., Finkelstein A., Las L., and Ulanovksy N. (2017). Vectorial representation of spatial goals in the hippocampus of bats. Science 355, 176–180. [DOI] [PubMed] [Google Scholar]
  • 91.Driscoll L.N., Petit N.L., Minderer M., Chettih S.N., and Harvey C.D. (2017). Dynamic reorganization of neuronal activity patterns in parietal cortex. Cell 170, 986–999. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92.Driscoll L.N., Duncker L., and Harvey C.D. (2022). Representational drift: Emerging theories for continual learning and experimental future directions. Current Opinion in Neurobiology 76, 102609. 10.1016/j.conb.2022.102609. [DOI] [PubMed] [Google Scholar]
  • 93.Deitch D., Rubin A., and Ziv Y. (2021). Representational drift in the mouse visual cortex. Curr Biol 31, 4327–4339.e6. 10.1016/j.cub.2021.07.062. [DOI] [PubMed] [Google Scholar]
  • 94.Benna M.K., and Fusi S. (2021). Place cells may simply be memory cells: Memory compression leads to spatial tuning and history dependence. Proceedings of the National Academy of Sciences 118, e2018422118. 10.1073/pnas.2018422118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95.Marr D. (1971). Simple memory: a theory for archicortex. Philos.Trans.R.Soc.Lond. 262, 23–81. [DOI] [PubMed] [Google Scholar]
  • 96.Fusi S., and Abbott L.F. (2007). Limits on the memory storage capacity of bounded synapses. Nature Neuroscience 10, 485–493. 10.1038/nn1859. [DOI] [PubMed] [Google Scholar]
  • 97.Susman L., Brenner N., and Barak O. (2019). Stable memory with unstable synapses. Nat Commun 10, 4441. 10.1038/s41467-019-12306-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98.Mankin E.A., Sparks F.T., Slayyeh B., Sutherland R.J., Leutgeb S., and Leutgeb J.K. (2012). Neuronal code for extended time in the hippocampus. Proceedings of the National Academy of Sciences of the United States of America 109, 19462–19467. 10.1073/pnas.1214107109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99.Jin S.-W., and Lee I. (2021). Differential encoding of place value between the dorsal and intermediate hippocampus. Current Biology 31, 3053–3072.e5. 10.1016/j.cub.2021.04.073. [DOI] [PubMed] [Google Scholar]
  • 100.Jarzebowski P., Hay Y.A., Grewe B.F., and Paulsen O. (2022). Different encoding of reward location in dorsal and intermediate hippocampus. Curr Biol 32, 834–841.e5. 10.1016/j.cub.2021.12.024. [DOI] [PubMed] [Google Scholar]
  • 101.Kinsky N.R., Mau W., Sullivan D.W., Levy S.J., Ruesch E.A., and Hasselmo M.E. (2020). Trajectory-modulated hippocampal neurons persist throughout memory-guided navigation. Nature Communications 11. 10.1038/s41467-020-16226-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 102.Grieves R.M., Wood E.R., and Dudchenko P.A. (2016). Place cells on a maze encode routes rather than destinations. Elife 5. 10.7554/eLife.15986. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 103.Lee I., Griffin A.L., Zilli E.A., Eichenbaum H., and Hasselmo M.E. (2006). Gradual translocation of spatial correlates of neuronal firing in the hippocampus toward prospective reward locations. Neuron 51, 639–650. [DOI] [PubMed] [Google Scholar]
  • 104.Wood E.R., Dudchenko P.A., Robitsek R.J., and Eichenbaum H. (2000). Hippocampal neurons encode information about different types of memory episodes occurring in the same location. Neuron 27, 623–633. [DOI] [PubMed] [Google Scholar]
  • 105.Frank L.M., Brown E.N., and Wilson M. (2000). Trajectory encoding in the hippocampus and entorhinal cortex. Neuron 27, 169–178. [DOI] [PubMed] [Google Scholar]
  • 106.Ferbinteanu J., and Shapiro M.L. (2003). Prospective and retrospective memory coding in the hippocampus. Neuron 40, 1227–1239. [DOI] [PubMed] [Google Scholar]
  • 107.Ormond J., and O’Keefe J. (2022). Hippocampal place cells have goal-oriented vector fields during navigation. Nature 607, 741–746. 10.1038/s41586-022-04913-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 108.Wikenheiser A.M., and Redish A.D. (2011). Changes in reward contingency modulate the trial-to-trial variability of hippocampal place cells. Journal of neurophysiology 106, 589–598. 10.1152/jn.00091.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109.Wikenheiser A.M., and Redish A.D. (2015). Hippocampal theta sequences reflect current goals. Nat Neurosci 18, 289–294. 10.1038/nn.3909. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 110.McKenzie S., Robinson N.T., Herrera L., Churchill J.C., and Eichenbaum H. (2013). Learning causes reorganization of neuronal firing patterns to represent related experiences within a hippocampal schema. The Journal of neuroscience : the official journal of the Society for Neuroscience 33, 10243–10256. 10.1523/JNEUROSCI.0879-13.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 111.Priestley J.B., Bowler J.C., Rolotti S.V., Fusi S., and Losonczy A. (2022). Signatures of rapid plasticity in hippocampal CA1 representations during novel experiences. Neuron 110, 1978–1992.e6. 10.1016/j.neuron.2022.03.026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 112.Pettit N.L., Yuan X.C., and Harvey C.D. (2022). Hippocampal place codes are gated by behavioral engagement. Nat Neurosci 25, 561–566. 10.1038/s41593-022-01050-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 113.Krishnan S., Heer C., Cherian C., and Sheffield M.E.J. (2022). Reward expectation extinction restructures and degrades CA1 spatial maps through loss of a dopaminergic reward proximity signal. Nat Commun 13, 6662. 10.1038/s41467-022-34465-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 114.Musall S., Kaufman M.T., Juavinett A.L., Gluf S., and Churchland A.K. (2019). Single-trial neural dynamics are dominated by richly varied movements. Nat Neurosci 22, 1677–1686. 10.1038/s41593-019-0502-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 115.Sadeh S., and Clopath C. (2022). Contribution of behavioural variability to representational drift. eLife 11, e77907. 10.7554/eLife.77907. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 116.Liberti W.A., Schmid T.A., Forli A., Snyder M., and Yartsev M.M. (2022). A stable hippocampal code in freely flying bats. Nature 604, 98–103. 10.1038/s41586-022-04560-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 117.Qian F.K., Li Y., and Magee J.C. (2023). Experience-dependent place-cell referencing in hippocampal area CA1. Preprint at bioRxiv, 10.1101/2023.11.23.568469 10.1101/2023.11.23.568469. [DOI] [Google Scholar]
  • 118.Knierim J.J., Kudrimoti H.S., and McNaughton B.L. (1996). Neuronal mechanisms underlying the interaction between visual landmarks and path integration in the rat. Int.J.Neural Syst. 7, 213–218. [DOI] [PubMed] [Google Scholar]
  • 119.Etienne A.S., and Jeffery K. (2004). Path integration in mammals. Hippocampus 14, 180–192. [DOI] [PubMed] [Google Scholar]
  • 120.Wang C., Chen X., Lee H., Deshmukh S.S., Yoganarasimha D., Savelli F., and Knierim J.J. (2018). Egocentric coding of external items in the lateral entorhinal cortex. Science 362, 945–949. 10.1126/science.aau4940. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 121.Alexander A.S., Robinson J.C., Stern C.E., and Hasselmo M.E. (2023). Gated transformations from egocentric to allocentric reference frames involving retrosplenial cortex, entorhinal cortex, and hippocampus. Hippocampus 33, 465–487. 10.1002/hipo.23513. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 122.Rosen Z.B., Cheung S., and Siegelbaum S.A. (2015). Midbrain dopamine neurons bidirectionally regulate CA3-CA1 synaptic drive. Nat Neurosci 18, 1763–1771. 10.1038/nn.4152. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 123.Palacios-Filardo J., and Mellor J.R. (2019). Neuromodulation of hippocampal long-term synaptic plasticity. Curr Opin Neurobiol 54, 37–43. 10.1016/j.conb.2018.08.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 124.Loughlin S.E., Foote S.L., and Grzanna R. (1986). Efferent projections of nucleus locus coeruleus: morphologic subpopulations have different efferent targets. Neuroscience 18, 307–319. 10.1016/0306-4522(86)90156-9. [DOI] [PubMed] [Google Scholar]
  • 125.Kaufman A.M., Geiller T., and Losonczy A. (2020). A Role for the Locus Coeruleus in Hippocampal CA1 Place Cell Reorganization during Spatial Reward Learning. Neuron 105, 1018–1026 e4. 10.1016/j.neuron.2019.12.029. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 126.Mamad O., Stumpp L., McNamara H.M., Ramakrishnan C., Deisseroth K., Reilly R.B., and Tsanov M. (2017). Place field assembly distribution encodes preferred locations. PLoS Biol 15, e2002365. 10.1371/journal.pbio.2002365. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 127.Gasbarri A., Packard M.G., Campana E., and Pacitti C. (1994). Anterograde and retrograde tracing of projections from the ventral tegmental area to the hippocampal formation in the rat. Brain Res. Bull. 33, 445–452. [DOI] [PubMed] [Google Scholar]
  • 128.Michon F., Krul E., Sun J.-J., and Kloosterman F. (2021). Single-trial dynamics of hippocampal spatial representations are modulated by reward value. Curr Biol 31, 4423–4435.e5. 10.1016/j.cub.2021.07.058. [DOI] [PubMed] [Google Scholar]
  • 129.Euston D.R., Gruber A.J., and McNaughton B.L. (2012). The role of medial prefrontal cortex in memory and decision making. Neuron 76, 1057–1070. 10.1016/j.neuron.2012.12.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 130.Tang W., and Jadhav S.P. (2022). Multiple-Timescale Representations of Space: Linking Memory to Navigation. Annual Review of Neuroscience 45, 1–21. 10.1146/annurev-neuro-111020-084824. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 131.Ito H.T., Zhang S., Witter M.P., Moser E.I., and Moser M.B. (2015). A prefrontal-thalamo-hippocampal circuit for goal directed spatial navigation. Nature 522, 50–55. [DOI] [PubMed] [Google Scholar]
  • 132.Srinivasan A., Srinivasan A., Riceberg J.S., Goodman M.R., Guise K.G., and Shapiro M.L. (2023). Hippocampal and medial prefrontal ensemble spiking represents episodes and rules in similar task spaces. Cell Rep 42, 113296. 10.1016/j.celrep.2023.113296. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 133.Jadhav S.P., Rothschild G., Roumis D.K., and Frank L.M. (2016). Coordinated Excitation and Inhibition of Prefrontal Ensembles during Awake Hippocampal Sharp-Wave Ripple Events. Neuron 90, 113–127. 10.1016/j.neuron.2016.02.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 134.Tang W., Shin J.D., and Jadhav S.P. (2023). Geometric transformation of cognitive maps for generalization across hippocampal-prefrontal circuits. Cell Rep 42, 112246. 10.1016/j.celrep.2023.112246. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 135.El-Gaby M., Harris A.L., Whittington J.C.R., Dorrell W., Bhomick A., Walton M.E., Akam T., and Behrens T.E.J. (2023). A Cellular Basis for Mapping Behavioural Structure. Preprint at bioRxiv, 10.1101/2023.11.04.565609 10.1101/2023.11.04.565609. [DOI] [Google Scholar]
  • 136.Basu R., Gebauer R., Herfurth T., Kolb S., Golipour Z., Tchumatchenko T., and Ito H.T. (2021). The orbitofrontal cortex maps future navigational goals. Nature 599, 449–452. 10.1038/s41586-021-04042-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 137.Boccara C.N., Nardin M., Stella F., O’Neill J., and Csicsvari J. (2019). The entorhinal cognitive map is attracted to goals. Science 363, 1443–1447. 10.1126/science.aav4837. [DOI] [PubMed] [Google Scholar]
  • 138.Butler W.N., Hardcastle K., and Giocomo L.M. (2019). Remembered reward locations restructure entorhinal spatial maps. Science 363, 1447–1452. 10.1126/science.aav5297. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 139.Bowler J.C., and Losonczy A. (2023). Direct cortical inputs to hippocampal area CA1 transmit complementary signals for goal-directed navigation. Neuron 111, 4071–4085.e6. 10.1016/j.neuron.2023.09.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 140.Issa J.B., Radvansky B.A., Xuan F., and Dombeck D.A. (2024). Lateral entorhinal cortex subpopulations represent experiential epochs surrounding reward. Nat Neurosci, 1–11. 10.1038/s41593-023-01557-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 141.Lee J.Y., Jun H., Soma S., Nakazono T., Shiraiwa K., Dasgupta A., Nakagawa T., Xie J.L., Chavez J., Romo R., et al. (2021). Dopamine facilitates associative memory encoding in the entorhinal cortex. Nature 598, 321–326. 10.1038/s41586-021-03948-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 142.Lavoie A.M., and Mizumori S.J. (1994). Spatial, movement- and reward-sensitive discharge by medial ventral striatum neurons of rats. Brain Res. 638, 157–168. [DOI] [PubMed] [Google Scholar]
  • 143.Trouche S., Koren V., Doig N.M., Ellender T.J., El-Gaby M., Lopes-Dos-Santos V., Reeve H.M., Perestenko P.V., Garas F.N., Magill P.J., et al. (2019). A Hippocampus-Accumbens Tripartite Neuronal Motif Guides Appetitive Memory in Space. Cell 176, 1393–1406 e16. 10.1016/j.cell.2018.12.037. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 144.Sosa M., Joo H.R., and Frank L.M. (2020). Dorsal and Ventral Hippocampal Sharp-Wave Ripples Activate Distinct Nucleus Accumbens Networks. Neuron 105, 725–741 e8. 10.1016/j.neuron.2019.11.022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 145.Floresco S.B. (2015). The nucleus accumbens: an interface between cognition, emotion, and action. Annu Rev Psychol 66, 25–52. 10.1146/annurev-psych-010213-115159. [DOI] [PubMed] [Google Scholar]
  • 146.McGinty V.B., Lardeux S., Taha S.A., Kim J.J., and Nicola S.M. (2013). Invigoration of reward seeking by cue and proximity encoding in the nucleus accumbens. Neuron 78, 910–922. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 147.van der Meer M.A., and Redish A.D. (2011). Theta phase precession in rat ventral striatum links place and reward information. J. Neurosci. 31, 2843–2854. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 148.Harvey R.E., Robinson H.L., Liu C., Oliva A., and Fernandez-Ruiz A. (2023). Hippocampo-cortical circuits for selective memory encoding, routing, and replay. Neuron, S0896–6273(23)00300–8. 10.1016/j.neuron.2023.04.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 149.Barnstedt O., Mocellin P., and Remy S. (2023). A hippocampus-accumbens code guides goal-directed appetitive behavior. Preprint at bioRxiv, 10.1101/2023.03.09.531869 10.1101/2023.03.09.531869. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 150.Danielson N.B., Zaremba J.D., Kaifosh P., Bowler J., Ladow M., and Losonczy A. (2016). Sublayer-Specific Coding Dynamics during Spatial Navigation and Learning in Hippocampal Area CA1. Neuron 91, 652–665. 10.1016/j.neuron.2016.06.020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 151.Bartolo R., and Averbeck B.B. (2021). Inference as a fundamental process in behavior. Curr Opin Behav Sci 38, 8–13. 10.1016/j.cobeha.2020.06.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 152.Hennig J.A., Pinto S.A.R., Yamaguchi T., Linderman S.W., Uchida N., and Gershman S.J. (2023). Emergence of belief-like representations through reinforcement learning. Preprint at bioRxiv, 10.1101/2023.04.04.535512 10.1101/2023.04.04.535512. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 153.Sun W., Winnubst J., Natrajan M., Lai C., Kajikawa K., Michaelos M., Gattoni R., Stringer C., Flickinger D., Fitzgerald J.E., et al. (2023). Learning produces a hippocampal cognitive map in the form of an orthogonalized state machine. Preprint at bioRxiv, 10.1101/2023.08.03.551900 10.1101/2023.08.03.551900. [DOI] [Google Scholar]
  • 154.Beron C.C., Neufeld S.Q., Linderman S.W., and Sabatini B.L. (2022). Mice exhibit stochastic and efficient action switching during probabilistic decision making. Proc Natl Acad Sci U S A 119, e2113961119. 10.1073/pnas.2113961119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 155.Sanders H., Wilson M.A., and Gershman S.J. (2020). Hippocampal remapping as hidden state inference. eLife 9, e51140. 10.7554/eLife.51140. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 156.Eichenbaum H. (2000). A cortical-hippocampal system for declarative memory. Nat Rev Neurosci 1, 41−-50. [DOI] [PubMed] [Google Scholar]
  • 157.Stachenfeld K.L., Botvinick M.M., and Gershman S.J. (2017). The hippocampus as a predictive map. Nat Neurosci 20, 1643–1653. 10.1038/nn.4650. [DOI] [PubMed] [Google Scholar]
  • 158.Tanni S., de Cothi W., and Barry C. (2022). State transitions in the statistically stable place cell population correspond to rate of perceptual change. Curr Biol 32, 3505–3514.e7. 10.1016/j.cub.2022.06.046. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 159.Bakermans J.J.W., Warren J., Whittington J.C.R., and Behrens T.E.J. (2023). Constructing future behaviour in the hippocampal formation through composition and replay. Preprint at bioRxiv, 10.1101/2023.04.07.536053 10.1101/2023.04.07.536053. [DOI] [Google Scholar]
  • 160.Kurth-Nelson Z., Behrens T., Wayne G., Miller K., Luettgau L., Dolan R., Liu Y., and Schwartenbeck P. (2023). Replay and compositional computation. Neuron 111, 454–469. 10.1016/j.neuron.2022.12.028. [DOI] [PubMed] [Google Scholar]
  • 161.Duncker L., Driscoll L., Shenoy K.V., Sahani M., and Sussillo D. (2020). Organizing recurrent network dynamics by task-computation to enable continual learning. In Advances in Neural Information Processing Systems (Curran Associates, Inc.), pp. 14387–14397. [Google Scholar]
  • 162.McNamee D.C., Stachenfeld K.L., Botvinick M.M., and Gershman S.J. (2021). Flexible modulation of sequence generation in the entorhinal-hippocampal system. Nat. Neurosci. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 163.Herweg N.A., and Kahana M.J. (2018). Spatial Representations in the Human Brain. Frontiers in Human Neuroscience 12. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 164.Umbach G., Kantak P., Jacobs J., Kahana M., Pfeiffer B.E., Sperling M., and Lega B. (2020). Time cells in the human hippocampus and entorhinal cortex support episodic memory. Proceedings of the National Academy of Sciences 117, 28463–28474. 10.1073/pnas.2013250117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 165.Constantinescu A.O., O’Reilly J.X., and Behrens T.E. (2016). Organizing conceptual knowledge in humans with a gridlike code. Science 352, 1464–1468. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 166.Theves S., Fernández G., and Doeller C.F. (2020). The Hippocampus Maps Concept Space, Not Feature Space. J Neurosci 40, 7318–7325. 10.1523/JNEUROSCI.0494-20.2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 167.Muhle-Karbe P.S., Sheahan H., Pezzulo G., Spiers H., Chien S., Schuck N.W., and Summerfield C. (2023). Goal-seeking compresses neural codes for space in the human hippocampus and orbitofrontal cortex. Preprint at bioRxiv, 10.1101/2023.01.12.523762 10.1101/2023.01.12.523762. [DOI] [PubMed] [Google Scholar]
  • 168.Behrens T.E.J., Muller T.H., Whittington J.C.R., Mark S., Baram A.B., Stachenfeld K.L., and Kurth-Nelson Z. (2018). What Is a Cognitive Map? Organizing Knowledge for Flexible Behavior. Neuron 100, 490–509. 10.1016/j.neuron.2018.10.002. [DOI] [PubMed] [Google Scholar]
  • 169.Bellmund J.L.S., de Cothi W., Ruiter T.A., Nau M., Barry C., and Doeller C.F. (2020). Deforming the metric of cognitive maps distorts memory. Nat Hum Behav 4, 177–188. 10.1038/s41562-019-0767-3. [DOI] [PubMed] [Google Scholar]
  • 170.Scoville W.B., and Milner B. (1957). Loss of recent memory after bilateral hippocampal lesions. J.Neurol.Neurosurg.Psychiatry 20, 11–21. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 171.Zeithamova D., Gelman B.D., Frank L., and Preston A.R. (2018). Abstract Representation of Prospective Reward in the Hippocampus. J Neurosci 38, 10093–10101. 10.1523/JNEUROSCI.0719-18.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 172.Braun E.K., Wimmer G.E., and Shohamy D. (2018). Retroactive and graded prioritization of memory by reward. Nat Commun 9, 4886. 10.1038/s41467-018-07280-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 173.Dombeck D.A., Harvey C.D., Tian L., Looger L.L., and Tank D.W. (2010). Functional imaging of hippocampal place cells at cellular resolution during virtual navigation. Nat Neurosci 13, 1433–1440. 10.1038/nn.2648. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 174.Pachitariu M., Stringer C., Dipoppa M., Schröder S., Rossi L.F., Dalgleish H., Carandini M., and Harris K.D. (2017). Suite2p: beyond 10,000 neurons with standard two-photon microscopy. Preprint at bioRxiv, 10.1101/061507 10.1101/061507. [DOI] [Google Scholar]
  • 175.Friedrich J., Zhou P., and Paninski L. (2017). Fast online deconvolution of calcium imaging data. PLOS Computational Biology 13, e1005423. 10.1371/journal.pcbi.1005423. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 176.Berens P. (2009). CircStat: a MATLAB toolbox for circular statistics. Journal of Statistical Software 31, 1–21. [Google Scholar]
  • 177.Kempter R., Leibold C., Buzsaki G., Diba K., and Schmidt R. (2012). Quantifying cicular-linear associations: hippocampal phase precession. J Neurosci Methods 207, 113–124. [DOI] [PubMed] [Google Scholar]
  • 178.Skaggs W.E., McNaughton B.L., Gothard K., and Markus E. (1993). An information theoretic approach to deciphering the hippocampal code. In Advanced in Neural Information Processing Systems, Hanson S., Cowan J. D., and Giles C. L., eds. (Morgan Kaufmann Publishers; ), pp. 1030–1037. [Google Scholar]
  • 179.Mao D., Neumann A.R., Sun J., Bonin V., Mohajerani M.H., and McNaughton B.L. (2018). Hippocampus-dependent emergence of spatial sequence coding in retrosplenial cortex. Proc Natl Acad Sci U S A 115, 8015–8018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 180.Grijseels D.M., Shaw K., Barry C., and Hall C.N. (2021). Choice of method of place cell classification determines the population of cells identified. PLOS Computational Biology 17, e1008835. 10.1371/journal.pcbi.1008835. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 181.Jammalamadaka S.R., Sengupta A., and Sengupta A. (2001). Topics in Circular Statistics (World Scientific; ). [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplement 1

Articles from bioRxiv are provided here courtesy of Cold Spring Harbor Laboratory Preprints

RESOURCES