Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Nov 1.
Published in final edited form as: Hippocampus. 2020 Aug 18;30(11):1194–1208. doi: 10.1002/hipo.23250

Spatial encoding in dorsomedial prefrontal cortex and hippocampus is related during deliberation

Brendan M Hasz 1, A David Redish 2,*
PMCID: PMC7606780  NIHMSID: NIHMS1635357  PMID: 32809246

Abstract

Deliberation is thought to involve the internal simulation of the outcomes of candidate actions, the valuation of those outcomes, and the selection of the actions with the highest expected value. While it is known that deliberation involves prefrontal cortical areas, specifically the dorsomedial prefrontal cortex (dmPFC), as well as the hippocampus (HPC) and other brain regions, how these areas process prospective information and select actions is not well understood. We recorded simultaneously from ensembles in dmPFC and CA1 of dorsal HPC in rats during performance of a spatial contingency switching task, and examined the relationships between spatial and reward encoding in these two areas during deliberation at the choice point. We found that CA1 and dmPFC represented either goal locations or the current position simultaneously, but that when goal locations were encoded, HPC and dmPFC did not always represent the same goal location. Ensemble activity in dmPFC predicted when HPC would represent goal locations, but on a broad timescale on the order of seconds. Also, reward encoding in dmPFC increased during hippocampal theta cycles where CA1 ensembles represented the goal location. These results suggest that dmPFC and HPC share prospective information during deliberation, that dmPFC may influence whether HPC represents prospective information, and that information recalled about goal locations by HPC may be integrated into dmPFC reward representations on fast timescales.

Keywords: dorsomedial prefrontal cortex, hippocampus, deliberation, spatial encoding, value

1. Introduction

Deliberation is an approach to decision-making where the values of potential actions are estimated and compared to select the expected best action. Value-based decision making is thought to involve a deliberative process where outcomes of potential actions are estimated, the value of those hypothetical outcomes are evaluated and compared, and then the action with the highest expected value is taken (Gilbert and Wilson, 2007; Rangel et al., 2008; Padoa-Schioppa, 2011; Redish, 2013). However, the neural processes by which the brain performs deliberation are complex and poorly understood.

Deliberation is known to depend on the prefrontal cortex (PFC), hippocampus (HPC), ventral and dorsomedial striatum, among other areas (Doll et al., 2012; van der Meer et al., 2012). In rodents, the representation of prospective information by ensembles in HPC has been observed during deliberation at choice points (Johnson and Redish, 2007; Kay et al., 2020), which is thought to correspond to the simulation of the outcomes of candidate actions using an internal model stored in part by HPC (Redish, 2016). Furthermore, neural activity thought to correspond to the estimation of the value of these simulated outcomes has been observed in ventral striatum and in orbitofrontal cortex (van der Meer and Redish, 2010; Steiner and Redish, 2012; Stott and Redish, 2014; Steiner and Redish, 2014; Rich and Wallis, 2016; Wallis, 2018). However, it is unclear how these valuations are used to select between candidate actions, where and how these internal simulations are initiated, as well as where the information corresponding to the predicted outcomes are integrated.

Candidate brain regions for performing some or all of these roles include the various subregions of the prefrontal cortex (PFC). The prefrontal cortex has long been thought to support executive function (Miller and Cohen, 2001; Dalley et al., 2004; Kesner and Churchwell, 2011; Fuster, 2015). Areas of the dorsomedial prefrontal cortex (dmPFC) such as anterior cingulate cortex, have been implicated in conflict detection (Botvinick et al., 2001; Haddon and Killcross, 2005, 2006; Marquis et al., 2007; Dwyer et al., 2010; Shenhav et al., 2013), suggesting it may be responsible for detecting the need for deliberative control. Also, dmPFC has been found to play an active role in the storage and recall of working memories (Fuster and Alexander, 1971; Tronel and Sara, 2003; Ragozzino and Kesner, 1998; Delatour and Gisquest-Verrier, 1999; Cowen and McNaughton, 2007; Yoon et al., 2008; Horst and Laubach, 2009; Euston et al., 2012; Preston and Eichenbaum, 2013; Urban et al., 2014), which may also translate to the storage of internally simulated outcome valuations. Some have theorized that dmPFC may initiate the internal construction of hypothetical situations, and use valuations of hypothetical outcomes to perform action selection (Hassabis and Maguire, 2009; van der Meer et al., 2012; Wang et al., 2015).

HPC and dmPFC, along with other structures, are thought to form an information-processing loop where top-down signals from dmPFC influence encoding in HPC, and information retrieved by HPC informs prefrontal representations (Wang et al., 2015; Jai and Frank, 2015; Redish, 2016; Shin and Jadhav, 2016; Eichenbaum, 2017). Lesion and inactivation studies have found that interrupting communication between PFC and HPC disrupts animals’ spatial working memory (Floresco et al., 1997; Wang and Cai, 2006; Barker et al., 2017; Maharjan et al., 2018). Also, theta oscillations in HPC and PFC are coherent during deliberation and appear to be important for spatial working memory (Siapas et al., 2005; Jones and Wilson, 2005; Hyman et al., 2005; Benchenane et al., 2010; O’Neill et al., 2013). Furthermore, inactivating the medial PFC changes information representation by hippocampal place cells (Hok et al., 2013; Navawongse and Eichenbaum, 2013; Guise and Shapiro, 2017; Schmidt et al., 2019). There are both monosynaptic and polysynaptic projections from CA1 to dmPFC (Swanson, 1981; Ferino et al., 1987; Jay and Witter, 1991; Verwer et al., 1997; Delatour and Witter, 2002; Floresco and Grace, 2003; Hoover and Vertes, 2007), as well as a bisynaptic and bidirectional connections between CA1 and dmPFC via the nucleus reuniens and other thalamic nuclei (Vertes, 2002, 2004; McKenna and Vertes, 2004; Vertes et al., 2006; Di Prisco and Vertes, 2006; Vertes et al., 2007; Hoover and Vertes, 2012; Cassel et al., 2013; Dolleman-Van der Weel et al., 2017; Dolleman-van der Weel et al., 2019). Therefore, it is very likely that HPC and dmPFC share information relevant to deliberative decision-making on fast timescales.

However, it is unclear what specific information HPC and dmPFC share, and how quickly any information transfer occurs. Theories of the deliberative system suggest that prefrontal areas detect a need for deliberative control, instigate internal simulations of the outcomes of candidate actions (perhaps via HPC), keep track of the valuations of those outcomes (valuations which are perhaps generated by ventral stratum or orbitofrontal cortex), and use that information to decide which candidate action to enact. Therefore, likely candidates for information being passed between dmPFC and HPC include information about candidate actions, location, and reward or value. Previous work has discovered that hippocampal ensembles represent non-local spatial information which appears to correspond to internal simulations of candidate actions (Johnson and Redish, 2007; Kay et al., 2020), and other areas such as ventral striatum and orbitofrontal cortex represent value in ways suggesting they may be estimating the value of these internally simulated outcomes (van der Meer and Redish, 2010; Steiner and Redish, 2012; Stott and Redish, 2014; Steiner and Redish, 2014; Rich and Wallis, 2016; Wallis, 2018). But what instigates these internal simulations? If dmPFC plays this role, then it should be possible to predict from activity in dmPFC whether non-local information is about to be represented in HPC. Also, if dmPFC is keeping track of the predicted values of candidate actions, then when the outcomes of these candidate actions are represented in HPC, downstream estimates of the value of these outcomes should affect reward encoding in dmPFC.

Here, we recorded simultaneously from ensembles in dmPFC and in CA1 of dorsal HPC in rats during performance of a spatial decision-making task. This task was a contingency switching task, where rats were required to choose actions at the choice point of the maze which was consistent with the current rule. The rule switched occasionally, preventing rats from simply automating their behavior on the task. We investigated the relationships between information encoded about spatial location and reward in dmPFC and CA1 as rats deliberated at the choice point of the maze. Spatial encoding by ensembles in dmPFC and CA1 was correlated during deliberation. While both dmPFC and CA1 encoded either prospective information or local information simultaneously, the prospective information encoded was not always identical. We found that activity in dmPFC predicted whether ensembles in HPC represented goal locations or the current location, but that the time scale of this relationship was broad, on the order of seconds. We also found that reward representation in dmPFC increased as CA1 ensembles represented the goal location, but that this increase in reward encoding was restricted to the hippocampal theta cycle during which prospective information was encoded in HPC. These results suggest that any influence of dmPFC on hippocampal prospective spatial representations may be temporally diffuse, while information about goal locations recalled by hippocampus may be integrated into prefrontal reward representations more quickly, on the order of single theta cycles.

2. Methods

2.1. Subjects

Six FBNF-1 rats aged 8–14 months at the beginning of behavior were subjects for the experiment (4 male, 2 female), bred from Fischer and Brown Norway rats (Envigo). Rats were housed on a 12 h light-dark cycle, and experimental sessions were conducted at the same time each day during the light phase. All experimental and animal care procedures complied with US National Institutes of Health guidelines for animal care and were approved by the Institutional Animal Care and Use Committee and the University of Minnesota.

2.2. Training and behavior

The contingency switching task was a spatial reversal task where rats were required to adjust their behavioral strategies after uncued rule changes. The maze consisted of several low-cost choice points followed by a high-cost choice point between two actions: left or right (Figure 1A). The maze was constructed using LEGO Duplo blocks on a white canvas. The configuration of the low-cost choice points at the center of the maze was determined by a single wall in the middle of the maze, which switched back and forth from the left to right side randomly each day. Each lap, if rats chose the action at the high-cost choice point which was consistent with the current contingency, they received one unflavored food pellet (45 mg each, Test Diet, Richmond, IN) at one of two reward sites on the side of the maze, and an additional food pellet at the rear of the maze. If their choice was inconsistent with the current contingency, no reward was delivered and rats had to circle around to the start of the maze to initiate a new lap. Two different auditory cues also signalled to the rats whether their decision was correct or incorrect: a swept-frequency sinewave “chirp” from 1kHz to 3kHz for correct, and two shorter 1kHz square wave tones for incorrect. The contingency on any given lap was either Left (only left choices at the choice point lead to reward), Right (only right choices), or Alternate (the opposite choice from the previous lap was required for reward).

Figure 1 :

Figure 1 :

Rat behavior on the contingency switching task. (A) The contingency switching task is a spatial maze consisting of a central maze segment, a choice point, and reward zones where rats receive rewards dependent on making choices consistent with the current contingency. (B) Example of a pass through the choice point where the rat displayed VTE and (C) a non-VTE pass. Bold lines show the position of the rat’s head during a pass through the choice point. Gray dots show the position of the rat’s head throughout the entire session. (D) Contingencies are presented in blocks of laps lasting 30 ± 5 trials. (E) Example behavioral data from a single session.

The first contingency type of a session was selected randomly, and contingencies were presented in blocks: every 30 ± 5 laps, the contingency changed randomly to one of the other two contingencies. Rats were allowed to run laps freely on the task for one hour each day, and their daily food allowance came only from performing the task. However, rats were fed extra food after running the task if their weight dropped below 80% of their free-feeding weight. During the experiment, none of the 6 rats dropped below 80% of their free-feeding weight, and so no post-feeding was required, but rats were post-fed after 7 out of 212 sessions during training (3%).

Rat positions on the maze were tracked using a video camera (Cohu, Inc., San Diego, CA) placed above the maze. Custom Matlab (The MathWorks, Natick, MA) software determined animal position from the video, and controlled the state of the task (the current contingency, food pellet release, the presentation of audio cues, etc). The Matlab software also interfaced with an Arduino Uno Rev3 which ran custom software and triggered the release of food pellets from food pellet dispensers (MedAssociates, St. Albans, VT).

Rats were trained over the course of four weeks. Starting on the first week, rats were deprived of the freely available food in their home cages, but continued to have free access to water. Rats were handled and offered up to 15g of food pellets each day for half an hour, to train them to eat the food pellets which would be available while performing the contingency switching task. For the second week, rats performed a simplified version of the task where the contingency was either Left or Right, and the contingency stayed constant throughout each session but changed randomly from session to session. Rats were rewarded with 2 food pellets per feeder at all feeder sites for the second week. For the third week, again there were no within-session contingency switches, but all three contingencies were possible (including Alternate), and only 1 food pellet was delivered at the rear food delivery site. For the final week of training, only 1 food pellet was delivered at all feeder sites, but the task was otherwise the same as during week 3.

After training, rats were given free access to food for at least 3 days, and then surgerized. After 3 days of post-surgery recovery, rats were again food deprived and re-trained for 1–2 weeks on the final training phase of the task (all 3 contingencies possible, but no within-session contingency switches, and 1 pellet per feeder). Finally, rats performed the full version of the task including within-session contingency switches and neural recordings for 2–3 weeks.

2.3. Surgery and electrode targets

After training, rats were given free access to food for at least 3 days, and then chronically implanted with a hyperdrive containing 24 tetrodes, and a separate drive containing a 32-site silicon probe (Cambridge NeuroTech, Cambridge, England). The hyperdrives contained two bundles of 12 tetrodes each, targeting the CA1 region of dorsal hippocampus bilaterally (3.8 mm posterior and ± 3.0 mm lateral from bregma). The hyperdrive for one rat contained a single bundle of 24 tetrodes targeting the right hippocampus. The silicon probes consisted of two 16-site shanks which were implanted 3.8 mm anterior to and 0.7 mm lateral from bregma at a 25 degree angle (targeting dmPFC on the right hemisphere, such that the final target was 2.3 mm A/P, −0.7 mm M/L, and 3.9mm D/V, all coordinates relative to bregma). The hyperdrives and silicon probe drives were made in-house, and protective shrouds around the drives and amplifier boards were printed on a Form 2 3D printer (Formlabs, Somerville, MA).

Animals were anaesthetized with and maintained on isoflurane (0.5 – 2% isoflurane vaporized in O2) for the duration of the surgery. Rats were placed in a sterotaxic apparatus (Kopf, Tujunga, CA) and were given penicillin (Combi-Pen-48) intramuscularly in each hindlimb, and carprofen (Rimadyl) subcutaneously. Rats’ heads were shaved and disinfected with Betadine (Purdue Rederick, Norwalk, CT) before making an incision to reveal the skull. 3–5 jewlers’ screws were used to anchor the drives to the skull, one of which was used as ground for the tetrodes, and a separate screw used as ground for the probe.

Three craniotomies were opened: two for the bilateral tetrode bundles using a surgical trephine, and one for the silicon probes using a burr. The dura was removed with forceps, the probe and tetrode drives were positioned with the sterotax, and then silicone gel (Dow Corning, Midland, MI) was applied to the craniotomies. A layer of MetaBond (Parkell, Edgewood, NY) and then dental acrylic (The Hygenic Corporation, Cuyahoga, OH) was applied to secure the drives to the skull. After surgery, the probes and tetrodes were turned down 640 μm. Rats were subcutaneously injected with carprofen on the day of surgery and for 2 days after surgery, as well as enrofloxacin (Enroflox) the day of surgery and for 5 days post-surgery.

2.4. Data acquisition and electrophysiology

Neural data from all rats was acquired on an Intan RHD2000 recording system (Intan Technologies, Los Angeles, CA), using four RHD2132 amplifier boards (three for the tetrodes and one for the silicon probe). The digitized signals were passed through a 24-channel commutator (Moog, East Aurora, NY) to allow the rats to move freely throughout recording sessions. To synchronize behavior with the neural recordings, timestamps were sent from the Matlab (Version 2017a, The MathWorks, Inc., Natick, MA) software running the task to digital input ports on the Intan RHD2000 USB Interface Board via an Arduino Uno.

Tetrodes were slowly advanced toward the hippocampal pyramidal layer, and the probes toward dmPFC, over the course of around 2 weeks, as the rats recovered and were re-trained on the task. The pyramidal layer was identified by the size of ripples and the direction of sharp wave deflection, as well as spike bursts during these SWRs.

Signals were filtered and spikes and LFP signals were extracted using in-house software written in Matlab and C. Spikes recorded on tetrodes in the hippocampus were manually clustered using the MClust 4.4 software package (Redish, 2017). Only well-separated clusters were kept and used for analysis. The median isolation distance was 21.2, and the median L-Ratio was 0.0899 (Schmitzer-Torbert et al., 2005). Spikes recorded on silicon probes were sorted offline using Kilosort (Pachitariu et al., 2016) into putative clusters, and then manually refined using Phy (Rossant et al., 2016).

2.5. Histology

After rats were finished running the experiment, both tetrode and silicon probe recording locations were marked with electrolytic lesions. 10μA was passed through a channel on each tetrode, and every fourth channel on the silicon probes, for 10s. At least two days after the lesions were made, the rats were anesthetized with a pentobarbital sodium solution (150 mg/kg, Fatal-Plus) and then perfused transcardially with saline followed by 10% formalin. Brains were stored in formalin, and then in a 30% sucrose formalin solution until slicing. Coronal slices were made through the hippocampus and prefrontal cortex (sagittal slices were made instead in PFC for 4 rats) using a cryostat, and the slices were stained with cresyl violet and imaged to determine tetrode and silicon probe recording locations (Figure 2).

Figure 2 :

Figure 2 :

Electrode recording locations and histology. (A) Silicon probe recording locations in dmPFC for each rat. (B) Example photo of a cresyl violet stained coronal slice through PFC, showing the electrolytic lesion created to mark the recording location. (C) Tetrode recording region (highlighted in orange) was in the pyramidal layer of CA1 of dorsal HPC. (D) Example photo of a cresyl violet stained coronal slice through HPC, showing an electrolytic lesion and electrode track. Anatomy diagrams in (A) and (C) are from Paxinos and Watson (2006). IL, infralimbic cortex. PrL, prelimbic cortex. Cg1, cingulate cortex area 1. M2, secondary motor cortex. M1, primary motor cortex. fmi, forceps minor of the corpus callosum. hf, hippocampal fissure. Py, pyramidal layer. cc, corpus callosum. DG, dentate gyrus.

2.6. Statistics and general data analysis

All analyses were performed using Matlab (Version 2017a, The MathWorks, Inc., Natick, MA), except the binomial tests which were performed using the SciPy package for Python (Jones et al., 2001–2019). For all neural analyses, we included only sessions where ≥ 10 cells were recorded in dmPFC and ≥ 10 cells were recorded in CA1 simultaneously.

2.7. Vicarious trial and error (VTE)

To measure vicarious trial and error behaviors, we used zIdPhi, a metric which captures both how long the rat hesitates at the choice point, and how quickly it moves its head back and forth (Papale et al., 2012). When x and y are the x and y position of the rat’s head,

IdPhi =zone entry zone exit δδt atan 2(δyδt,δxδt)δt

and zIdPhi is the IdPhi, z-scored within session.

To distinguish VTE events from non-VTE events, we fit a half-Gaussian distribution to values less than the mode of the zIdPhi distribution. We then assumed that zIdPhi values under a full Gaussian distribution with the same mean and standard deviation as the fit half-Gaussian corresponded to non-VTE events, and passes through the choice point with greater zIdPhi values corresponded to VTE events.

2.8. Bayesian decoding

We used cross-validated Bayesian decoding (Zhang et al., 1998) to decode several variables from ensemble activity in dmPFC or CA1. In Figures 3, 4, and 6, we decoded spatial position from ensemble firing rates in CA1 or dmPFC, using a 32 × 32 grid of spatial bins. In Figure 4, we trained a binary Bayesian decoder on ensemble activity in dmPFC to predict whether CA1 represented the choice point or the reward zone. In Figure 5A, we trained a binary decoder on ensemble activity in dmPFC to predict whether reward was received on the current trial or not. In Figure 6, we also trained a binary decoder on ensemble activity in dmPFC to predict whether reward was received on the current trial or not. However, the training period used for this decoder was from 1–3 seconds after reward zone entry, and the test period was during rats’ deliberation at the choice point.

Figure 3 :

Figure 3 :

Correlations between location encoding in dmPFC and CA1. (A) Examples of Bayesian decoding of spatial location was performed on ensemble activity in dmPFC and CA1 using time bins corresponding to hippocampal theta cycles. Shown are example decoded posterior distributions from four consecutive theta cycles. Green dots indicate the actual position of the animal’s head. Red arrows indicate discrepancies between the animal’s true and the decoded locations. (B) Correlation coefficients of the distance relative to the rats true position decoded from ensemble activity between dmPFC and CA1, for each session and for each rat. (C) Difference in the average decoded position relative to the actual position of the rat at the choice point between VTE and non-VTE passes. Higher values indicate positions further ahead of the rat were represented during VTE passes. Units of the y-axis are laps (full revolutions around the maze). (D) Examples of Bayesian decoding and how it was used to compute the posterior probability of three different zones: the choice point (green dotted box), the left reward zone (red dotted box), and the right reward zone (blue dotted box). The top row shows decoding from CA1, the bottom row shows decoding from dmPFC, and each column shows decoding from the same theta cycle. (E) Correlation between dmPFC and CA1 encoding of the choice point vs either reward site. The probability that CA1 represented either reward zone as opposed to the choice point was computed from CA1 2D spatial posteriors for each theta cycle (left panel, y axis of the 2D histogram). The probability that dmPFC represented either reward zone as opposed to the choice point was also computed from dmPFC 2D spatial posteriors for each theta cycle (left panel, x axis of the 2D histogram). These values, where each datapoint corresponded to a theta cycle during a pass through the choice point, were correlated to produce the correlation coefficients grouped by session (middle panel) or rat (rightmost panel). (F) Correlation between dmPFC and CA1 encoding of the reward site identity. Same as (E), except the probability that either area represented the right reward zone as opposed to the left reward zone (instead of either reward zone vs. the choice point). Supplemental Figures S2, S3, S4, and S5 show these analyses for the first 10 and last 10 laps of each block.

Figure 4 :

Figure 4 :

Predictions of prospective representation in CA1 from activity in dmPFC. (A) Area under the receiver operating characteristic curve per session for the classifier predicting whether CA1 is representing local or prospective spatial information, trained on dmPFC firing rates. (B) The AUROC relative to the AUROC of models fit to shuffled dmPFC firing rates. The x-axis is the number of standard deviations above the shuffle distribution. (C) AUROC of the classifier as a function of when dmPFC firing rates were used to predict representation in CA1. Negative values on the horizontal axis correspond to when dmPFC firing rates were used to predict upcoming representation in CA1, and positive values on the horizontal axis correspond to when dmPFC firing rates were used to predict past representation in CA1. Line and shaded area shows the mean ± standard error, N = 40 sessions, while the gray dots show the AUROC for each individual session. Horizontal dotted line shows 0.5, corresponding to the AUROC expected from noise. (D) Same as in panel C, but showing the standard deviations of the AUROC above the corresponding shuffle distributions as in panel B. The horizontal dotted lines show the median of the shuffle distribution and 3σ. (E) The AUROC relative to shuffle distributions for all sessions combined. (F) The AUROC of Bayesian decoders trained to predict whether HPC ensembles were representing nonlocal information from dmPFC ensemble firing rates, when dmPFC firing rates were shuffled within or between laps. Supplemental Figure S6 shows these analyses for the first 10 and last 10 laps of each block.

Figure 6 :

Figure 6 :

Reward encoding in dmPFC and its dependence on goal encoding in CA1. (A) Decoded reward probability from dmPFC as a function of hippocampal theta cycles since the reference theta cycle, split by whether hippocampal ensembles represented the goal location or the current location during the reference theta cycle. (B) Same as in panel A, but theta cycles where CA1 represented non-local information have been further split by whether CA1 represented the goal location the rat ultimately chose on that lap (“Chosen”) or the opposite goal location (“Unchosen”). (C) Change in the decoded reward probability as a function of lag (the derivative of the reward signal in panel A). (D) Change in the decoded reward probability as a function of lag (the derivative of the reward signal shown in panel B). Shown in all panels is the mean ± standard error, N = 40 sessions. p(R) indicates the posterior probability of reward, as decoded from ensemble firing rates in dmPFC. (E) Same as in (C) but showing only data from the first 10 laps of contingency blocks. (F) Same as in (C) but showing only data from the last 10 laps of contingency blocks.

Figure 5 :

Figure 5 :

Reward representation in dmPFC. (A) Accuracy of Bayesian decoding of reward (vs the lack thereof) from dmPFC ensembles. Line shows the mean and shaded area shows the standard deviation across N = 40 sessions. Vertical dotted line shows the time of reward zone entry. (B) z-scored firing rates of individual units following reward zone entry, split by whether reward was received on that lap or not. Vertical dotted line shows the time of reward zone entry. (C) Difference in z-scored firing rates between reward and lack thereof. Panels B and C show the mean ± standard error across N = 824 units. (D) z-scored firing rates upon reward zone entry for individual example cells. Color of the line in each panel corresponds to the colored dots in (E). (E) Difference in z-scored firing rates between reward and lack thereof for all units. Greater values indicate greater firing rates following reward than following a lack of reward, while lesser values indicate a lesser firing rate following reward than following a lack of reward. Note that the median is only slightly less than 0 (the averaging effect seen in panel B and C), but that this difference is small compared to the spread of the entire distribution.

The time bins used for decoding corresponded to peak-to-peak times of the local field potential recorded in the hippocampal fissure, band-passed between 6 and 11 Hz (theta oscillations). We used these hippocampal theta time bins even for the decoders trained on ensemble activity in dmPFC. In Figure 5A, we instead used 200ms time bins aligned to the time of rats’ entry into the goal zone.

To compute the decoded posterior probability of different zones, we took the average posterior probability within the different zones shown in Figure 3D. The zones used were the choice point (green dotted box in Figure 3D), the left goal site (red dotted box), the right goal site (blue dotted box), and either goal site (the area within both the red and blue dotted boxes).

3. Results

3.1. Rat behavior on the contingency switching task

We recorded from dmPFC and CA1 simultaneously in six rats as they performed the contingency switching task, in which rats ran through a spatial maze with one main choice point between two actions: left or right (Figure 1A). Rats received a food pellet reward if they chose the action at the choice point which was correct under the current contingency. If their choice was inconsistent with the current contingency, no reward was delivered and rats had to circle around to the start of the maze to initiate a new lap. The contingency on any given lap was either Left (only left choices led to reward), Right (only right choices), or Alternate (the opposite choice from the previous lap was required for reward). Contingencies were presented in blocks of laps: every 30 ± 5 laps, the contingency switched to one of the other two contingencies (Figure 1D).

Rats ran 137.7 ± 31.7 laps per session (mean ± standard deviation), and there were 4.1 ± 1.2 contingency switches per session. Rats made choices which were consistent with the current contingency on 78.0 ± 3.0 percent of laps. Rats sometimes displayed vicarious trial and error (VTE) behaviors as they passed through the choice point, where rats moved their heads back and forth as if considering which path to take (Figure 1B). VTE is considered a behavioral marker of deliberation in rodents (Redish, 2016). For more details of rat behavior during the contingency switching task see Hasz and Redish (2020).

3.2. Non-local spatial representations were correlated between CA1 and dmPFC

To determine how spatial information represented in dmPFC and CA1 was related during deliberation, we performed cross-validated Bayesian decoding of spatial position from ensemble spiking activity in dmPFC, and also separately for simultaneously recorded ensembles in CA1. Using the decoded spatial posterior distributions, we computed how far ahead of or behind rats’ actual positions CA1 and dmPFC represented, and how the spatial representations in each area were related (Figure 3A).

Previous work has found that HPC represents locations further ahead of the rat’s actual location during vicarious trial and error (Johnson and Redish, 2007; Amemiya and Redish, 2016; Papale et al., 2016; Amemiya and Redish, 2018; Kay et al., 2020). Replicating this previous work, we found that during passes through the choice point on the contingency switch task, indeed CA1 encoded positions further ahead of the rat during VTE than during non-VTE choice point passes (p = 0.0156, one-sided Wilcoxon signed rank test, N = 6 rats, Figure 3C). Similarly, the dmPFC also encoded positions further ahead of the rat during VTE at the choice point (p = 0.0469, one-sided Wilcoxon signed rank test, N = 6 rats), though the size of this effect was less prominent than that observed in CA1 (Figure 3C). This suggests that both CA1 and dmPFC represented prospective spatial information during deliberation.

However, simply because both areas represented prospective information during deliberation does not necessarily mean they represented identical spatial information at the same time. To determine whether spatial representations in dmPFC and CA1 were tightly locked on a fast timescale, we correlated the distance ahead or behind rats being represented by ensembles over the course of single theta cycles. We found that there was a small correlation between the relative position represented in CA1 and the relative position represented in dmPFC (a mean Spearman’s correlation coefficient of 0.0458, which was significantly greater than zero, p = 0.0469, one-sided Wilcoxon signed rank test, N = 6 rats, Figure 3B).

To determine whether these correlations were due to both areas representing goal locations simultaneously, as opposed to simply being a result of minute positional differences in the spatial representations across theta cycles, we analyzed the decoded spatial locations by which zone was being represented. Because the Bayesian decoding resulted in a decoding posterior across the entire maze, we were able to define three zones of interest: the choice point, the reward zone on the left side of the maze, and the reward zone on the right side of the maze (Figure 3D). To perform the following correlations, we computed the sum of the decoding posterior within each zone (normalized by area) per theta cycle. Encoding of the choice point versus either reward zone was correlated between dmPFC and CA1 (a median Spearman’s correlation coefficient of 0.129, which was significantly greater than zero, p = 0.0313, Wilcoxon signed rank test, N = 6 rats, Figure 3E). However the identity of the reward site being encoded by dmPFC and CA1 was not significantly correlated (p = 0.156, Wilcoxon signed rank test, N = 6 rats, Figure 3F). The correlation coefficients for choice point and either reward zone representation were greater than the correlation coefficients for the identity of the encoded reward site (paired Wilcoxon signed rank test, p = 0.0313, N = 6 rats, Figure 3E vs Figure 3F).

These results suggest that while both the dmPFC and HPC may have been in deliberative modes simultaneously, and both seemed to represent either local or prospective information within the same theta cycles, that they were not necessarily representing identical information at the same time.

3.3. dmPFC predicted non-local spatial representation in CA1

If dmPFC instigates internal simulations and evaluations of candidate actions, carried out by the hippocampus and other structures, then neural signatures corresponding to that initiation should be present in dmPFC ensemble activity. Therefore, it should be possible to predict from ensemble activity in PFC whether hippocampus will represent prospective information.

To determine whether activity in dmPFC carried information about whether CA1 was representing prospective information, we first used cross-validated Bayesian decoding to decode the location represented by CA1 on a per-theta-cycle basis. We categorized hippocampal theta cycles while rats passed through the choice point as either “local” (the top 10% of theta cycles with the highest posterior density in the choice point) or “prospective” (the top 10% of theta cycles with the highest posterior density in the reward zones – either the left or right reward zone, as in Figure 3E). Then, we trained a classifier (also a cross-validated Bayesian decoder) to predict from ensemble activity in dmPFC whether simultaneous hippocampal activity was representing local or prospective information. We measured the performance of the classifier using the area under the receiver operating characteristic curve (AUROC) metric. An AUROC value of 0.5 indicates a classifier is performing at chance, while an AUROC value of 1.0 would indicate the decoder is correctly classifying every single theta cycle.

We found that this classifier trained on dmPFC activity performed well above chance, with AUROC values generally in the 0.6 to 0.8 range (Figure 4A). To ensure that the classifier was performing above chance, we repeatedly trained models on shuffled dmPFC firing rates (we used 1000 shuffles per session). The performance of the classifier trained on actual dmPFC firing rates was better than shuffles for all but 3 sessions, and was significantly above the shuffle distributions for the vast majority of sessions (Figure 4B). The AUROC value for the decoder across all sessions was over 20 standard deviations above the distribution resulting from applying the decoder to shuffled firing rates (Figure 4E, lag of 0).

However, the predictive power of dmPFC activity did not seem to be highly temporally specific. We repeated the classification analysis, but using neural activity from dmPFC at different time lags relative to the HPC theta cycle for which the classifier was trying to predict. If, say, activity in dmPFC were causing the next theta cycle in HPC to represent non-local information, but had no influence over hippocampal representations during preceding or subsequent theta cycles, then we would expect to see a sharp peak in the classifier’s performance at a lag of −1 theta cycle. In contrast to that hypothesis, the performance of the classifier had a very broad performance profile as a function of lag (Figure 4C,D,E). Furthermore, the performance of classifiers trained on dmPFC firing rates which were shuffled within-lap was better than the performance of classifiers trained on dmPFC firing rates which were shuffled between-lap (Figure 4F). These results suggest that medial prefrontal cortex may be affecting the hippocampal state in a way which corresponds to strategy in general, such as whether to deliberate or not (Powell and Redish, 2014; Hasz and Redish, 2020), and that this influence may last over longer time periods such as entire trials.

3.4. Non-local representation in CA1 corresponded to an increase in reward encoding in dmPFC

Our results thus far suggest that dmPFC could play some role in influencing hippocampal circuitry to enter into prospective modes, but it is also possible that hippocampal activity has an effect on prefrontal representations. Next, we investigated if and how the encoding of reward in prefrontal cortex is affected by spatial representations in hippocampus.

First, we examined whether ensembles in dmPFC encoded whether animals received reward or not. We performed cross-validated Bayesian decoding on ensemble activity in dmPFC as rats entered the reward zone, and decoded whether rats received reward on that lap or whether there was a lack of reward. The task included an audio cue as to whether a choice was correct or incorrect upon reward zone entry, and so rats had multiple sources of information as to whether their choice was correct or incorrect (both the audio cue and the presence or absence of food reward at the feeder site). The presence or absence of reward was most accurately decoded from dmPFC activity between around 1–3 seconds after reward zone entry (Figure 5A). This indicates that dmPFC does in fact carry information about reward, at least at the time of reward delivery.

One potential confound, however, is that rats’ running velocity may have differed between reward receipt and the lack thereof: rats stopped to consume food when they received reward, but often continued on to start a new lap when they did not. In this case, dmPFC may have been simply representing running velocity, and not reward per se, but decoding of reward from ensemble activity could still be possible. Therefore, we performed reward decoding from ensemble activity in dmPFC, but using only times when rats were running (which we defined as times when rats were moving more quickly than 10 cm/sec). We found that we were still able to decode reward receipt from the lack thereof with similar accuracy during the same time period following reward delivery (Figure S1). This suggests that dmPFC represents information about reward independently of velocity.

How was this reward information represented? Firing rates were, on average, significantly higher after an incorrect choice than after reward was delivered (Figure 5B,C). However, to say that firing rates decreased following reward receipt would be a mischaracterization of the data, as the effect was mostly due to averaging. While cells showed a decreased firing rate upon reward on average, there was a wide distribution of responses to reward across individual units. Some cells’ firing rates increased in response to reward while other cells had greater firing rates after incorrect choices (Figure 5E, and see Figure 5D for examples of cells’ responses across the distribution). This indicates that the reward encoding we observed in dmPFC was not simply due to “reward cells” or cells which decreased their firing rate in the same direction upon reward receipt, but rather that the representation of reward was due to encoding across the entire ensemble. This is consistent with previous work which finds that single units in prefrontal areas tend to have highly mixed selectivity (Asaad et al., 1998; Mansouri et al., 2006; Mante et al., 2013; Rigotti et al., 2013).

Did prefrontal representations of reward change in response to the encoding of non-local information in hippocampus? We performed cross-validated Bayesian decoding of reward from ensemble activity in dmPFC and also cross-validated Bayesian decoding of spatial location from simultaneously recorded ensemble activity in CA1, while rats deliberated at the choice point of the contingency switch task. The CA1 decoder was trained on location and neural activity across the entire session, with 100-fold cross validation, using time bins corresponding to hippocampal theta cycles. The decoder used for dmPFC was trained to decode reward vs the lack thereof, using only epochs 1–3s after reward zone entry, which was the time during which we found dmPFC to be most reliably encoding reward (Figure 5A). This decoder was then used to predict the probability of reward from ensemble activity in dmPFC while rats passed through the choice point. The time bins used for dmPFC decoding were the same as those used for hippocampal decoding (theta cycles of the hippocampal LFP). We split hippocampal theta cycles into those where hippocampus was representing “local” spatial information and those where hippocampus was representing non-local, reward zone information. “Local” theta cycles were those where the decoded spatial posterior distribution had a greater posterior probability in the choice point than anywhere else on the maze. In contrast, the “non-local” theta cycles were those during which the reward zone had a greater posterior probability than the rest of the maze combined (so, they were not just non-local in a broad sense, but specifically those theta cycles during which hippocampus represented the reward zone most strongly).

We found that during theta cycles during which CA1 was representing the reward zone, there was not a significantly different amount of reward encoding in dmPFC than during theta cycles when CA1 was representing the choice point (Figure 6A). However, there was a significant change in the reward representation in dmPFC. Reward encoding in dmPFC significantly increased during theta cycles where CA1 represented the reward zone (Figure 6C, orange line at zero lag; p = 0.0021, two-sided Wilcoxon signed rank test comparing the change in decoded reward probabilities from the previous theta cycle, N = 40 sessions). The rate of this change was approximately 1.02% per theta cycle, where a 100% change would correspond to a transition from complete certainty that reward would not be received to complete certainty that reward would be received within one theta cycle. In contrast, dmPFC reward encoding did not change during theta cycles where CA1 represented only local information (Figure 6C, gray line at zero lag; p = 0.85, two-sided Wilcoxon signed rank test, N = 40 sessions). However, this change in dmPFC reward encoding was not significantly greater during theta cycles in which HPC represented the reward zone than during theta cycles in which HPC represented locally (Figure 6C, gray vs orange line at zero lag; p = 0.12, two-sided Wilcoxon rank sum test, N = 40 sessions).

To determine the temporal specificity of this effect, we performed a lag analysis similar to that used in Figures 4C and D. We examined the change in reward representation in dmPFC as a function of the number of theta cycles by which the reference theta cycle in CA1 led or lagged the decoding time bin used for reward decoding from dmPFC. Unlike the results in the previous section (Figures 4C and D), we found that the change in reward representation in dmPFC was very tightly locked to the theta cycle during which CA1 represented the reward zone (Figure 6C). The only offset during which the reward encoding in dmPFC significantly increased was when the same time bin was used for reward decoding from dmPFC and location decoding from CA1, during which CA1 represented the reward zone (Figure 6C, orange line; two-sided Wilcoxon signed rank tests, p = 0.0021 for a theta cycle lag of zero; for all other lags p > 0.0039, the Bonferroni-corrected significance level for α = 0.05). Theta cycles during which CA1 represented only local information showed no systematic change in reward encoding in dmPFC, at any lag (Figure 6C, gray line; two-sided Wilcoxon signed rank tests, for all lags including zero theta cycle lag p > 0.0039, the Bonferroni-corrected significance level for α = 0.05). Although on average the reward encoding in dmPFC increased by a greater amount on theta cycles during which CA1 represented specifically the reward zone on the side which the rat ended up choosing on that lap, this was not significantly different between theta cycles during which CA1 represented the chosen vs the unchosen side (Figure 6B and D, two-sided Wilcoxon signed rank test, p = 0.158).

How does this change in reward representation in dmPFC correspond to the changing contingencies? If non-local representation in HPC is driving reward expectancy changes in dmPFC, then we may expect to see these changes occurring most strongly when the task contingencies are stable and any hippocampal prospective information reaching dmPFC is most confident about its predictions. While our previous results did not seem to be related to whether the rat was early in a contingency block or late in a contingency block (Supplemental Figures 26), the increase in reward encoding in dmPFC concurrent with nonlocal representation in CA1 was especially prominent late within contingency blocks (Figures 6E and F). In the first 10 laps of contingency blocks, there was no significant increase in reward encoding in dmPFC on theta cycles when HPC representations were the most nonlocal (Figure 6E at a lag of 0 theta cycles, an average reward encoding increase of 0.62% per lap, p = 0.202, two-sided Wilcoxon signed rank test), and the change in dmPFC reward encoding was not significantly greater during theta cycles when HPC representations were the most nonlocal than during theta cycles when HPC representations were the most local (Figure 6E at a lag of 0 theta cycles, gray vs. orange lines, p = 0.881, two-sided Wilcoxon rank sum test). In contrast, during the last 10 laps of contingency blocks, there was a significant increase in reward encoding in dmPFC on theta cycles when HPC representations were the most nonlocal (Figure 6F at a lag of 0 theta cycles, an average reward encoding increase of 2.2% per lap, p = 0.00285, two-sided Wilcoxon signed rank test), and this change in reward encoding by dmPFC was significantly greater during theta cycles when HPC represented the reward zone most strongly than during theta cycles when HPC representations were local (Figure 6F at a lag of 0 theta cycles, gray vs. orange lines, p = 0.030, two-sided Wilcoxon rank sum test). Taken together, these results suggest that hippocampal spatial representations may have a fast, temporally precise effect on reward encoding in dmPFC, specifically at times when the environment is stable.

4. Discussion

We examined the relationship between hippocampal and prefrontal representations during deliberation, and found that they were related in complex ways. While CA1 and dmPFC spatial representations were correlated in that they represented either local information or prospective information together, the two areas did not appear to always represent identical prospective information at the same time. Furthermore, activity in dmPFC predicted whether CA1 was representing local or prospective information, but this relationship occurred across a very broad timescale, on the order of seconds. On the other hand, representations of goal location in CA1 appeared to have very temporally specific effects on reward encoding in dmPFC, across timescales on the order of a single theta cycle.

Furthermore, we found that hippocampal goal encoding corresponded to an increase in reward encoding in dmPFC late in contingency blocks (Figure 6F), when animals were generally more confident as to the contingency identity and deliberated less, but not early in contingency blocks where deliberation was more common (Figure 6E). This is consistent with previous observations that hippocampal representations predict goal locations when animals show no signs of deliberation, but not when animals display markers of deliberation such as vicarious trial and error (Johnson and Redish, 2007; Papale et al., 2016; Kay et al., 2020). Taken together with these findings, our result suggests that hippocampal representations of goal locations may have a more robust effect on reward representations in dmPFC when hippocampal predictions of future outcomes are more confident.

If the prefrontal cortex uses estimates of reward or value associated with candidate actions to perform action selection, then presumably it or other areas must keep track of these candidate actions and the corresponding value estimates. However, there are different mechanisms the brain may use to perform this computation, and different brain areas may play different roles in this process. One possibility is that the prefrontal cortex uses a similar mechanism for value-based decision making as sensory areas do for sensory-based decision making, by integrating evidence until some decision threshold is reached. These so-called “drift diffusion models” or “sequential sampling models” have been found to explain both behavior and neural activity in sensory areas during sensory decision-making tasks (Stone, 1960; Ratcliff, 1978; Hanes and Schall, 1996; Ratcliff and McKoon, 2008; Forstmann et al., 2016). However, it is unclear whether the brain uses a similar mechanism for making value-based decisions, where the properties of the options being decided between are entirely internal, as opposed to sensory decision-making where those properties are external and directly observable. Some work suggests that, at least behaviorally, reaction times during value-based decision making tasks can be explained using drift-diffusion models (Krajbich and Rangel, 2011).

However, an alternative possibility is that there is no slow integration process, but rather the brain considers options serially and discretely, and makes a decision without evidence accumulation per se. Recent work indicates that, at least in the orbitofrontal cortex, value signals switch back and forth suddenly, suggesting that options are being considered serially during value-based decisions (Rich and Wallis, 2016; Wallis, 2018). Such stepwise representational jumps during decision-making have been observed even during sensory-based decisions (Yang and Shadlen, 2007; Latimer et al., 2015; Sadacca et al., 2016). Yet a third possibility is that both of these mechanisms occur simultaneously: certain brain areas could consider and evaluate options simultaneously, while others accumulate evidence for each decision and trigger the corresponding action to be taken when some decision threshold is reached. While our results cannot distinguish these three possibilities, they do indicate that valuations arising from the representation of prospective information in HPC change reward encoding in dmPFC.

Our results are consistent with recent work by Zielinski et al. (2019), who also find that representations of location in PFC and CA1 are correlated during a spatial decision-making task. Our data further indicates that HPC and dmPFC spatial representations are coherent not only in general, but specifically that they encode prospective information simultaneously. However, our data suggests that dmPFC and CA1 may not be encoding identical prospective spatial information simultaneously during deliberation. Instead, we found that the goal zone represented by dmPFC and HPC was uncorrelated during prospective encoding by both structures, but that reward encoding in dmPFC changed during hippocampal theta cycles where CA1 was encoding goal locations. This suggests that while dmPFC and HPC represent prospective information together, the relationship between prospective information represented in these areas may be more nuanced than a simple duplication of spatial information, and also involve decision making variables such as reward or value.

Supplementary Material

Supplemental

Acknowledgments

We thank Christopher Boldt, Kelsey Seeland, and Ayaka Sheehan for technical support and for building the tetrode drives, Christopher Boldt for building the silicon probe drives, Ayaka Sheehan for performing the histology, as well as Onni Rauhala and Daniel Min for help training rats. This work was funded by NSF IGERT Neuroengineering grant DGE-1069104, NIH R01-MH080318, and NIH R01-MH112688.

Footnotes

Declaration of Interests

The authors declare no competing interests.

Data Availability

Data is available upon request.

References

  1. Amemiya S and Redish AD. Manipulating decisiveness in decision making: effects of clonidine on hippocampal search strategies. Journal of Neuroscience, 36(3):814–827, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Amemiya S and Redish AD. Hippocampal theta-gamma coupling reflects state-dependent information processing in decision making. Cell reports, 22(12):3328–3338, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Asaad WF, Rainer G, and Miller EK. Neural activity in the primate prefrontal cortex during associative learning. Neuron, 21(6):1399–1407, 1998. [DOI] [PubMed] [Google Scholar]
  4. Barker GRI, Banks PJ, Scott H, Ralph GS, Mitrophanous KA, Wong L, Bashir ZI, Uney JB, and Warburton EC. Separate elements of episodic memory subserved by distinct hippocampal–prefrontal connections. Nature neuroscience, 20 (2):242, 2017. [DOI] [PubMed] [Google Scholar]
  5. Benchenane K, Peyrache A, Khamassi M, Tierney PL, Gioanni Y, Battaglia FP, and Wiener SI. Coherent theta oscillations and reorganization of spike timing in the hippocampal-prefrontal network upon learning. Neuron, 66(6):921–936, 2010. [DOI] [PubMed] [Google Scholar]
  6. Botvinick MM, Braver TS, Barch DM, Carter CS, and Cohen JD. Conflict monitoring and cognitive control. Psychological review, 108(3):624, 2001. [DOI] [PubMed] [Google Scholar]
  7. Cassel J, De Vasconcelos AP, Loureiro M, Cholvin T, Dalrymple-Alford JC, and Vertes RP. The reuniens and rhomboid nuclei: neuroanatomy, electrophysiological characteristics and behavioral implications. Progress in neurobiology, 111:34–52, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Cowen SL and McNaughton BL. Selective delay activity in the medial prefrontal cortex of the rat: contribution of sensorimotor information and contingency. Journal of neurophysiology, 98(1):303–316, 2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Dalley JW, Cardinal RN, and Robbins TW. Prefrontal executive and cognitive functions in rodents: neural and neurochemical substrates. Neuroscience & Biobehavioral Reviews, 28(7):771–784, 2004. [DOI] [PubMed] [Google Scholar]
  10. Delatour B and Gisquest-Verrier P. Lesions of the prelimbic–infralimbic cortices in rats do not disrupt response selection processes but induce delay-dependent deficits: evidence for a role in working memory? Behavioral neuroscience, 113(5):941, 1999. [DOI] [PubMed] [Google Scholar]
  11. Delatour B and Witter MP. Projections from the parahippocampal region to the prefrontal cortex in the rat: evidence of multiple pathways. European Journal of Neuroscience, 15 (8):1400–1407, 2002. [DOI] [PubMed] [Google Scholar]
  12. Di Prisco GV and Vertes RP. Excitatory actions of the ventral midline thalamus (rhomboid/reuniens) on the medial prefrontal cortex in the rat. Synapse, 60(1):45–55, 2006. [DOI] [PubMed] [Google Scholar]
  13. Doll BB, Simon DA, and Daw ND. The ubiquity of model-based reinforcement learning. Current opinion in neurobiology, 22(6):1075–1081, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Dolleman-Van der Weel MJ, Da Silva FHL, and Witter MP. Interaction of nucleus reuniens and entorhinal cortex projections in hippocampal field ca1 of the rat. Brain Structure and Function, 222(5):2421–2438, 2017. [DOI] [PubMed] [Google Scholar]
  15. Dolleman-van der Weel MJ, Griffin AL, Ito HT, Shapiro ML, Witter MP, Vertes RP, and Allen TA. The nucleus reuniens of the thalamus sits at the nexus of a hippocampus and medial prefrontal cortex circuit enabling memory and behavior. Learning & Memory, 26(7):191–205, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Dwyer DM, Dunn MJ, Rhodes SEV, and Killcross AS. Lesions of the prelimbic prefrontal cortex prevent response conflict produced by action–outcome associations. Quarterly Journal of Experimental Psychology, 63(3):417–424, 2010. [DOI] [PubMed] [Google Scholar]
  17. Eichenbaum H. Prefrontal–hippocampal interactions in episodic memory. Nature Reviews Neuroscience, 18(9):547, 2017. [DOI] [PubMed] [Google Scholar]
  18. Euston DR, Gruber AJ, and McNaughton BL. The role of medial prefrontal cortex in memory and decision making. Neuron, 76(6):1057–1070, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Ferino F, Thierry AM, and Glowinski J. Anatomical and electrophysiological evidence for a direct projection from ammon’s horn to the medial prefrontal cortex in the rat. Experimental brain research, 65(2):421–426, 1987. [DOI] [PubMed] [Google Scholar]
  20. Floresco SB and Grace AA. Gating of hippocampal-evoked activity in prefrontal cortical neurons by inputs from the mediodorsal thalamus and ventral tegmental area. Journal of Neuroscience, 23(9):3930–3943, 2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Floresco SB, Seamans JK, and Phillips AG. Selective roles for hippocampal, prefrontal cortical, and ventral striatal circuits in radial-arm maze tasks with or without a delay. Journal of Neuroscience, 17(5):1880–1890, 1997. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Forstmann BU, Ratcliff R, and Wagenmakers E. Sequential sampling models in cognitive neuroscience: Advantages, applications, and extensions. Annual review of psychology, 67:641–666, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Fuster J. The prefrontal cortex. Academic Press, 2015. [Google Scholar]
  24. Fuster JM and Alexander GE. Neuron activity related to short-term memory. Science, 173(3997):652–654, 1971. [DOI] [PubMed] [Google Scholar]
  25. Gilbert DT and Wilson TD. Prospection: Experiencing the future. Science, 317(5843): 1351–1354, 2007. [DOI] [PubMed] [Google Scholar]
  26. Guise KG and Shapiro ML. Medial prefrontal cortex reduces memory interference by modifying hippocampal encoding. Neuron, 94(1):183–192, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Haddon JE and Killcross AS. Medial prefrontal cortex lesions abolish contextual control of competing responses. Journal of the experimental analysis of behavior, 84 (3):485–504, 2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Haddon JE and Killcross S. Prefrontal cortex lesions disrupt the contextual control of response conflict. Journal of Neuroscience, 26(11):2933–2940, 2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Hanes DP and Schall JD. Neural control of voluntary movement initiation. Science, 274(5286):427–430, 1996. [DOI] [PubMed] [Google Scholar]
  30. Hassabis D and Maguire EA. The construction system of the brain. Philosophical Transactions of the Royal Society B: Biological Sciences, 364(1521):1263–1271, 2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Hasz BM and Redish AD. Dorsomedial prefrontal cortex and hippocampus represent strategic context even while simultaneously changing representation throughout a task session. Neurobiology of Learning and Memory, page 107215, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Hok V, Chah E, Save E, and Poucet B. Prefrontal cortex focally modulates hippocampalplace cell firing patterns. Journal of Neuroscience, 33(8):3443–3451, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Hoover WB and Vertes RP. Anatomical analysis of afferent projections to the medialprefrontal cortex in the rat. Brain Structure and Function, 212(2):149–179, 2007. [DOI] [PubMed] [Google Scholar]
  34. Hoover WB and Vertes RP. Collateral projections from nucleus reuniens of thalamus to hippocampus and medial prefrontal cortex in the rat: a single and double retrograde fluorescent labeling study. Brain Structure and Function, 217(2):191–209, 2012. [DOI] [PubMed] [Google Scholar]
  35. Horst NK and Laubach M. The role of rat dorsomedial prefrontal cortex in spatial working memory. Neuroscience, 164(2):444–456, 2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Hyman JM, Zilli EA, Paley AM, and Hasselmo ME. Medial prefrontal cortex cells show dynamic modulation with the hippocampal theta rhythm dependent on behavior. Hippocampus, 15(6):739–749, 2005. [DOI] [PubMed] [Google Scholar]
  37. Jai YY and Frank LM. Hippocampal–cortical interaction in decision making. Neurobiology of learning and memory, 117:34–41, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Jay TM and Witter MP. Distribution of hippocampal ca1 and subicular efferents in the prefrontal cortex of the rat studied by means of anterograde transport of phaseolus vulgaris-leucoagglutinin. Journal of Comparative Neurology, 313(4):574–586, 1991. [DOI] [PubMed] [Google Scholar]
  39. Johnson A and Redish AD. Neural ensembles in ca3 transiently encode paths forward of the animal at a decision point. Journal of Neuroscience, 27(45):12176–12189, 2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Jones E, Oliphant T, Peterson P, et al. SciPy: Open source scientific tools for Python. Version 1.3.1, 2001–2019. URL http://www.scipy.org.
  41. Jones MW and Wilson MA. Theta rhythms coordinate hippocampal–prefrontal interactions in a spatial memory task. PLoS biology, 3(12):e402, 2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Kay K, Chung JE, Sosa M, Schor JS, Karlsson MP, Larkin MC, Liu DF, and Frank LM. Constant sub-second cycling between representations of possible futures in the hippocampus. Cell, 180(3):552–567, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Kesner RP and Churchwell JC. An analysis of rat prefrontal cortex in mediating executive function. Neurobiology of learning and memory, 96(3):417–431, 2011. [DOI] [PubMed] [Google Scholar]
  44. Krajbich I and Rangel A. Multialternative drift-diffusion model predicts the relationship between visual fixations and choice in value-based decisions. Proceedings of the National Academy of Sciences, 108(33):13852–13857, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Latimer KW, Yates JL, Meister MLR, Huk AC, and Pillow JW. Single-trial spike trains in parietal cortex reveal discrete steps during decision-making. Science, 349 (6244):184–187, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Maharjan DM, Dai YY, Glantz EH, and Jadhav SP. Disruption of dorsal hippocampal–prefrontal interactions using chemogenetic inactivation impairs spatial learning. Neurobiology of learning and memory, 155:351–360, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Mansouri FA, Matsumoto K, and Tanaka K. Prefrontal cell activities related to monkeys’ success and failure in adapting to rule changes in a wisconsin card sorting test analog. Journal of Neuroscience, 26(10):2745–2756, 2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Mante V, Sussillo D, Shenoy KV, and Newsome WT. Context-dependent computation by recurrent dynamics in prefrontal cortex. nature, 503(7474):78, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Marquis J, Killcross S, and Haddon JE. Inactivation of the prelimbic, but not infralimbic, prefrontal cortex impairs the contextual control of response conflict in rats. European Journal of Neuroscience, 25(2):559–566, 2007. [DOI] [PubMed] [Google Scholar]
  50. McKenna JT and Vertes RP. Afferent projections to nucleus reuniens of the thalamus. Journal of comparative neurology, 480(2):115–142, 2004. [DOI] [PubMed] [Google Scholar]
  51. Miller EK and Cohen JD. An integrative theory of prefrontal cortex function. Annual review of neuroscience, 24(1):167–202, 2001. [DOI] [PubMed] [Google Scholar]
  52. Navawongse R and Eichenbaum H. Distinct pathways for rule-based retrieval and spatial mapping of memory representations in hippocampal neurons. Journal of Neuroscience, 33(3):1002–1013, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. O’Neill P, Gordon JA, and Sigurdsson T. Theta oscillations in the medial prefrontal cortex are modulated by spatial working memory and synchronize with the hippocampus through its ventral subregion. Journal of Neuroscience, 33(35):14211–14224, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Pachitariu M, Steinmetz NA, Kadir SN, Carandini M, and Harris KD. Fast and accurate spike sorting of high-channel count probes with kilosort. In Advances in Neural Information Processing Systems, pages 4448–4456, 2016. [Google Scholar]
  55. Padoa-Schioppa C. Neurobiology of economic choice: a good-based model. Annual review of neuroscience, 34:333–359, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Papale AE, Stott JJ, Powell NJ, Regier PS, and Redish AD. Interactions between deliberation and delay-discounting in rats. Cognitive, Affective, & Behavioral Neuroscience, 12(3):513–526, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Papale AE, Zielinski MC, Frank LM, Jadhav SP, and Redish AD. Interplay between hippocampal sharp-wave-ripple events and vicarious trial and error behaviors in decision making. Neuron, 92(5):975–982, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Paxinos G and Watson C. The rat brain in stereotaxic coordinates: hard cover edition. Elsevier, 2006. [Google Scholar]
  59. Powell NJ and Redish AD. Complex neural codes in rat prelimbic cortex are stable across days on a spatial decision task. Frontiers in behavioral neuroscience, 8:120, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Preston AR and Eichenbaum H. Interplay of hippocampus and prefrontal cortex in memory. Current Biology, 23(17):R764–R773, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Ragozzino ME and Kesner RP. The effects of muscarinic cholinergic receptor blockade in the rat anterior cingulate and prelimbic/infralimbic cortices on spatial working memory. Neurobiology of learning and memory, 69(3):241–257, 1998. [DOI] [PubMed] [Google Scholar]
  62. Rangel A, Camerer C, and Montague PR. A framework for studying the neurobiology of value-based decision making. Nature reviews neuroscience, 9(7):545, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Ratcliff R. A theory of memory retrieval. Psychological review, 85(2):59, 1978. [Google Scholar]
  64. Ratcliff R and McKoon G. The diffusion decision model: theory and data for two-choice decision tasks. Neural computation, 20(4):873–922, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Redish AD. The mind within the brain: How we make decisions and how those decisions go wrong. Oxford University Press, 2013. [Google Scholar]
  66. Redish AD. Vicarious trial and error. Nature Reviews Neuroscience, 17(3):147, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Redish AD. Mclust. Version 4.4.07, 2017. URL http//redishlab.neuroscience.umn.edu/MClust/MClust.html.
  68. Rich EL and Wallis JD. Decoding subjective decisions from orbitofrontal cortex. Nature neuroscience, 19(7):973, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Rigotti M, Barak O, Warden MR, Wang X, Daw ND, Miller EK, and Fusi S. The importance of mixed selectivity in complex cognitive tasks. Nature, 497(7451):585, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Rossant C, Kadir S, Goodman D, Hunter M, and Harris K. Phy. Version 1.0.9, 2016. URL http//github.com/cortex-lab/phy.
  71. Sadacca BF, Mukherjee N, Vladusich T, Li JX, Katz DB, and Miller P. The behavioral relevance of cortical neural ensemble responses emerges suddenly. Journal of Neuroscience, 36(3):655–669, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Schmidt B, Duin AA, and Redish AD. Disrupting the medial prefrontal cortex alters hippocampal sequences during deliberative decision-making. Journal of neurophysiology, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Schmitzer-Torbert N, Jackson J, Henze D, Harris K, and Redish AD. Quantitative measures of cluster quality for use in extracellular recordings. Neuroscience, 131(1): 1–11, 2005. [DOI] [PubMed] [Google Scholar]
  74. Shenhav A, Botvinick MM, and Cohen JD. The expected value of control: an integrative theory of anterior cingulate cortex function. Neuron, 79(2):217–240, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Shin JD and Jadhav SP. Multiple modes of hippocampal–prefrontal interactions in memory-guided behavior. Current opinion in neurobiology, 40:161–169, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Siapas AG, Lubenov EV, and Wilson MA. Prefrontal phase locking to hippocampal theta oscillations. Neuron, 46(1):141–151, 2005. [DOI] [PubMed] [Google Scholar]
  77. Steiner AP and Redish AD. The road not taken: neural correlates of decision making in orbitofrontal cortex. Frontiers in neuroscience, 6:131, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Steiner AP and Redish AD. Behavioral and neurophysiological correlates of regret in rat decision-making on a neuroeconomic task. Nature neuroscience, 17(7):995, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Stone M. Models for choice-reaction time. Psychometrika, 25(3):251–260, 1960. [Google Scholar]
  80. Stott JJ and Redish AD. A functional difference in information processing between orbitofrontal cortex and ventral striatum during decision-making behaviour. Philosophical Transactions of the Royal Society B: Biological Sciences, 369(1655): 20130472, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Swanson LW. A direct projection from ammon’s horn to prefrontal cortex in the rat. Brain research, 217(1):150–154, 1981. [DOI] [PubMed] [Google Scholar]
  82. Tronel S and Sara SJ. Blockade of nmda receptors in prelimbic cortex induces an enduring amnesia for odor–reward associative learning. Journal of Neuroscience, 23 (13):5472–5476, 2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Urban KR, Layfield DM, and Griffin AL. Transient inactivation of the medial prefrontal cortex impairs performance on a working memory-dependent conditional discrimination task. Behavioral neuroscience, 128(6):639, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. van der Meer M, Kurth-Nelson Z, and Redish AD. Information processing in decision-making systems. The Neuroscientist, 18(4):342–359, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. van der Meer MAA and Redish AD. Expectancies in decision making, reinforcement learning, and ventral striatum. Frontiers in neuroscience, 3:6, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Vertes RP. Analysis of projections from the medial prefrontal cortex to the thalamus in the rat, with emphasis on nucleus reuniens. Journal of Comparative Neurology, 442(2): 163–187, 2002. [DOI] [PubMed] [Google Scholar]
  87. Vertes RP. Differential projections of the infralimbic and prelimbic cortex in the rat. Synapse, 51(1):32–58, 2004. [DOI] [PubMed] [Google Scholar]
  88. Vertes RP, Hoover WB, Do Valle AC, Sherman A, and Rodriguez JJ. Efferent projections of reuniens and rhomboid nuclei of the thalamus in the rat. Journal of comparative neurology, 499(5):768–796, 2006. [DOI] [PubMed] [Google Scholar]
  89. Vertes RP, Hoover WB, Szigeti-Buck K, and Leranth C. Nucleus reuniens of the midline thalamus: link between the medial prefrontal cortex and the hippocampus. Brain research bulletin, 71(6):601–609, 2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  90. Verwer RWH, Meijer RJ, Van Uum HFM, and Witter MP. Collateral projections from the rat hippocampal formation to the lateral and medial prefrontal cortex. Hippocampus, 7(4):397–402, 1997. [DOI] [PubMed] [Google Scholar]
  91. Wallis JD. Decoding cognitive processes from neural ensembles. Trends in cognitive sciences, 22(12):1091–1102, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Wang G and Cai J. Disconnection of the hippocampal–prefrontal cortical circuits impairs spatial working memory performance in rats. Behavioural brain research, 175(2):329–336, 2006. [DOI] [PubMed] [Google Scholar]
  93. Wang JX, Cohen NJ, and Voss JL. Covert rapid action-memory simulation (crams): A hypothesis of hippocampal–prefrontal interactions for adaptive behavior. Neurobiology of learning and memory, 117:22–33, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Yang T and Shadlen MN. Probabilistic reasoning by neurons. Nature, 447(7148):1075, 2007. [DOI] [PubMed] [Google Scholar]
  95. Yoon T, Okada J, Jung MW, and Kim JJ. Prefrontal cortex and hippocampus subserve different components of working memory in rats. Learning & memory, 15(3):97–105, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Zhang K, Ginzburg I, McNaughton BL, and Sejnowski TJ. Interpreting neuronal population activity by reconstruction: unified framework with application to hippocampal place cells. Journal of neurophysiology, 79(2):1017–1044, 1998. [DOI] [PubMed] [Google Scholar]
  97. Zielinski MC, Shin JD, and Jadhav SP. Coherent coding of spatial position mediated by theta oscillations in the hippocampus and prefrontal cortex. Journal of Neuroscience, 39(23):4550–4565, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplemental

RESOURCES