Abstract
When participants performed a visual search task, fMRI responses in entorhinal cortex (EC) exhibited a 6-fold periodic modulation by gaze movement direction. The orientation of this modulation was determined by the shape and orientation of the bounded search space. These results indicate that human EC represents visual space using a boundary-anchored grid, analogous to that used to represent navigable space in rodents.
Main Text
During spatial navigation in rodents, grid cells fire when the body of the animal occupies a hexagonal lattice of spatial locations tiling the floor of the environment1. These cells are believed to support a metric for navigational space that is anchored to environmental boundaries2, 3. Recent work with monkeys has expanded the variety of spaces that might be represented by grid cells, by demonstrating the existence of neurons in EC that fire in a hexagonal lattice of positions on a screen while animals explore visual space4. However, it is currently unknown whether a similar grid-like coding of visual space exists in humans, or whether putative grid representations of visual space obey the same boundary-anchoring principles as grid representations of navigational space.
To address these issues, we used fMRI methods previously developed for identifying grid signals in humans during virtual navigation5. These methods are motivated by the observation that firing patterns for grid cells within an individual tend to have the same orientation1, 5. Because of this common orientation, movements along a shared grid axis will yield a stronger grid-driven fMRI signal in EC than movements between grid axes, resulting in 60° periodic modulation by movement direction. We reasoned that if grid cells represent visual space in humans, then we should observe a similar 60° periodic fMRI signal as a function of gaze movement direction while participants visually explored the environment. To test this idea, participants (n=36) were scanned with fMRI and had their gaze tracked while they performed an unconstrained visual search task in which they had to find a target letter (‘L’) among numerous distractors letters (‘T’s) (Fig. 1A). A square border surrounded the search display for half the participants (n=18) and a rectangular border surrounded the display for the other half (n=18).
For each participant, we split the fMRI data into halves, identified the orientation of the 60° periodic signal as a function of gaze movement direction within EC in one half of the data, and tested the reliability of this visual grid orientation in the independent second half (Supplementary Fig. 1). Consistent with our prediction, we observed significant reliable 6-fold modulation of the fMRI signal as a function of gaze movement direction bilaterally in EC (Fig. 1B). This result reflects greater fMRI response when gaze movement directions were aligned with the three grid axes than misaligned (Fig. 1C; Supplementary Fig. 2). Conducting the same analyses for other rotational symmetries, we found no evidence of reliable 90° or 45° periodic signals across independent halves of the data in EC (Fig. 1D). Notably, across participants, the magnitude of the 60° periodic EC signal significantly correlated with self-reported navigational ability, suggesting that the same population of grid cells might support both vision and navigation (Supplementary Fig. 2). All gaze movement directions were sampled during the visual search task, and we detected no 6-fold biases in gaze behavior that could explain the presence of a 6-fold symmetric fMRI signal (Supplementary Fig. 3). Thus these results are evidence of a grid representation in human EC that codes for locations in visual space, complementing previous findings of grid representations in navigable space5, 6. Grid-like coding of visual space was also observed in a medial prefrontal region of interest previously reported to exhibit a grid-like response during navigation5 (Supplementary Fig. 4).
We next explored the coordinate system that EC uses to encode visual space. For grid cells to provide useful information about environmental locations, grid cell firing patterns must be stably anchored to features of the external world, such as environmental boundaries. In previous work examining grid cells tiling visual space in monkey EC, the head of the animal was fixed relative to the visual display, making it difficult to determine whether these cells coded locations in egocentric (head-centered) or allocentric (world-centered) coordinates. To address this issue, we tested whether EC grid representations of visual space exhibit two signatures of boundary-anchored coding previously observed in rodent grid cells.
First, we asked whether grid orientations are reliably aligned by search display shape. When rodents explore square environments, the grid lattice aligns to ±7.5° from the cardinal axes of the borders2, 3 (Fig. 2A). We looked for a similar effect in the participants who searched square displays (Fig. 2B). Across these participants, the average EC visual grid orientations were significantly clustered around ±7.5° offset from the cardinal axes of the square display border (Fig. 2C). Moreover, of the 14/18 participants who showed significant clustering of grid angles across voxels in bilateral EC, 12/14 had grid angles significantly clustered across voxels 6°–9° offset from the display borders (Fig. 2D; Supplementary Fig. 5). By contrast, visual grid orientations for the rectangular display participants were not clustered around ±7.5° from the rectangular borders across participants (Fig. 2E). Indeed, grid orientations were closer to 7.5° offset from the display borders in the square display participants than in the rectangular display participants (t-test, t(34)=2.26, p=0.015, one-tailed). Because the shape of the display was the only stable environmental feature that differed between these participants, these results confirm that visual grid orientations were affected by the geometry of the visual environment.
Second, we examined whether rotation of the search display would induce a corresponding rotation of the visual grid. To address this question, each participant who performed the search task with the upright rectangular search displays also completed two additional scan runs in which the displays were rotated 30° clockwise (Fig. 3A). If the visual grid code is anchored to the borders of the search display, then rotation of the search display should yield a corresponding 30° rotation of the visual grid orientation, as observed in navigating rodents when chamber boundaries are rotated2. We found that the fMRI signal in EC during rotated runs was better predicted by a grid angle that was rotated 30° relative to the upright-display-fit grid orientation than it was by a grid angle that was not rotated (Fig. 3B; Supplementary Fig. 6). Furthermore, the average grid orientation during rotated runs was offset 28.33°±2.87° (mean angle ± s.e.m.) relative to the grid orientation during upright runs (Fig. 3C). Interesting, 6 participants showed little grid angle rotation (Fig. 3D). Surprisingly, these non-rotating participants were faster at finding the target letter during the rotated-display runs than the participants whose grid orientations rotated (Fig. 3E). Thus, although visual grids were anchored to the borders of the search display on average, there were individual differences in which external reference frame was selected, and these differences had consequences for search behavior.
In sum, we report the first evidence that human EC represents locations in visual space using a grid code. This visual grid code exhibited two signatures of boundary-anchoring previously observed in rodent grid cells—alignment to boundaries based on the shape of the environment, and rotation when the environmental borders are rotated—indicating that similar computational principles anchor primate and rodent grid cells to the external world, even across different spatial domains. These results may illuminate a longstanding controversy over the representation of visual space. Previous evidence suggests that the mammalian visual system represents space in retinotopic coordinates7–9, which are updated before each eye movement based on information about the intended direction of the upcoming saccade10. Although non-retinotopic spatial codes are observed under some circumstances11, 12, it is often unclear whether these codes are egocentric (head-centered) or allocentric (world-centered), and evidence for an allocentric map that represents where a viewer is looking relative to stable visual environmental cues has remained sparse (although see refs. [13, 14]). The current results provide evidence for such a map and suggest a mechanism by which it might be generated. During navigation, grid cells are thought to perform path integration by using self-motion inputs1, 15 to update allocentric representations of location2, 3. We hypothesize that visual grid cells may use a similar path integration mechanism to update an allocentric representation of the current gaze position based on eye motion signals present in the hippocampus and EC16, 17. Beyond navigation, recent work has also shown that a grid-like code is used to represent both imagined and conceptual spaces18–20. Our data add to this growing body of work by showing that grid cells may provide the mechanism by which locations in visual space are coded, thus allowing us to form durable visuospatial representations that are stable across eye movements.
Online Methods
Participants
36 participants (14 male) took part in this experiment (mean age: 23; range: 18–32). All participants gave written consent and were paid for participating, in compliance with procedures approved by the University of Pennsylvania Internal Review Board. All had normal vision and reported to be in good health with no history of neurological disease. Data from 7 additional participants were collected but discarded before analysis of fMRI data due to poor eye tracking quality (6 because of inaccurate gaze reconstructions; one because of poor sampling of all gaze angles). Data from one additional participant was discarded due to excessive head motion during scanning (>3 mm average absolute head motion). Following scanning, each participant completed the Santa Barbara Sense of Direction (SBSOD) questionnaire21, which provides a standardized measure of self-reported navigational ability.
Visual search task
Participants completed a series of 6.5 min fMRI scan runs during which they performed a visual search task. Square display participants completed four runs and rectangular display participants completed six runs. Participants were randomly assigned to display shape groups. During each run, participants viewed visual search displays consisting of a single target letter ‘L’ shown amongst distractor letters ‘T’s (letter height = 0.74°). Participants were instructed to use their eyes to search for the target, and to press a button when they found the target letter. Each trial was self-paced, and lasted an average of 7.50±0.58 seconds (mean±s.e.m.). Stimuli were presented using Matlab (2016a, The MathWorks Inc., Massachusetts) and the Psychtoolbox22 (Version 3.0.11). A pseudo-random search display was generated on each trial, such that all letters had a random location within the borders of the search display shape, subject to the constraint that only partial overlap between the letters was permitted, and a random orientation. Note that this meant that the shape implied by the array of letters was the same as the shape defined by the drawn border. Each search display had one of three possible densities ([100, 144, 169] or [81, 100, 121] letters total in the square and rectangle conditions, respectively). The search display density was randomly selected on each trial, with the constraint that each of the three possible densities was presented once before repeating. Search displays subtended a visual angle of 17.0° × 17.0° (square participants) or 11.0° × 17.0° (rectangular participants), and the search display border line thickness was 0.21°. There was a variable inter-trial interval of 2–6 seconds, randomly selected on each trial, during which participants fixated on a centrally located fixation cross. The onset of each trial was time-locked to the onset of an fMRI acquisition.
For the rectangular display participants, four scan runs consisted of upright rectangular displays, and two runs consisted of rectangular displays rotated by 30° clockwise. For these participants, the presentation order of the displays was URUURU, where U and R correspond to upright and rotated displays, respectively. This ordering ensured that any effect of display rotation could not be due to general drift across runs.
Eye tracking methods and preprocessing
Participant’s gaze position during scanning was monitored and recorded using a LiveTrack AV MR-compatible eye tracking camera (Cambridge Research Systems, Rochester England). The gaze position of the right eye was recorded at 30 Hz. Prior to each scan run, gaze position was calibrated using a series of nine fixation points evenly spaced between −8° and +8° in the horizontal and vertical dimensions relative to screen center. The average calibration error across all runs was 0.332° ± 0.018° (mean ± s.e.m.). In order to separate periods of gaze movements from periods of fixations, periods of gaze movement were defined by a movement-velocity-thresholding procedure, as follows. To reduce misattribution of gaze movements to eye tracking noise, the gaze position time course was first temporally smoothed with a boxcar filter (half width = 0.185 seconds). Gaze movements were then identified based on a median split of the smoothed gaze movement instantaneous velocity. Gaze position measurements in the bottom half of gaze movement velocities were treated as no movement, as were samples during which participants blinked. Gaze movements with velocities in the upper median half tended to be long saccades relative to the size of the search displays, with an average ballistic gaze trajectory length of 1.60° ± 0.60° (mean ± standard deviation). Note that this velocity-thresholding procedure is conservative in that it excludes short gaze movements during which we would not expect to observe a strong grid-like fMRI signal, based on previous fMRI studies of human navigation5, 18. Based on this method of classifying gaze movements, 7.1% ± 0.57% (mean ± s.e.m.) of all fMRI acquisitions contained no gaze movements whatsoever for the entire duration of the acquisition, which served as the implicit baseline relative to which fMRI signal change was measured.
fMRI acquisition
Scanning was performed at the Center for Functional Imaging at the University of Pennsylvania using a 3T Siemens Prisma scanner equipped with a 64-channel head coil. High-resolution T1-weighted images for anatomical localization were acquired using a 3-dimensional magnetization-prepared rapid-acquisition gradient-echo pulse sequence (repetition time [TR], 1620 ms; echo time [TE], 3.09 ms; inversion time, 950 ms; voxel size, 1x1x1 mm; matrix size, 192x256x160). T2*-weighted images sensitive to blood oxygenation level-dependent contrasts were acquired using a gradient-echo echoplanar pulse sequence (TR, 1000 ms; TE, 25 ms; flip angle 45°; voxel size, 2x2x2 mm; field of view, 192; matrix size, 96x96x78; multiband acceleration factor of 4). Ten additional fMRI volumes were also collected at the start of each scan run that were excluded from data analysis to account for signal steady-state transition. Visual stimuli were displayed at the rear bore face on an InVivo SensaVue Flat Panel Screen at 1920 × 1080 pixel resolution (diag = 80.0 cm, w × h = 69.7 × 39.2 cm). Participants viewed the stimuli through a mirror attached to the head coil. Behavioral responses were collected using a fiber-optic button box.
fMRI analysis – preprocessing
FMRI data analysis was carried out using FSL FEAT (FMRIB’s Software Library, version 6.00, www.fmrib.ox.ac.uk/fsl). The following standard data preprocessing was performed: motion correction using MCFLIRT23, non-brain removal using BET24; spatial smoothing using a Gaussian kernel of FWHM 8mm; grand-mean intensity normalization of the 4D dataset by a single multiplicative factor for each scan run; highpass temporal filtering (Gaussian-weighted least-squares straight line fitting, with sigma=50.0s). For second-level group analyses, EPI images were registered to the high-resolution anatomical image using boundary-based reconstruction and then normalized into standard space (MNI305) using non-linear registration. All data normalization was performed using Freesurfer (version 5.3.0, http://surfer.nmr.mgh.harvard.edu/).
fMRI analysis – Identifying grid-like coding of visual space
We performed a split-half analysis to estimate the orientation of the visual grid code during periods of gaze movement, following procedures used previously to identify grid-like codes during virtual navigation5, 18, 25, 26 (Supplementary Fig. 1). Data were first split into halves by run (runs [2,4] and [1,3] for square participants; runs [1,4] and [3,6] for rectangle participants, so that only the upright rectangle runs were used to identify grid-like coding in this initial analysis). For each half of the data, we identified the angular orientation of the putative visual grid axes in each participant’s bilateral EC. The grid orientation thus obtained was then subsequently used to predict a grid signal during the other independent half of the runs.
To fit the orientation of the 6-fold gaze movement direction-modulated signal within EC, we constructed a general linear model (GLM) with two parametric modulators (PMs) for periods of gaze movement. These two PMs were cos(6a(t)) and sin(6a(t)), where a(t) is the gaze movement direction sampled at time t (30 Hz). Each PM was down-sampled to the TR (1 Hz) by summing the values of the PM within each TR. The weights (b1 and b2) on these PMs were fitted to the fMRI time series for each voxel within the anatomically defined bilateral EC ROI. This EC ROI was constructed uniquely for each participant based on the automatic anatomical parcellation of the EC derived from FreeSurfer structural reconstruction. We then calculated the orientation of the 6-fold gaze movement direction-modulation from the mean weights across all voxels in the EC ROI as φ = [arctan(<b2>/<b1>)]/6, separately for each run, where arctan was mapped into 360° space, varying between 0° and 60°, according to signs of b2 and b1. Finally we computed the circular average orientation across runs for each separate half of the data. Grid orientations were quantitatively similar if we averaged the beta weights voxel-wise across runs before calculating the orientation instead of averaging the orientations across runs (circular correlation across participants: c=0.84, p<10−5).
To test whether the fit orientations predicted the analogous 6-fold periodic signal in the other independent half of the data, we constructed a GLM with a PM modeling the effect of gaze movement direction on the fMRI signal. The value of this PM at each timepoint was the cosine of gaze movement direction at that timepoint aligned to the orientation predicted by the first half of the data, cos(6(a(t)- φ)), where a(t) is the gaze movement direction sampled at time t (30 Hz). This PM was down sampled to the TR by summing the values of the PM within each TR. Each beta from this analysis reflects the extent of reliable split-half φ-oriented 6-fold gaze movement direction-modulated fMRI signal (which we call “orientation consistency”). The beta weights for this PM were averaged across all scan runs within each participant. The group-level test of the significance of these weights was small-volume FWE-corrected (SVC) within a group-level bilateral EC ROI, which was defined as the union of all individual-participant anatomically-defined EC ROIs projected into MNI space. To confirm that the gaze movement direction-modulated signal in EC exhibited a specifically 60° periodicity, we conducted this same split-half analysis for 90° (i.e., 4-fold) and 45° (i.e., 8-fold) periodicities.
Analyses were performed in FSL using FILM with local autocorrelation correction27. Included in all GLMs was a binary boxcar regressor of no interest corresponding to periods of visual search and its temporal derivative, as well as six nuisance PMs to account for head motion-related artifacts. All regressors were convolved with double gamma hemodynamic response function and filtered by the same high pass filter as the fMRI data before entry into the GLM.
fMRI analysis – reliable offset of the grid-like representation orientation from the search display shape
To test whether grid orientations across participants consistently cluster around an offset of ±7.5° from the cardinal axes of the search display borders, we first computed the average of the grid orientations across all EC voxels and runs within each participant. We then folded the grid orientations of all participants by φ mod 15°, which would align all hypothesis-consistent alignments to 7.5° in a circular 0° to 15° space. Next we performed a V-test for nonuniformity centered around 7.5°. The V-test is similar to Rayleigh’s test for circular uniformity with the difference that under the alternative hypothesis the distribution is non-uniform centered at a particular hypothesized angle (in this case, 7.5°)28, 29.
To test whether grid orientations in voxels within individual participants clustered around 7.5° offset from cardinal axes, we first evaluated whether EC voxels in each participant showed orientation clustering around any angle. To do so, we averaged the grid angle derived from each voxel across runs, yielding a distribution of voxel-wise grid orientations. Next we tested these voxel-wise grid orientations for non-uniformity using Rayleigh’s test for circular data. Note that because grid orientations were averaged voxel-wise across fMRI runs for this analysis, significant orientation clustering also requires temporal stability across runs of the grid orientation across voxels. This analysis identified participants who had significant nonuniformity, i.e. orientation clustering, of grid angles in EC (p<0.05, accounting for spatial smoothing). Finally, we tested whether the voxel-wise grid orientations in participants with significant clustering were specifically clustered around 6.0°–9.0° degrees in 0.5° increments (via folding and V-test, as described above; p-values were Bonferonni corrected for the seven grid angles tested).
fMRI analysis – rotation of the grid-like representation orientation with rotation of the search display
To test whether the visual grid orientations of rectangular-display participants rotated in concert with the rotated displays, we first computed the circular average of the grid orientation derived from each upright-display run. Next we rotated this average grid orientation, φ, by 30° and used this rotated orientation to predict the fMRI signal during rotated-display runs with a GLM. Specifically, a single PM to was used to model the effect of gaze movement direction on the fMRI signal during the rotated rectangle runs: a cosine of gaze movement angle aligned to the 30° rotated grid orientation, cos(6(a(t)-(φ+30°))). Positive weights from this analysis indicate that the 60° periodic fMRI signal is better predicted when the orientation of the grid axes is rotated 30° from φ during rotated-display runs, whereas negative weights indicate that rotated display runs are better predicted by the original grid orientation φ without rotation. The weights for this PM were first combined across both rotated-display runs in each participant, and then tested across participants with small-volume FWE-correction within the group-level bilateral EC ROI.
To examine the distribution of rotation effects across EC voxels, we first identified the grid orientation for each voxel during the rotated scan runs in the same fashion as we did for the upright scan runs. We then compared the distribution of grid orientations across all EC voxels when the display was upright to the distribution when the display was rotated. Specifically, for each participant, we subtracted φ from each EC voxel’s grid orientation, separately for the upright and rotated display runs, so that the average grid orientation across voxels were aligned relative to φ in each participant. We then calculated the distribution of voxels with grid orientations occurring from 0°–60° in 2° increments separately for the upright and rotated scan runs, and subtracted the upright distribution from the rotated distribution. If grid orientations across voxels rotate in concert with rotation of the search display, then there should be a higher percentage of voxels with grid angles around φ+30° when the display is rotated than upright.
Statistics
No statistical methods were used to pre-determine sample sizes, but our sample size was similar to those reported in previous publications18–20. Parametric t-tests and non-parametric sign-tests were used throughout the paper. For each parametric test, unless otherwise noted, data values met normality assumptions (Lilliefors test, p>0.05). If data did not meet normality assumptions, only sign-tests are reported. Rayleigh’s tests and V-tests were also used, as described in detail in the two preceding fMRI analysis methods sections. Data collection and analysis were not performed blind to the conditions of the experiment. A Life Sciences Reporting Summary is available.
Code availability
The code that support the findings of this study are available from the corresponding authors upon request.
Data availability
The data that support the findings of this study are available from the corresponding authors upon request.
Supplementary Material
Acknowledgments
This work was supported by US NIH Grants EY-022350 and EY-027047 (R.A.E.), NSF Grant SBE-0541957 (R.A.E.) and Graduate Research Fellowship (J.B.J.). We thank G. Aguirre for providing the eye tracking camera, and J. Ryan and S. Liapis for help with data collection.
Footnotes
Author Contributions
J.B.J., A.T.K., and R.A.E. designed study. Data collection was performed by J.B.J., A.T.K., and G.F., and J.B.J. and A.T.K. analyzed data. The manuscript was drafted by J.B.J., A.T.K., and R.A.E.
Competing Financial Interests Statement
The authors declare no competing financial interests.
References
- 1.Hafting T, Fyhn M, Molden S, Moser MB, Moser EI. Microstructure of a spatial map in the entorhinal cortex. Nature. 2005;436:801–806. doi: 10.1038/nature03721. [DOI] [PubMed] [Google Scholar]
- 2.Krupic J, Bauza M, Burton S, Barry C, O’Keefe J. Grid cell symmetry is shaped by environmental geometry. Nature. 2015;518:232–235. doi: 10.1038/nature14153. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Stensola T, Stensola H, Moser MB, Moser EI. Shearing-induced asymmetry in entorhinal grid cells. Nature. 2015;518:207–212. doi: 10.1038/nature14151. [DOI] [PubMed] [Google Scholar]
- 4.Killian NJ, Jutras MJ, Buffalo EA. A map of visual space in the primate entorhinal cortex. Nature. 2012;491:761–764. doi: 10.1038/nature11587. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Doeller CF, Barry C, Burgess N. Evidence for grid cells in a human memory network. Nature. 2010;463:657–661. doi: 10.1038/nature08704. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Jacobs J, et al. Direct recordings of grid-like neuronal activity in human spatial navigation. Nature neuroscience. 2013;16:1188–1190. doi: 10.1038/nn.3466. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Irwin DE, Yantis S, Jonides J. Evidence against visual integration across saccadic eye movements. Attention, Perception, & Psychophysics. 1983;34:49–57. doi: 10.3758/bf03205895. [DOI] [PubMed] [Google Scholar]
- 8.Gardner JL, Merriam EP, Movshon JA, Heeger DJ. Maps of visual space in human occipital cortex are retinotopic, not spatiotopic. Journal of Neuroscience. 2008;28:3988–3999. doi: 10.1523/JNEUROSCI.5476-07.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Golomb JD, Kanwisher N. Higher level visual cortex represents retinotopic, not spatiotopic, object location. Cerebral Cortex. 2011:bhr357. doi: 10.1093/cercor/bhr357. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Duhamel JR, Colby CL, Goldberg ME. The updating of the representation of visual space in parietal cortex by intended eye movements. Science. 1992;255:90. doi: 10.1126/science.1553535. [DOI] [PubMed] [Google Scholar]
- 11.Galletti C, Battaglini PP, Fattori P. Parietal neurons encoding spatial locations in craniotopic coordinates. Experimental brain research. 1993;96:221–229. doi: 10.1007/BF00227102. [DOI] [PubMed] [Google Scholar]
- 12.Snyder LH, Grieve KL, Brotchie P, Andersen RA. Separate body-and world-referenced representations of visual space in parietal cortex. Nature. 1998;394:887–891. doi: 10.1038/29777. [DOI] [PubMed] [Google Scholar]
- 13.Dean HL, Platt ML. Allocentric spatial referencing of neuronal activity in macaque posterior cingulate cortex. Journal of Neuroscience. 2006;26:1117–1127. doi: 10.1523/JNEUROSCI.2497-05.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Rolls ET. Spatial view cells and the representation of place in the primate hippocampus. Hippocampus. 1999;9:467–480. doi: 10.1002/(SICI)1098-1063(1999)9:4<467::AID-HIPO13>3.0.CO;2-F. [DOI] [PubMed] [Google Scholar]
- 15.McNaughton BL, Battaglia FP, Jensen O, Moser EI, Moser MB. Path integration and the neural basis of the ‘cognitive map’. Nature Reviews Neuroscience. 2006;7:663–678. doi: 10.1038/nrn1932. [DOI] [PubMed] [Google Scholar]
- 16.Meister ML, Buffalo EA. Getting directions from the hippocampus: the neural connection between looking and memory. Neurobiology of learning and memory. 2016;134:135–144. doi: 10.1016/j.nlm.2015.12.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Killian NJ, Potter SM, Buffalo EA. Saccade direction encoding in the primate entorhinal cortex during visual exploration. Proceedings of the National Academy of Sciences. 2015;112:15743–15748. doi: 10.1073/pnas.1417059112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Horner AJ, Bisby JA, Zotow E, Bush D, Burgess N. Grid-like processing of imagined navigation. Current Biology. 2016;26:842–847. doi: 10.1016/j.cub.2016.01.042. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Constantinescu AO, O’Reilly JX, Behrens TE. Organizing conceptual knowledge in humans with a gridlike code. Science. 2016;352:1464–1468. doi: 10.1126/science.aaf0941. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Bellmund JL, Deuker L, Schröder TN, Doeller CF. Grid-cell representations in mental simulation. eLife. 2016;5:e17089. doi: 10.7554/eLife.17089. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Hegarty M, Richardson AE, Montello DR, Lovelace K, Subbiah I. Development of a self-report measure of environmental spatial ability. Intelligence. 2002;30:425–447. [Google Scholar]
- 22.Brainard DH. The psychophysics toolbox. Spatial vision. 1997;10:433–436. [PubMed] [Google Scholar]
- 23.Jenkinson M, Bannister P, Brady M, Smith S. Improved optimization for the robust and accurate linear registration and motion correction of brain images. Neuroimage. 2002;17:825–841. doi: 10.1016/s1053-8119(02)91132-8. [DOI] [PubMed] [Google Scholar]
- 24.Smith SM. Fast robust automated brain extraction. Human brain mapping. 2002;17:143–155. doi: 10.1002/hbm.10062. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Kunz L, et al. Reduced grid-cell-like representations in adults at genetic risk for Alzheimer’s disease. Science. 2015;350:430–433. doi: 10.1126/science.aac8128. [DOI] [PubMed] [Google Scholar]
- 26.Stangl M, Shine J, Wolbers T. The GridCAT: A toolbox for automated analysis of human grid cell codes in fMRI. Front Neuroinform. 2017;11:47. doi: 10.3389/fninf.2017.00047. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Woolrich MW, Ripley BD, Brady M, Smith SM. Temporal autocorrelation in univariate linear modeling of FMRI data. Neuroimage. 2001;14:1370–1386. doi: 10.1006/nimg.2001.0931. [DOI] [PubMed] [Google Scholar]
- 28.Berens P. CircStat: a MATLAB toolbox for circular statistics. J Stat Softw. 2009;31:1–21. [Google Scholar]
- 29.Zar JH. Biostatistical analysis. Pearson Education India; 1999. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The data that support the findings of this study are available from the corresponding authors upon request.