Skip to main content
The Journal of Neuroscience logoLink to The Journal of Neuroscience
. 2023 Aug 2;43(31):5723–5737. doi: 10.1523/JNEUROSCI.2043-22.2023

Scene Perception and Visuospatial Memory Converge at the Anterior Edge of Visually Responsive Cortex

Adam Steel 1,, Brenda D Garcia 1, Kala Goyal 1, Anna Mynick 1, Caroline E Robertson 1
PMCID: PMC10401646  PMID: 37474310

Abstract

To fluidly engage with the world, our brains must simultaneously represent both the scene in front of us and our memory of the immediate surrounding environment (i.e., local visuospatial context). How does the brain's functional architecture enable sensory and mnemonic representations to closely interface while also avoiding sensory-mnemonic interference? Here, we asked this question using first-person, head-mounted virtual reality and fMRI. Using virtual reality, human participants of both sexes learned a set of immersive, real-world visuospatial environments in which we systematically manipulated the extent of visuospatial context associated with a scene image in memory across three learning conditions, spanning from a single FOV to a city street. We used individualized, within-subject fMRI to determine which brain areas support memory of the visuospatial context associated with a scene during recall (Experiment 1) and recognition (Experiment 2). Across the whole brain, activity in three patches of cortex was modulated by the amount of known visuospatial context, each located immediately anterior to one of the three scene perception areas of high-level visual cortex. Individual subject analyses revealed that these anterior patches corresponded to three functionally defined place memory areas, which selectively respond when visually recalling personally familiar places. In addition to showing activity levels that were modulated by the amount of visuospatial context, multivariate analyses showed that these anterior areas represented the identity of the specific environment being recalled. Together, these results suggest a convergence zone for scene perception and memory of the local visuospatial context at the anterior edge of high-level visual cortex.

SIGNIFICANCE STATEMENT As we move through the world, the visual scene around us is integrated with our memory of the wider visuospatial context. Here, we sought to understand how the functional architecture of the brain enables coexisting representations of the current visual scene and memory of the surrounding environment. Using a combination of immersive virtual reality and fMRI, we show that memory of visuospatial context outside the current FOV is represented in a distinct set of brain areas immediately anterior and adjacent to the perceptually oriented scene-selective areas of high-level visual cortex. This functional architecture would allow efficient interaction between immediately adjacent mnemonic and perceptual areas while also minimizing interference between mnemonic and perceptual representations.

Keywords: fMRI, memory, scene perception, virtual reality, visual cortex, visuospatial context

Introduction

Perception is shaped by our immediate sensory input and our memories formed through prior experience (Albright, 2012; Rust and Palmer, 2021). Yet, how memory influences ongoing perception in the brain is poorly understood. A particular puzzle is how the brain enables sensory and mnemonic representations to functionally interface while avoiding interference that occurs when the same neural population concurrently represents both sensory and mnemonic information (Rust and Palmer, 2021). One elegant system for studying these questions is scene perception, which relies heavily on concurrent mnemonic-sensory interactions: as we explore our environment, memory of the visuospatial content immediately outside of the current FOV informs what we will see next as we shift our gaze (Haskins et al., 2020; Draschkow et al., 2022).

How does the brain's functional architecture support coexisting representations of the current visual scene and memory of the surrounding environment? Recent studies suggest that scene-related mnemonic and sensory information converges at the anterior edge of high-level visual cortex (Baldassano et al., 2013, 2016; Silson et al., 2016; Bainbridge et al., 2021; Popham et al., 2021; Steel et al., 2021). For example, we recently localized brain regions that preferentially activated when participants recall personally familiar real-world places (e.g., their house) versus faces (e.g., their mother). This revealed three patches of cortex located anterior and adjacent to one of the three classic “scene perception areas” of high-level visual cortex that selectively activated when recalling familiar places (Steel et al., 2021). These “place memory areas” constituted a distinct functional network from the scene perception areas and cofluctuated with both spatial memory structures (i.e., the hippocampus) and scene perception areas, bridging mnemonic and perceptual systems. Based on these results, we proposed that these paired perceptually and mnemonically oriented functional networks may, respectively, represent scenes and their remembered visuospatial context. This possibility is attractive because it would allow efficient interaction between adjacent mnemonic and perceptual areas while minimizing mnemonic-perceptual interference (Steel et al., 2021).

However, while previous work identified place-recall driven activity anterior to each of the scene-perception areas, the content of the representations in these anterior regions is currently unclear. This is partly because prior studies investigating place memory, including our own, used participants' personally familiar real-world places as stimuli (e.g., “your mother's kitchen”) (Peer et al., 2015, 2019; Steel et al., 2021). While personal memories elicit strong responses and are appropriate for some experimental questions, these memories contain rich associations that go beyond visuospatial content (i.e., the appearance of your mother's kitchen), including semantic knowledge (e.g., where her kitchen is located), multisensory details (e.g., the smell of baking cake), and episodic context (e.g., blowing our birthday candles). As a result, it is unclear whether visuospatial memory per se is sufficient to engage the anterior memory areas. Likewise, whether the anterior areas process information about the visuospatial context of a scene is currently unknown.

To address this knowledge gap, we used head-mounted virtual reality (VR) to teach participants a set of real-world environments. This approach allowed us (1) to ensure participants' memories contained no confounding episodic information and limited semantic content, and (2) to precisely manipulate the visuospatial context associated with each environment. We defined visuospatial context as the spatial extent, configuration, and visual details of the local environment. Next, we used individualized, within-subject fMRI to test whether, across the brain, the place memory areas play a special role in processing visuospatial memory. Specifically, we asked whether these areas represent the extent of the visuospatial context associated with a visual scene during recall (Experiment 1) and visual recognition (Experiment 2) of single scene views from each place (Fig. 1B,C). We hypothesized that, if the place memory areas process visuospatial context associated with a scene view, (1) their activity will reflect the degree of associated visuospatial context (Bar and Aminoff, 2003; Baumann and Mattingley, 2016), and (2) they should contain multivariate representations of the scene's specific visuospatial content.

Figure 1.

Figure 1.

Experimental design to test processing of visuospatial context associated with a real-world scene. The experiment took place across three sessions: one study session and two fMRI sessions. These sessions took place on 3 separate days. A, In the VR study session, participants studied 20 real-world visual scenes with varying amounts of visuospatial context in head-mounted VR. Each scene was associated with one of three levels of spatial visuospatial context: single images (45° visible, 315° occluded), panoramas (270° visible, 90° occluded), and streets (three contiguous, navigable 360° photospheres). During the two fMRI sessions, we tested the neural response during recall (B) and recognition (C) of the learned scenes using fMRI.

Materials and Methods

Participants

Seventeen subjects (10 female; mean ± SD age = 21.6 ± 2.9 years) participated in this experiment. This number was based on a power calculation from our prior work, specifically based on the average effect size observed when participants viewed panning movies of familiar versus unfamiliar scenes across ROIs (Cohen's d = 0.76 with 80% power and α = 0.05 requires a minimum of 16 subjects; Experiment 2) (Steel et al., 2021). Specifically, from our prior work, we calculated the difference in activation between the place memory and scene perception areas on the lateral, ventral, and medial surfaces to familiar versus unfamiliar videos, as this was our best proxy for visuospatial context. We then took the average difference score across cortical surfaces for each subject and used these values for our power calculation. We felt confident assuming this large effect size because of the consistency in the effect across subjects (all subjects showed greater familiarity effects in the place memory areas compared with the scene perception areas).

Participants were Dartmouth students and post docs recruited via posters and word-of-mouth advertising. They received monetary compensation for taking part in the study. Participants had normal or corrected-to-normal vision, were not colorblind, and were free from neurologic or psychiatric conditions. Written consent was obtained from all participants in accordance with the Declaration of Helsinki with a protocol approved by the Dartmouth College Committee for the Protection of Human Subjects.

Experimental design

The study took place over three sessions (each on a different day): one study session and two fMRI sessions. During the study session, participants studied 20 real-world places using a head-mounted VR system (see Stimulus presentation). The 20 places comprised three spatial context conditions (see Stimuli). During the first fMRI session (Experiment 1), we investigated which brain areas represent the level of known spatial context by assessing the neural response across the three spatial context conditions using fMRI. During the second fMRI session (Experiment 2), we tested whether the same regions that represent context during recall also represent known spatial context during visual recognition by evaluating neural responses when viewing images of the studied places. Seventeen participants took part in Experiment 1, and 16 participants took part in Experiment 2.

Stimuli: study phase

Stimuli consisted of 20, 360° panoramic photospheres of outdoor real-world places from five cities (Oslo, Philadelphia, Padua, Quebec, Seville), sourced from Google Maps using iStreetView (www.istreetview.com). Participants were not told about the real-world locations from which the photospheres were taken. Stimuli can be found on Open Science Framework in the “panoramas” folder: https://osf.io/3ygqb/?view_only=5f436411c021448abf5157df3f2b2f30. At the end of the second fMRI session, we asked participants whether they recognized any photospheres from real-life (i.e., had they been to any of these locations outside of VR). Importantly, no participant recognized any stimuli, suggesting that real-world familiarity could not explain our results.

Example videos from the study phase on Open Science Framework in the “trial videos” folder: https://osf.io/3ygqb/?view_only=5f436411c021448abf5157df3f2b2f30. Stimuli were applied to a virtual environment created in Unity (www.unity3d.com), and the experimental routine was programmed using custom scripts written in C#.

The photospheres were divided into three spatial context conditions: (1) single images (45° from a photosphere, 315° occluded by a gray screen); (2) panoramas (270° visible, 90° occluded); and (3) streets (three contiguous 360° photospheres tiling a city street, navigable via button press) (Fig. 2). The cities were represented equally in all conditions (i.e., from each city, 2 photospheres were used in as image trials, 1 photosphere was a panorama trial, and 3 contiguous photospheres comprised a street trial). During the study session, we occluded a portion of the photosphere in the Image and Panorama conditions to obscure the wider visuospatial context. Because of the unique demands of curating and presenting photospheres for the street condition, all photospheres were assigned to the same conditions across participants. Therefore, it is possible that different levels of memory for the photospheres could contribute to a main effect of condition in Experiments 1 and 2. Importantly, however, our a priori hypotheses deal with interactions between conditions and ROI (i.e., differences in activation magnitude between perceptual and mnemonic areas), which would not be impacted by this main effect of condition. Moreover, we found no differences across conditions in our behavioral assessments of memory, suggesting that condition-related differences we observed in neural activity were not driven by stimulus-related effects.

Figure 2.

Figure 2.

Aerial view of three study conditions. The “key scene views” were used as the stimuli for Experiment 2. The key scene view was always looking “down a street” to control for visual information across the conditions.

Unbeknownst to the participants, the image condition was comprised of two sets of five images derived from (1) the occluded portion of the panorama condition and (2) an isolated photosphere from an independent location within the same city. We found no difference between the image conditions in any ROIs during the recall and recognition experiments (Experiment 1 (recall): OPA: t(16) = 0.95, p = 0.35; PPA: t(16) = 0.81, p = 0.43; MPA: t(16) = 0.58, p = 0.56; LPMA: t(16) = 0.83, p = 0.42; VPMA: t(16) = 0.84, p = 0.41; MPMA: t(16) = 0.75; p = 0.46; Experiment 2 (recognition): OPA: t(15) = 1.84, p = 0.086; PPA: t(15) = 0.0007, p = 0.99; MPA: t(15) = 0.75, p = 0.46; LPMA: t(15) = 0.013, p = 0.99; VPMA: t(15) = 0.86, p = 0.40; MPMA: t(15) = 0.63, p = 0.53), so we considered these images together as a single Image condition.

The Street condition consisted of three contiguous photospheres representing a city street. After entering the trial, participants rotated 180° from their initial facing direction to see an arrow on the floor, which indicated that they could advance forward. This rotation ensured that the key scene view presented during fMRI (i.e., initial viewpoint) did not depict the extra visuospatial context experienced by the participant. Participants advanced forward by pressing a button on a handheld controller. Participants could advance forward two positions, allowing them to view around a street corner unobservable from the initial position. Participants could not go backward to revisit photospheres during the street trial (i.e., they could only travel “one-way”).

Stimulus presentation

During the study phase, experimental stimuli were presented to participants via an immersive, head-mounted VR display (16 Oculus Quest 2, 1 Oculus Go; Fast-switch LED display panel 1832 × 1920 resolution per eye; 98° FOV; 120 Hz refresh rate). Participants stood wearing the head-mounted VR display during the experiment. This setup offered participants a self-directed opportunity to explore the naturalistic environment from an egocentric perspective naturally via eye movements and head turns. All fMRI stimuli were presented using Psychopy 3 (version 3.2.3) (Peirce, 2007).

Procedure: study phase

During the study session, participants studied all 20 scenes (20 s each). Study trials automatically terminated after 20 s. On each trial, participants were instructed to look naturally around each environment, just as they would in real life. Importantly, although the view was restricted in the Image and Panorama conditions, participants could still turn their heads and move their eyes to naturally explore the environment. Therefore, the conditions only differed in the amount of spatial context that participants learned.

Before and after each study trial, participants saw an associated place “name” (e.g., “Scallop path”) that later served as a cue during the fMRI recall experiment. We instructed participants to learn these name-place pairings for a later memory test. After each experimental trial, participants returned to a virtual home screen (a platform surrounded by clouds), where they were instructed to take a break. Time between trials was self-paced (participants initiated the beginning of a new trial via a button press).

During the study session, participants saw the 20 scenes grouped in blocks. No scene was repeated during a study block. Scene order was randomized in each study block. All participants completed at least 5 study blocks (∼1 h of training, depending on subjective confidence and availability). After the fifth block, we asked participants if they were confident that they could recall each place; if participants were not confident, they completed one additional study block. Importantly, all analyses below are within-subject, minimizing the potential impact of across-subject differences in study session duration or stimulus familiarity.

All participants completed a “refresh” study block immediately before entering the scanner on both scan sessions.

fMRI

Experiment 1: memory recall

The goal of Experiment 1 was to determine what brain areas represent the known visuospatial context associated with a scene view. To this end, we compared BOLD activation when participants recalled places that varied in their degree of known spatial context.

Experiment 1 was performed across eight runs. Each place stimulus was cued once in each run (20 trials/run). On each trial, participants saw the name of a place from the study phase (1 s), followed by a dynamic mask of random alphanumeric characters that rapidly changed (1 s). After the mask, the recall cue (“oooo”) appeared on the screen for 10 s. We instructed participants to perform mental imagery of a single FOV from the prompted place in as much detail as possible while the recall cue was on the screen. The specific instructions given to the participant are reproduced below:

“During scanning you will perform the memory recall task. In this task, you will see the name of a place that you studied, followed by some scrambled letters and numbers. Then four circles will appear on the screen. When the four circles are on the screen, we want you to visually recall a view from the cued place as vividly as you can. When you are doing the imagery, you should only imagine a single image, whatever the first image that comes to mind when you are cued for that place. So, for Church Street, you might immediately recall the church at the top of the hill — you should hold that image in your mind the whole time and not try to look around the place. When the cue goes away, a cross will appear. When the cross is on the screen, clear your mind and prepare for the next trial.”

Imagery trials lasted 10 s and were separated by a jittered interstimulus interval (range 4-8 s). The stimulus order was pseudo-randomized within each run so that conditions could only repeat 2 times consecutively in an imaging run. Although stimuli were not presented in an optimized sequence, we confirmed that our regressors were minimally collinear (correlation between regressors < 0.2 for 100% of task regressors across participants) post hoc using afni's 1d_tool.py.

Experiment 2: spatial context during recognition

In Experiment 2, we assessed whether the same regions that represent known spatial context during recall also represent contextual information during visual recognition. For this experiment, we compared neural activity when participants viewed briefly presented images while they performed an orthogonal familiarity judgment task. We hypothesized that by having subjects perform a task that was unrelated to spatial context, any activation that was modulated by spatial context would reflect implicit reactivation of contextual knowledge.

Experiment 2 was comprised of eight runs. Participants performed a familiar/novel judgment task. On each trial, participants saw a single 45° portion of a photosphere (500 ms), and then indicated whether they had studied the image in VR (“familiar”) or not (“novel”) (1500 ms). Trials were separated by a jittered interstimulus interval (range 4-8 s). Each familiar stimulus was presented once in each run, along with four novel stimuli (24 trials/run). The stimulus order was pseudo-randomized so that conditions could only repeat 2 times per run. For consistency, all stimuli were images looking down the center of a street, which limited the amount of identifying visual information the participant saw. Because overall accuracy was high (>92% correct), all trials were included in the analysis.

Novel stimuli for each run were randomly sampled from a set of 10 images participants had not experienced during the study session (2 novel images per city). One image was taken from the unseen portion of the photosphere from the panorama condition (180° from the studied view). The other was taken from a photosphere that had not been experienced by the participant and always comprised a view looking down the street. Novel photospheres were chosen to be visually similar to the other stimuli from their respective city.

Although stimuli were not presented in an optimized sequence, we confirmed that our stimulus time series were minimally collinear (collinearity < 0.2 for 99.9% of task regressors across participants) post hoc using afni's 1d_tool.py.

Localizer tasks

In addition to the main experimental tasks, participants underwent localizers to establish individualized ROIs for the scene perception areas (PPA, OPA, MPA) and place memory areas (VPMA, LPMA, MPMA). We used the same localization procedure defined in our prior work (Steel et al., 2021), briefly described below.

Scene perception localizer

To localize the scene perception areas, participants passively viewed blocks of scene, face, and object images presented in rapid succession (500 ms stimulus, 500 ms interstimulus interval). Blocks were 24 s long, and each run consisted of 12 blocks (four blocks/condition). There was no interval between blocks. Participants performed two runs of the scene perception localizer.

Place memory localizer

The place memory areas are defined as regions that selectively activate when an individual recalls personally familiar places (i.e., their kitchen) compared with personally familiar people (i.e., their mother). To establish individualized stimuli, before fMRI scanning, we had participants generate a list of 36 personally familiar people and places (72 stimuli total). These stimuli were generated based on the following instructions.

“For your scan, you will be asked to visualize people and places that are personally familiar to you. So, we need you to provide these lists for us. For personally familiar people, please choose people that you know in real life (no celebrities) that you can visualize in great detail. You do not need to be in contact with these people now, as long as you knew them personally and remember what they look like. So, you could choose a childhood friend even if you are no longer in touch with this person. Likewise, for personally familiar places, please list places that you have been to and can richly visualize. You should choose places that are personally relevant to you, so you should avoid choosing places that you have only been to one time. You should not choose famous places where you have never been. You can choose places that span your whole life, so you could do your current kitchen, as well as the kitchen from your childhood home.”

During fMRI scanning, participants recalled these people and places. On each trial, participants saw the name of a person or place and recalled them in as much detail as possible for the duration that the name appeared on the screen (10 s, no dynamic mask). Trials were separated by a variable interstimulus interval (4-8 s). Place memory areas were localized by contrasting activity when participants recalled personally familiar places compared with people (see ROI definitions).

MRI acquisition

All data were collected at Dartmouth College on a Siemens Prisma 3T scanner (Siemens) equipped with a 32-channel head coil. Images were transformed from dicom to nifti format using dcm2niix (version 1.0. 20190902) (Li et al., 2016).

T1 image

For registration purposes, on day 1, a high-resolution T1-weighted MPRAGE imaging sequence was acquired (TR = 2300 ms, TE = 2.32 ms, inversion time = 933 ms, flip angle = 8°, FOV = 256 × 256 mm, slices = 255, voxel size = 1 × 1 × 1 mm). T1 images were segmented, and surfaces were generated using Freesurfer (version 6.0) (Dale et al., 1999; Fischl et al., 2002) and aligned to the fMRI data using align_epi_anat.py and @SUMA_AlignToExperiment (Argall et al., 2006; Saad and Reynolds, 2012).

On day 2, we collected a fast anatomic scan (MPRAGE; TR = 2300 ms, TE = 2.32 ms, inversion time = 933 ms, flip angle = 8°, FOV = 256 × 256 mm, slices = 255, voxel size = 1 × 1 × 1 mm, GRAPPA factor = 4). The Freesurfer reconstructions from the scan collected in day 1 were registered to this scan for analysis of Experiment 2 using @SUMA_AlignToExperiment.

fMRI acquisition

To mitigate the dropout artifacts and allow advanced BOLD preprocessing, fMRI data were acquired using a multiecho T2*-weighted sequence (Poser et al., 2006; Kundu et al., 2012, 2017; Posse, 2012). The sequence parameters were as follows: TR = 2000 ms, TEs = [14.6, 32.84, 51.08], GRAPPA factor = 2, flip angle = 70°, FOV = 240 × 192 mm, matrix size = 90 × 72, slices = 52, multiband factor = 2, voxel size = 2.7 mm isotropic. The initial two frames of data acquisition were discarded by the scanner to allow the signal to reach steady state.

Preprocessing

Multiecho data processing was implemented based on the multiecho preprocessing pipeline from afni_proc.py in AFNI (Cox, 1996). Signal outliers in the data were attenuated (3dDespike) (Jo et al., 2013). Motion correction was calculated based on the second echo, and these alignment parameters were applied to all runs.

The data were then denoised using multiecho ICA denoising (tedana.py) (Kundu et al., 2012; Evans et al., 2015; DuPre et al., 2019, 2021) to improve overall data quality in ventral temporal and parietal cortex (Steel et al., 2022, 2023). In brief, the optimal combination of the three echoes was calculated, and the echoes were combined to form a single, optimally weighted time series (T2smap.py, included in the tedana package). PCA was applied, and thermal noise was removed using the Kundu decision tree method. Subsequently, ICA was performed, and the resulting components were classified as signal and noise based on the known properties of the T2* signal decay of the BOLD signal versus noise. Components classified as noise were discarded, and the remaining components were recombined to construct the optimally combined, denoised time series.

Following denoising, images were smoothed with a 5 mm Gaussian kernel (3dBlurInMask), and signals were normalized to percent signal change.

ROI definitions

Scene perception and place memory areas (functionally defined)

Scene perception regions were established using the same criterion used in our prior work (Steel et al., 2021). Scene and face areas were drawn based on a general linear test comparing the coefficients of the GLM during scene versus face blocks. These contrast maps were then transferred to the SUMA standard mesh (std.141) using @SUMA_Make_Spec_FS and @Suma_AlignToExperiment (Argall et al., 2006; Saad and Reynolds, 2012). A vertex-wise significance of p < 0.001 along with expected anatomic locations was used to define the ROIs (Julian et al., 2012; Weiner et al., 2018). As a control region, the fusiform face area (FFA) was defined using the contrast of scenes versus faces using the same procedure.

To define category-selective memory areas, the familiar people/places memory data were modeled by fitting a γ function of the trial duration for trials of each condition (people and places) using 3dDeconvolve. Estimated motion parameters were included as additional regressors of no interest. Polynomial regressors were not included in multiecho data analysis. Activation maps were then transferred to the SUMA standard mesh (std.141) using @SUMA_Make_Spec_FS and @Suma_AlignToExperiment. People and place-memory areas were drawn based on a general linear test comparing coefficients of the GLM for people and place memory. A vertex-wise significance threshold of p < 0.001 was used to draw ROIs.

To ensure our ROIs were comparable in size, we further restricted all functional ROIs to the top 300 most selective vertices (Steel et al., 2021). Importantly, our results were consistent using 300 vertex ROIs defined using the ROI center of mass (rather than most selective vertices) as well as 600 vertices rather than 300, suggesting that the results did not depend on the ROI definition.

Anatomical ROIs

The hippocampus and primary visual cortex (occipital pole) were defined for each participant using FreeSurfer's automated segmentation (aparc + aseg).

Statistical analysis

fMRI analysis

Experiment 1 Deconvolution

For Experiment 1, we used two standard GLM analyses: (1) group whole-brain analysis and (2) univariate and multivariate ROI analyses.

For the group-based whole-brain analysis, our GLM included regressors for each condition (image, panorama, street) that considered each trial using a γ function of duration 10 s. In addition, a regressor for the memory cues and dynamic masks of each trial was included as a regressor of no interest. These regressors, along with motion parameters and fourth-order polynomials, were fit to the data. Activation maps were then transferred to the SUMA standard mesh (std.141) using @SUMA_Make_Spec_FS and @SUMA_AlignToExperiment (Argall et al., 2006; Saad and Reynolds, 2012).

For ROI univariate and multivariate analyses, we analyzed data using a GLM fit with each trial modeled individually for each run using 3dDeconvolve. The memory recall period for each trial was modeled as a single regressor using a γ function of duration 10 s. In addition, a regressor for the memory cue and dynamic mask was included as a regressor of no interest using a γ function of duration 1 s. These regressors, along with motion parameters and fourth-order polynomials were fit to the data. Activation maps were then transferred to the SUMA standard mesh (std.141) using @SUMA_Make_Spec_FS and @SUMA_AlignToExperiment (Argall et al., 2006; Saad and Reynolds, 2012).

Whole-brain analysis

To locate where in the brain neural activity during recall was modulated by spatial context, we conducted a whole-brain analysis on the cortical surface. Beta values for each condition from all subjects were compared using a within-subject ANOVA (AFNI 3dANOVA2, Type 3) with condition (image, panorama, street) as a factor and participant as a random effect. Data were thresholded at a vertex-wise FDR-corrected p value (q) = 0.05 (F = 6.4, p < 0.0001). The result of this analysis (i.e., main effect of condition) was compared with the location of the place memory and scene perception areas from an independent group of participants (Steel et al., 2021).

To determine whether the neural activity of any subcortical or allocortical structures (i.e., hippocampus) was modulated by spatial context, we conducted a volumetric whole-brain analysis. Participants' anatomic data were registered to standard space (MNI template) using auto_warp.py, and the registration parameters were applied to their statistical maps. The β values for each condition were then compared using a within-subject ANOVA (AFNI 3dANOVA2, Type 3) with condition (image, panorama, street) as a factor and participant as a random effect. No subcortical or allocortical regions showed an effect of condition, even at a modest threshold (p = 0.05, uncorrected).

ROI analysis: univariate

The t values for all trials in all runs were extracted from all ROIs (scene perception areas [PPA, OPA, MPA], place memory areas [VPMA, LPMA, MPMA], hippocampus, V1, FFA; see ROI definition). These t statistics were averaged across runs and conditions, and these average t statistics were compared using an ANOVA implemented in MATLAB (The MathWorks).

For comparison between the scene perception and place memory areas, ANOVAs were conducted separately for each surface (ventral, lateral, medial) with condition (image, panorama, street), ROI (scene perception, place memory), and hemisphere (left, right) as factors. For all other regions (hippocampus, V1, FFA), separate ANOVAs were used, with condition (image, panorama, street) and hemisphere (left, right) as factors. There was no significant effect of hemisphere in any test (all p values > 0.05); thus, data are presented collapsed across hemisphere. Post hoc tests were implemented using Bonferroni-corrected paired t tests.

ROI analysis: multivariate

We sought to determine whether the scene perception and place memory areas contained representations of the specific places that participants recalled. In other words, we tested whether we could decode stimulus identity during trials of the recall experiment. To this end, we performed an iterative split-half correlation analysis and looked for correspondence in the spatial pattern of neural activity between each place across the two halves of the data. We focused on identity decoding indices because condition (i.e., image, panorama, street) decoding could be impacted by the differing response magnitude.

For this analysis, we iteratively half-split our runs, such that we created all possible combinations of half-split data (total half-splits = 70; iteration 1: 1, 2, 3, 4 vs 5, 6, 7, 8; iteration 2: 1, 2, 3, 5 vs 4, 6, 7, 8; and so on). For each iteration, we averaged t values for each place from half of the runs (e.g., odd vs even) together for each voxel within each ROI. We then calculated the correlation matrix (Spearman's ρ) across halves. This yielded a similarity matrix of all images for each ROI. We then calculated the identity decoding index as the difference in correlation between the pattern of response to a place with itself (across independent halves of data) compared with the correlation of the places with the other members of its visuospatial context condition (i.e., images were correlated with images, panoramas with panoramas, and streets with streets). This ensured that condition-related effects could not explain our decoding results. This was iteratively calculated for all combinations of runs.

Fisher-transformed correlation values for the scene perception and place memory areas on each surface were separately compared with hippocampus and control regions (V1 and FFA) using an ANOVA, with ROI (SPA, PMA, hippocampus, FFA, V1) as a factor and subject as a random effect. Post hoc tests were implemented using Bonferroni-corrected paired t tests.

Experiment 2

Data were analyzed using a GLM fit separately for each run using 3dDeconvolve. Each familiar image trial was modeled as a single regressor using a γ function of duration 2 s. In addition, a regressor for unfamiliar images was included as a regressor of no interest using a γ function of duration 2 s. Because unfamiliar images were not trial unique (i.e., some unfamiliar stimuli were repeated across runs) and differed in number from the familiar images, we could not analyze these data. These regressors, along with motion parameters and third-order polynomials, were fit to the data. Activation maps were then transferred to the SUMA standard mesh (std.141) using @SUMA_Make_Spec_FS and @SUMA_AlignToExperiment.

Whole-brain analysis

To locate where in the brain neural activity during recognition was modulated by spatial context, we conducted a whole-brain analysis on the cortical surface. Beta values for each condition from all subjects were compared using a within-subject ANOVA (AFNI 3dANOVA2, Type 3) with condition (image, panorama, street) as a factor and participant as a random effect. Data were thresholded at a vertex-wise FDR-corrected p value (q) = 0.05 (F = 8.1, p < 0.0001). The result of this analysis (i.e., main effect of condition) was compared with the location of the place memory and scene perception areas from an independent group of participants (Steel et al., 2021).

To determine whether the neural activity of any subcortical or allocortical structures (i.e., hippocampus) was modulated by spatial context, we conducted a volumetric whole-brain analysis. Participants' anatomic data were registered to standard space (MNI template) using auto_warp.py, and the registration parameters were applied to their statistical maps. The β values for each condition were then compared using a within-subject ANOVA (AFNI 3dANOVA2, Type 3) with condition (image, panorama, street) as a factor and participant as a random effect. As we found in Experiment 1, no subcortical or allocortical regions showed an effect of condition, even at a modest threshold (p = 0.05, uncorrected).

ROI analysis

In Experiment 2, we sought to determine whether the regions in which activation reflected spatial context during recall also reflect this information during visual recognition. So, we used the same univariate ROI analysis approach taken in Experiment 1.

The t values for all trials in all runs were extracted from all ROIs (scene perception areas [PPA, OPA, MPA], place memory areas [VPMA, LPMA, MPMA], hippocampus, V1, FFA; see ROI definition). These t statistics were averaged across runs and conditions, and we compared these average values using an ANOVA implemented in MATLAB.

For comparison between the scene perception and place memory areas, ANOVAs were conducted separately for each surface (ventral, lateral, medial) with condition (image, panorama, street), ROI (scene perception, place memory), and hemisphere (left, right) as factors. For all other regions (hippocampus, V1, FFA), separate ANOVAs were used, with condition (image, panorama, street) and hemisphere (left, right) as factors. There was no significant effect of hemisphere in any test (all p values > 0.05); and thus, data are presented collapsed across hemisphere. Post hoc tests were implemented using Bonferroni-corrected paired t tests.

Post-scan questionnaires

Vividness

To ensure that any differences in neural responses across the conditions during recall were not because of a difference in vividness of imagery, we queried participants on their experience after the scan. Specifically, participants rated each stimulus on the overall vividness of their visual imagery on a 1-5 scale (1: not able to visualize; 5: as though they were looking at the image). We compared the average ratings across conditions using an ANOVA with condition (image, panorama, street) as a factor and subject as a random effect.

Image naming test

To ensure participants could remember the trained images, participants performed an image naming test after completing Experiment 2. Eleven participants completed the recognition test; the remaining participants were not able to be contacted. One participant performed the naming task after substantial delay (>2 weeks after fMRI scanning) and was therefore excluded from this analysis. Participants were shown the images used in the Experiment 2 scan session and had to report the place name. We compared the accuracy across the conditions using an ANOVA with condition (image, panorama, street) as a factor and subject as a random effect.

Data and code availability

All data and code reported in this paper will be shared by the lead contact upon request.

Results

Participants (N = 17) learned a set of 20 immersive, real-world visuospatial environments in head-mounted VR. Each environment depicted a small portion of a city (e.g., Padua, Seville, etc.) where participants had never been. These environments comprised three levels of visuospatial context: (1) single images (45° from a photosphere, 315° occluded by a gray screen); (2) panoramas (270° visible, 90° occluded); and (3) streets (three contiguous 360° photospheres tiling a city street, navigable via button press) (Fig. 1A, see Materials and Methods). In the street condition, the final photosphere allowed the participant to look around a corner to view visuospatial information that could not be seen from their original position, which ensured that the known visuospatial context in this condition was greater than the panorama condition. The cities were equally balanced across the conditions to control for visual and semantic components unique to each city.

Participants explored each environment from a first-person perspective using head turns and eye movements for 20 s and learned: (1) each visuospatial environment and (2) a name associated with each environment. After studying, participants could vividly imagine each scene and accurately recall each scene-name pairing (see Post-scan behavioral results confirmed memory success). Then, we used fMRI to assess which brain areas process visuospatial context during recall (Experiment 1; Fig. 1B) and visual recognition (Experiment 2, Fig. 1C). In these tasks, participants either visually recalled or viewed pictures of a single scene view from each place (the key scene view, Fig. 1).

We began by identifying any brain areas where visuospatial context associated with scenes learned in VR modulated recall-related activation using an exploratory whole-brain ANOVA with condition (image, panorama, street) as a factor and subject as a random effect. In other words, we tested where in the brain activity was modulated by the extent of visuospatial context associated with a scene. The logic of this test builds on studies in the object memory literature showing that association cortex responds more strongly to objects with strong associated context (e.g., a bowling pin, which is strongly associated with a bowling alley) versus weak associated context (e.g., a chair) (Bar and Aminoff, 2003; Baumann and Mattingley, 2016). These univariate effects are thought to reflect activation of brain areas that harbor information from memory associated with a percept, in our case, visuospatial context associated with a scene.

This group-level, whole-brain analysis revealed several distinct clusters in posterior cerebral cortex that were modulated by known visuospatial context (Fig. 3). Three clusters, in particular, were located immediately anterior to the three scene perception areas (OPA, PPA, and MPA), as identified in an independent group of participants (Steel et al., 2021). Indeed, the location of these clusters closely corresponded to the location of the “place memory areas,” three brain areas on the brain's lateral (LMPA), ventral (VMPA), and medial surfaces (MPMA) that selectively respond when participants recall or perceive familiar places, as defined by a group mask from our previous work (Steel et al., 2021). On the ventral surface, the cluster fell at the anterior edge of the VPMA, which largely overlaps with the parahippocampal cortex (Glasser et al., 2016), a region known to represent an animal's relative position within an environment (Julian et al., 2018; LaChance et al., 2019; LaChance and Taube, 2022). These results confirm that the place memory areas, which we previously showed are activated during recall of personally familiar places (Steel et al., 2021), are active during recall of a purely visuospatial environment (i.e., a photosphere learned in VR, which is devoid of episodic or semantic information). Having found that the place memory areas are engaged during visuospatial recall, even when episodic and semantic information is controlled, we went on to test whether these areas play a role in processing visuospatial context.

Figure 3.

Figure 3.

Regions in posterior cerebral cortex are modulated by the extent of known visuospatial context during recall in a group analysis. We compared neural activation across the visuospatial context conditions (image, panorama, and street) using a vertex-wise ANOVA. We found that three clusters showed significant modulation by known spatial context. By comparing these clusters with our prior work mapping perception and memory-related responses (Steel et al., 2021), we determined that these regions fell anterior to group-average locations of the scene perception areas (white) and within the place memory areas (black). Whole-brain analysis thresholded at FDR-corrected p value (q) < 0.05 (vertex-wise F > 6.4, p < 0.0001). Scene perception areas (white) and place memory areas (black) were defined using data from an independent group of participants (Steel et al., 2021).

Place memory area activity reflects known visuospatial context associated with a scene view during recall

While whole-brain analyses are powerful exploratory tools, they obscure the highly detailed and idiosyncratic location of functional areas present at the individual participant level (Braga and Buckner, 2017; Gordon et al., 2017; Braga et al., 2019; DiNicola et al., 2020). So, to determine whether the place memory and scene perception areas have dissociable roles when processing visuospatial context, we adopted an individualized ROI-based approach. We independently localized each participant's OPA, PPA, MPA, and place memory areas on the lateral, ventral, and medial surfaces (LMPA, VPMA, MPMA) (see Materials and Methods). We then compared regional activation when participants recalled (Experiment 1) or visually recognized (Experiment 2) the scene views across the three visuospatial context conditions. For each cortical surface, we used an ANOVA with condition (image, panorama, street), ROI (scene perception/place memory area), and hemisphere (lh/rh) as factors and subject as a random effect. Because we found no effect or interaction with hemisphere, results are presented averaged across hemispheres.

This analysis revealed a remarkable dissociation between the scene perception areas and their anterior, memory-driven counterparts (the place memory areas) (Fig. 4A). Namely, on all cortical surfaces, visuospatial context modulated place memory area activity, while the scene perception areas were largely unaffected.

Figure 4.

Figure 4.

The extent of visuospatial context associated with a scene view modulates place memory area activation during recall. A, For each participant, we compared activation of individually localized scene perception and place memory areas across the visuospatial context conditions during recall using an ANOVA with ROI (perception/memory area), visuospatial context (image [i], panorama [p], street [s]), and hemisphere as factors. Across all cortical surfaces, the place memory area activation was modulated by spatial context to a greater degree than the scene perception areas. Gray dots indicate individual participants. See Figure 5 for control areas. B, Time course of activation during recall (in seconds) shows differential responses of the scene perception and place memory areas. On all cortical surfaces, the place memory areas showed robust activation during recall (4-12 s), while the scene perception areas on the ventral and lateral surfaces did not. Gray lines indicate SEM at each TR (2 s). *p < 0.05. **p < 0.01. ***p < 0.005.

On the lateral surface, LPMA activation increased concomitantly with visuospatial context, while OPA did not (ROI × condition interaction: F(2203) = 14.7, p < 0.0001; post hoc ANOVAs – OPA: F(2101) = 0.38, pcorr = 0.99; LPMA: F(2101) = 10.48, pcorr = 0.0006; difference across conditions, i.e., difference of differences: LPMA vs OPA – panorama vs image: t(16) = 2.771, pcorr = 0.0408, Cohen's d = 0.672; panorama vs street: t(16) = 2.767, pcorr = 0.0411, d = 0.671; street vs image: t(16) = 5.21, pcorr < 0.001, d = 1.26). The same pattern was present on the ventral surface – VPMA's activation increased across the conditions, but PPA's did not (ROI × condition interaction F(2203) = 5.29, p = 0.01; post hoc ANOVAs – PPA: F(2101) = 3.58, p = 0.078; LPMA: F(2101) = 8.19, pcorr = 0.0078; difference across conditions: VPMA vs PPA – panorama vs image: t(16) = 0.13, pcorr = 0.99, Cohen's d = 0.078; panorama vs street: t(16) = 2.68, pcorr = 0.048, d = 0.065; street vs image: t(16) = 2.83, pcorr = 0.036, d = 0.69). On the medial surface, both the MPA and MPMA activity increased with visuospatial context, which likely reflects the high degree of overlap between these regions (Steel et al., 2021) (MPA: F(2101) = 12.95, p < 0.001; post hoc test: LPMA vs OPA – panorama vs image: t(16) = 4.68, pcorr < 0.001, d = 0.74; panorama vs street: t(16) = 2.55, pcorr = 0.06, d = 0.64; street vs image: t(16) = 4.12, pcorr < 0.001, d = 1.17; MPMA: F(2101) = 19.18, p < 0.001; post hoc test: panorama vs image: t(16) = 5.61, pcorr < 0.001, d = 0.88; panorama vs street: t(16) = 3.15, pcorr = 0.018, d = 0.84; street vs image: t(16) = 5.02, pcorr < 0.001, d = 1.51). However, consistent with the lateral and ventral surfaces, activity in MPMA was modulated by spatial context to a greater degree than MPA (ROI × condition interaction: F(2203) = 4.24, p = 0.023; post hoc test: MPMA vs MPA – panorama vs image: t(16) = 2.01, pcorr = 0.18, d = 0.49; panorama vs street: t(16) = 1.4, pcorr = 0.51, d = 0.35; street vs image: t(16) = 2.75, pcorr = 0.042, d = 0.69).

To summarize, in contrast to the scene perception areas, all three place memory areas showed a strong effect of visuospatial context. Additionally, control regions occipital pole (V1) and FFA (Kanwisher et al., 1997), were not modulated by visuospatial context (V1: F(2101) = 1.68, p = 0.20; FFA: F(2,50) = 0.78, p = 0.46; Fig. 5). Further, activity of the hippocampus, which is involved in scene construction from memory (Spiers and Maguire, 2006; Hassabis and Maguire, 2009; Zeidman et al., 2015; Zeidman and Maguire, 2016; Barry et al., 2019) and represents the spatial scale of large environments (Baumann and Mattingley, 2013; Brunec et al., 2018; Peer et al., 2019) was not modulated by visuospatial context (F(2101) = 0.13, p = 0.87; Fig. 5). Together, these results emphasize the specificity of visuospatial context activity modulation to the place memory areas.

Figure 5.

Figure 5.

Analysis of control areas suggests that modulation by visuospatial context during recall is specific to the place memory areas. To determine whether visuospatial context was processed outside of the place memory areas, we considered additional regions for analysis: an early visual area (occipital pole, anatomically defined), a high-level visual area that selectively responds to faces rather than scenes (FFA; functionally defined), and an area known to be involved in mnemonic processing (hippocampus, anatomically defined). No region was modulated by visuospatial context during recall. I, Image; P, panorama; S, street.

In addition to exhibiting different activation across the conditions, the place memory and scene perception areas showed markedly different overall activity magnitude during recall. Scene perception areas OPA and PPA were not activated above baseline in any condition (OPA – all t values < 1.30, p values > 0.21; PPA – all t values < 1.11, p values > 0.28). On the other hand, their respective place memory areas were significantly active in all conditions (LPMA – all t values > 2.95, p values > 0.009, d values > 0.72; VPMA – all t values > 6.99, p values > 0.0001, d values > 1.69). This pattern can be readily observed by examining each region's activity time series (Fig. 4B). This result, which is consistent with our prior work (Steel et al., 2021), suggests that the scene-selective perceptual areas may not play an active role in memory reinstatement but instead are principally involved in processing visual input.

These univariate results suggest that the place memory areas process the extent of visuospatial context associated with a scene. But do the place memory areas actually represent the specific environment being recalled? To answer this question, we turned to multivariate analyses. We used an iterative split-half correlation analysis to determine whether we could decode the recalled environment's identity from the place memory areas. Importantly, to ensure that condition-related differences did not account for our ability to discriminate across the recalled environments, we calculated identity discrimination within each condition separately (i.e., only image vs image, panorama vs panarama, and street vs street discrimination was considered). We found that the place memory areas represented the stimulus identity during recall (LPMA: t(16) = 3.13, p = 0.0065, d = 0.89; VPMA: t(16) = 3.57, p = 0.0026, d = 0.95; MPMA: t(16) = 4.71, p < 0.001, d = 1.14; Fig. 6). Intriguingly, the scene perception areas also represented the scene being recalled (lateral – OPA: t(16) = 3.517, p = 0.009, d = 0.85; ventral – PPA: t(16) = 5.78, p < 0.001, d = 1.40; medial – MPA: t(16) = 4.558, p < 0.01, d = 1.11). On average, the scene perception areas contained significantly more identity-related information compared with our control areas V1, FFA, and hippocampus (perception areas vs V1: t(16) = 2.77, p = 0.012, d = 0.67; vs FFA: t(16) = 2.55, p = 0.02, d = 0.62; hippocampus: t(16) = 4.93, p < 0.0001, d = 1.20), and the place-memory areas showed a similar trend (memory areas vs V1: t(16) = 1.89, pone-tailed = 0.0378, d = 0.4607; vs FFA: t(16) = 1.76, pone-tailed = 0.043, d = 0.4274; hippocampus: t(16) = 4.36, p = 0.0005, d = 1.06). Because the OPA and PPA showed no univariate modulation during recall, we suggest that the stimulus-relevant information in these areas arises via feedback from the place memory areas.

Figure 6.

Figure 6.

The place memory and scene perception areas represent the identity of the stimulus being recalled. Top, Representational similarity matrix showing stimulus × stimulus correlation of the pattern of activity within each ROI. Each cell represents the similarity of activation patterns between two stimuli derived from an iterative half-split multivoxel pattern analysis. The matrix diagonals represent the average similarity of a stimulus with itself across each data split (identity). Stimuli are grouped by condition. Colored axis labels represent image, panorama, street. Bottom, Identity discrimination index for the place memory and scene perception areas. To ensure that condition-related differences could not explain identity decoding, we examined the ability to discriminate stimulus identity by comparing the average similarity between a stimulus and itself versus other stimuli within its condition. Both the scene perception and place memory areas evidenced significant identity discrimination: *p < 0.05; **p < 0.01; ***p < 0.005.

Together, the results of Experiment 1 confirmed that the place memory areas' activity: (1) reflects the extent of visuospatial context associated with the key scene view and (2) contains stimulus-specific information, as would be required if they represent the remembered local visuospatial environment.

Place memory area activity reflects visuospatial context during visual recognition

Investigating neural activity during memory recall (Experiment 1) allowed us to look directly for regions processing unseen visuospatial information. Yet focusing on an explicit recall task has limitations. First, hypothetically, acting effectively in the real-world requires representing visuospatial context online along with incoming visual information (Haskins et al., 2020; Ryan and Shen, 2020; Rust and Palmer, 2021; Draschkow et al., 2022). Thus, we would expect a region that represents visuospatial context should also be involved in perception by implicitly reactivating contextual information from memory during recognition of perceived scene views. Second, methodologically, we do not know the precise content of participants' imaginations, so magnitude differences in our explicit recall task could arise from the amount of information being recalled if participants recalled multiple scene views on each trial, or different scene views in different trials. Furthermore, although participants were instructed to recall a single FOV on each trial, internal reorientation during recall could activate the cortical head-direction network (Taube, 2007; Baumann and Mattingley, 2010; Shine et al., 2016; Nau et al., 2020) and potentially drive the apparent effect of visuospatial context observed in the place memory areas.

To address these limitations, we next turned to a visual recognition paradigm (Experiment 2). In Experiment 2, we briefly (500 ms) presented the key scene view from each learned environment and participants performed an old/new judgment (Fig. 7). We presented stimuli briefly to discourage participants from explicitly recalling unseen details and to ensure that the task primarily engaged recognition. In addition, we chose a recognition task to ensure that participants were not performing any mental reorientation that could activate the head direction network. Therefore, any difference in activation across the conditions would likely reflect implicit reactivation of visuospatial context rather than explicit, constructive mnemonic processes. We hypothesized that activity in the place memory areas would reflect the amount of known spatial context when performing a recognition task on a visually recognized scene. On the other hand, the scene perception areas would be active during the recognition task (because scenes are being perceived), but their activation magnitude would not reflect the amount of spatial context associated with the perceived scene.

Figure 7.

Figure 7.

The extent of visuospatial context associated with a scene view modulates place memory area activation during the perceptual recognition task. We compared activation of the scene perception and place memory areas across visuospatial context conditions during recognition using an ANOVA with ROI (perception/memory area), visuospatial context (image [i], panorama [p], street [s]), and hemisphere as factors. On the lateral and ventral surfaces, the place memory area activation was modulated by visuospatial context to a greater degree than the scene perception areas. On the medial surface, activity of both the scene perception and place memory area was modulated by known visuospatial context. *p < 0.05. **p < 0.01. ***p < 0.005.

Do the place memory areas process visuospatial context during recognition? Consistent with their roles in visual scene processing, all scene perception and place memory areas responded significantly during the recognition task (all t values > 2.64, p values < 0.0187, d values > 0.65). Importantly, however, we again found that visuospatial context specifically modulated activity in the place memory areas. On the lateral surface, visuospatial context impacted LPMA activity significantly more than OPA (ROI × condition interaction: F(2191) = 6.95, p = 0.003; post hoc ANOVAs – OPA: F(2,95) = 3.51, p = 0.085; LPMA: F(2,95) = 11.93, p = 0.0004; difference across conditions: LPMA vs OPA – panorama vs image: t(15) = 2.41, pcorr = 0.084, d = 0.60; panorama vs street: t(15) = 1.35, pcorr = 0.594, d = 0.34; street vs image: t(15) = 3.567, pcorr = 0.006, d = 0.89). On the ventral surface, both PPA and VPMA activation was modulated by the amount of known visuospatial context. However, VPMA was modulated to a significantly greater degree (F(2191) = 9.95, p = 0.005; post hoc ANOVAs – PPA: F(2,95) = 6.19, p = 0.01; VPMA: F(2,95) = 13.23, p = 0.0002; difference across conditions: VPMA vs PPA – panorama vs image: t(15) = 2.55, pcorr = 0.06, d = 0.63; panorama vs street: t(15) = 2.21, pcorr = 0.129, d = 0.55; street vs image: t(15) = 4.19, pcorr < 0.001, d = 1.05). On the medial surface, both MPA and MPMA activation reflected visuospatial context, and these regions did not differ in their scaling across conditions (ROI × condition – F(2191) = 1.75, p = 0.19; main effect of condition – MPA: F(2,95) = 12.94, p < 0.001; post hoc test: panorama vs image: t(15) = 2.18, pcorr = 0.135, d = 0.746; panorama vs street: t(15) = 3.15, pcorr = 0.018, d = 1.01; street vs image: t(15) = 4.50, pcorr = 0.0012, d = 1.62; MPMA: F(2,95) = 14.31, p < 0.001; post hoc test: panorama vs image: t(15) = 2.385, pcorr = 0.123, d = 0.98; panorama vs street: t(15) = 2.92, pcorr = 0.033, d = 1.20; street vs image: t(15) = 5.32, pcorr < 0.0001, d = 2.24). Importantly, control ROIs FFA and V1 were not modulated by spatial context (V1: F(2,95) = 0.29, p = 0.74; FFA: F(2,47) = 3.02, p = 0.06; Fig. 8). Although FFA approached significance, this was driven by lower activation to street images, which is opposite to the place memory areas. Likewise, hippocampal activity was not significantly modulated by visuospatial context during recognition (F(2,95) = 3.02, p = 0.06; Fig. 8). While the effect of visuospatial context during recognition approached significance, this was driven by lower activation to panoramas compared with streets and images.

Figure 8.

Figure 8.

Analysis of control areas suggests that modulation by visuospatial context during perceptual recognition is specific to the place memory areas. To determine whether visuospatial context was processed outside of the place memory areas, we considered additional regions for analysis: an early visual area (occipital pole, anatomically defined), a high-level visual area that selectively responds to faces rather than scenes (FFA; functionally defined), and an area known to be involved in mnemonic processing (hippocampus, anatomically defined). No regions outside the place memory areas were significantly modulated by visuospatial context during recognition. I, Image; P, panorama; S, street.

Finally, we wanted to determine whether the modulation by known visuospatial context was specific to the place memory areas. To answer this question, we turned to a whole-brain analysis (Fig. 9). As we found in Experiment 1, the modulation by visuospatial context was remarkably specific: within posterior cerebral cortex, significant vertices fell largely within the place memory areas and did not appear in other category-selective areas. Notably, at the whole-brain level, the lateral place memory area contained only a small number of significant vertices compared with the ventral and medial place memory areas; this is likely because of the heterogeneity in the location of the lateral place memory area across participants. Thus, the effect of visuospatial context on activation during recognition is largely restricted to the place memory areas.

Figure 9.

Figure 9.

Modulation by spatial context across the brain during recognition. We compared neural activation across the visuospatial context conditions (Image, Panorama, and Street) using a vertex-wise ANOVA. We found that three clusters showed significant modulation by known spatial context, which fell largely within the place memory areas (black). Whole-brain analysis thresholded at FDR-corrected p value (q) < 0.05 (vertex-wise F > 8.1, p < 0.001; q < 0.05). Scene perception areas (white) and place memory areas (black) were defined using data from an independent group of participants (Steel et al., 2021).

The results of Experiment 2 suggest that, during visual recognition, the lateral and ventral place memory areas represent the visuospatial context associated with a visual scene to a greater degree than their perceptual counterparts, while MPA and MPMA both appear to process visuospatial context. Together with the result of Experiment 1, these data suggest that perceptual information from the visual system and contextual information from memory converge anterior to each of the scene perception areas, in the place memory areas.

Post-scan behavioral results confirmed memory success

To ensure that systematic differences in (1) subjective vividness and (2) memory strength could not explain the difference in neural activity across conditions, we had participants perform two behavioral assessments after MRI scanning. First, participants rated their ability to vividly imagine each stimulus on a 1-5 scale, with 5 being the most vivid. Participants reported strong overall ability to imagine the learned scenes (4.31 ± 0.5 SD) and reported no difference in imagery ability across the conditions (F(2,50) = 3.03, p = 0.06). Although this analysis approached significance, it reflects a decrease in vividness in the street condition compared with the image and panorama condition, which is in the opposite direction to our magnitude analysis results. Second, a subset of participants completed a naming test. In the naming test, participants were shown each key scene view and provided the name of the associated environment. Participants' responses were highly accurate on this difficult memory assessment (0.86 ± 0.20 SD), and their accuracy did not differ across the conditions (F(2,32) = 2.48, p = 0.109). These results suggests that the observed differences in activation across conditions in the place memory areas reflect visuospatial context rather than other behavioral factors.

Discussion

A central debate in neuroscience concerns the neural mechanisms facilitating the interplay of mnemonic and sensory information in the brain. This topic is particularly relevant for real-world scene perception, which requires representing properties of the environment that are outside the current FOV to inform ongoing perception and behavior (Robertson et al., 2016; Haskins et al., 2020; Berens et al., 2021; Draschkow et al., 2022). Yet, while the neural systems representing visual scenes are relatively well understood (Epstein and Kanwisher, 1998; Maguire et al., 1998; Hasson et al., 2002; Malach et al., 2002; Epstein et al., 2007a; Dilks et al., 2013, 2022; Silson et al., 2015; Kamps et al., 2016; Malcolm et al., 2016; Bonner and Epstein, 2017; Epstein and Baker, 2019; Lescroart and Gallant, 2019), how memory for the surrounding environment interfaces with scene processing is largely unknown. On one hand, visual cortex could jointly represent mnemonic and sensory information to facilitate memory-based perception (D'Esposito et al., 1997; O'Craven and Kanwisher, 2000; Rademaker et al., 2019). On the other hand, dedicated systems might respectively harbor mnemonic and sensory information to prevent mnemonic-perceptual interference (Spreng et al., 2009; Ranganath and Ritchey, 2012; Ritchey et al., 2015; Barnett et al., 2020; Ritchey and Cooper, 2020).

Our results suggest an intriguing third possibility. Here, we demonstrate that recalling a visuospatial environment activates a functional network immediately anterior and adjacent to the three scene-selective areas of high-level visual cortex, the “place memory network” (Steel et al., 2021). This arrangement could allow efficient mnemonic-perceptual interaction while minimizing mnemonic-perceptual interference. Although we previously identified the place memory network and described its responses to familiar stimuli and its connectivity, the nature of mnemonic processing in these areas was not clear (Steel et al., 2021). The present study offers three insights into the place memory areas' mnemonic representations. First, by using VR, we show that visuospatial information alone may be sufficient to drive activity in the place memory areas, implying that the episodic and multisensory details present in personally familiar memories may not be required. Second, we observed that the amount of known visuospatial context associated with a scene modulates the anterior areas' activity during recall and recognition, suggesting that these areas process visuospatial content associated with a scene view. Third, using multivoxel pattern analysis, we show that the place memory areas represent the specific environment being recalled, which we had not demonstrated in our prior work (Steel et al., 2021). Together with prior work showing an anterior shift for mnemonic compared with visual information for scenes (Baldassano et al., 2013, 2016; Rugg and Thompson-Schill, 2013; Silson et al., 2016, 2019; Peer et al., 2019; Bainbridge et al., 2021; Steel et al., 2021; Srokova et al., 2022), these results suggest that the place memory areas likely play a key role in memory-guided scene perception by processing the remembered visuospatial context of a scene.

The conventional view of scene perception is that integration of mnemonic and perceptual information occurs in the MPA (also known as retrosplenial complex) (Malcolm et al., 2016; Epstein and Baker, 2019; Dilks et al., 2022). This view is based on reports showing that mnemonic tasks engage MPA. For example, unlike OPA and PPA, MPA responds preferentially to familiar versus unfamiliar images, and MPA activates more than OPA or PPA in memory tasks that involve linking a scene view with its environment context (Epstein et al., 2007b). Further, unlike OPA and PPA, MPA jointly represents both egocentric viewpoint of the current scene and the viewer's allocentric position within the environment (Vass and Epstein, 2013; Marchette et al., 2014; Shine et al., 2016; Nau et al., 2020). In contrast, PPA and OPA are implicated in perceptually oriented functions: scene categorization (Persichetti and Dilks, 2019; Dilks et al., 2022) and visually guided navigation (Julian et al., 2016; Bonner and Epstein, 2017; Lescroart and Gallant, 2019), respectively (Dilks et al., 2022). Thus, the scene perception areas appeared to differ in their degree of mnemonic function.

Our results significantly broaden our understanding of how mnemonic and perceptual information interface during scene processing. Although the amount of known visuospatial context largely did not impact PPA and OPA activation (consistent with the conventional view), we found that activation of the cortical territory immediately anterior to each of these regions is modulated by remembered visuospatial context. Thus, while the function of OPA and PPA may be more directly linked to perception, mnemonic information likely influences representations in these regions via their anterior mnemonic counterparts. Consistent with this view, we found that OPA and PPA represent the identity of the recalled scene, which we hypothesize arises via feedback from the place memory areas. Therefore, memory information likely plays a larger role in the overall function of OPA and PPA than the conventional view of scene processing suggests. A second, related hypothesis concerns the rigidity of visual representations in the scene perception versus place memory areas. Specifically, it could be that the scene perception areas focus only on the present image (i.e., the scope of information processed by these areas is limited to the current FOV). In contrast, place memory areas may have more malleable FOVs, such that the amount of information they represent may depend both on the extent of memory for one's surroundings and the scope of one's internal attention based on current task demands (i.e., are you remembering the visuospatial context around the immediate corner, or planning a longer journey through a city). This could relate to the slight decrease in identity decoding within the place memory areas compared with the scene perception areas during recall if the amount of visuospatial information within the place memory areas is not fixed across trials. This may also relate to boundary transformations, such as extension and contraction (Intraub and Richardson, 1989; Bainbridge and Baker, 2020; Hafri et al., 2022; Gandolfo et al., 2023). Thus, our results show that a new network of brain areas processing mnemonic information regarding visuospatial context associated with a scene is interposed between structures implicated in scene memory (e.g., the hippocampus or MPA) and the perceptually oriented scene areas (i.e., OPA and PPA), and future work should consider these alternative accounts of the memory area's activity.

Our findings in MPA largely agree with the conventional understanding of this region's function. Unlike the OPA and PPA, the degree of associated visuospatial context modulated MPA activation in both memory tasks. Prior work has implicated MPA in representing visuospatial context. For example, boundary extension, considered a proxy for representing visual information out of view, has been linked to MPA (Park et al., 2007). Likewise, panoramic content associated with a scene can be decoded from MPA (Robertson et al., 2016; Berens et al., 2021). In addition, MPA also represents world-centered (allocentric) facing direction (Vass and Epstein, 2013; Marchette et al., 2014; Julian et al., 2018), which requires knowledge of the environment's spatial layout (Marchette et al., 2014). Activity in medial parietal cortex near MPA becomes more correlated with the hippocampus after learning a new environment, suggesting that this region is important for encoding new place memories (Steel et al., 2019). Further, damage to this area causes severe topographic agnosia while leaving visual recognition abilities intact (Takahashi et al., 1997; Aguirre and D'Esposito, 1999). Notably, unlike OPA and PPA, which lie immediately posterior to a place memory area, MPA is largely contained within its corresponding place-memory area (i.e., MPMA) (Steel et al., 2021). Together, our findings support the conclusion that MPA could be considered a visually responsive subregion of a largely mnemonic cortical area and may be different in kind from PPA and OPA, which play an active role in visual analysis.

More broadly, our findings raise important questions about how perceptual and mnemonic neural signals converge. Several studies have noted anterior shifts in activation when people visualize versus perceive stimuli (Rugg and Thompson-Schill, 2013; Favila et al., 2018, 2020; Silson et al., 2019; Bainbridge et al., 2021; Popham et al., 2021; Steel et al., 2021; Srokova et al., 2022). What representational transformation underlies this shift? One account posits that this shift reflects a transformation from sensory information (i.e., the visual properties of a house) in high-level visual areas to more abstract, amodal, semantic representations (i.e., the mental concept of a house) (Chao et al., 1999; Huth et al., 2012; Margulies et al., 2016; Huntenburg et al., 2018; Popham et al., 2021). For example, one recent study compared a computational language model fit to movie watching versus listening to stories and concluded that the anterior shift reflected a transformation from visual features to representations grounded in language (Popham et al., 2021). However, listening to stories engages many processes, including top-down imagery, working memory, and narrative understanding, so pinpointing the nature of these representations using naturalistic paradigms is challenging.

In contrast, our results argue against a purely semantics-based account of the anterior shift. By focusing on one visual process (visual scene analysis) and isolating visuospatial context, we found that the anteriorly shifted activity reflects visuospatial context currently out of sight. As such, our findings are not well accounted for by language-based accounts of the anterior shift. Instead, we hypothesize that these anterior areas contain visuospatial information relevant to making memory-based predictions of the visuospatial environment in service of perception (Steel et al., 2023). Notably, because we focused only on visuospatial context, we cannot say whether the anteriorly shifted representations are exclusively visuospatial. One limitation of our approach (i.e., modulating the amount of visuospatial content) is that increasing visuospatial context necessarily entails the addition of other information as well as increased ecological naturalism. For example, visual information on the left and right side of the street could include details, such as houses or shops that provide semantic details about the environment, and this information would be more prevalent in the panorama and street conditions (vs the image condition). Relatedly, the presence of the occluder in the image condition diminishes its naturalistic and immersive quality compared with the panorama and street conditions. Therefore, while our results demonstrate that episodic and multisensory experiences are not necessary to elicit the anterior shift for mnemonic information, we cannot rule out that other information could elicit this anterior shift, such as semantic associations, multisensory context, and motor memories related to head movements made during stimulus encoding, and the feeling of immersion. Could anteriorly shifted representations even reflect amodal, semantic-based information (Popham et al., 2021)? Future studies will be needed to understand the nature of the representational transformation anterior to visually responsive cortex, and how these principles generalize outside the case of scenes.

In conclusion, by using VR to precisely manipulate the visuospatial context associated with a scene view, we found that perceptual representation of the visual scene in front of us and our memory of the surrounding environment converge at the anterior edge of visually responsive cortex in the place memory areas. This functional architecture could allow efficient interaction between the representation of the current visual scene with our memory of the wider visuospatial context.

Footnotes

This work was supported by National Science Foundation Grant 2144700 and NVIDIA to C.E.R. A.S. was supported by the Neukom Institute for Computational Science. We thank Yeo Bi Choi and Terry Sackett for assistance with data collection.

The authors declare no competing financial interests.

References

  1. Aguirre GK, D'Esposito M (1999) Topographical disorientation: a synthesis and taxonomy. Brain 122:1613–1628. 10.1093/brain/122.9.1613 [DOI] [PubMed] [Google Scholar]
  2. Albright TD (2012) On the perception of probable things: neural substrates of associative memory, imagery, and perception. Neuron 74:227–245. 10.1016/j.neuron.2012.04.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Argall BD, Saad ZS, Beauchamp MS (2006) Simplified intersubject averaging on the cortical surface using SUMA. Hum Brain Mapp 27:14–27. 10.1002/hbm.20158 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bainbridge WA, Baker CI (2020) Boundaries extend and contract in scene memory depending on image properties. Curr Biol 30:537–543. e3. 10.1016/j.cub.2019.12.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bainbridge WA, Hall EH, Baker CI (2021) Distinct representational structure and localization for visual encoding and recall during visual imagery. Cereb Cortex 31:1898–1913. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Baldassano C, Beck DM, Fei-Fei L (2013) Differential connectivity within the parahippocampal place area. Neuroimage 75:228–237. 10.1016/j.neuroimage.2013.02.073 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Baldassano C, Esteva A, Fei-Fei L, Beck DM (2016) Two distinct scene-processing networks connecting vision and memory. eNeuro 3:89ENEURO.0178-16.2016. 10.1523/ENEURO.0178-16.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Bar M, Aminoff E (2003) Cortical analysis of visual context. Neuron 38:347–358. 10.1016/s0896-6273(03)00167-3 [DOI] [PubMed] [Google Scholar]
  9. Barnett AJ, Reilly W, Dimsdale-Zucker H, Mizrak E, Reagh Z, Ranganath C (2020) Organization of cortico-hippocampal networks in the human brain. bioRxiv 142166. 10.1101/2020.06.09.142166. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Barry DN, Barnes GR, Clark IA, Maguire EA (2019) The neural dynamics of novel scene imagery. J Neurosci 39:4375–4386. 10.1523/JNEUROSCI.2497-18.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Baumann O, Mattingley JB (2010) Medial parietal cortex encodes perceived heading direction in humans. J Neurosci 30:12897–12901. 10.1523/JNEUROSCI.3077-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Baumann O, Mattingley JB (2013) Dissociable representations of environmental size and complexity in the human hippocampus. J Neurosci 33:10526–10533. 10.1523/JNEUROSCI.0350-13.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Baumann O, Mattingley JB (2016) Functional organization of the parahippocampal cortex: dissociable roles for context representations and the perception of visual scenes. J Neurosci 36:2536–2542. 10.1523/JNEUROSCI.3368-15.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Berens SC, Joensen BH, Horner AJ (2021) Tracking the emergence of location-based spatial representations in human scene-selective cortex. J Cogn Neurosci 33:445–462. 10.1162/jocn_a_01654 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Bonner MF, Epstein RA (2017) Coding of navigational affordances in the human visual system. Proc Natl Acad Sci USA 114:4793–4798. 10.1073/pnas.1618228114 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Braga RM, Buckner RL (2017) Parallel interdigitated distributed networks within the individual estimated by intrinsic functional connectivity. Neuron 95:457–471. e5. 10.1016/j.neuron.2017.06.038 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Braga RM, van Dijk KR, Polimeni JR, Eldaief MC, Buckner RL (2019) Parallel distributed networks resolved at high resolution reveal close juxtaposition of distinct regions. J Neurophysiol 121:1513–1534. 10.1152/jn.00808.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Brunec IK, Bellana B, Ozubko JD, Man V, Robin J, Liu ZX, Grady C, Rosenbaum RS, Winocur G, Barense MD, Moscovitch M (2018) Multiple scales of representation along the hippocampal anteroposterior axis in humans. Curr Biol 28:2129–2135. e6. 10.1016/j.cub.2018.05.016 [DOI] [PubMed] [Google Scholar]
  19. Chao LL, Haxby JV, Martin A (1999) Attribute-based neural substrates in temporal cortex for perceiving and knowing about objects. Nat Neurosci 2:913–919. 10.1038/13217 [DOI] [PubMed] [Google Scholar]
  20. Cox RW (1996) AFNI: software for analysis and visualization of functional magnetic resonance neuroimages. Comput Biomed Res 29:162–173. 10.1006/cbmr.1996.0014 [DOI] [PubMed] [Google Scholar]
  21. Dale AM, Fischl B, Sereno MI (1999) Cortical surface-based analysis: I. Segmentation and surface reconstruction. Neuroimage 9:179–194. 10.1006/nimg.1998.0395 [DOI] [PubMed] [Google Scholar]
  22. D'Esposito M, Detre JA, Aguirre GK, Stallcup M, Alsop DC, Tippet LJ, Farah MJ (1997) A functional MRI study of mental image generation. Neuropsychologia 35:725–730. 10.1016/s0028-3932(96)00121-2 [DOI] [PubMed] [Google Scholar]
  23. Dilks DD, Julian JB, Paunov AM, Kanwisher N (2013) The occipital place area is causally and selectively involved in scene perception. J Neurosci 33:1331–1336a. 10.1523/JNEUROSCI.4081-12.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Dilks DD, Kamps FS, Persichetti AS (2022) Three cortical scene systems and their development. Trends Cogn Sci 26:117–127. 10.1016/j.tics.2021.11.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. DiNicola LM, Braga RM, Buckner RL (2020) Parallel distributed networks dissociate episodic and social functions within the individual. J Neurophysiol 123:1144–1179. 10.1152/jn.00529.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Draschkow D, Nobre AC, van Ede F (2022) Multiple spatial frames for immersive working memory. Nat Hum Behav 6:536–544. 10.1038/s41562-021-01245-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. DuPre E, Salo T, Markello R, Kundu P, Whitaker K, Handwerker D (2019) ME-ICA/tedana: 0.0.6. https://zenodo.org/record/2558498 [Accessed February 7, 2020].
  28. DuPre E, Salo T, Ahmed Z, Bandettini P, Bottenhorn K, Caballero-Gaudes C, Dowdle L, Gonzalez-Castillo J, Heunis S, Kundu P, Laird A, Markello R, Markiewicz C, Moia S, Staden I, Teves J, Uruñuela E, Vaziri-Pashkam M, Whitaker K, Handwerker D (2021) TE-dependent analysis of multiecho fMRI with *tedana*. J Syst Softw 6:3669. 10.21105/joss.03669 [DOI] [Google Scholar]
  29. Epstein R, Kanwisher N (1998) A cortical representation the local visual environment. Nature 392:598–601. 10.1038/33402 [DOI] [PubMed] [Google Scholar]
  30. Epstein RA, Baker CI (2019) Scene perception in the human brain. Annu Rev Vis Sci 5:373–397. 10.1146/annurev-vision-091718-014809 [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Epstein RA, Higgins JS, Jablonski K, Feiler AM (2007a) Visual scene processing in familiar and unfamiliar environments. J Neurophysiol 97:3670–3683. 10.1152/jn.00003.2007 [DOI] [PubMed] [Google Scholar]
  32. Epstein RA, Parker WE, Feiler AM (2007b) Where am I now? Distinct roles for parahippocampal and retrosplenial cortices in place recognition. J Neurosci 27:6141–6149. 10.1523/JNEUROSCI.0799-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Evans JW, Kundu P, Horovitz SG, Bandettini PA (2015) Separating slow BOLD from non-BOLD baseline drifts using multiecho fMRI. Neuroimage 105:189–197. 10.1016/j.neuroimage.2014.10.051 [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Favila SE, Samide R, Sweigart SC, Kuhl BA (2018) Parietal representations of stimulus features are amplified during memory retrieval and flexibly aligned with top-down goals. J Neurosci 38:7809–7821. 10.1523/JNEUROSCI.0564-18.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Favila SE, Lee H, Kuhl BA (2020) Transforming the concept of memory reactivation. Trends Neurosci 43:939–950. 10.1016/j.tins.2020.09.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Fischl B, Salat DH, Busa E, Albert M, Dieterich M, Haselgrove C, van der Kouwe A, Killiany R, Kennedy D, Klaveness S, Montillo A, Makris N, Rosen B, Dale AM (2002) Whole brain segmentation: automated labeling of neuroanatomical structures in the human brain. Neuron 33:341–355. 10.1016/s0896-6273(02)00569-x [DOI] [PubMed] [Google Scholar]
  37. Gandolfo M, Nägele H, Peelen MV (2023) Predictive processing of scene layout depends on naturalistic depth of field. Psychol Sci 34:394–405. 10.1177/09567976221140341 [DOI] [PubMed] [Google Scholar]
  38. Glasser MF, Coalson TS, Robinson EC, Hacker CD, Harwell J, Yacoub E, Ugurbil K, Andersson J, Beckmann CF, Jenkinson M, Smith SM, van Essen DC (2016) A multi-modal parcellation of human cerebral cortex. Nature 536:171–178. 10.1038/nature18933 [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Gordon EM, Laumann TO, Gilmore AW, Newbold DJ, Greene DJ, Berg JJ, Ortega M, Hoyt-Drazen C, Gratton C, Sun H, Hampton JM, Coalson RS, Nguyen AL, McDermott KB, Shimony JS, Snyder AZ, Schlaggar BL, Petersen SE, Nelson SM, Dosenbach NU (2017) Precision functional mapping of individual human brains. Neuron 95:791–807. e7. 10.1016/j.neuron.2017.07.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Hafri A, Wadhwa S, Bonner MF (2022) Perceived distance alters memory for scene boundaries. Psychol Sci 33:2040–2058. 10.1177/09567976221093575 [DOI] [PubMed] [Google Scholar]
  41. Haskins AJ, Mentch J, Botch TL, Robertson CE (2020) Active vision in immersive, 360° real-world environments. Sci Rep 10:14304. 10.1038/s41598-020-71125-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Hassabis D, Maguire EA (2009) The construction system of the brain. Philos Trans R Soc Lond B Biol Sci 364:1263–1271. 10.1098/rstb.2008.0296 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Hasson U, Levy I, Behrmann M, Hendler T, Malach R (2002) Eccentricity bias as an organizing principle for human high-order object areas. Neuron 34:479–490. 10.1016/s0896-6273(02)00662-1 [DOI] [PubMed] [Google Scholar]
  44. Huntenburg JM, Bazin PL, Margulies DS (2018) Large-scale gradients in human cortical organization. Trends Cogn Sci 22:21–31. 10.1016/j.tics.2017.11.002 [DOI] [PubMed] [Google Scholar]
  45. Huth AG, Nishimoto S, Vu AT, Gallant JL (2012) A continuous semantic space describes the representation of thousands of object and action categories across the human brain. Neuron 76:1210–1224. 10.1016/j.neuron.2012.10.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Intraub H, Richardson M (1989) Wide-angle memories of close-up scenes. J Exp Psychol Learn Mem Cogn 15:179–187. 10.1037//0278-7393.15.2.179 [DOI] [PubMed] [Google Scholar]
  47. Jo HJ, Gotts SJ, Reynolds RC, Bandettini PA, Martin A, Cox RW, Saad ZS (2013) Effective preprocessing procedures virtually eliminate distance-dependent motion artifacts in resting state fMRI. J Appl Math 2013:10.1155/2013/935154. 10.1155/2013/935154 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Julian JB, Fedorenko E, Webster J, Kanwisher N (2012) An algorithmic method for functionally defining regions of interest in the ventral visual pathway. Neuroimage 60:2357–2364. [DOI] [PubMed] [Google Scholar]
  49. Julian JB, Ryan J, Hamilton RH, Epstein RA (2016) The occipital place area is causally involved in representing environmental boundaries during navigation. Curr Biol 26:1104–1109. 10.1016/j.cub.2016.02.066 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Julian JB, Keinath AT, Marchette SA, Epstein RA (2018) The neurocognitive basis of spatial reorientation. Curr Biol 28:R1059–R1073. 10.1016/j.cub.2018.04.057 [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Kamps FS, Julian JB, Kubilius J, Kanwisher N, Dilks DD (2016) The occipital place area represents the local elements of scenes. Neuroimage 132:417–424. 10.1016/j.neuroimage.2016.02.062 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Kanwisher N, McDermott J, Chun MM (1997) The fusiform face area: a module in human extrastriate cortex specialized for face perception. J Neurosci 17:4302–4311. 10.1523/JNEUROSCI.17-11-04302.1997 [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Kundu P, Inati SJ, Evans JW, Luh WM, Bandettini PA (2012) Differentiating BOLD and non-BOLD signals in fMRI time series using multiecho EPI. Neuroimage 60:1759–1770. 10.1016/j.neuroimage.2011.12.028 [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Kundu P, Voon V, Balchandani P, Lombardo MV, Poser BA, Bandettini PA (2017) Multiecho fMRI: a review of applications in fMRI denoising and analysis of BOLD signals. Neuroimage 154:59–80. 10.1016/j.neuroimage.2017.03.033 [DOI] [PubMed] [Google Scholar]
  55. LaChance PA, Taube JS (2022) Spatial context and the functional role of the postrhinal cortex. Neurobiol Learn Mem 189:107596. 10.1016/j.nlm.2022.107596 [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. LaChance PA, Todd TP, Taube JS (2019) A sense of space in postrhinal cortex. Science 365:eaax4192. 10.1126/science.aax4192 [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Lescroart MD, Gallant JL (2019) Human scene-selective areas represent 3D configurations of surfaces. Neuron 101:178–192. e7. 10.1016/j.neuron.2018.11.004 [DOI] [PubMed] [Google Scholar]
  58. Li X, Morgan PS, Ashburner J, Smith J, Rorden C (2016) The first step for neuroimaging data analysis: DICOM to NIfTI conversion. J Neurosci Methods 264:47–56. 10.1016/j.jneumeth.2016.03.001 [DOI] [PubMed] [Google Scholar]
  59. Maguire EA, Burgess N, Donnett JG, Frackowiak RS, Frith CD, O'Keefe J (1998) Knowing where and getting there: a human navigation network. Science 280:921–924. 10.1126/science.280.5365.921 [DOI] [PubMed] [Google Scholar]
  60. Malach R, Levy I, Hasson U (2002) The topography of high-order human object areas. Trends Cogn Sci 6:176–184. 10.1016/s1364-6613(02)01870-3 [DOI] [PubMed] [Google Scholar]
  61. Malcolm GL, Groen IIA, Baker CI (2016) Making sense of real-world scenes. Trends Cogn Sci 20:843–856. 10.1016/j.tics.2016.09.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Marchette SA, Vass LK, Ryan J, Epstein RA (2014) Anchoring the neural compass: coding of local spatial reference frames in human medial parietal lobe. Nat Neurosci 17:1598–1606. 10.1038/nn.3834 [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Margulies DS, Ghosh SS, Goulas A, Falkiewicz M, Huntenburg JM, Langs G, Bezgin G, Eickhoff SB, Castellanos FX, Petrides M, Jefferies E, Smallwood J (2016) Situating the default-mode network along a principal gradient of macroscale cortical organization. Proc Natl Acad Sci USA 113:12574–12579. 10.1073/pnas.1608282113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Nau M, Navarro Schröder T, Frey M, Doeller CF (2020) Behavior-dependent directional tuning in the human visual-navigation network. Nat Commun 11:3247. 10.1038/s41467-020-17000-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. O'Craven KM, Kanwisher N (2000) Mental imagery of faces and places activates corresponding stimulus-specific brain regions. J Cogn Neurosci 12:1013–1023. 10.1162/08989290051137549 [DOI] [PubMed] [Google Scholar]
  66. Park S, Intraub H, Yi DJ, Widders D, Chun MM (2007) Beyond the edges of a view: boundary extension in human scene-selective visual cortex. Neuron 54:335–342. 10.1016/j.neuron.2007.04.006 [DOI] [PubMed] [Google Scholar]
  67. Peer M, Salomon R, Goldberg I, Blanke O, Arzy S (2015) Brain system for mental orientation in space, time, and person. Proc Natl Acad Sci USA 112:11072–11077. 10.1073/pnas.1504242112 [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Peer M, Ron Y, Monsa R, Arzy S (2019) Processing of different spatial scales in the human brain. Elife 8:e47492. 10.7554/eLife.47492 [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Peirce JW (2007) PsychoPy: Psychophysics software in Python. J Neurosci Methods 162:8–13. 10.1016/j.jneumeth.2006.11.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Persichetti AS, Dilks DD (2019) Distinct representations of spatial and categorical relationships across human scene-selective cortex. Proc Natl Acad Sci USA 116:21312–21317. 10.1073/pnas.1903057116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Popham SF, Huth AG, Bilenko NY, Deniz F, Gao JS, Nunez-Elizalde AO, Gallant JL (2021) Visual and linguistic semantic representations are aligned at the border of human visual cortex. Nat Neurosci 24:1628–1636. 10.1038/s41593-021-00921-6 [DOI] [PubMed] [Google Scholar]
  72. Poser BA, Versluis MJ, Hoogduin JM, Norris DG (2006) BOLD contrast sensitivity enhancement and artifact reduction with multiecho EPI: parallel-acquired inhomogeneity-desensitized fMRI. Magn Reson Med 55:1227–1235. 10.1002/mrm.20900 [DOI] [PubMed] [Google Scholar]
  73. Posse S (2012) Multiecho acquisition. Neuroimage 62:665–671. 10.1016/j.neuroimage.2011.10.057 [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Rademaker RL, Chunharas C, Serences JT (2019) Coexisting representations of sensory and mnemonic information in human visual cortex. Nat Neurosci 22:1336–1344. 10.1038/s41593-019-0428-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Ranganath C, Ritchey M (2012) Two cortical systems for memory-guided behaviour. Nat Rev Neurosci 13:713–726. 10.1038/nrn3338 [DOI] [PubMed] [Google Scholar]
  76. Ritchey M, Cooper RA (2020) Deconstructing the posterior medial episodic network. Trends Cogn Sci 24:451–465. 10.1016/j.tics.2020.03.006 [DOI] [PubMed] [Google Scholar]
  77. Ritchey M, Libby LA, Ranganath C (2015) Cortico-hippocampal systems involved in memory and cognition: the PMAT framework 3. Prog Brain Res 219:45–64. [DOI] [PubMed] [Google Scholar]
  78. Robertson CE, Hermann KL, Mynick A, Kravitz DJ, Kanwisher N (2016) Neural representations integrate the current field of view with the remembered 360° panorama in scene-selective cortex. Curr Biol 26:2463–2468. 10.1016/j.cub.2016.07.002 [DOI] [PubMed] [Google Scholar]
  79. Rugg MD, Thompson-Schill SL (2013) Moving forward with fMRI data. Perspect Psychol Sci 8:84–87. 10.1177/1745691612469030 [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Rust NC, Palmer SE (2021) Remembering the past to see the future. Annu Rev Vis Sci 7:349–365. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Ryan JD, Shen K (2020) The eyes are a window into memory. Curr Opin Behav Sci 32:1–6. 10.1016/j.cobeha.2019.12.014 [DOI] [Google Scholar]
  82. Saad ZS, Reynolds RC (2012) SUMA. Neuroimage 62:768–773. 10.1016/j.neuroimage.2011.09.016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Shine JP, Valdés-Herrera JP, Hegarty M, Wolbers T (2016) The human retrosplenial cortex and thalamus code head direction in a global reference frame. J Neurosci 36:6371–6381. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Silson EH, Chan AW, Reynolds RC, Kravitz DJ, Baker CI (2015) A retinotopic basis for the division of high-level scene processing between lateral and ventral human occipitotemporal cortex. J Neurosci 35:11921–11935. 10.1523/JNEUROSCI.0137-15.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Silson EH, Steel AD, Baker CI (2016) Scene-selectivity and retinotopy in medial parietal cortex. Front Hum Neurosci 10:412. 10.3389/fnhum.2016.00412 [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Silson EH, Steel A, Kidder A, Gilmore AW, Baker CI (2019) Distinct subdivisions of human medial parietal cortex support recollection of people and places. Elife 8:e47391. 10.7554/eLife.47391 [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Spiers HJ, Maguire EA (2006) Thoughts, behaviour, and brain dynamics during navigation in the real world. Neuroimage 31:1826–1840. 10.1016/j.neuroimage.2006.01.037 [DOI] [PubMed] [Google Scholar]
  88. Spreng RN, Mar RA, Kim AS (2009) The common neural basis of autobiographical memory, prospection, navigation, theory of mind, and the default mode: a quantitative meta-analysis. J Cogn Neurosci 21:489–510. 10.1162/jocn.2008.21029 [DOI] [PubMed] [Google Scholar]
  89. Srokova S, Hill PF, Rugg MD (2022) The retrieval-related anterior shift is moderated by age and correlates with memory performance. J Neurosci 42:1765–1776. [DOI] [PMC free article] [PubMed] [Google Scholar]
  90. Steel A, Thomas C, Trefler A, Chen G, Baker CI (2019) Finding the baby in the bath water: evidence for task-specific changes in resting state functional connectivity evoked by training. Neuroimage 188:524–538. 10.1016/j.neuroimage.2018.12.038 [DOI] [PubMed] [Google Scholar]
  91. Steel A, Billings MM, Silson EH, Robertson CE (2021) A network linking scene perception and spatial memory systems in posterior cerebral cortex. Nat Commun 12:263213. 10.1038/s41467-021-22848-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Steel A, Garcia BD, Silson EH, Robertson CE (2022) Evaluating the efficacy of multiecho ICA denoising on model-based fMRI. Neuroimage 264:119723. 10.1016/j.neuroimage.2022.119723 [DOI] [PubMed] [Google Scholar]
  93. Steel A, Silson EH, Garcia BD, Robertson CE (2023) A retinotopic code structures the interaction between perception and memory systems. bioRxiv 2023.05.15.540807. [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Takahashi N, Kawamura M, Shiota J, Kasahata N, Hirayama K (1997) Pure topographic disorientation due to right retrosplenial lesion. Neurology 49:464–469. 10.1212/wnl.49.2.464 [DOI] [PubMed] [Google Scholar]
  95. Taube JS (2007) The head direction signal: origins and sensory-motor integration. Annu Rev Neurosci 30:181–207. 10.1146/annurev.neuro.29.051605.112854 [DOI] [PubMed] [Google Scholar]
  96. Vass LK, Epstein RA (2013) Abstract representations of location and facing direction in the human brain. J Neurosci 33:6133–6142. 10.1523/JNEUROSCI.3873-12.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Weiner KS, Barnett MA, Witthoft N, Golarai G, Stigliani A, Kay KN, Gomez J, Natu VS, Amunts K, Zilles K, Grill-Spector K (2018) Defining the most probable location of the parahippocampal place area using cortex-based alignment and cross-validation. Neuroimage 170:373–384. 10.1016/j.neuroimage.2017.04.040 [DOI] [PMC free article] [PubMed] [Google Scholar]
  98. Zeidman P, Maguire EA (2016) Anterior hippocampus: the anatomy of perception, imagination and episodic memory. Nat Rev Neurosci 17:173–182. 10.1038/nrn.2015.24 [DOI] [PMC free article] [PubMed] [Google Scholar]
  99. Zeidman P, Mullally SL, Maguire EA (2015) Constructing, perceiving, and maintaining scenes: hippocampal activity and connectivity. Cereb Cortex 25:3836–3855. 10.1093/cercor/bhu266 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

All data and code reported in this paper will be shared by the lead contact upon request.


Articles from The Journal of Neuroscience are provided here courtesy of Society for Neuroscience

RESOURCES