Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2021 Apr;137:1–17. doi: 10.1016/j.cortex.2020.12.018

Characterising the hippocampal response to perception, construction and complexity

Cornelia McCormick 1,1, Marshall A Dalton 1,2, Peter Zeidman 1, Eleanor A Maguire 1,
PMCID: PMC8048772  PMID: 33571913

Abstract

The precise role played by the hippocampus in supporting cognitive functions such as episodic memory and future thinking is debated, but there is general agreement that it involves constructing representations comprised of numerous elements. Visual scenes have been deployed extensively in cognitive neuroscience because they are paradigmatic multi-element stimuli. However, questions remain about the specificity and nature of the hippocampal response to scenes. Here, we devised a paradigm in which we had participants search pairs of images for either colour or layout differences, thought to be associated with perceptual or spatial constructive processes respectively. Importantly, images depicted either naturalistic scenes or phase-scrambled versions of the same scenes, and were either simple or complex. Using this paradigm during functional MRI scanning, we addressed three questions: 1. Is the hippocampus recruited specifically during scene processing? 2. If the hippocampus is more active in response to scenes, does searching for colour or layout differences influence its activation? 3. Does the complexity of the scenes affect its response? We found that, compared to phase-scrambled versions of the scenes, the hippocampus was more responsive to scene stimuli. Moreover, a clear anatomical distinction was evident, with colour detection in scenes engaging the posterior hippocampus whereas layout detection in scenes recruited the anterior hippocampus. The complexity of the scenes did not influence hippocampal activity. These findings seem to align with perspectives that propose the hippocampus is especially attuned to scenes, and its involvement occurs irrespective of the cognitive process or the complexity of the scenes.

Keywords: Hippocampus, Scene perception, Scene construction, Visual complexity

1. Introduction

The hippocampus makes a crucial contribution to episodic memory (Scoville & Milner, 1957), spatial navigation (O'Keefe & Dostrovsky, 1971), and a range of other cognitive domains including imagining the future (Addis, Wong, & Schacter, 2007; Hassabis, Kumaran, Vann, & Maguire, 2007), mind-wandering (Karapanagiotidis, Bernhardt, Jefferies, & Smallwood, 2016; McCormick, Rosenthal, Miller, & Maguire, 2018) and dreaming (Spano et al., 2020). Theoretical accounts differ on precisely how the hippocampus supports these diverse cognitive functions. Nevertheless, across perspectives there is a common thread, namely that its contribution involves constructing representations composed of numerous elements (Cohen & Eichenbaum, 1993; Lee et al., 2005; Yonelinas, Ranganath, Ekstrom, & Wiltgen, 2019; Hassabis & Maguire, 2007; Schacter & Addis, 2007). Visual scenes are paradigmatic multi-element stimuli and consequently have been deployed extensively to test hippocampal function in perceptual, associative, recognition, recall and imagination tasks (Aly, Ranganath, & Yonelinas, 2013; Barry, Chadwick, & Maguire, 2018; Lee, Brodersen, & Rudebeck, 2013; McCormick, Rosenthal, Miller, & Maguire, 2017). A scene is defined as a naturalistic three-dimensional spatially-coherent representation of the world typically populated by objects and viewed from an egocentric perspective (Dalton, Zeidman, McCormick, & Maguire, 2018).

Patients with hippocampal damage show scene-related perceptual, imagination and mnemonic impairments (Aly et al., 2013; Hassabis et al., 2007; Mullally, Intraub, & Maguire, 2012; Lee et al., 2005), and functional MRI (fMRI) studies have consistently reported hippocampal engagement during scene processing as part of a wider set of brain areas that includes ventromedial prefrontal cortex (vmPFC), parahippocampal, retrosplenial and parietal cortices (Robin, Buchsbaum, & Moscovitch, 2018; Zeidman, Mullally, & Maguire, 2015). It is unclear precisely why the hippocampus responds to scenes, and theoretical perspectives differ in the emphasis they place on specific features of scenes and their processing in order to explain hippocampal engagement, and to make inferences about its function. However, there are several gaps in our knowledge of scene processing which, if filled, may help to clarify the role of the hippocampus. Here we sought to increase our understanding by asking three questions using functional MRI (fMRI).

First, is the hippocampus especially attuned to scenes? Some accounts argue that scenes and spatial contexts merely exemplify relational processing where elements are bound together, and it is this fundamental associative processing that the hippocampus provides (Aly et al., 2013; Eichenbaum, 2006; Eichenbaum & Cohen, 2014; Erez, Cusack, Kendall, & Barense, 2016; Lee et al., 2005; Yonelinas et al., 2019). A contrasting perspective proposes that the hippocampus is specifically concerned with constructing scene representations, and more so than other types of multi-feature representations (Hassabis & Maguire, 2007; Maguire & Mullally, 2013).

It is challenging to compare these differing views in a controlled way. One approach, devised by Dalton, Zeidmann, McCormick, and Maguire (2018; see also Monk, Dalton, Barnes, & Maguire, 2020), had participants gradually build scene imagery from three successive auditorily-presented object descriptions and an imagined 3D space during fMRI. This was contrasted with constructing mental images of non-scene arrays that were composed of three objects and an imagined 2D space. The scene and array stimuli were, therefore, highly matched in terms of content and the associative and constructive processes they evoked. Moreover, the objects in each triplet were not contextually related, and for half the participants an object triplet was in a scene, and for the other half of participants it was in an array, thus controlling for semantic elements across conditions. Constructing scenes compared to arrays was associated with increased activity in the anterior medial hippocampus. Consequently, Dalton et al. (2018) concluded that it is representations that combine objects with specifically a 3D space that consistently engage the hippocampus, and that the anterior hippocampus may be especially attuned to constructing these scene representations (Dalton & Maguire, 2017; Zeidman & Maguire, 2016).

However, these simple scene representations are far removed from the naturalistic scenes that we experience in the real world. But if we want to examine naturalistic scenes, then it is still important to compare them to similar non-scene stimuli. One possibility, which we pursued here, is to create phase-scrambled versions of scenes (Yoonessi & Kingdom, 2008). The resulting images possess the same spatial frequency and colour scheme as the original scenes, but their phase is randomized such that any meaning is removed from the image. By combining the scenes and their phase-scrambled versions with the manipulation of a key feature of interest (complexity – more on this later), and matched cognitive task requirements, we predicted that, in line with Dalton et al. (2018), naturalistic scene stimuli would preferentially engage the anterior medial hippocampus.

If the hippocampus were more involved in scene processing, the second question we asked is whether the cognitive process engaged at the time would influence hippocampal recruitment. Lee et al. (2005; see also Lee et al., 2013; Barense, Henson, Lee, & Graham, 2010) used an odd-one-out paradigm where patients with bilateral hippocampal damage were shown three scenes from different viewpoints and were unable to select the one scene that was different from the two others. The authors interpreted this result as a scene perception deficit since all scenes were visible throughout each trial. However, it has been argued that this odd-one-out task also requires the construction of internal models of the scenes which are needed to mentally rotate the scenes in order to compare them to one another (Zeidman & Maguire, 2016). Consequently, findings such as these could reflect a hippocampal role in scene perception and/or the construction of scene imagery.

Zeidman, Lutti, Maguire (2015) examined this issue further by having participants view visual scenes or construct scenes in their imagination, where there was the potential to be asked subsequently to hold the perceived or imagined scene in working memory. They found that perceiving scenes was associated with extensive activation in posterior hippocampus and the anterior medial hippocampus, whereas scene construction engaged the anterior medial hippocampus. This suggests that while posterior hippocampus might be particularly engaged during scene perception, anterior medial hippocampus might play a role in constructing representations of scenes, whether perceived or imagined, when there is a need to retain them in memory. While it is difficult to separate perception and construction completely, here we sought to disentangle perception and construction processes, where visual input was identical, and there was no memory requirement.

Extending previous experimental designs (Aly et al., 2013; Lee et al., 2013), on each trial participants had to examine two images that were displayed side by side and judge whether or not the two images were the same or subtly different. Images could have either a colour or layout difference. We reasoned, based on pilot testing, that a very subtle change in global colour between two images would engage participants in comparing the perceptual qualities of images to one another, while minimising the processing of layout within the image. For the constructive task, we manipulated the spatial relationships between elements within an image (see Aly et al., 2013 for a similar approach). Here, we expected that participants would focus on mentally constructing the spatial layout of one image in order to compare it to the other image. Participants were cued before each image pair whether they should focus on the colour or layout of the images. Importantly, most of the image pairs, and those that were the focus of data analysis, were identical; therefore, participants focused on colour or layout differences in the absence of visual differences. This manipulation allowed us to counterbalance the stimuli across participants so that half of the participants searched a particular image pair for colour differences and the other half of the participants searched the same pair for layout differences. We predicted that the scene colour conditions, which would most likely engage perception and would not require scene construction, would engage the posterior hippocampus. By contrast, we expected that the scene layout conditions, which were more likely to require scene construction, would recruit the anterior hippocampus.

If the hippocampus were more involved in scene processing, the third question we asked was whether its engagement would be affected by the complexity of the scenes. We used Snodgrass and Vanderwart’s (1980) definition of visual complexity as the amount of detail or intricacy in an image (see also Donderi, 2006 for a review). Thus, for example, simple scenes have a very limited number of conjunctions in the image (e.g., a deserted beach, a sky with a single bird). By contrast, complex scenes have many conjunctions (e.g., a crowded supermarket, an amusement park). Complexity is central to several perspectives on hippocampal function with high complexity, increased level of detail, number of associations or conjunctions, held to be linked to hippocampal engagement (Barense et al., 2010; Eichenbaum & Cohen, 2014; Graham, Barense, & Lee, 2010; Lee, Yeung, & Barense, 2012; Yonelinas, 2013). Therefore, the prediction of these accounts would be that complex scenes would activate the hippocampus more than simple scenes. By contrast if, as we predicted here, the hippocampus is particularly attuned to scene processing (Hassabis & Maguire, 2007; Maguire & Mullally, 2013), then any scene, simple or complex, should recruit the hippocampus.

Overall, therefore, this study had a 2 × 2 × 2 factorial design which enabled us to examine the main effects of three factors and their interactions: 1. Image type (naturalistic scenes or phase-scrambled images); 2. Task (colour or layout); and 3. Complexity (simple or complex images). To reiterate, data analysis focused on the trials where the image pairs were identical (which were the majority of trials), with stimuli counterbalanced across participants. We employed two approaches, one data driven and the other involving pre-specified contrasts. While our main focus was on the hippocampus, we examined the whole brain in order to contextualise and further inform the research questions. In addition, we recorded eye-tracking data during scanning, performed a post-scan surprise memory task to examine potential effects of incidental encoding, collected complexity ratings for the stimuli, and asked participants about the cognitive strategies they used to perform the tasks. While we had clear hypotheses, as outlined above, our paradigm also permitted evaluation of other perspectives, given the clearly contrasting predicted outcomes.

2. Materials and methods

2.1. Participants

Aligning with previous fMRI studies involving scene processing, twenty healthy, right-handed participants (8 males, mean age 27.6 years, SD 5.5, range 21–38 years) participated in the study. None had a history of neurological or psychiatric disorders. Given the pictorial nature of the stimuli, we excluded individuals engaging (as professionals, students or hobbyists) in any intensive art or design-related activities. Colour-blind individuals were also excluded as one of the tasks involved detecting subtle colour differences. All inclusion and exclusion criteria were established before data collection commenced. All participants gave informed written consent in accordance with the University College London research ethics committee.

2.2. Stimuli and conditions

Five hundred and six pairs of images were used in this study (26 in a pre-scan practice session, 320 during scanning, and 160 as lures in a post-scan surprise memory test). Images were all in colour, and adjusted in Adobe Photoshop CS6 to an image size of 300 dpi and cropped to the same square size (450 × 450 pixels). Given our three main factors (image type, task and complexity), there were eight main conditions: 1. Simple scene colour, 2. Complex scene colour, 3. Simple scrambled colour, 4. Complex scrambled colour, 5. Simple scene layout, 6. Complex scene layout, 7. Simple scrambled layout, 8. Complex scrambled layout. In addition, we included a number of image pairs that were of medium complexity. Their function was to act as distractors for the participants so that overall the stimuli seemed to reflect a range of complexity rather than two extremes. This resulted in four more conditions: 9. Middle scene colour, 10. Middle scrambled colour, 11. Middle scene layout, and 12. Middle scrambled layout, although the fMRI data from these middle complexity conditions were not considered as they comprised fewer stimuli than the main eight conditions. Lastly, there was also a low-level baseline task that involved viewing a fixation cross and counting from one onwards until the next cue. This was designed to allow participants to disengage from the other cognitively challenging tasks. For each of the main eight conditions, there were 25 target stimuli (that is, those with no difference between the two images in a pair) and 8 catch images (those with a difference between the two images in a pair). For each of the four middle complexity conditions, there were 10 target trials and 3 catch trials. There were 25 low-level baseline trials.

2.3. Image manipulations

Although we only analysed target trials in which the two images in a pair were identical and the participants identified them correctly as such, the manipulation of the catch trials was crucial to ensure that participants would engage in the different tasks, i.e., focused either on colour or layout. We, therefore, carefully created the catch trials, whereby naturalistic and scrambled as well as simple and complex images underwent the same manipulations (see Fig. 1).

Fig. 1.

Fig. 1

Image manipulations and experimental design. The upper panel illustrates the main image manipulations that were made in order to create the catch trials for the colour and layout conditions. The middle panel shows example scene and scrambled stimuli. Simple scenes had only a few details whereas complex scenes had many details. The scrambled versions of these scenes led to simple scrambled and complex scrambled images. The lower panels illustrate example trials, first where the participant received a cue (orange background) for 1.5sec indicating that, on this trial, they should search the upcoming image pair for a colour difference. After the cue, the image pair was displayed for 5 sec, after which there was up to 2 sec in which to make a decision. Below this is an example of a layout trial. There are no differences between the images in either of these examples. The images shown here are free to use and were obtained from: www.stock.adobe.com/uk/search/free, www.unsplash.com, www.mydailymagazine.com, www.littleguidedetroit.com.

2.3.1. Colour catch trials

For the catch trials in the colour conditions, we manipulated the global colour balance of one image of a pair. For this effect, we selected one image and used Adobe Photoshop's colour balance feature on the entire image to change the balance very slightly to either red, green or blue. Hence, when displayed, one image was shown with its original colour balance next to the altered image, creating a catch trial in which the images were different. Pilot testing (including in the MRI scanner) ensured that the colour changes were detectable but sufficiently subtle to engage the participants for the trial duration. Furthermore, the analyses reported in this study were based solely on pairs in which there were no differences between the images of a pair, yet all participants reported that they kept searching the images for colour differences for the entire trial length.

2.3.2. Layout catch trials

For the catch trials in the layout conditions, we manipulated the spatial relationship between features of one image of a pair. Of note, we chose not to employ the fisheye distortion used by Aly et al. (2013) because we found that this could on occasion result in objects or lines (e.g., people, horizons), bending unnaturally. Instead, for each layout catch trial, we divided the catch image into six identical strips either vertically or horizontally. We then stretched or pinched each strip into a new dimension using Adobe Photoshop's content-aware stretching feature which preserves lines or naturally occurring objects. Thus, whereas each strip originally occupied 1/6 of the original image, in a manipulated catch trial image, the first strip could occupy 2/6 and the sixth strip only 1/12 of the resulting image. Together, this procedure allowed us to selectively manipulate the spatial configuration of the images in a natural and global manner (i.e., the detection of errors could not be achieved by single features). At display, one image was shown with its original spatial layout next to the altered image, creating a catch trial in which the spatial layout of the images was different.

2.3.3. Phase scrambling

In order to examine whether hippocampal engagement was scene-specific or not, we created phase scrambled images from the stimuli used in the scene conditions using Matlab (2014a, Mathworks), adapting a script from www.visionscience.com. This technique produced images with the same spatial frequency and colour scheme as the original scene images but because the phase was random, any meaning was removed from the images (Yoonessi & Kingdom, 2008).

2.3.4. Complexity

We selected scene images that were freely available from the internet and which had varying degrees of visual complexity, basing our selection on Snodgrass and Vanderwart’s (1980) definition of visual complexity as the amount of detail or intricacy in an image (see also Donderi, 2006 for a review). There is no agreed-upon measure of visual complexity, including where to place the dividing line between simple and complex images. Much research in this domain focuses on the number of lines, objects and conjunctions that help to define the subjective feeling of image complexity (Snodgrass & Vanderwart, 1980). Here we included simple scenes that had a very limited number of conjunctions in the image (e.g., a single bird in the sky), whereas complex scenes had many conjunctions (e.g., multiple exploding fireworks in the sky). As a general rule, our complex images had over 100 conjunctions and more than 20 objects, and our simple images had under 20 conjunctions and fewer than 4 objects. Importantly, the complexity of the images did not alter significantly when simple scenes were converted into simple scrambled images and complex scenes were converted into complex scrambled images. Before commencing data collection, pilot testing endorsed our complexity classification in that simple images were judged to be visually simpler than the complex images. The participants in the fMRI study also rated scene complexity post-scanning in a very similar manner (Fig. 2). Our categorisation of complexity was used in the fMRI analyses.

Fig. 2.

Fig. 2

Participants' stimulus complexity ratings. The means and standard errors of the complexity ratings made by the participants (1 = very simple to 6 = very complex) are shown for all conditions. Sim = simple, mid = middle, com = complex.

2.4. Tasks and procedure

Before scanning, participants had a short introduction and practice session. They were told that on each trial they would see a pair of images on the screen. They were instructed to look carefully at each pair because some of them would show two images that were identical whereas for others, the images would be slightly different, and that the main question they would have to answer after viewing each pair was whether the two images were the same or different. Participants were further told that a pair could differ in two ways, either in terms of the colour or the layout, and there would be cues to inform them about the type of difference they would encounter in the upcoming trial. Participants were then shown an example of a pair of scene images with one image being slightly different in colour. They were alerted to the fact that the colour change was not specific to a single object in the image but rather would affect the whole image. The participants were then shown an example of a pair of scene images with one image containing layout differences. They were told to focus specifically on the spatial relationships between features of the images and that, in the case of the example, the layout differed subtly between the two images. Participants were then shown examples of pairs with scrambled images and it was explained that in some cases, just as with the scenes, they would be asked to detect either colour or layout differences in scrambled image pairs. It was stressed to the participants that in all cases it was important to follow the cue and only look for colour differences if the cue had indicated that it was a colour trial, and only for layout differences if the cue had indicated that it was a layout trial. Moreover, they were told they should follow these instructions regardless of whether the trial involved a pair of scenes or scrambled images.

To help focus on the different tasks, pairs of images were also surrounded by an orange frame for colour trials or a blue frame for layout trials. Importantly, we counterbalanced target images (pairs of images that were identical) across participants, so that for any one image pair, 12 participants looked for colour differences and 8 for layout differences. Participants were further instructed to indicate with a button press after each pair whether they thought the pair had been the same (key 1) or different (key 3). Lastly, participants were informed that occasionally a fixation cross would appear on the screen and they were asked to empty their minds from any images and instead to count from one onwards until the next cue appeared on the screen (the low-level baseline condition).

Following these instructions, participants completed a practice session on a computer. There were 2 blocks with 13 trials each and involved stimuli that were not used in the experiment proper. The experiment was run using Cogent 2000 version 125 (Wellcome Centre for Human Neuroimaging, UCL, London, UK). Each trial started with the cue (either “Colour” or “Layout”) being displayed for 1.5 sec. Next, a pair of images was presented for 5 sec after which the decision question “Same (1) or (3) Different” appeared on the screen. Participants then responded in a self-paced manner (up to a maximum of 2 sec) by pressing the first button on the MRI button box if they thought the current pair was identical, and the third button if they thought the pair was different. After each trial of the practice session, the experimenter would give verbal feedback and if there were any mistakes, the experimenter would bring up the relevant image pair on the computer screen again after completion of the practice session for closer inspection.

After the practice session, participants were set up in the scanner, and the main experiment began. The experiment proper was completed in four blocks with 80 trials each. The trials were presented in pseudo-randomised order so that no more than two trials of the same condition were presented consecutively. The timings of the main experiment were identical to the practice session (Fig. 1). Completion of the practice and main experiment took approximately 120 min.

2.5. Eye-tracking during fMRI scanning

To examine whether and how patterns of eye movements changed depending on the image type, task or image complexity, we acquired eye-tracking data during the fMRI experiment. We used an MRI compatible Eyelink 1000 Plus (SR Research) eye-tracker during scanning and the Eyelink Data Viewer (SR Research) to examine fixation durations and fixation counts. We used the built-in online data parser of the Eyelink software whereby fixation duration was parsed automatically with fixations exceeding 100 msec. The right eye was used for a 9-point grid calibration, recording and analyses. During the visual exploration of the image pairs, we recorded x and y coordinates of all fixations at a sampling rate of 1000 Hz.

2.6. MR image acquisition

Structural and functional MRI data were acquired using a 3 T Magnetom Trio scanner (Siemens Healthcare, Erlangen, Germany). The structural images were collected using a T1-weighted fast low angle shot (FLASH) sequence with 1 mm isotropic resolution (Weiskopf et al., 2013). Functional T2∗-weighted images were acquired over four sessions each lasting ~15 min. The sequence was optimised to minimise signal dropout in the ventromedial prefrontal cortex (vmPFC) and medial temporal lobes using a slice tilt of −30° and a z-shim of −.4 (Weiskopf, Hutton, Josephs, & Deichmann, 2006). The volume TR was 3.36 sec, with a TE of 30 msec and echo spacing of .5 msec. Per volume, 48 slices were collected in transverse orientation, resulting in a matrix size of 64 × 74 and a 3 mm isotropic voxel size. Following the first functional session, we also acquired a fieldmap with the following parameters: short TE = 10 msec, long TE = 12.46 msec, polarity of phase-encode blips = −1, applied Jacobian modulation = no, total EPI readout time = 37 msec, in an ascending slice order.

2.7. Post-scan surprise memory test and complexity ratings

After the scan, participants underwent a surprise memory test. In addition, we asked participants to rate the complexity of each image. Visual complexity was explained to the participants as the level of detailedness or intricacy of an image, and an example scale of simple and complex scenes and scrambled images was provided. On a computer screen, one at a time, they saw the 320 images (scenes and scrambled) from the fMRI experiment plus 160 lures (40 simple scenes, 40 complex scenes, 40 simple scrambled, and 40 complex scrambled images). On each trial, participants responded in a self-paced manner but with a maximum of 5 sec response time to each of three questions: 1. Recognition memory: “(1) Old or (3) New”; 2. Confidence: “1 = very sure, 2 = somewhat sure, and 3 = not at all sure”; and 3. Complexity: “(1) very simple, (2) simple, (3) middle simple, (4) middle complex, (5) complex, (6) very complex”.

2.8. Strategies

In a debriefing session, we asked participants to describe what strategies they had used during the experiment to search for colour or layout differences in simple and complex, scene and scrambled images. We also asked whether participants had seen any of the images before, but none had.

2.9. Data analysis

2.9.1. Behavioural

Behavioural data collected during the fMRI scan and during the post-scan memory test were assessed using separate 2 × 2 × 2 repeated measures analysis of variance (3way-RM-ANOVA) with factor 1 being the image type with two levels (scenes, scrambled images), factor 2 being the task (colour, layout), and factor 3 being image complexity with two levels (simple, complex). Where 3way-RM-ANOVAs yielded significant effects (at p < .05), we report the main and interaction effects. We examined significant effects further using Sidak's multiple comparisons test and report significant results if p < .05.

2.9.2. MRI pre-processing

All MRI pre-processing was performed using SPM12 (Statistical Parametric Mapping 12; Wellcome Centre for Human Neuroimaging, London, UK). The anatomical images were segmented into grey matter, white matter and CSF maps and normalized to the Montreal Neurological Institute (MNI) template. The first five functional images were discarded to allow for signal equilibrium. Functional data were then realigned and unwarped (including distortion correction with fieldmaps) and coregistered to the anatomical image. Forward deformation fields from the anatomical image were then used to normalise the functional images into MNI space. Finally, functional images were smoothed with an 8 × 8 × 8mm kernel FWHM.

2.9.3. Partial least squares (PLS) analyses

We used PLS to analyse the fMRI data. This is a multivariate, correlational technique that allowed us to examine associations between brain activity and the experimental conditions in two ways – in a contrast free, data driven manner, and also using pre-specified contrasts (Krishnan, Williams, McIntosh, & Abdi, 2011; McIntosh, Chau, & Protzner, 2004; McIntosh & Lobaugh, 2004). Detailed descriptions of PLS can be found elsewhere (Krishnan et al., 2011). In brief, PLS uses singular value decomposition to extract ranked latent variables (LVs) from the covariance matrix of brain activity and conditions in a data driven manner. These LVs express patterns of brain activity associated with each condition. Statistical significance of the LVs was assessed using permutation testing. In this procedure, each participant's data was randomly reassigned (without replacement) to different experimental conditions, and a null distribution was derived from 500 permutated solutions. We considered LV as significant if p < .05. Furthermore, we assessed the reliability of each voxel that contributed to a specific LV's activity pattern using a bootstrapped estimation of the standard error (bootstrap ratio, BSR). For each bootstrapped solution (100 in total), participants were sampled randomly with replacement and a new analysis was performed. In the current study, we considered clusters of 50 or more voxels with BSRs greater than 2 (approximately equal to a p < .05) to represent reliable patterns of activation. Of note, PLS uses two re-sampling techniques that (1) scramble the data of each participant's conditions so that small but reliable differences between true experimental conditions can be detected, and (2) exclude whole datasets of participants, so that outliers who may drive significant effects can be detected. Therefore, even with the current sample size (n = 20), we have confidence in the robustness of the results.

In a first pass, we used a mean-centred version of PLS for block fMRI data which maximises the correlation between brain data and experimental conditions in a data driven way. Importantly, this approach allowed us to examine the fMRI data without specifying a priori contrasts.

In a second pass, we used the non-rotated version of PLS for block fMRI data to specify contrasts that would test our hypotheses. While there is the potential to examine many different contrasts with this data set, we restricted our multiple comparisons to three contrast-driven PLS analyses that corresponded to our three research questions: 1. Is the hippocampus specifically engaged in scene processing? To test this, we contrasted brain activity correlated with all scene trials (regardless of task and complexity) versus all scrambled images (regardless of task and complexity). 2. If the hippocampus is specifically engaged in scene processing, does the task matter? Here, we contrasted simple and complex scene colour versus simple and complex scene layout. 3. If the hippocampus is specifically engaged in scene processing, does the complexity of the scenes matter? Here, we contrasted all simple scenes versus complex scenes (regardless of the task). To account for the multiple PLS analyses, we corrected the p-value of the LV's using Bonferroni's multiple comparison correction for four LV's, resulting in a statistical threshold of p < .017.

2.9.4. Signal intensity extraction

In order to assist the reader with appreciating the specific contributions of a given brain region across all analyses and conditions, we extracted signal intensities from a number of brain regions that are typically associated with scene processing, including anterior (MNI -32 -2 -22) and posterior (-28 -36 4) hippocampus, vmPFC (2 50 -22), and occipito-temporal cortex (-6 -92 -4). Additional signal intensities (superior parietal lobule, parahippocampal gyrus and fusiform gyrus) can be found in the Supplementary Material (Fig. S1). The coordinates were chosen based on the highest boot strap ratios within these regions from the contrast driven PLS #1 (i.e., scenes versus scrambled images). Signal extraction for each condition for each participant was performed within the PLS toolbox using a sphere around the MNI coordinates of 3 adjacent voxels. Signal intensities in PLS can be positive and negative dependent on the averaged signal intensity of all the fMRI data, and do not represent percent signal change associated with experimental conditions. Therefore, these values do not reflect fMRI activation or deactivation compared to a baseline.

3. Results

3.1. In-scanner behavioural measures

In general, participants performed the in-scanner task (same or different) with high accuracy (mean of the corrected hit rate over all conditions = 90.6, SD = 8.8, see Table 1 for more details). The 3way-RM-ANOVA yielded a significant main effect of complexity [F(1,152) = 16.5, p = .001], as well as interaction effect of image type and complexity [F(1,152) = 9.1, p = .007], and an interaction effect of task and complexity [F(1,152) = 6.5, p = .02]. However, the three way interaction between image type, task, and complexity was not significant [F(1,152) = .2, p = .64]. Post hoc statistics revealed that these effects were driven by a lower accuracy for simple scenes than complex scenes during the colour detection task [t(152) = 3.5, p = .005]. Importantly, there was no main effect of task [F(1,152) = 1.4, p = .71], nor was there an image type by task interaction [F(1,152) = 3.6, p = .07], indicating that there were no systematic differences in behavioural performance that would impact the interpretation of the fMRI data regarding our research questions. We also examined reaction times and found that there were no significant differences across conditions [all F's(1,152)<2.6, p's>.11].

Table 1.

Summary of in-scanner accuracy and eye-movement measures.

Image category Accuracy RT Fixation duration Fixation counts
Simple scene colour 83.4 12.4 .7 .1 298.5 60.7 27.7 4.0
Complex scene colour 95.5 5.7 .7 .1 280.9 51.3 29.2 4.3
Simple scrambled colour 89.2 10.0 .7 .2 311.7 75.4 26.4 4.6
Complex scrambled colour 95.6 5.0 .7 .1 294.1 72.8 26.2 4.0
Simple scene layout 87.6 12.3 .8 .1 287.1 61.0 28.8 3.9
Complex scene layout 93.2 7.2 .8 .2 279.6 72.6 30.8 4.5
Simple scrambled layout 89.4 10.5 .7 .1 289.7 57.9 28.0 4.2
Complex scrambled layout 91.1 7.0 .8 .1 278.0 54.6 28.3 3.9

Means are shown for all eight conditions of interest with standard deviations adjacently displayed in italics. In-scanner accuracy of task performance is expressed as a percentage of corrected hit rate, and reaction times (RT) in seconds. Also shown are the means and standard deviations for the in-scanner eye-tracking - fixation duration (in msec) and fixation counts.

3.2. In-scanner eye-tracking

Next, we examined eye-movements while participants were searching for colour or layout differences in simple and complex scene and scrambled images. Eye-tracking was not possible for two participants due to technical difficulties, so the following analyses are based upon data from 18 participants. We focussed on two eye-tracking measures, fixation duration and fixation counts (see Table 1 for more details).

3.2.1. Fixation duration

The 3way-RM-ANOVA yielded a significant main effect of task [F(1,136) = 7.1, p = .02] and of complexity [F(1,136) = 23.0, p = .001]. Post hoc analyses revealed that in general participants spent longer fixating during the colour than during the layout conditions [simple scrambled colour versus simple scrambled layout: t(17) = 4.1, p < .02; complex scene colour versus complex scene layout: t(17) = 4.5, p = .01], and on simple compared to complex images [simple scene colour versus complex scene colour: t(17) = 4.0, p < .02; simple scrambled colour versus complex scrambled colour: t(17) = 4.1, p < .02].

3.2.2. Fixation counts

We observed a different pattern for fixation counts. Here, the 3way-RM-ANOVA yielded significant main effects for all three factors, image type [F(1,136) = 53.5, p = .001], task [F(1,136) = 27.7, p = .001], and complexity [F(1,136) = 29.6, p = .001]. In addition, we found an interaction effect of image type and complexity [F(1,136) = 16.9, p = .001]. Post hoc analyses revealed that searching images for layout differences resulted in more fixation counts than searching images for colour differences [simple scrambled colour versus simple scrambled layout: t(17) = 4.2, p = .01; complex scrambled colour versus complex scrambled layout: t(17) = 4.3, p = .002]. This effect was more pronounced for scenes than scrambled images [simple scrambled colour versus simple scene colour: t(17) = 4.7, p = .003; complex scrambled colour versus complex scene colour: t(17) = 9.6, p < .0001; complex scrambled layout versus complex scene layout: t(17) = 6.9, p = .0001], and more pronounced for complex compared to simple scenes [simple scene colour versus complex scene colour: t(17) = 4.0, p = .01; simple scene layout versus complex scene layout: t(17) = 4.3, p = .007]. Overall, these results indicate that participants had more fixation counts during the complex scene layout condition.

3.3. Post-scan surprise memory test and complexity ratings

3.3.1. Memory accuracy

Due to the large number of different images and the short encoding time, recognition memory was, unsurprisingly, poor and barely exceeded chance level (mean over all conditions = 60.7, SD = 17.9, see Table 2 for further details). The 3way-RM-ANOVA across all conditions yielded a significant main effect of complexity [F(1, 152) = 19.0, p = .001] and an interaction effect of image type and complexity [F(1, 152) = 21.4, p = .001]. Simple scrambled images for both colour[(t(17) = 4.7, p = .004] and layout [t(17) = 5.2, p = .001] tasks were less well remembered than the complex scrambled images for both colour and layout conditions. This result did not come as a surprise since simple scrambled images were particularly featureless. Importantly, there was no main effect of task [F(1,152) = .4, p = .52], indicating that there were no systematic differences in encoding success between the colour and layout conditions that could have impacted the interpretation of the fMRI data in relation to our research questions.

Table 2.

Summary of post-scan behavioural measures.

Image category Recognition memory Confidence Complexity
Simple scene colour 55.1 17.3 1.7 .3 2.2 .4
Complex scene colour 60.3 21.8 1.6 .3 4.1 .5
Simple scrambled colour 49.9 14.5 2.2 .6 2.6 .5
Complex scrambled colour 72.3 17.2 2.1 .7 4.2 .6
Simple scene layout 68.2 20.1 1.7 .3 2.0 .4
Complex scene layout 65.5 23.8 1.6 .3 4.2 .5
Simple scrambled layout 45.9 16.0 2.1 .6 2.5 .4
Complex scrambled layout 68.6 12.9 2.1 .7 4.2 .9

Means are shown for all eight conditions of interest with standard deviations adjacently displayed in italics. Incidental encoding was evaluated using percentage of corrected hit rates of recognition memory, as well as confidence ratings (1 = very sure to 3 = not sure at all). The means and standard deviations of participants' complexity ratings (1 = very simple to 6 = very complex) are also shown (see also Fig. 2).

3.3.2. Memory confidence ratings

In general, participants were not confident about whether or not they had seen a particular image in the scanner, showing that they had insight into their poor memory performance on the surprise memory test (mean over all conditions = 2.0, SD = .5, see Table 2 for further details). The 3way-RM-ANOVA across all conditions yielded a main effect of image type [F(1,152) = 17.7, p < .001], with participants, unsurprisingly, less confident about their memory judgements for scrambled images compared to scenes [simple scrambled colour versus simple scene colour: t(17) = 3.2, p = .02; complex scrambled colour versus complex scene colour: t(17) = 3.4, p = .01; complex scrambled layout versus complex scene layout: t(17) = 3.6, p = .005].

3.3.3. Complexity ratings

To examine whether participants' ratings of complexity (shown in Fig. 2) accorded with our designations, we calculated the mean complexity rating for each of our designated conditions but now based on the participants' ratings, and entered these into a 3way-RM-ANOVA. This yielded a significant main effect of complexity [F(1,152) = 428.8, p < .0001]. Post hoc analyses revealed significant differences between all simple and complex stimuli [simple scrambled colour versus complex scrambled colour: t(17) = 10.6, p < .0001; simple scene colour versus complex scene colour: t(17) = 9.1, p < .0001; simple scrambled layout versus complex scrambled layout: t(17) = 12.1, p < .0001; simple scene layout versus complex scene layout: t(17) = 9.7, p < .0001]. Therefore, the participants’ ratings accorded well with our classification of complexity which was mirrored across image types and task.

3.4. Strategies

Also after scanning, we asked participants about how they had decided whether two images were the same or different. Generally, participants reported different strategies for colour and layout conditions, but did not report different strategies based on the image type or complexity.

For the colour task, participants indicated that they mostly focused on selected parts of the images without paying much attention to the content of the image. For example, they would compare corners, brightly lit or especially dark areas between the images. Participants indicated that they followed this strategy whether the images were simple or complex, or scenes or scrambled images. In contrast, for the layout task, participants reported examining and mentally constructing the spatial relationships within one image and then comparing these relationships to the second image. Again, participants described the same constructive strategy for simple and complex images, and for scenes and scrambled images. The strategies of each participant are summarised in the Supplementary Material (Table S1).

3.5. Data driven mean-centred PLS

The fMRI data driven PLS included all eight conditions of interest and revealed three significant LV's (Fig. 3). As noted previously, this approach allowed us to examine the fMRI data without specifying a priori contrasts, enabling us to explore the dominant patterns of activity in the brain elicited by the tasks. Below we unpack each LV in turn.

Fig. 3.

Fig. 3

Summary of the significant latent variables (LVs) detected by the data driven PLS. Bar graphs depict means and confidence intervals for all conditions. Sim = simple, com = complex. Activations are displayed on a T1-weighted MRI scan (MNI template); L = left, R = right, BSR = boot strap ratio. (A) LV1 explained 50% of the variance. This pattern contrasted simple and complex scene layout against all other conditions, a pattern which likely reflects scene construction processes. (B) LV2 explained 15% of the variance. This pattern contrasted most simple images against complex images, regardless whether they were scenes or scrambled images, or whether participants searched images for colour or layout differences. (C) LV3 also explained 15% of the variance. This pattern highlighted simple and complex scene colour, likely reflecting scene perception processes. See also Tables S2-S4 for full details.

3.5.1. LV1 – scene construction

The first significant LV (LV1, p < .0001, explaining 50% of the variance, Fig. 3A, see Table S2 for all the brain regions that were engaged and their MNI coordinates), identified a contrast between the two conditions that we argued might be most dependent upon scene construction (i.e., simple scene layout and complex scene layout) and the other conditions. The correlated brain pattern yielded widespread activation of brain regions which are typically engaged during the construction of scene imagery, such as bilateral anterior medial hippocampus, parahippocampal gyrus, fusiform gyrus, vmPFC, bilateral precuneus and inferior parietal lobules as well as occipital cortices. Together, this finding suggests that constructing internal models of scene layouts is a dominant cognitive process associated with a distributed brain activation pattern.

3.5.2. LV2 – main effect of complexity

The second significant LV (LV2, p < .0001, explaining 15% of the variance, Fig. 3B, see Table S3 for all the brain regions that were engaged and their MNI coordinates) reflected the main effect of complexity. Interestingly, we found distinct patterns associated with simple and complex images. While simple images seemed to engage more medial posterior brain regions (e.g., medial occipital cortices), complex images engaged more lateral posterior brain regions (e.g., lateral occipital, temporal and parietal cortices). An exception to this rule was the medial subgenual vmPFC which was more activate for complex than simple images. Of note, hippocampal activity was not modulated by image complexity. Overall, this LV suggests that complexity as a general image feature engages mostly posterior visual brain regions. Nevertheless, one has to interpret this result with caution since this main effect reflects a combination of multiple, very different conditions. In our follow-up contrast driven PLS analyses, we specified more tailored contrasts to examine the effect of stimulus-specific complexity.

3.5.3. LV3 – scene perception

The third significant LV (LV3, p < .02, explaining 15% of the variance, Fig. 3C, see Table S4 for the relevant brain regions and their MNI coordinates) identified a contrast between conditions that we argued mostly depend upon scene perception (i.e., simple scene colour and complex scene colour) and the search for layout differences in scrambled images (i.e., simple scrambled layout and complex scrambled layout). While searching scrambled images for layout differences involved mainly lateral temporal, parietal and dorsomedial prefrontal cortices, the associated brain pattern for examining scenes for colour differences involved the posterior hippocampi, as well as several brain regions along the ventral visual pathway, such as the fusiform gyrus, parahippocampal gyrus and the inferior temporal gyrus. Of note, a number of regions previously found to be associated with scene construction, such as the anterior medial hippocampi and the vmPFC, were absent for this brain activity pattern. Again, this LV resulted from a data driven method, hence the interpretation of contrasting conditions involving the search for colour differences in scenes against scrambled layout is not immediately clear. Therefore, in a second pass, we conducted non-rotated PLS analyses where conditions that involved searching scenes for colour or layout differences were directly contrasted.

3.6. Contrast driven non-rotated PLS

We had three specific research questions which we focussed on in the analyses described below: 1. Is the hippocampus specifically engaged in scene processing in this experiment?; 2. If hippocampal engagement is specific to scene processing, does the task (i.e., colour or layout) matter?; and 3. If the hippocampus is responsive during scene processing, does the complexity of the scenes matter?

3.6.1. Contrast driven PLS #1: Is the hippocampus specifically engaged in scene processing?

To answer this question, we contrasted all four conditions involving scene processing (simple scene colour, complex scene colour, simple scene layout, and complex scene layout) against the other four conditions of scrambled images (simple scrambled colour, complex scrambled colour, simple scrambled layout, and complex scrambled layout). The PLS revealed a significant LV (p < .0001, Fig. 4A, see Table S5 for the relevant brain regions and their MNI coordinates) separating all scene conditions from all scrambled conditions. Confirming our hypotheses, the brain pattern associated with scene processing included bilateral hippocampi (both anterior and posterior segments), as well as the usual ventral visual brain regions that are typically associated with scene processing. In addition, the subgenual vmPFC also showed greater activation during scene processing than for the scrambled conditions. In contrast, processing of scrambled images was associated with a much more restricted pattern of brain activity that included engagement of the precuneus and the anterior cingulate cortex.

Fig. 4.

Fig. 4

Brain activity patterns associated with the contrast driven PLS. Activations are displayed on a T1-weighted MRI scan (MNI template); L = left, R = right, BSR = boot strap ratio. (A) The brain activity pattern associated with simple and complex scene colour and layout conditions, shown in the upper panel, reflected the well-established set of brain regions associated with scene processing including increased activity in the vmPFC, bilateral hippocampus and along the ventral visual stream. This is in contrast to the brain pattern associated with simple and complex scrambled colour and layout conditions, shown in the lower panel, which included lateral and medial parietal cortices and anterior cingulate cortex. (B) The brain activity pattern associated with simple and complex scene layout (most likely depending on scene construction), shown in the upper panel, included increased activity along the ventral visual stream, bilateral anterior medial hippocampus and vmPFC. The brain pattern associated with simple and complex scene colour (most likely depending on scene perception), shown in the lower panel, included medial and lateral parietal cortices and anterior cingulate cortex. In addition, we observed increased bilateral posterior hippocampal activity for the scene colour conditions. See also Tables S5-S6 for full details.

3.6.2. Contrast driven PLS #2: If the hippocampus is specifically engaged in scene processing, does the task matter?

Given that in the first contrast we identified greater hippocampal involvement in scene processing compared to scrambled images, we next asked whether the brain activity patterns, including hippocampal engagement, was specific to conditions that most likely involve scene construction. We, therefore, contrasted conditions that involved searching scenes for layout differences (simple scene layout and complex scene layout) against those involving the search for colour differences in scenes (simple scene colour and complex scene colour), which most likely involved scene perception. The resulting LV was highly significant (p < .0001, all conditions contributing, Fig. 4B, see Table S6 for all relevant brain regions and their MNI coordinates). The hippocampus was among the areas engaged during both scene construction and scene perception. Interestingly, there was a clear dissociation between its anterior and posterior segments in terms of their responsiveness to the different types of scene processing. Whereas searching for colour differences engaged the posterior hippocampus, searching scenes for layout differences engaged bilateral anterior hippocampus. Furthermore, while the scene colour conditions engaged the precuneus and angular gyrus, as well as anterior cingulate cortex, the scene layout conditions were associated with several brain regions along the ventral visual pathway and superior parietal lobule. In addition, the subgenual vmPFC was also more activated during the scene layout compared to scene colour conditions.

3.6.3. Contrast driven PLS #3: If the hippocampus is specifically engaged in scene processing, does the complexity of the scenes matter?

Here, we contrasted simple (simple scene colour and simple scene layout) versus complex scenes (complex scene colour and complex scene layout). However, this PLS analysis did not reveal a significant LV (p = .06 with the Bonferroni cut-off being p < .017).

3.6.4. Regional signal intensities

We extracted signal intensities from several of the brain which had the highest bootstrap ratios in the contrast driven PLS #1 analyses (Fig. 5). The resulting graphs merely illustrate what has already been detected by the PLS pattern and do not offer any new information per se, however, they do provide a convenient overview of the activity patterns across all eight experimental conditions. Activity in vmPFC (x y z: 2 50 -22) had a multifaceted pattern, reflecting a preference for scenes, especially scene layout conditions, whilst also keeping track of the visual complexity of all stimuli. The anterior (-32 -2 -22) and posterior (-28 -36 4) hippocampus had distinct patterns of activity that were more clear-cut, with the former engaged by conditions tapping into scene construction, and the latter by conditions tapping into scene perception. Activity within the occipito-temporal cortex (-6 -92 -4) reflected predominantly visual complexity. Additional signal intensities (superior parietal lobule, parahippocampal gyrus and fusiform gyrus) can be found in the Supplementary Material (Fig. S1).

Fig. 5.

Fig. 5

Extracted signal intensities from brain regions associated with scene processing. Bar graphs depict means and standard errors of the eight conditions for the vmPFC, anterior and posterior hippocampus and occipito-temporal cortex. Sim = simple, com = complex. These regions were chosen based on the highest bootstrap ratio in the associated contrast driven PLS #1 analysis. Of note, signal intensities are compared to an arbitrary fMRI baseline, hence negative values do not necessarily represent deactivations. Activity in vmPFC had a multifaceted pattern, reflecting a preference for scenes, especially scene layout conditions, whilst also keeping track of the visual complexity of all stimuli. The anterior and posterior hippocampus had distinct patterns of activity that were more clear-cut, with the former engaged by conditions associated with scene construction, and the latter by conditions associated with scene perception. Activity within the visual cortex reflected predominantly visual complexity. Additional signal intensities (superior parietal lobule, parahippocampal gyrus and fusiform gyrus) are provided in the Supplementary Material (Fig. S1).

4. Discussion

In this study we sought to deepen our understanding of hippocampal processing by addressing three questions: 1. Is the hippocampus recruited specifically during scene processing? 2. If the hippocampus is active in response to scenes, does the task, and the cognitive process it likely engages, influence its activation? 3. If the hippocampus is upregulated during scene processing, does the complexity of the scenes affect its response? We found that, compared to phase-scrambled versions of the scenes, the hippocampus was more responsive to scene stimuli. Moreover, there was a clear distinction in terms of which parts of the hippocampus were engaged, with conditions that likely relate to scene perception associated with the posterior hippocampus and conditions that tend to depend on scene construction involving the anterior hippocampus. The complexity of the scenes did not influence hippocampal activity. We discuss each of these results in turn.

4.1. The hippocampus is upregulated during scene processing

The hippocampi (anterior and posterior segments) were more activated for scenes than scrambled images. This echoes previous work using simplified representations of scenes and non-scene arrays which also showed preferential engagement of the hippocampus for scenes (Dalton et al., 2018; Monk, Dalton, Banes, & Maguire, 2020), in that case also controlling for semantic elements. Our findings align in particular with accounts of hippocampal function that propose the hippocampus is especially attuned to scene processing (Hassabis & Maguire, 2007; Maguire & Mullally, 2013). We acknowledge that it is challenging to devise non-scene stimuli for comparison with naturalistic scenes. Here, we contrasted the scenes with images that were phase-scrambled versions of the same scenes, thus preserving the spatial frequency and colours. The scenes and scrambled stimuli were rated comparably by participants in terms of the various levels of complexity. There was no main effect of image type for the same/different judgements during scanning, or in eye movement fixation duration. The pattern of memory performance in the surprise post-scan test was similar, in particular for complex scenes and complex scrambled stimuli. Moreover, the strategies participants reported using during the tasks did not differ as a function of stimulus type. Nevertheless, despite all of these similarities, the hippocampus was engaged preferentially for scenes.

It is perhaps not surprising that scenes are particularly stimulating for the hippocampus, as they mirror how people experience and perceive the world. In addition, scenes are a highly efficient means of packaging information. Clark et al. (2019) recently reported that the ability to construct scene imagery explained the relationships between episodic memory, imagining the future and spatial navigation task performance. The prominence of scenes was further emphasised in another study involving the same sample, where the explicit strategies people used to perform episodic memory recall, future thinking and spatial navigation tasks was assessed (Clark, Monk, & Maguire, 2020). In each case, the use of scene visual imagery strategies was significantly higher than for all other types of strategies (see also Andrews-Hanna, Reidler, Sepulcre, Poulin, & Buckner, 2010). The apparent utility and prevalence of scene processing has led to the suggestion that scene imagery may be the currency of cognition (Maguire & Mullally, 2013).

4.2. A hippocampal distinction between scene perception and scene construction

Given that scenes activated the hippocampus more than scrambled versions of the same scenes, we next considered whether the task, and by extension the cognitive process that was likely being engaged, influenced hippocampal recruitment. As outlined previously, the conditions that involved searching images for layout differences were held to tap into constructive cognitive processes, while examining images for colour differences involved perceptual processes. The scene construction and perception tasks were well-matched in a number of respects. Importantly, we counterbalanced images across participants so that all analyses dissociating perception and construction processes were conducted on the same images. The accuracy during scanning where participants correctly identified pairs of images as being identical did not differ between colour and layout trials. In addition, recognition memory assessed during a surprise memory task after scanning was similar for both tasks, indicating that there was no disparity in terms of incidental encoding. Despite these similarities, there were differences in the hippocampal response to perception and construction.

A striking result of the data driven mean centred PLS (LV1) was a clear anterior medial hippocampus preference for the two conditions thought to weigh most heavily on scene construction, namely the tasks involving the processing of simple and complex scene layouts. This finding accords with the numerous previous reports associating scene construction with the anterior hippocampus (reviewed in Zeidman & Maguire, 2016; see also Zeidman, Mullally, & Maguire, 2015,b; Dalton et al., 2018; Dalton & Maguire, 2017). By contrast, another data driven LV, this time involving the posterior hippocampus, was associated with conditions relating to the perception of scenes, simple and complex, with no evidence of the anterior hippocampus in this brain pattern. A direct contrast between the scene layout and scene colour conditions confirmed that the two processes involved different hippocampal segments, anterior and posterior respectively. These findings are in line with research stressing functional dissociations along the anterior–posterior axis of the hippocampus during scene construction and scene perception (e.g., Zeidman, Mullally, & Maguire, 2015; Zeidman & Maguire, 2016), and more generally (Poppenk, Evensmoen, Moscovitch, & Nadel, 2013; Sekeres, Winocur, & Moscovitch, 2018; Strange, Witter, Lein, & Moser, 2014).

The anterior medial hippocampus is well suited to scene-based cognition given its anatomical (Dalton & Maguire, 2017) and functional (Dalton, McCormick, & Maguire, 2019,b) connectivity with regions in the parieto-medial temporal pathway. While the resolution of the current fMRI data do not permit hippocampal subfield analyses, it is important to acknowledge that subfields might further differentiate between perceptual and constructive processes, as well as scenes and other types of stimuli. For example, recent high-resolution fMRI studies indicate that the anterior medial portion of the subiculum or pre-/parasubiculum may be specifically engaged during scene construction (Dalton et al., 2018; Hodgetts et al., 2017; Zeidman, Lutti, & Maguire, 2015). In contrast, several previous fMRI studies showed that posterior hippocampal activity was associated with tasks involving perceptual discrimination between visually similar scenes (Aly et al., 2013; Barense et al., 2010; Lee, Buckley, et al., 2005; Lee, Bussey, et al., 2005). The posterior hippocampus has strong anatomical and functional connections with the ventral visual stream and early visual cortices (Chadwick, Mullally, & Maguire, 2013; Kahn, Andrews-Hanna, Vincent, Snyder, & Buckner, 2008). Therefore, it may be that the posterior hippocampus is involved in guiding ongoing scene perception while the anterior hippocampus supports online construction into a coherent mental model of the world.

Interestingly, eye-tracking data recorded during scanning revealed a difference between the colour and layout tasks. Whereas the colour conditions were associated with longer fixation durations, the layout conditions had more fixation counts. This effect was most pronounced during processing of scenes, especially complex scenes. These results generally align with extant studies linking rapid visual sampling to the construction of mental events (El Haj & Lenoble, 2018; Hannula & Ranganath, 2009; Liu, Shen, Olsen, & Ryan, 2017). Of note, since there were no recognition memory differences between colour and layout trials, our eye-tracking results speak against a proposal that exploratory visual sampling is purely memory-guided (Voss, Bridge, Cohen, & Walker, 2017). Rather, our findings seem to indicate that the pattern of eye-movements relates to an interaction between the dominant cognitive process and the image type (naturalistic versus scrambled, complex versus simple) during a particular task.

We also asked participants directly about their cognitive strategies during the colour and layout tasks. All participants reported distinct cognitive approaches to searching image pairs for colour versus layout differences. They indicated that they mostly focused on selected parts of the images without paying much attention to the content of the image during colour trials. For example, they would compare corners, brightly lit or especially dark areas between the images. In contrast, for the layout task, participants reported examining and mentally constructing the spatial relationships within one image and then comparing these relationships to the second image, also without paying much attention to the content of the scenes. As already noted, the images in the scene (and scrambled) conditions were counterbalanced across participants, such that half the participants searched a specific scene for colour differences and the other half of the participants searched the same scene for layout differences. Hence, the same scene content (and semantic meaning) was present in both conditions, and so is unlikely to explain our results. Furthermore, we have recently shown that while patients with hippocampal damage have difficulty detecting spatial-constructive impossibilities in scenes (e.g., an endless stair case), they did not have any problem detecting semantic impossibilities in scenes (e.g., an elephant with butterfly ears, see McCormick et al., 2017).

Overall, therefore, our results suggest that both cognitive processes, scene perception and scene construction, engaged the hippocampus, but with long-axis differences in the portion most involved. The next question we addressed was whether or not the complexity of scenes affected hippocampal recruitment.

4.3. No effect of scene complexity on hippocampal engagement

In the current study, we operationalised visual complexity in terms of the amount of detail or intricacy of an image (Snodgrass & Vanderwart, 1980; see also; Donderi, 2006), and the participants showed high agreement with our designations of simple, middle and complex images. A number of current hippocampal theories argue that visual complexity (or the number of associations or conjunctions), is an important driver of hippocampal activity (Eichenbaum & Cohen, 2014; Lee et al., 2012; Yonelinas, 2013). Thus, more complex images should evoke greater hippocampal response compared to simpler images. This is in contrast to another perspective that suggests a primary function of the hippocampus is to construct scene imagery, irrespective of whether the scenes are simple or complex (Hassabis & Maguire, 2007; Maguire & Mullally, 2013). Complexity might also be relevant for another issue, namely the assertion by Poppenk et al. (2013; see also Brunec et al., 2018) that representations in the hippocampus vary from fine to coarse grained in a posterior to anterior direction. This leads to the prediction that complex scenes should engage the posterior hippocampus.

However, the data driven and contrast driven analyses showed that hippocampal activity was not influenced by scene complexity. Moreover, the cognitive strategies used by participants did not differ for simple and complex scenes. Instead, complexity as a general image feature engaged mostly posterior visual brain regions. The simple and complex scenes in this study differed vastly in terms of their complexity, and so we believe they offered a credible test of the effect of complexity in terms of naturalistic scenes. Our finding accords with the view that the hippocampus processes scenes regardless of whether they are simple or complex (Hassabis & Maguire, 2007; Maguire & Mullally, 2013). How perspectives advocating complexity as a key feature of hippocampal processing (e.g., Lee et al., 2012; Yonelinas, 2013), or the view that specifically fine grained (e.g., complex) representations would engage posterior hippocampus (Brunec et al., 2018; Poppenk et al., 2013), which they did not, can account for our results is unclear. It may be that such theories need to define notions of complexity more precisely, to stipulate the specific processes or features of real-world perception and mental representations that might be subject to this purported effect. Certainly we can conclude from the current study that the number of objects, associations and conjunctions in naturalistic scenes did not influence hippocampal engagement.

4.4. Beyond the hippocampus

The focus of the current study was the hippocampus. Nevertheless, our analyses revealed that the hippocampus was part of a wider set of activated brain areas, including many that have been previously implicated in scene and event processing, such as the parahippocampal and fusiform gyri, and parietal cortex. Among these areas the vmPFC had perhaps the most interesting profile. While it was recruited to a greater extent during layout compared to colour conditions, it also seemed to be more responsive to scenes than scrambled images and to complex than simple stimuli. We speculate that this result might suggest that the vmPFC is a hierarchically superordinate structure that keeps track of scene processing (for a related idea see Robin & Moscovitch, 2017; Sekeres et al., 2018). In fact, we have suggested that the vmPFC may initiate scene construction processes in the hippocampus (Ciaramelli, De Luca, Monk, McCormick, & Maguire, 2019; McCormick, Ciaramelli, De Luca, & Maguire, 2018; see also; De Luca et al., 2018; De Luca, McCormick, Ciaramelli, & Maguire, 2019). In support of this proposal, recent magnetoencephalography studies have found vmPFC activity preceded and then drove that of the hippocampus during both scene imagination (Barry, Barnes, Clark, & Maguire, 2019; Monk et al., 2020) and the recall of autobiographical memories (McCormick, Barry, Jafarian, Barnes, & Maguire, 2020).

5. Conclusions

In this study we sought to probe hippocampal function by manipulating three factors. We found evidence that the hippocampus was engaged by naturalistic scenes compared to scrambled images. Furthermore the posterior hippocampus was activated to a greater extent during tasks relating to scene perception and the anterior hippocampus during tasks associated with scene construction, regardless of the complexity of scenes. In-scanner task performance and incidental encoding could not explain these findings. Overall, these results seem to fit best with the view that the hippocampus may be attuned to processing scenes, be they simple or complex (Hassabis & Maguire, 2007; Maguire & Mullally, 2013). This conclusion could be investigated further in future studies by testing patients with bilateral hippocampal damage, whereby the prediction would be that they should be impaired on tasks involving scenes, be they simple or complex, but unimpaired on tasks involving scrambled stimuli.

Funding

This work was supported by a Wellcome Principal Research Fellowship to E.A.M (210567/Z/18/Z) and the Centre by a Strategic Award from Wellcome (203147/Z/16/Z).

CRediT authorship contribution statement

Cornelia McCormick: Conceptualization, Methodology, Investigation, Formal analysis, Writing - original draft, Writing - review & editing. Marshall A. Dalton: Conceptualization, Writing - review & editing. Peter Zeidman: Conceptualization, Writing - review & editing. Eleanor A. Maguire: Conceptualization, Methodology, Funding acquisition, Supervision, Formal analysis, Writing - original draft, Writing - review & editing.

Notes

No part of the study procedures or analyses was preregistered prior to the research being conducted. We report how we determined our sample size, all data exclusions (if any), all inclusion/exclusion criteria, whether inclusion/exclusion criteria were established prior to data analysis, all manipulations, and all measures in the study. The presentation code is publicly available via the OSF: https://osf.io/K9XD3/. The raw data (neuroimaging, behavioural, eye tracking) cannot be shared publicly due to ethical restrictions relating to General Data Protection Regulation. Data will be released to researchers on the following conditions: approval from the local research ethics committee and with appropriate safeguards to protect from identification of individuals. Legal restrictions relating to copyright preclude the stimuli being shared in a public repository. Please contact the corresponding author via email for access to the stimuli (which will be granted unconditionally), and for any other questions.

Declaration of Competing Interest

None.

Acknowledgements

We are grateful to David Bradbury and Eric Featherstone for technical assistance.

Action editor Paul Reber

Reviewed 19 October 2020

Footnotes

Appendix A

Supplementary data to this article can be found online at https://doi.org/10.1016/j.cortex.2020.12.018.

Appendix A. Supplementary data

The following is the Supplementary data to this article:

Multimedia component 1
mmc1.docx (253.6KB, docx)

References

  1. Addis D.R., Wong A.T., Schacter D.L. Remembering the past and imagining the future: Common and distinct neural substrates during event construction and elaboration. Neuropsychologia. 2007;45(7):1363–1377. doi: 10.1016/j.neuropsychologia.2006.10.016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Aly M., Ranganath C., Yonelinas A.P. Detecting changes in ccenes: The hippocampus is critical for strength-based perception. Neuron. 2013;78:1127–1137. doi: 10.1016/j.neuron.2013.04.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Andrews-Hanna J.R., Reidler J.S., Sepulcre J., Poulin R., Buckner R.L. Functional-anatomic fractionation of the brain's default network. Neuron. 2010;65(4):550–562. doi: 10.1016/j.neuron.2010.02.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Barense M.D., Henson R.N., Lee A.C., Graham K.S. Medial temporal lobe activity during complex discrimination of faces, objects, and scenes: Effects of viewpoint. Hippocampus. 2010;20(3):389–401. doi: 10.1002/hipo.20641. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Barry D.N., Barnes G.R., Clark I.A., Maguire E.A. The neural dynamics of novel scene imagery. Journal of Neuroscience. 2019;39(22):4375–4386. doi: 10.1523/JNEUROSCI.2497-18.2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Barry D.N., Chadwick M.J., Maguire E.A. Nonmonotonic recruitment of ventromedial prefrontal cortex during remote memory recall. Plos Biology. 2018;16(7) doi: 10.1371/journal.pbio.2005479. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Brunec I.K., Bellana B., Ozubko J.D., Man V., Robin J., Liu Z.X., Moscovitch M. Multiple scales of representation along the hippocampal anteroposterior axis in humans. Current Biology. 2018;28(13):2129–2135. doi: 10.1016/j.cub.2018.05.016. e2126. [DOI] [PubMed] [Google Scholar]
  8. Chadwick M.J., Mullally S.L., Maguire E.A. The hippocampus extrapolates beyond the view in scenes: An fMRI study of boundary extension. Cortex. 2013;49(8):2067–2079. doi: 10.1016/j.cortex.2012.11.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Ciaramelli E., De Luca F., Monk A.M., McCormick C., Maguire E.A. What "wins" in VMPFC: Scenes, situations, or schema? Neuroscience and Biobehavioral Reviews. 2019;100:208–210. doi: 10.1016/j.neubiorev.2019.03.001. [DOI] [PubMed] [Google Scholar]
  10. Clark I.A., Hotchin V., Monk A., Pizzamiglio G., Liefgreen A., Maguire E.A. Identifying the cognitive processes underpinning hippocampal-dependent tasks. Journal of Experimental Psychology General. 2019;148(11):1861–1881. doi: 10.1037/xge0000582. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Clark I.A., Monk A., Maguire E.A. Characterising strategy use during the performance of hippocampal-dependent tasks. Frontiers in Psychology. 2020;11:2119. doi: 10.3389/fpsyg.2020.02119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Cohen N.J., Eichenbaum H. Memory, amnesia, and the hippocampal system. MIT Press; Cambridge, Mass: 1993. [Google Scholar]
  13. Dalton M.A., Maguire E.A. The pre/parasubiculum: A hippocampal hub for scene-based cognition? Current Opinion in Behavioral Sciences. 2017;17:34–40. doi: 10.1016/j.cobeha.2017.06.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Dalton M.A., McCormick C., De Luca F., Clark I.A., Maguire E.A. Functional connectivity along the anterior-posterior axis of hippocampal subfields in the ageing human brain. Hippocampus. 2019;29:1049–1062. doi: 10.1002/hipo.23097. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Dalton M.A., McCormick C., Maguire E.A. Differences in functional connectivity along the anterior-posterior axis of human hippocampal subfields. Neuroimage. 2019;192:38–51. doi: 10.1016/j.neuroimage.2019.02.066. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Dalton M.A., Zeidmann P., McCormick C., Maguire E.A. Differentiable processing of objects, associations, and scenes within the hippocampus. Journal of Neuroscience. 2018;38(38):8146–8159. doi: 10.1523/JNEUROSCI.0263-18.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. De Luca F., McCormick C., Ciaramelli E., Maguire E.A. Scene processing following damage to the ventromedial prefrontal cortex. Neuroreport. 2019;30:828–833. doi: 10.1097/WNR.0000000000001281. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. De Luca F., McCormick C., Mullally S.L., Intraub H., Maguire E.A., Ciaramelli E. Boundary extension is attenuated in patients with ventromedial prefrontal cortex damage. Cortex. 2018;108:1–12. doi: 10.1016/j.cortex.2018.07.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Donderi D.C. Visual complexity. A review. Psychological Bulletin. 2006;132(1):1–33. doi: 10.1037/0033-2909.132.1.73. [DOI] [PubMed] [Google Scholar]
  20. Eichenbaum H. Oxford University Press Inc; NY: 2006. Memory binding in hippocampal relational networks. [Google Scholar]
  21. Eichenbaum H., Cohen N.J. Can we reconcile the declarative memory and spatial navigation views on hippocampal function? Neuron. 2014;83(4):764–770. doi: 10.1016/j.neuron.2014.07.032. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. El Haj M., Lenoble Q. Eying the future: Eye movement in past and future thinking. Cortex. 2018;105:97–103. doi: 10.1016/j.cortex.2017.05.024. [DOI] [PubMed] [Google Scholar]
  23. Erez J., Cusack R., Kendall W., Barense M.D. Conjunctive coding of complex object features. Cerebral Cortex. 2016;26(5):2271–2282. doi: 10.1093/cercor/bhv081. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Graham K.S., Barense M.D., Lee A.C. Going beyond LTM in the MTL: A synthesis of neuropsychological and neuroimaging findings on the role of the medial temporal lobe in memory and perception. Neuropsychologia. 2010;48(4):831–853. doi: 10.1016/j.neuropsychologia.2010.01.001. [DOI] [PubMed] [Google Scholar]
  25. Hannula D.E., Ranganath C. The eyes have it: Hippocampal activity predicts expression of memory in eye movements. Neuron. 2009;63(5):592–599. doi: 10.1016/j.neuron.2009.08.025. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Hassabis D., Kumaran D., Vann S.D., Maguire E.A. Patients with hippocampal amnesia cannot imagine new experiences. Proceedings of the National Academy of Sciences USA. 2007;104(5):1726–1731. doi: 10.1073/pnas.0610561104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Hassabis D., Maguire E.A. Deconstructing episodic memory with construction. Trends in Cognitive Sciences. 2007;11(7):299–306. doi: 10.1016/j.tics.2007.05.001. [DOI] [PubMed] [Google Scholar]
  28. Hodgetts C.J., Voets N.L., Thomas A.G., Clare S., Lawrence A.D., Graham K.S. Ultra-high-field fMRI reveals a role for the subiculum in scene perceptual discrimination. Journal of Neuroscience. 2017;37(12):3150–3159. doi: 10.1523/JNEUROSCI.3225-16.2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Kahn I., Andrews-Hanna J.R., Vincent J.L., Snyder A.Z., Buckner R.L. Distinct cortical anatomy linked to subregions of the medial temporal lobe revealed by intrinsic functional connectivity. Journal of Neurophysiology. 2008;100(1):129–139. doi: 10.1152/jn.00077.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Karapanagiotidis T., Bernhardt B.C., Jefferies E., Smallwood J. Tracking thoughts: Exploring the neural architecture of mental time travel during mind-wandering. Neuroimage. 2016;147:272–281. doi: 10.1016/j.neuroimage.2016.12.031. [DOI] [PubMed] [Google Scholar]
  31. Krishnan A., Williams L.J., McIntosh A.R., Abdi H. Partial least squares (PLS) methods for neuroimaging: A tutorial and review. Neuroimage. 2011;56(2):455–475. doi: 10.1016/j.neuroimage.2010.07.034. [DOI] [PubMed] [Google Scholar]
  32. Lee A.C., Brodersen K.H., Rudebeck S.R. Disentangling spatial perception and spatial memory in the hippocampus: A univariate and multivariate pattern analysis fMRI study. Journal of Cognitive Neuroscience. 2013;25(4):534–546. doi: 10.1162/jocn_a_00301. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Lee A.C., Buckley M.J., Pegman S.J., Spiers H., Scahill V.L., Gaffan D., Graham K.S. Specialization in the medial temporal lobe for processing of objects and scenes. Hippocampus. 2005;15(6):782–797. doi: 10.1002/hipo.20101. [DOI] [PubMed] [Google Scholar]
  34. Lee A.C., Yeung L.K., Barense M.D. The hippocampus and visual perception. Frontiers in Human Neuroscience. 2012;6:91. doi: 10.3389/fnhum.2012.00091. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Liu Z.X., Shen K., Olsen R.K., Ryan J.D. Visual sampling predicts hippocampal activity. Journal of Neuroscience. 2017;37(3):599–609. doi: 10.1523/JNEUROSCI.2610-16.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Maguire E.A., Mullally S.L. The hippocampus: A manifesto for change. Journal of Experimental Psychology General. 2013;142(4):1180–1189. doi: 10.1037/a0033650. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. McCormick C., Barry D.N., Jafarian A., Barnes G.R., Maguire E.A. vmPFC drives hippocampal processing during autobiographical memory recall regardless of remoteness. Cerebral Cortex. 2020;30:5972–5987. doi: 10.1093/cercor/bhaa172. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. McCormick C., Ciaramelli E., De Luca F., Maguire E.A. Comparing and contrasting the cognitive effects of hippocampal and ventromedial prefrontal cortex damage: A review of human lesion studies. Neuroscience. 2018;15(374):295–318. doi: 10.1016/j.neuroscience.2017.07.066. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. McCormick C., Rosenthal C.R., Miller T.D., Maguire E.A. Deciding what is possible and impossible following hippocampal damage in humans. Hippocampus. 2017;27(3):303–314. doi: 10.1002/hipo.22694. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. McCormick C., Rosenthal C.R., Miller T.D., Maguire E.A. Mind-wandering in people with hippocampal damage. Journal of Neuroscience. 2018;38(11):2745–2754. doi: 10.1523/JNEUROSCI.1812-17.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. McIntosh A.R., Chau W.K., Protzner A.B. Spatiotemporal analysis of event-related fMRI data using partial least squares. Neuroimage. 2004;23(2):764–775. doi: 10.1016/j.neuroimage.2004.05.018. [DOI] [PubMed] [Google Scholar]
  42. McIntosh A.R., Lobaugh N.J. Partial least squares analysis of neuroimaging data: Applications and advances. Neuroimage. 2004;23(Suppl 1):S250–S263. doi: 10.1016/j.neuroimage.2004.07.020. [DOI] [PubMed] [Google Scholar]
  43. Monk A.M., Dalton M.A., Barnes G.R., Maguire E.A. The role of hippocampal-vmPFC neural dynamics in building mental representations. Journal of Cognitive Neuroscience. 2020 doi: 10.1162/jocn_a_01634. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Mullally S.L., Intraub H., Maguire E.A. Attenuated boundary extension produces a paradoxical memory advantage in amnesic patients. Current Biology. 2012;22:261–268. doi: 10.1016/j.cub.2012.01.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. O'Keefe J., Dostrovsky J. The hippocampus as a spatial map. Preliminary evidence from unit activity in the freely-moving rat. Brain Research. 1971;34(1):171–175. doi: 10.1016/0006-8993(71)90358-1. [DOI] [PubMed] [Google Scholar]
  46. Poppenk J., Evensmoen H.R., Moscovitch M., Nadel L. Long-axis specialization of the human hippocampus. Trends in Cognitive Sciences. 2013;17(5):230–240. doi: 10.1016/j.tics.2013.03.005. [DOI] [PubMed] [Google Scholar]
  47. Robin J., Buchsbaum B.R., Moscovitch M. The primacy of spatial context in the neural representation of events. Journal of Neuroscience. 2018;38(11):2755–2765. doi: 10.1523/JNEUROSCI.1638-17.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Robin J., Moscovitch M. Details, gist and schema: Hippocampal-neocortical interactions underlying recent and remote episodic and spatial memory. Current Opinion in Behavioral Sciences. 2017;17:114–123. [Google Scholar]
  49. Schacter D.L., Addis D.R. The cognitive neuroscience of constructive memory: remembering the past and imagining the future. Philosophical Transactions of the Royal Society London B. Biological Sciences. 2007;362:773–786. doi: 10.1098/rstb.2007.2087. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Scoville W.B., Milner B. Loss of recent memory after bilateral hippocampal lesions. Journal of Neurology, Neurosurgery and Psychiatry. 1957;20(1):11–21. doi: 10.1136/jnnp.20.1.11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Sekeres M.J., Winocur G., Moscovitch M. The hippocampus and related neocortical structures in memory transformation. Neuroscience Letters. 2018;680:39–53. doi: 10.1016/j.neulet.2018.05.006. [DOI] [PubMed] [Google Scholar]
  52. Snodgrass J.G., Vanderwart M. A standardized set of 260 pictures: Norms for name agreement, image agreement, familiarity, and visual complexity. Journal of Experimental Psychology Human Learning and Memory. 1980;6(2):174–215. doi: 10.1037//0278-7393.6.2.174. [DOI] [PubMed] [Google Scholar]
  53. Spano G., Pizzamiglio G., McCormick C., Clark I.A., De Felice S., Miller T.D. Dreaming with hippocampal damage. Elife. 2020;9 doi: 10.7554/eLife.56211. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Strange B.A., Witter M.P., Lein E.S., Moser E.I. Functional organization of the hippocampal longitudinal axis. Nature Reviews Neuroscience. 2014;15(10):655–669. doi: 10.1038/nrn3785. [DOI] [PubMed] [Google Scholar]
  55. Voss J.L., Bridge D.J., Cohen N.J., Walker J.A. A closer look at the hippocampus and memory. Trends in Cognitive Sciences. 2017;21(8):577–588. doi: 10.1016/j.tics.2017.05.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Weiskopf N., Hutton C., Josephs O., Deichmann R. Optimal EPI parameters for reduction of susceptibility-induced BOLD sensitivity losses: A whole-brain analysis at 3 T and 1.5 T. Neuroimage. 2006;33(2):493–504. doi: 10.1016/j.neuroimage.2006.07.029. [DOI] [PubMed] [Google Scholar]
  57. Weiskopf N., Suckling J., Williams G., Correia M.M., Inkster B., Tait R. Quantitative multi-parameter mapping of R1, PD(∗), MT, and R2(∗) at 3T: A multi-center validation. Front Neuroscience. 2013;7:95. doi: 10.3389/fnins.2013.00095. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Yonelinas A.P. The hippocampus supports high-resolution binding in the service of perception, working memory and long-term memory. Behavioural Brain Research. 2013;254:34–44. doi: 10.1016/j.bbr.2013.05.030. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Yonelinas A.P., Ranganath C., Ekstrom A.D., Wiltgen B.J. A contextual binding theory of episodic memory: Systems consolidation reconsidered. Nature Reviews Neuroscience. 2019;20(6):364–375. doi: 10.1038/s41583-019-0150-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Yoonessi A., Kingdom F.A. Comparison of sensitivity to color changes in natural and phase-scrambled scenes. Journal of the Optical Society of America. A. Optics, Image Science, and Vision. 2008;25(3):676–684. doi: 10.1364/josaa.25.000676. [DOI] [PubMed] [Google Scholar]
  61. Zeidman P., Lutti A., Maguire E.A. Investigating the functions of subregions within anterior hippocampus. Cortex. 2015;73:240–256. doi: 10.1016/j.cortex.2015.09.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Zeidman P., Maguire E.A. Anterior hippocampus: The anatomy of perception, imagination and episodic memory. Nature Reviews Neuroscience. 2016;17(3):173–182. doi: 10.1038/nrn.2015.24. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Zeidman P., Mullally S.L., Maguire E.A. Constructing, perceiving, and maintaining scenes: Hippocampal activity and connectivity. Cerebral Cortex. 2015;25(10):3836–3855. doi: 10.1093/cercor/bhu266. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia component 1
mmc1.docx (253.6KB, docx)

RESOURCES