Abstract
When we remember events, we often do not only recall individual events, but also the connections between them. However, extant research has focused on how humans segment and remember discrete events from continuous input, with far less attention given to how the structure of connections between events impacts memory. Here we conduct a functional magnetic resonance imaging study in which participants watch and recall a series of realistic audiovisual narratives. By transforming narratives into networks of events, we demonstrate that more central events—those with stronger semantic or causal connections to other events—are better remembered. During encoding, central events evoke larger hippocampal event boundary responses associated with memory formation. During recall, high centrality is associated with stronger activation in cortical areas involved in episodic recollection, and more similar neural representations across individuals. Together, these results suggest that when humans encode and retrieve complex real-world experiences, the reliability and accessibility of memory representations is shaped by their location within a network of events.
Subject terms: Cognitive neuroscience, Long-term memory, Human behaviour
Naturalistic experiences often have complex structure, consisting of multiple inter-related events. Here, the authors show that the semantic and causal interconnectedness of events in narratives positively predicts memory performance and neural responses associated with memory encoding and recall.
Introduction
Remembering the experiences of our lives requires collecting and connecting the pieces and reasons for what transpired. When we tell each other about the minutes and hours leading up to this moment, the tale will be composed of a string of time periods, “events“1,2, distinguished by properties such as their locale or mood, and by our companions or goals at the time. Traditional experimental memory paradigms3,4 rely on isolated stimuli in which meaningful connections between memoranda across time are removed via trial randomization. Yet in reality, each event exists within, and is to some extent defined by, a dense network of connections across time. These connections come in multiple forms: different timepoints could share properties to greater or lesser degrees, and actions earlier may have consequences later. When remembering and retelling, we often need to recapitulate not only the most important individual events, but also the overall structure of the experience, i.e., the pattern of connections across time5,6. Thus, it is important to understand in what ways the web of interrelations between events contributes to our memories of those experiences.
In order to test how inter-event structure relates to later memory, experimenters must use study material which contains inter-event structure. Recently, researchers have sought to incorporate the complex, multi-event nature of real-world input into laboratory experiments by using auditory and/or visual narratives7,8. Since narratives are temporally continuous, a major question in the literature has been how the human brain identifies and remembers discrete events from continuous experiences2,9,10. As input arrives from the world, the perceiver constructs a mental model of the situation, which consists of agents, objects, spatiotemporal contexts, and the relations between these components11. Changes in the ongoing situation trigger the registration of the just-concluded event into long-term memory, evoking transient responses in the hippocampus and its cortical partners12,13. The boundaries between events are also associated with shifts in neural activation patterns in higher associative areas in the default mode network (DMN14)15. DMN activity patterns specific to individual events are thought to represent situation models16, and are reinstated during narrated memory recall17,18. However, these studies focus on how each event is segmented from its temporally adjacent neighbors. How do the myriad connections between events, both temporally proximal and distal, impact the cognitive and neural underpinnings of naturalistic memory?
Inter-event connections could benefit both memory encoding and retrieval. At encoding, events with strong connections to numerous other events might be frequently reactivated by these links to form robust and integrated representations19,20. At retrieval, events with many connections might be more likely to be cued by other events, enhancing their accessibility. These enhancing effects of inter-event connections on memory have been demonstrated in the reading comprehension literature, which focused on casual relations in relatively short and carefully designed text passages21,22. For example, statements that form causal chains are better remembered than isolated statements, and memory accuracy for a statement increases with the number of causal connections that it has22,23. Causal connectivity between statements also predicts how important readers will deem a given statement to be, and what they will judge to be the gist of the narrative22,24. The current study aims to examine the mnemonic benefits of inter-event connections in light of the burgeoning cognitive neuroscience of memories for events. Using previously unavailable neuroimaging approaches, we investigate the effect of inter-event structure on brain functions supporting the encoding and retrieval of event representations. In addition to testing the influence of causal relations, we take advantage of natural language processing techniques which allow effortless quantification of semantic similarity between text descriptions of complex events25–27. These non-causal (semantic) relations, based on shared meaning and overlapping components between events, may constitute a previously underexplored pathway through which inter-event connections enhance memory.
Here, we propose that when people view and recall realistic, continuous audiovisual stimuli (e.g., movies), events with stronger and more numerous semantic or causal connections to other events will be better remembered, with concomitant hippocampal and DMN activity reflecting enhanced encoding and retrieval-related processing for these events. We conducted a functional magnetic resonance imaging (fMRI) study in which participants watched a series of movies and then verbally recounted the movie plots. To quantify and assess the semantic relationship between events within a movie, we employed an approach scalable and easily generalizable to different types of narratives (Fig. 1). In this method, each narrative is transformed into a network of interconnected events based on semantic similarity measured from sentence embedding distances (the semantic narrative network). We then calculate semantic centrality for each event as the node degree, a graph metric which quantifies the number and strength of connections that a node (event) has to other nodes in the network. Behavioral results revealed that events with higher semantic centrality are more likely to be recalled, without showing primacy and recency effects typical in traditional random list memory experiments3,28. High centrality events are also associated with the neural signatures of stronger and more accurate recall: greater activation and more consistent neural patterns across individuals in the DMN areas including the posterior medial cortex (PMC). The hippocampus shows higher activation following the offset of high centrality events, suggesting that stronger hippocampus-mediated encoding contributes to the high centrality advantages. In parallel, we created a causal narrative network for each movie based on causal relations between events defined by human judgments. Causal centrality of events, again defined as node degree in the network, predicts memory success and neural responses in a similar way to semantic centrality, but also makes an independent contribution to each. Overall, our findings demonstrate that memories for events are shaped by their location within a narrative network, highlighting the importance of considering inter-event structure when studying the cognitive and neural mechanisms of complex and continuous real-world memory.
Results
Behavioral characteristics of unguided narrative recall
We first examined the behavioral characteristics of free spoken narrative recall. Participants watched a series of short movies with unique narratives (Supplementary Table 1) and then verbally recalled the movie plots while undergoing functional MRI. Participants were instructed to describe what they remembered from the movies in their own words in as much detail as they could, regardless of the order of presentation. No external cues or experimenter guidance were provided during recall.
Two example participants’ recall behaviors are depicted in Fig. 2a. On average, participants recalled 9 out of the 10 movies (s.d. 1.2) and the recall lasted 32.4 min in total (s.d. 14.5 min). Each movie was divided into 10–35 events by an independent coder based on major shifts in the narrative (e.g., time, location, action). Participants on average recalled 77.6% of the events within each recalled movie (s.d. 11.2%). Movies tended to be recalled in the original presentation order (mean Spearman’s ρ between the presentation order and the recalled order = .52, s.d. across participants = .55; Fig. 2b, top panel). Although participants were not explicitly instructed to perform serial recall, events were typically recalled strictly in the order in which they occurred within each movie (mean ρ = .97, s.d. = .03; Fig. 2c). Thus, recalling an event likely served as a strong cue for the following event which was often semantically/causally related.
Contrary to traditional random list memory experiments3,28, we did not observe the classic primacy and recency effects on recall probability3 either at the movie level or the event level. The proportion of participants who successfully recalled a movie was not higher for the first or last few movies compared to the movies presented in the middle of the list (Fig. 2b, bottom panel). Likewise, the recall probability of the first/last few events was not higher than that of the events presented in the middle, either within each movie or across all movies (Fig. 2d). Specifically, we did not find a significant difference between the mean recall probabilities of the first/middle/last three events of each movie (F(2,18) = .78, p = .47, η2 = .05). These results suggest that memorability of a movie event was largely influenced by narrative properties beyond the serial position of events.
Narrative network centrality predicts what people will remember later
One important factor that may have affected the behavioral characteristics of movie event recall is the inter-event structure inherent in narratives. We quantified narrative structure by transforming each movie plot into a network of events (Fig. 1), in which the connections between events were determined by their similarity based on semantic contents. To measure semantic similarity between movie events, we first converted the text descriptions of the events, generated by independent annotators, into vectors of 512 numbers using Google’s Universal Sentence Encoder (USE25). Consistent with a recent study26, the trajectories of movie annotations in the high-dimensional vector space were highly consistent across annotators (Supplementary Fig. 1), demonstrating that the text embeddings captured the semantic gist despite the differences in specific words used to describe the events. Likewise, the USE vectors of recall transcripts were similar to those of movie annotations and were also similar across participants (Supplementary Fig. 2). Semantic similarity between events was defined as the cosine similarity between their USE vectors.
Our main variable of interest reflecting the inter-event narrative structure was the centrality of individual events within a narrative network (Fig. 1c). An event’s centrality was computed as its degree (i.e., the sum of the weights of all connections to the event) normalized within each movie. Thus, events with stronger (higher semantic similarity) and greater numbers of connections with other events had higher centrality. Critically, semantic centrality positively predicted subsequent event recall probability, measured as the proportion of participants who recalled each event (r(202) = .20, p = .004, 95% confidence interval (CI) = [.07, .33]; Fig. 3a). To further test the effect of semantic centrality in individual participants, we grouped events into high or low centrality conditions within each movie (i.e., events whose semantic centrality values are within the top/bottom 40%), and measured the proportion of successfully recalled events in each condition. The recall probability averaged across movies was higher for the high than for the low semantic centrality condition (t(14) = 6.12, p < .001, Cohen’s dz = 1.58, 95% CI of the difference = [.06, .12]; Fig. 3b).
We next demonstrated that inter-event semantic relations and causal relations made overlapping as well as unique contributions to narrative memory performance. Classic studies on story comprehension have reported that the number of causal connections with other events predicts the perceived importance and memorability of an event22,24. To test the effect of causal relations, we created the causal narrative networks of the movies (Supplementary Fig. 4) by having independent coders identify causally related events within each movie (see Supplementary Fig. 5 and Supplementary Methods for detailed descriptions of causality responses and the instructions given to the coders). The connection strength between a pair of events was defined as the proportion of coders who responded that the pair is causally related. The centrality (i.e., normalized degree) of each event was then computed within each causal narrative network. Causal centrality was positively correlated with semantic centrality (r(202) = .28, p < .001, 95% CI = [.15, .41]) and recall probability (r(202) = .29, p < .001, 95% CI = [.16, .42]; Fig. 3c). Recall probability was also higher for high than for low causal centrality events within each participant (t(14) = 8.23, p < .001, Cohen’s dz = 2.12, 95% CI of the difference = [.1, .17]; Fig. 3d), consistent with earlier studies22,23. Importantly, a mixed-effects logistic regression analysis revealed that semantic centrality explains successful event recall even after controlling for causal centrality (β = .17, standard error (SE) = .05, χ2(1) = 12.24, p < .001) and vice versa (β = .38, SE = .05, χ2(1) = 55.04, p < .001).
We conducted a preregistered online experiment (N = 393) and replicated the same behavioral characteristics of narrative recall using a new set of 10 short movies (Supplementary Fig. 7). Each participant watched one of the movies and then performed a free written recall of the movie plot. Consistent with the behavioral results from the fMRI experiment, semantic centrality (β = .17, SE = .03, χ2(1) = 48.52, p < .001) and causal centrality (β = .44, SE = .03, χ2(1) = 255.67, p < .001) each uniquely predicted the successful recall of an event, without any clear evidence of serial position effects (i.e., no statistically significant difference between the mean recall probabilities of the first/middle/last three events of each movie, F(2,18) = .85, p = .44, η2 = .04).
High centrality events more strongly activate DMN during recall
Narrative network centrality predicted successful memory recall of movie events. Does it also predict brain responses associated with movie watching and recall? We first identified brain regions whose activation scaled with the semantic centrality of events. In this and all following analyses, we excluded the first event of each movie from the movie watching data. This was to minimize the influence of transient changes in activation associated with the boundaries between narratives12,29. The movie boundary-related responses also disrupted event-specific neural patterns by creating similar activation patterns across all movies (Supplementary Fig. 8).
We performed a whole-brain general linear model (GLM) analysis designed to predict the mean activation of individual events with their semantic centrality. Group-level analysis of the participant-specific beta maps showed that, at a liberal threshold (uncorrected p < .001), higher semantic centrality of an event was associated with stronger activation in several regions including visual and auditory association cortices and precuneus during movie watching (Supplementary Fig. 9a). The involvement of sensory areas may reflect high-level perceptual differences between the high and low centrality events, although low-level visual and auditory features including luminance, contrast, and audio amplitude were not significantly modulated by semantic centrality (all χ2(1)s < 1.94, ps > .16). More importantly, during recall, events with higher semantic centrality more strongly activated default mode network (DMN) areas including the angular gyrus and PMC (Supplementary Fig. 9b). DMN areas have been strongly associated with episodic recollection16,30. We also observed higher activation during recall for high than low semantic centrality events in the bilateral hippocampus (t(14) = 2.71, p = .017, Cohen’s dz = .7, 95% CI of the difference = [.01, .05]). These results are in accordance with the positive relationship between recall performance and semantic centrality, and may suggest that high centrality events were more strongly recollected with rich episodic details.
We used causal centrality as a regressor in the GLM analysis and again found greater activation in the same DMN areas for higher centrality events during recall (Supplementary Fig. 9d). In this and following fMRI analyses, the effects of causal centrality were generally comparable to those of semantic centrality, except that causal centrality effects were weaker in analyses involving intersubject similarity. Thus, we focus on the semantic centrality effects and report the causal centrality effects in Supplementary Fig. 10. We consider potential differences between semantic and causal centrality in the Discussion.
Neural patterns in DMN reflect both event-specific representations and narrative network structure
Prior studies have shown that narrative events are represented as distributed patterns of activation in DMN17,18. How does inter-event structure relate to the neural representations of events during movie watching and recall? To answer this question, we performed an intersubject pattern correlation (pISC) analysis17. Within a brain region, event-specific pISC was computed as the mean spatial similarity (i.e., Pearson correlation) between a participant’s activation pattern of a given event and each of the other participants’ activation patterns of the same event (Fig. 4a). By measuring neural signals shared across participants, the intersubject correlation method was expected to reduce the influence of task-unrelated idiosyncratic noise31,32.
We first created whole-brain pISC maps to identify brain regions that showed robust event representations shared across participants. For each cortical parcel of an atlas33, we computed the mean pISC averaged across events and participants. We then performed a nonparametric randomization test to determine whether the mean pISC was significantly different from a null distribution generated by randomly shuffling event labels across all movies. Replicating our prior study17, positive pISC was observed in widespread cortical regions during both movie watching and recall (FDR corrected q < .05). During movie watching (Supplementary Fig. 11a), the strongest pISC was found in sensory cortices, as all participants processed the same audiovisual stimuli. During recall (Fig. 4d), DMN areas, especially PMC, showed the strongest pISC, consistent with the view that PMC and functionally connected areas are engaged in the episodic construction and representation of events or situation models16.
We next demonstrated that neural patterns in DMN areas reflect not only the situations specific to individual events within each movie, but also the semantic relationships between them during recall. We used whole-brain representational similarity analysis (RSA34): for each cortical parcel and movie, we correlated the event-by-event similarity matrix based on the text descriptions of events (i.e., USE vectors from the movie annotations) and the similarity matrix based on neural responses during recall (Fig. 5). The neural similarity was again computed as intersubject pattern correlation, but here the pISC was computed between different events rather than matching events. Statistical significance was determined by randomization tests using event labels randomly shuffled within each movie, and then corrected for multiple comparisons across parcels (FDR q < .05). We found positive correlations between the semantic similarity and neural similarity in parcels mostly within DMN, especially those in and around PMC (Fig. 5). We also observed similar but stronger effects in DMN using the semantic similarity matrix generated from participants’ recall transcripts rather than movie annotations (Supplementary Fig. 12b).
Narrative network centrality predicts the between-brain similarity of event representations
Our next key question was whether the centrality of events modulates the quality of event-specific neural representations in DMN measured as pISC. Here, we used a region-of-interest (ROI) approach (Fig. 4b) and focused on PMC, which showed the strongest effects in the whole-brain pISC and RSA analyses above. As a lower-level control region, we used the early visual cortex (EVC). Both regions showed event-specific neural patterns (i.e., significantly positive pISC) during movie watching (pISC in PMC = .12, EVC = .3, one-tailed randomization ps < .001; Fig. 4c, left panel) and recall (pISC in PMC = .06, EVC = .01, ps < .001; Fig. 4c, right panel).
For each ROI, we compared the mean pISC of high vs. low semantic centrality events defined within each movie. Randomization tests were used to test the statistical significance of the difference between conditions. During recall, higher semantic centrality of an event was associated with higher pISC in PMC (high − low difference = .019, two-tailed randomization p = .037; Fig. 4e), whereas lower semantic centrality was associated with higher pISC in EVC (difference = −.013, p = .012; Fig. 4f). These findings indicate that high semantic centrality events were represented in a more reliable and convergent manner across brains within a higher associative region supporting situation model representations, but not within a sensory control area. In contrast, no significant difference between conditions was observed in either ROI during movie watching, although the direction of effect was consistent with that during recall in both ROIs (PMC difference = .019, p = .17; EVC difference = −.031, p = .12; Supplementary Fig. 11b). While speculative, the diminished effect of centrality on pISC during movie watching may reflect the fact that the structure of the whole narrative becomes apparent only after participants finish watching the movies (i.e., during recall).
In this and all the above analyses involving pISC during recall, twelve events recalled by fewer than five participants were excluded. However, our main pISC analysis results remained qualitatively identical when all events were included in the analysis (Supplementary Fig. 13).
Narrative network centrality modulates hippocampal encoding signals
Hippocampus has been known to play a crucial role in encoding continuous narratives as discrete events35. Hippocampus activation increases at the offset of a movie event, and the magnitude of the activation predicts subsequent remembering and neural reactivation of the event12,15,36. This boundary response has been interpreted as the registration of the just-concluded event into long-term memory. We tested whether the centrality of events influences the offset-triggered hippocampal encoding signal during movie watching, potentially mediating the behavioral effect of narrative network centrality. We measured the time courses of hippocampal blood oxygenation level dependent (BOLD) responses locked to the boundaries between events, and found that hippocampal responses were higher following the offset of high than low semantic centrality events (Fig. 6a). In contrast, hippocampal responses following the onset of high vs. low centrality events (i.e., before the events fully unfold and diverge in terms of their semantic contents) were not significantly different from each other (Fig. 6b), confirming that semantic centrality specifically affected the encoding of information accumulated during just-concluded events. Stronger hippocampal event offset responses (averaged across 10 – 13 TRs from each offset; TR = 1.5 s) also predicted the successful recall of individual events in a mixed-effects logistic regression analysis (β = .26, SE = .1, χ2(1) = 6.37, p = .012), consistent with prior studies12,36. Moreover, hippocampal offset responses significantly mediated the effects of semantic centrality on event recall (average causal mediation effects = .001, p = .016, 95% CI = [.0002, .003]); the effect of semantic centrality was still significant after controlling for hippocampal responses (β = .2, SE = .05, χ2(1) = 13.91, p < .001), indicating a partial mediation. These results suggest that rich connections between events lead to stronger hippocampus-mediated encoding.
Hippocampus also interacts with higher associative cortices when encoding naturalistic events37, and increased hippocampus-cortex connectivity during encoding is associated with successful learning and memory formation38,39. Does the centrality of events affect hippocampal-cortical coupling as well? We used intersubject functional connectivity analysis (ISFC32) to measure the interaction between the hippocampus and cortical ROIs during movie watching. ISFC computes correlations between activation time courses of different brain regions across participants rather than within participants, which makes it possible to isolate stimulus-locked activity from background noise32. We first computed ISFC between the hippocampus and PMC during the 26 movie events which were 22.5 s (15 TRs) or longer. Functional connectivity patterns computed within windows as short as 22.5 s have previously been shown to robustly predict cognitive states40. We then correlated the ISFC values with the semantic centrality of the events. We found that the hippocampal-PMC interaction was stronger for higher centrality events (r(26) = .49, p = .01, 95% CI = [.13, .74]). In contrast, the hippocampal-EVC interaction did not show a significant relationship with centrality (r(26) = .01, p = .95, 95% CI = [−.38, .4]), and the correlation was significantly lower than that between hippocampus-PMC ISFC and centrality (95% CI of the difference between correlations41 = [.05, .87]). Similar results were observed using different minimum event duration thresholds (Supplementary Table 5). The stronger hippocampal-PMC connectivity during higher centrality events might reflect greater reinstatement of other event representations cued by overlapping components (e.g., ref. 42). However, due to the limited number of movie events included in the analysis, it will be important to replicate these findings with a larger dataset.
Discussion
In this study, we found that the structure of inter-event connections in complex naturalistic experiences predicts the behavioral and neural signatures of their memory traces. We applied an approach of transforming audiovisual movies into networks, whose nodes are events and whose edges are based on semantic similarity or cause-effect relationships between events. Participants watched and recounted the movies in their own words; events highly connected with other events within the narrative network, i.e., high centrality events, were more likely to be recalled. Higher centrality was also associated with greater hippocampal activity at event boundaries, as well as with increased hippocampal-cortical interaction during movie watching. Furthermore, recalling high centrality events more strongly recruited high-order cortical regions in the DMN involved in episodic recollection, and the multivoxel patterns of high-centrality events were reinstated in a more convergent manner across individuals, relative to lower-centrality events. These findings demonstrate that the specific structure of relations between events in a natural experience predicts both what will be remembered and what the properties of hippocampal and DMN regions will be during later recollection.
Recent years have seen an explosion in the use of naturalistic stimuli such as movies and narratives in exploring the behavior and neuroscience of human memory, as they provide an engaging laboratory experience with strong ecological validity compared to isolated words or pictures7,8,43. These studies have suggested that findings from traditional random-item list paradigms, which have dominated the field for decades, do not always fully extend to naturalistic recall (e.g., ref. 26). In line with this, we observed that the recall probability of events from a movie does not show serial position effects typically reported in random-item list learning3,28 where the first and last few items in a list tend to be better remembered than items in the middle. This finding was consistent regardless of whether each participant watched a single movie (Supplementary Fig. 7c) or a series of movies in a row (Fig. 2d). The lack of clear primacy or recency advantages may be due to the inter-event dependencies which made each narrative a coherent structure, supporting memories for central events which did not necessarily occur at the beginning or end of the story; that is, inter-event connections may overshadow the existing effects of temporal positions. At the same time, centrality effects may not be specific to narratives; semantically related items in a random list trigger recall of each other4,44, which could lead to better memory performance for those items. Furthermore, the event complexity of narratives is not likely to be the main reason for the lack of serial position effects: in a highly realistic encoding setting (a real-world walk) where the events consisted of unrelated activities (e.g., visiting a number of artworks), naturalistic recall showed similar characteristics to random-item list recall45. Further studies are needed to determine the roles of structural coherence and event complexity in centrality effects on memory. Understanding when and how classic list-based memory effects fail to extend to narratives and other natural stimuli will be essential for evaluating the results of future studies which use complex realistic conditions to study memory.
Why does higher event centrality improve memory performance? Distinct benefits may be present at recall and at encoding. During recall, high centrality events may have greater opportunity to be cued during recall, as by definition they have higher association strengths with other events. In our experiment, recall itself took the form of a narrative; under these conditions, high centrality events may have been especially likely to be recounted, because omitting them might disproportionately affect the logical structure and coherence of the reconstructed story. Inter-event connections may also benefit encoding. During movie watching, events highly connected with other events are more likely to reactivate and be reactivated by the other events containing shared components42,46,47. Consistent with this, we found that the coupling between the hippocampus and a cortical region involved in representing events (PMC16,48, see below) was stronger when participants were watching events with higher semantic centrality. The reactivation of high centrality events during encoding may result in more robust memory for those events by functioning as repeated encoding, and/or integrating the interconnected events to form joint representations19,20,49. The benefit of high centrality during encoding is also reflected in the greater hippocampal responses following the offset of high than low centrality movie events (Fig. 6a). Such hippocampal event boundary responses have been linked to the successful registration of just-concluded episodes into long-term memory12,13,35, which was replicated in the current study. It has been shown that DMN connectivity during movie-viewing is modulated by surprise50; one possibility is that the conclusion of a higher centrality event produces greater uncertainty in the ongoing narrative, as higher centrality events are more likely to influence the main storyline of the narrative. This may result in a more salient boundary and stronger boundary-evoked encoding signals.
We demonstrated that DMN activity during remembering was modulated by the recollected event’s position in the narrative network. High-level associative areas in the DMN14, especially the PMC and its functionally connected subregions such as the angular gyrus, have been implicated in the episodic construction and representation of events16,17,48. In accordance with this view, we observed event-specific neural activation patterns in the medial and lateral DMN areas during recall (Fig. 4d), and representational similarity analysis revealed that the relational structure of these neural event patterns could be predicted by human-generated descriptions of the movie and by recall transcripts (Fig. 5, Supplementary Fig. 12b; for a similar approach, see ref. 26). Critically, activation in the PMC and angular gyrus scaled with the degree to which events had more connections with other events during recall (Supplementary Fig. 9b, 9d), consistent with prior studies showing that these areas are involved in combining and comprehending semantically connected information51–53.
Furthermore, higher semantic centrality predicted greater between-participant pattern convergence in PMC (Fig. 4e). This is likely to be a neural signature of stronger and more accurate recall of episodic details17,54 for high centrality events, dovetailing with the behavioral results. Additionally, higher intersubject similarity for high centrality events might arise from design pressure on narratives. Highly connected events are likely to be logically important in a story; indeed, we found that semantic centrality was positively correlated with the perceived importance of events as retrospectively rated by independent coders (r(202) = .22, p = .002, 95% CI = [.08, .34]). Thus, to aid the understanding of their linked events and eventually the whole story, high centrality events need to be designed in a way that minimizes the variability or ambiguity in how people interpret them. This adoption of a similar canonical interpretation of an event across people gives rise to more similar neural responses across individuals55–57. The design pressure may even produce unique characteristics associated with high centrality shared across events and narratives; high centrality events were semantically more similar to other high centrality events than to low centrality events across movies, although the difference was small (difference in r = .047, p < .001). Future work may investigate whether real-life everyday events without such design pressure would show similar centrality effects to what we observed here using fictional narratives.
In contrast with the pattern in the DMN, we observed that a low-level sensory region (EVC) showed higher between-brain convergence for low semantic centrality events during recall (Fig. 4f). This result should be interpreted with caution as the overall pISC was extremely low in EVC during recall due to the absence of shared visual stimulation (below the level typically considered reliable signal, in line with prior reports17,18,58). Nonetheless, we can speculate that the opposite effects obtained in PMC and EVC may reflect switching between internal and external modes of processing, primarily involving higher-order cortices in the DMN and sensory areas, respectively59,60. Participants are more likely to be in internal mode that prioritizes retrieval61,62 while watching high centrality events that reactivate associated events, whereas external mode is more likely to prevail during low centrality events as participants would focus primarily on the novel current input. This may result in more visually-driven memory reinstatement (e.g., involving salient visual fragments rather than the gist of the event) and thus stronger pISC in EVC for low centrality events. EVC indeed tended to show higher activation for low semantic centrality events during movie watching (Supplementary Fig. 9a), even though low-level visual features such as luminance and contrast were not modulated by semantic centrality. Similarly, a recent study63 reported that the visual sensory network is more activated when participants report a lower understanding of an ongoing narrative.
One might have expected that the effects of narrative structure would not be apparent in brain responses measured during ongoing movie watching, as the full structure of inter-event connections is only available after all movie events are completed. Still, as discussed above, event centrality significantly influenced hippocampal and cortical univariate responses during movie watching. A simple explanation for these results is that centrality based on partial narrative networks (i.e., a network that excluded events not-yet-presented) was sufficiently similar to the full-narrative centrality values, especially later in a movie. Indeed, semantic centrality computed from networks excluding not-yet-presented events was positively correlated with that based on full networks (r(192) = .76, p < .001, 95% CI = [.69, .81]). Another interesting, and not mutually exclusive, possibility is that participants were able to predict the full-narrative centrality of a current event by anticipating the potential connections with future events. In support of this interpretation, we found a strong positive correlation between the perceived importance of events obtained concurrently with watching the movies and those obtained retrospectively, after the movie ended (r(202) = .67, p < .001, 95% CI = [.58, .74]). Predictions of event centrality could be based on the learned schema of canonical story structures64,65 as well as on director’s cues used in popular movies such as luminance and shot motion66. Future work will explore how brain responses are driven by the temporally evolving, rather than static, inter-event structure when participants consume unpredictable stories, or actively engage in selecting upcoming narrative events. Future work may also explore the cognitive and neural mechanisms supporting the learning of novel narrative network structures, and whether they are similar to learning the network structure of simple isolated stimuli or actions (e.g., refs. 67,68).
What is the nature of information reflected in narrative network centrality? We believe that both semantic centrality and causal centrality primarily reflect aspects of how situation models for individual events are related to each other, rather than surface features (or “textbase”) of the stimuli or annotations as discussed in classic discourse theories69,70. With respect to semantic centrality, each event annotation included descriptions from three different annotators and often consisted of multiple sentences. All of these different sentences and phrasing choices were incorporated into the text embedding for each event, from which the semantic narrative network arises. Thus, the event embedding vectors capture information abstracted beyond the surface features of the original sentences. With respect to causal centrality, cause-effect relations between events are traditionally associated with the situation model, identified as crucial elements of knowledge structures69,71,72. Indeed, half of the movies in the current study contain no dialogue at all, and thus human raters cannot be relying on text or language provided in the stimulus to make causality judgments. Thus, while some surface features such as annotators’ word choices should be expected to reflect aspects of the situation model (see Supplementary Methods for semantic centrality based on word-level information), situation-level information, rather than low-level textual overlap, is likely to determine our centrality measures.
Causal relations have long been considered an important organizing factor for event and narrative memories22,64,73. Consistent with earlier work, we found that events with stronger causal connections with other events are better remembered (Fig. 3c, d), and these effects were not redundant to those of semantic connections. Yet, while the effects of causality on univariate responses during movie-viewing (see also ref. 63) were comparable to the effects of semantic centrality (Supplementary Figs. 9c and 10c), multivoxel pattern effects of causality during recall were not as clear as those of semantic similarity (Supplementary Figs. 10b, e). Several characteristics of causal relations in movie stimuli might have reduced the reliability of the effects of causal narrative network structure. First, causal relations were sparse and mostly identified between adjacent events (Supplementary Figs. 4a, 5d). In addition, causality judgments may be more idiosyncratic: average across-coder correlation was lower for causal (mean r(202) = .34) than semantic centrality (mean r(202) = .52) when centrality was computed from each individual coder’s causality rating or movie annotation. It is also noteworthy that semantic and causal connections were measured in distinct ways (text embeddings and human judgments, respectively) and reflect different types of information: semantic connections are based on similar or shared features such as people, places, and objects, whereas causal connections additionally require an action, its outcome, and internal models providing a logical dependency between the two74,75. For example, two events “Jill threw the ball” and “Jack fell to the ground, unconscious” may have a clear causal link for a reader with some background knowledge, but low semantic similarity according to text embeddings as they have no overlapping or similar-meaning words or topics. In this study, we did not focus strongly on dissociating semantic and causal centrality, as they were positively correlated in our movie stimuli. Future studies designed to orthogonalize different types of inter-event relations, including semantic and causal relations as well as other dimensions such as emotional similarity76,77, will be able to further clarify their unique influences on the behavioral and neural signatures of memory. Additionally, further investigations with more stimuli may examine the extent to which the two co-occur in naturalistic narratives, as well as in non-narrative real-world experiences.
In summary, we applied a recently developed natural language model and neuroimaging techniques to a universal and natural form of human memory recall: telling stories about the past. This approach allowed us to demonstrate that rich connections between events in complex realistic experiences protect against forgetting and predict the neural responses associated with successful memory encoding, as well as the properties of brain activity during spoken recall. Consideration of the effects of inter-event structure on real-world memory may benefit practical applications such as the development of memory interventions for clinical and healthy aging populations78 or promoting learning in educational settings79–81. In addition, our work demonstrates that holistic metrics which capture the interrelations of events within episodes may be important to incorporate into models of learning and comprehension, especially as these models grow in their sophistication and power to explain complex experiences in the real world82,83.
Methods
The current study complies with ethical regulations for research on human participants. The fMRI experiment was conducted following protocols as approved by the Princeton University Institutional Review Board. The preregistered online experiment was conducted following protocols as approved by the Johns Hopkins University Institutional Review Board.
Participants
Twenty-one healthy participants were recruited from the Princeton community (12 females, age 20–33 years, mean age 26.6 years). All participants were right-handed native English speakers and reported normal hearing and normal or corrected-to-normal vision. Informed consent was obtained in accordance with procedures approved by the Princeton University Institutional Review Board. Participants received monetary compensation for their time ($20 per hour). Data from 6 of the 21 participants were excluded from analyses due to excessive head motion (absolute displacement greater than 4 mm) in at least one scanning run.
Stimuli
The audiovisual stimuli consisted of 10 short movies including 3 animations and 7 live-action movies. The movies were on average 4.54 min long (ranged 2.15 to 7.75 min) and had narratives that varied in content and structure. Two of the movies consisted of short clips edited from longer full movies (Catch Me If You Can, The Prisoner). Detailed information about each movie is provided in Supplementary Table 1. Each movie was prepended with a 6-s long title scene in which the title in white letters appeared at the center of the black screen and then gradually disappeared. Five movies were presented in the first movie watching phase scanning run and the other five were presented in the second run. The movies were played consecutively within each scanning run without gaps in between other than the title scenes. The presentation order of the ten movies was fixed across participants. As in our prior study17, an additional 39-s audiovisual cartoon (Let’s All Go to the Lobby) unrelated to the movie stimuli was prepended at the beginning of each movie watching scanning run. The introductory cartoon was excluded from analyses.
Experimental procedures
The experiment consisted of three phases: movie watching, free spoken recall, and cued spoken recall. All three phases were performed inside the scanner. Before the movie watching phase, participants were told that they would be watching a series of short movies. As in our prior study17, we instructed participants to pay attention to the movies as they would normally do in real life. Participants were also told that they would be asked to verbally describe the movie plots later. The movie watching phase consisted of two consecutive scanning runs. Participants watched five movies in each run (first run video duration = 24.9 min, second run video duration = 22.9 min). No behavioral responses were required from the participants during scanning.
The free spoken recall phase immediately followed the movie watching phase. Participants were instructed to describe aloud what they remembered from the movies in as much detail as they could, regardless of the order of presentation. We encouraged participants to speak for at least 10 min and told them that if they chose to speak for longer, that would be even better. Participants were also allowed to return to a movie that they had described earlier in case they realized they had missed something while speaking about another movie. We instructed participants to verbally indicate that they were finished by saying “I’m done” after recalling everything they could remember. A white fixation dot was presented on the black screen while participants were speaking; participants were told that they did not need to fixate on this dot. In case participants needed to take a break or the duration of the scanning run exceeded the scanner limit (35 min), we stopped the scan in the middle and started a new scanning run where participants resumed from where they had stopped in the previous run. 4 of the 15 participants included in the analysis had such a break within their spoken recall session.
During the cued spoken recall phase immediately following the free spoken recall phase, participants viewed a series of titles of the ten movies they watched. For each movie, participants were instructed to first read the title out loud and then describe the movie. Participants were told to provide a short summary of a few sentences in case they previously described the movie during the free spoken recall, but describe the movie in as much detail as they could if the movie was previously forgotten. The cued spoken recall phase was not analyzed for the current study.
All visual stimuli were projected using an LCD projector onto a rear-projection screen located in the magnet bore and viewed with an angled mirror. The Psychophysics Toolbox (http://psychtoolbox.org/) for MATLAB was used to display the stimuli and to synchronize stimulus onset with MRI data acquisition. Audio was delivered via in-ear headphones. Participants’ speech was recorded using a customized MR-compatible recording system (FOMRI II; Optoacoustics Ltd.).
Behavioral data collection and preparation
Movie event segmentation
Each of the ten movie stimuli was segmented into 10–35 events (mean 20.2, excluding the title scenes) by an independent coder who was not aware of the experimental design or results. Following the method used in our previous study17, we instructed the coder to identify event boundaries based on major shifts in the narrative (e.g., location, topic, and/or time). Unlike in the prior study that used a 50-min movie17, we did not set the minimum event duration (10 s) because we used much shorter movie stimuli in the current study. The coder gave each event a descriptive label (for example, “girl inside room alone with a pizza”). The start and stop timestamps of each event were recorded. There were 202 movie events in total and the duration of events ranged from 2 to 42 s (s.d. = 7.4 s). The number and the mean duration of events for individual movies are summarized in Supplementary Table 2.
Movie annotations
Movie annotations were provided by three independent annotators who did not participate in the fMRI experiment. Each annotator identified finer-grained sub-event boundaries within each of the 202 movie events based on their subjective judgments. The beginning and end of the fine-grained sub-events were also timestamped. For each sub-event, the annotators provided written descriptions about what was happening in the movie at that moment in their own words. No edits were made on the written descriptions other than correcting typos and removing/replacing special characters not recognized by our text analysis scripts. Supplementary Table 2 summarizes the number of fine-grained sub-events and the number of words generated by individual annotators for each movie. An example movie annotation can be found in Supplementary Table 3.
Recall transcripts
The audio recording of each participant’s free spoken recall was transcribed manually. Each recall transcript was segmented into discrete utterances based on pauses and changes in the topic. The recall transcripts were segmented such that each utterance was not longer than 50 words. Timestamps were also identified for the beginning and end of each utterance. Each utterance was categorized as one of the followings based on its content: (1) recall of specific movie events, (2) general comment about the movie, (3) memory search attempt (e.g., “Let’s see…”), (4) end of recall (e.g, “I’m done.”), and (5) speech unrelated to the task (e.g., “Can I start now?”). In case an utterance was a recall of movie events, the specific movie events described in the utterance were identified. Among the different types of utterances, only the recall of specific movie events was used in the behavioral and fMRI analyses in the current study.
Importance ratings
Importance ratings for each of the 202 movie events were collected from four independent raters who did not participate in the fMRI experiment. The raters watched each movie and then retrospectively rated how important each event was for understanding what happened within the movie on a scale from 1 (not important at all) – 10 (very important). The ratings were averaged within each event across raters for analyses (range 1.5–10 across events, mean 6.09, s.d. 1.92). We additionally collected importance ratings from a separate group of four independent raters while they were watching the 10 movies for the first time. At the end of each movie event, the movie stopped playing, and the raters rated the importance of the just-played event on a scale from 1 to 10. These rate-as-you-go importance ratings averaged across the raters were positively correlated with the retrospective ratings (r(202) = .67, p < .001, 95% CI = [.58, .74]). Importance ratings were positively correlated across raters for both retrospective ratings and rate-as-you-go ratings (mean event-wise cross-rater correlation computed within each movie = .65 and .55, respectively).
Narrative networks
To quantify and assess the inter-event structure of the movie stimuli, we transformed each movie plot into a graph/network. In this narrative network, the events within a movie (nodes) form connections with each other (edges), and the connection strength between a pair of events (edge weight) is determined by their content similarity or causality. The narrative network edges were unthresholded (except for the visualization of semantic narrative networks) and undirected. The centrality of each individual event within a movie was defined as the degree of each node (i.e., the sum of the weights of all edges connected to the node) in the network, normalized by the sum of degrees and then z-scored within each movie. Events with stronger and greater numbers of connections with other events had higher centrality.
Semantic narrative networks
Movie annotations were used to generate narrative networks based on the semantic similarity between events (Fig. 1). For each annotator and movie, the text descriptions for the fine-grained sub-events were concatenated within each movie event. The text descriptions were then encoded into high-dimensional vectors with Google’s Universal Sentence Encoder (USE25), a natural language processing model built in TensorFlow (https://www.tensorflow.org), such that each movie event was represented as a 512-dimensional vector. The USE vectors from the three annotators were highly similar to each other (mean event-wise cross-annotator cosine similarity between all possible annotator pairs = .78; Supplementary Fig. 1); thus the USE vectors were averaged across annotators within each movie event. For each movie, the narrative network was generated by using the cosine similarity between the USE vectors of movie event pairs as the edge weights between nodes (events). The semantic centrality values based on USE sentence embedding vectors were correlated with those based on word-level overlap or word2vec embeddings (Supplementary Methods).
Causal narrative networks
To generate narrative networks based on the causal relationship between events (Supplementary Fig. 4), we had 18 independent coders identify causally related event pairs (the cause event and the effect event) within each movie. Each coder coded different subsets of the ten movies and each movie was coded by 12 (Catch Me If You Can) or 13 coders (all the other movies). The coders watched the movies and were given the movie annotation with sub-event segmentation by the annotator JL. The coders were instructed to consider two movie events as causally related if any fine-grained sub-event of an event is a strong cause of any (at least one) sub-event of the other event (see Supplementary Methods for the exact instructions given to the coders). Whether a causal relationship was strong enough to be identified depended on the coders’ subjective criteria; the coders were instructed to keep the criteria as consistent as possible. The coders were also told to ignore any causal relationship between the sub-events within the same event. Thus, an event pair always consisted of two different events. For each movie, the edge weights between nodes in the narrative network were defined as the proportion of coders who identified a movie event pair as causally related, regardless of the cause-effect direction. However, causal centrality computed from directed networks which accounted for the cause-effect direction showed highly similar behavioral effects as the centrality computed from undirected networks (Supplementary Fig. 6). The average Jaccard similarity between a pair of coders’ lists of causally related event pairs was .31 (computed within each movie and then averaged across movies).
Semantic similarity between events across movies
To examine whether there are semantic characteristics shared among high semantic centrality events across different movies, we computed similarity between event-specific USE vectors (averaged across annotators) across movies. Specifically, we tested whether the similarity between high centrality events was higher than the similarity between high and low centrality events. High and low centrality events were defined as the events whose semantic centrality values were within the top and bottom 40% in each movie, respectively. For each movie, we computed Pearson correlations between the USE vector of each high centrality event and the USE vectors of each of the other movies’ high centrality events. The correlation coefficients were averaged across events and movies to produce the mean similarity value for high centrality-high centrality event pairs. Likewise, we computed the mean similarity between each movie’s high centrality events and each of the other movies’ low centrality events. We then performed a randomization test to assess whether the difference between the mean similarities of high-high pairs and high-low pairs was significantly different from zero. A null distribution of the difference of mean USE vector similarities was generated by randomly shuffling the high or low centrality labels of the events within each movie and then computing the difference 1000 times. A two-tailed p-value was defined as the proportion of values from the null distribution equal to or more extreme than the actual difference.
Mixed-effects logistic regression analysis of recall behavior
We performed a mixed-effects logistic regression analysis implemented in R’s “lme4” package to test the unique effect of semantic centrality and causal centrality on recall performance after controlling for each other. Each event from each participant served as a data point. Data were concatenated across all participants. The dependent variable of each data point was the event recall success (1 = recalled, 0 = not recalled). Normalized semantic and causal centrality were included as fixed effects. Individual participants and movie stimuli were included as random effects. Statistical significance of the unique effect of each type of centrality was tested by performing a likelihood ratio test for the full model against a null model including all independent variables except for the variable of interest.
fMRI acquisition
fMRI scanning was conducted at Princeton Neuroscience Institute on a 3 T Siemens Prisma scanner with a 64-channel head/neck coil. Functional images were acquired using a T2*-weighted multiband accelerated echo-planar imaging (EPI) sequence (TR = 1.5 s; TE = 39 ms; flip angle = 50°; acceleration factor = 4; shift = 3; 60 oblique axial slices; grid size 96 × 96; voxel size 2 × 2 × 2 mm3). Fieldmap images were also acquired to correct for B0 magnetic field inhomogeneity (60 oblique axial slices; grid size 64 × 64; voxel size 3 × 3 × 2 mm3). Whole-brain high-resolution anatomical images were acquired using a T1-weighted MPRAGE pulse sequence. Scanning parameters for the anatomical images varied across participants (15 participants had 176 sagittal slices with voxel size 1 × 1 × 1 mm3; 6 participants had 192 sagittal slices with voxel size .9 × .86 × .86 mm3), as the anatomical images of a subset of participants were originally obtained for other projects unrelated to the current study.
fMRI preprocessing
Preprocessing of high-resolution anatomical images and cortical surface reconstruction were performed using FreeSurfer’s recon-all pipeline. For each scanning run, functional images were corrected for head motion and B0 magnetic inhomogeneity using FSL’s MCFLIRT and FUGUE, respectively. Functional images were then coregistered to the anatomical image, resampled to the fsaverage6 template surface (for cortical analysis) and the MNI 305 volume space (for subcortical analysis), and then smoothed (FWHM 4 mm) using the FreeSurfer Functional Analysis Stream. The smoothed functional data were then high-pass filtered within each scanning run (cutoff = 140 s). For intersubject functional connectivity analysis, we additionally projected out the following nuisance regressors from the filtered functional data: the average time courses (z-scored within each run) of 1) high s.d. voxels outside the grey matter mask (voxels in the top 1% largest s.d.), 2) cerebrospinal fluid, and 3) white matter32. The resulting time series were z-scored within each vertex or voxel across TRs. The first 2 TRs of movie watching scanning runs were discarded as the movies were played 2 TRs after the scanning onset. The first 3 TRs of both movie watching and free spoken recall scanning runs were additionally removed, shifting the time-courses by 4.5 s, to account for the hemodynamic response delay.
Cortical parcellation and region of interest (ROI) definition
For whole-brain pattern-based analyses, we used a cortical parcellation atlas based on fMRI functional connectivity patterns33. Specifically, we used the atlas where the cortical surface of the brain is divided into 400 parcels (200 parcels per hemisphere) which are clustered into previously reported 17 functional networks84. For region-of-interest analyses, we defined the bilateral posterior-medial cortex (PMC) and the bilateral early visual cortex (Fig. 4b) by combining parcels from the 400-parcel atlas that correspond to the areas of interest. The PMC ROI consisted of the posterior cingulate cortex and precuneus parcels in the default mode network. The early visual cortex ROI consisted of parcels around the primary visual cortex (see Supplementary Table 4 for the list of parcels used to create the ROIs). The bilateral hippocampus mask was extracted from FreeSurfer’s subcortical (Aseg) atlas on the MNI volume space.
Univariate activation analysis
We performed whole-brain univariate activation analysis on the cortical surface to identify regions whose activation scales with the narrative network centrality (Supplementary Fig. 9). The analysis was performed separately for the movie watching phase and the recall phase. For each vertex of each participant, we first computed the mean activation for each movie event by averaging the preprocessed BOLD signal across TRs that correspond to the event. The first event of each movie was excluded from this and all other univariate analysis of the movie watching phase (see Supplementary Fig. 8). For the recall phase, only the events successfully recalled by the participant were included in the analysis. We then performed a linear regression where the event-by-event activation (combined across all 10 movies) was explained by the semantic or causal centrality of the events, after regressing out the overall movie-level activation from the event-by-event activation. Finally, one-sample t-tests against zero (two-tailed) were applied on the participant-specific vertex-wise parameter estimate maps to generate the group-level t-statistic map.
We also compared the ROI-specific univariate activation for high vs. low centrality movie events during each experimental phase. High and low centrality events were defined as the events whose semantic/causal centrality metrics were within the top and bottom 40% in each movie, respectively. For each participant and event, the preprocessed BOLD signals were first averaged across voxels or vertices within an ROI and across all TRs corresponding to the event. The mean signal was then averaged across events in the same condition and then across movies, resulting in a single value per participant and condition. Two-tailed paired t-tests were used to test the statistical significance of the difference between the high vs. low centrality conditions.
Intersubject pattern correlation analysis
Whole-brain intersubject pattern correlation (pISC17,31) maps were generated for the movie phase (Movie-Movie similarity; Supplementary Fig. 11a) and the recall phase (Recall-Recall similarity; Fig. 4d) separately. pISC was calculated in a participant-pairwise manner using the following procedures. For each cortical parcel of each participant, first the mean activation pattern of each event was generated by averaging the preprocessed movie or recall phase BOLD data across TRs within the event in each vertex within the parcel. Note that as recall BOLD data existed only for successfully recalled events, each participant had a different subset of recall event patterns. For each participant and event, we computed the Pearson correlation between the event pattern of the participant and the pattern of the matching event from each of the remaining participants, which resulted in N – 1 correlation coefficients (N = the total number of participants who watched/recalled the event). The correlation coefficients were then averaged to create a single pISC (r) value per event per participant. These pISC values were averaged across events (combined across all 10 movies) and participants, resulting in a single pISC value for each parcel. We performed a randomization test for each parcel to test the statistical significance of the mean pISC. Parcel-wise mean pISC values were obtained using the same procedures as described above, except that we randomly shuffled the event labels before computing the between-participants pattern similarity. That is, one participant’s neural pattern of an event was correlated with another participant’s neural pattern of a non-matching event. This procedure was repeated 1000 times to generate a null distribution of pISC. A one-tailed p-value was defined as the proportion of values from the null distribution equal to or greater than the actual mean pISC. The p-values from the entire cortical surface were corrected for multiple comparisons across all 400 parcels using the Benjamini-Hochberg procedure (q < .05).
We also computed pISC in the PMC and early visual cortex to test the relationship between the semantic/causal narrative network centrality metrics and event-specific neural representations in the ROIs. The participant-specific, event-by-event pISC values were computed for each ROI in the same way we computed pISC for each parcel of the whole-brain pISC map above (Fig. 4a), separately for movie watching and recall. We compared the pISC for high vs. low centrality events, defined as the events whose centrality metrics were within the top or bottom 40% in each movie. The pISC values were first averaged across events within the high/low centrality condition for each movie and then across movies, resulting in a single pISC value per condition per participant. We then ran a randomization test to assess whether the difference of pISC between the high vs. low centrality conditions, averaged across participants, was significantly different from zero. A null distribution of the mean difference between the conditions was generated by randomly shuffling the event labels of the event-specific pISC values within each movie and then computing the difference 1000 times. A two-tailed p-value was defined as the proportion of values from the null distribution equal to or more extreme than the actual difference.
In all analyses involving intersubject neural similarity (including the representational similarity analysis and the intersubject functional connectivity analysis), the first event of each movie was excluded from movie phase analyses to minimize the effect of movie onset (Supplementary Fig. 8). For recall phase analyses, we excluded twelve events recalled by fewer than five participants (1–3 events per movie from 6 movies). However, we obtained qualitatively identical results when we included all events in the analyses (Supplementary Fig. 13).
Representational similarity analysis
We performed representational similarity analysis34 by comparing the event-by-event similarity matrices based on two different types of event representations: the text descriptions of events (i.e., movie annotations or recall transcripts) and neural activation patterns measured during recall (Fig. 5, Supplementary Fig. 12b). The similarity matrix based on movie annotations was generated for each movie by computing the pairwise cosine similarity between the USE vectors of all events within the movie. This matrix was identical to the adjacency matrix of the semantic narrative network. To create the similarity matrix based on recall, we first extracted the sentences from each participant’s recall transcript describing each event and then converted them into USE vectors. The similarity matrix was generated for each participant and movie by computing the cosine similarity between the USE vectors of all events recalled by the participant. The matrices were then averaged across all participants. As participants recalled different subsets of events, the number of participants averaged was different across event pairs.
The fMRI recall pattern similarity matrix was generated for each parcel of the Schaefer atlas. Within each of the ten movies, we computed pattern correlations (Pearson r) between all possible pairs of events between all pairs of participants. For each participant and movie, this resulted in N – 1 fMRI pattern similarity matrices with the size of M × M, where N is the total number of participants and M is the number of events within the movie. We took the average of each matrix and its transpose to make the similarity matrix symmetric (i.e., similarity between events a and b across participants i and j = average of corr(participant i event a, participant j event b) and corr(participant j event a, participant i event b)), and then averaged the N – 1 similarity matrices to generate a single fMRI similarity matrix per movie and participant.
The representational similarity between a text-based similarity matrix and an fMRI pattern-based similarity matrix was measured by computing the Pearson correlation between the lower triangles (excluding the diagonal values) of each matrix. The correlation coefficients were next averaged across movies and then across participants to create a single value per parcel. For each parcel, a randomization test was performed to test whether the mean representational similarity was significantly greater than zero. We randomly shuffled the event labels of the text-based similarity matrix within each movie and then computed the mean representational similarity as described above. This procedure was repeated 1000 times to generate a null distribution, and a one-tailed p-value was defined as the proportion of values from the null distribution equal to or greater than the actual mean representational similarity. The whole-brain p-values were corrected for multiple comparisons across parcels using the Benjamini–Hochberg procedure (q < .05).
Hippocampal event boundary responses
We compared hippocampal event boundary responses following the onset/offset of high vs. low centrality events during movie watching (Fig. 6), High and low centrality events were defined as the events whose centrality values were within the top or bottom 40% in each movie. We first averaged TR-by-TR BOLD signals across voxels within the bilateral hippocampus mask for each participant. We then extracted time series around the onset/offset (−2 – 15 TRs) of each high/low centrality event. The first and last events of each movie were excluded to minimize the effect of between-movie transitions. Each time series was baseline corrected by subtracting the mean activation of the two TRs immediately preceding the onset/offset of the event from each time point. The participant-specific time series were then averaged across events within each condition and then across movies. Two-tailed paired t-tests were used for each time point to compare the high vs. low centrality conditions. We applied the Benjamini-Hochberg procedure (q < .05) to correct for multiple comparisons across time points.
To test whether the effect of semantic centrality on event-by-event recall success (1 = recalled, 0=not recalled) was mediated via the hippocampal event offset responses, we performed a mediation analysis. Each event from each participant served as a data point, and data were concatenated across all participants. For each participant, the hippocampal offset response of each event was computed by averaging the BOLD time series measured from 10 to 13 TRs after the event offset. Again, the responses were baseline corrected for each event by subtracting the mean response of the two TRs immediately preceding the event offset from the time series. The first/last events of each movie and not recalled events were excluded from the analysis. Three mixed-effects linear or logistic regression models were defined to test 1) the total effect of semantic centrality on recall success (logistic), 2) the effect of semantic centrality on hippocampal offset responses (linear), and 3) the direct effect of semantic centrality on recall success, controlling for hippocampal offset responses (logistic). An additional mixed-effects logistic regression analysis was also performed to test the effect of hippocampal offset responses on recall success. In all models, participants were included as random effects. The significance of the indirect effect of hippocampal offset responses on the relationship between semantic centrality and recall success was tested via the quasi-Bayesian Monte Carlo simulation as implemented in the “mediation” package in R. Specifically, 1000 simulations were performed to compute the 95% confidence interval of the average causal mediation effects.
Intersubject functional connectivity analysis
We performed intersubject functional connectivity analysis (ISFC32) to test the relationship between narrative network centrality and the hippocampus-cortex interaction during movie watching. We first averaged the TR-by-TR time courses of the preprocessed (non-neuronal signals removed; see fMRI preprocessing) functional data across all voxels/vertices within each of the hippocampal and cortical ROIs (PMC, early visual cortex). For each movie event as long as 22.5 s or longer (total number of events used across all movies = 26), we computed the ISFC between the hippocampus and a cortical ROI. Functional connectivity patterns computed within windows as short as 22.5 s have previously been shown to robustly predict cognitive states40. For each participant, we correlated the participant’s hippocampal time series of the event and the cortical ROI time series averaged across all other participants. We then averaged the Pearson correlation coefficients across all participants. This procedure was repeated by correlating each participant’s cortical ROI time series and the hippocampal time series averaged across all other participants. Again, the correlation coefficients were averaged across participants. We then took the mean of the two averaged correlations to produce a single ISFC between the hippocampus and the cortical ROI for each event. Finally, we computed the Pearson correlation between the event-wise ISFC and the semantic/causal narrative network centrality.
Low-level sensory characteristics of movie events
We measured the low-level visual and auditory features of movie events to examine whether the sensory characteristics can explain the effects of centrality on neural responses during movie watching. For visual features, we measured luminance and contrast averaged across grayscale-converted movie frames within each event. In each frame, luminance was defined as the mean of pixel values, and contrast was defined as the difference between the maximum and minimum pixel values. For an auditory feature, we measured the mean amplitude of sounds played during each event. We extracted the single-channel downsampled (8000 Hz) version of audio signals from the movie clips. Within each event, the audio signals were divided into 100-ms segments, and each segment’s amplitude was computed as the difference between the maximum and minimum signal intensities. The amplitudes were then averaged across all segments within each event. The first events of each movie were excluded from the analysis to be consistent with the movie phase fMRI analyses. All sensory features were z-scored across events. We performed mixed-effects linear regression analyses to test whether each of the event-wise low-level sensory features was modulated by semantic centrality, using semantic centrality as a fixed effect and movies as a random effect. Statistical significance of the effect of centrality was tested via likelihood ratio tests for the full models against the null models including the random effect of movies only.
Preregistered online experiment
We conducted an online experiment to replicate and generalize the behavioral results of the fMRI experiment with a larger number of participants and a new set of movie stimuli (Supplementary Fig. 7). The online experiment was preregistered at AsPredicted (https://aspredicted.org/fw59g.pdf; Preregistration date: July 15, 2019). We recruited a total of 393 participants (194 female, 198 male, 1 other) on Amazon’s Mechanical Turk using the psiTurk system85. The final sample size was slightly larger than the preregistered sample size (N = 380) because one extra batch of participants was recruited due to a technical issue. Participants were aged between 18 and 71 years (mean age 38.2 years), excluding two participants who failed to report their ages. Each participant watched one of 10 short movies and then provided a written recall of the movie plot. The audiovisual movie stimuli were different from the ones used in the fMRI experiment and included both animations and live-action movies. The movies were on average 9.1 min long (ranged 5.9 to 12.7 min). Each movie was watched by 38–49 participants (mean = 39.3). Additional 99 participants were excluded from the analysis because their recall was too short (<150 words) or they had watched the movie before the experiment. All participants were provided with an informed consent form approved by the Johns Hopkins University Institutional Review Board. Participants received monetary compensation for their time ($10 per hour).
The experiment was run in web browsers using JavaScript. After reading instructions, participants watched a 2-min-long example video clip. An example recall of the example video clip was also provided to inform participants about the level of detail they need to produce during recall. Participants then watched a short movie and performed the written recall task by typing in a text box to retell the movie plot in their own words. To mimic the irreversible nature of the spoken recall used in the fMRI experiment, participants were not allowed to backspace beyond the current sentence in order to edit already-written sentences. Participants were encouraged to take as much time as needed to provide as much detail as they can remember. During the delay between movie watching and recall, participants completed a short demographic questionnaire and the Mind-Wandering Questionnaire86, and then practiced using the text box by providing written descriptions of simple shapes. Participants also completed a series of questionnaires at the end of the experiment, including the Survey of Autobiographical Memory87 and the Plymouth Sensory Imagery Questionnaire88. Findings from the delay period and post-experiment questionnaires are not reported in this paper.
As in the fMRI experiment, independent coders segmented each movie into discrete events (mean number of events per movie = 25.2) and provided written descriptions of each event. The written recall of each participant was also segmented into sentences, and the movie events that each sentence describes were identified. The semantic and causal narrative networks of the movies were generated using procedures identical to those used in the fMRI experiment, except that 1) the USE vectors were not averaged across annotators as each movie was annotated by a single coder, and 2) a total of 16 independent coders identified the causally related events and each movie was rated by 10 coders.
Testing statistical assumptions
To validate the normality assumption for paired t-tests comparing the recall probability or mean ROI activation for high vs. low semantic/causal centrality events, we performed Anderson-Darling tests on the differences between conditions. The distributions of the differences did not significantly deviate from a normal distribution (ps > .2). For the repeated-measures one-way ANOVAs testing the serial position effects in recall performance, we performed Mauchly’s tests of sphericity and confirmed that the assumption of sphericity was not violated (ps > .44). We also performed Anderson-Darling tests and confirmed that the distribution of recall probability for each condition did not significantly deviate from a normal distribution (ps > .17).
Citation diversity statement
Recent work in neuroscience and several other fields of science has identified a citation bias whereby papers from people from under-represented groups are under-cited89–93. Here we aimed to proactively cite references that reflect the diversity of the field in gender, race, and ethnicity. First, we predicted the gender of the first and last author of each reference based on the first names of the authors using online databases93,94. Excluding self-citations, the references of the current article contain 8.86% woman(first)/woman(last), 15.53% man/woman, 25.66% woman/man, and 49.95% man/man. Note that this method may not always be indicative of gender identity, and cannot account for intersex, non-binary, or transgender individuals. Similarity, we predicted the race of the first and last author of each reference based on the first and last names of the authors using online databases95,96. Excluding self-citations, the references of the current article contain 5.24% author of color (first)/author of color(last), 12.87% White author/author of color, 16.34% author of color/White author, and 65.55% White author/White author. Note that this method may not always be indicative of racial or ethnic identity, and cannot account for Indigenous and multiracial individuals, or those who may face differential biases due to the ambiguous racialization or ethnicization of their names.
Reporting summary
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
Supplementary information
Acknowledgements
We thank Kenneth A. Norman, Uri Hasson, Christopher J. Honey, Qihong Lu, and Yuan Chang Leong for scientific discussions and comments on earlier versions of the manuscript. We thank Savannah Born and Elly Yeom for their assistance with collecting behavioral ratings. We thank Buddhika Bellana and Yoonjung Lee for their assistance with online experiment preparation and data analysis. This work was supported by the Sloan Research Fellowship to J.C. (FG-2018-10490) and Google Faculty Research Award to J.C. (Google LLC 3-9-2020).
Source data
Author contributions
H.L. and J.C. conceived and designed the experiments. H.L. and J.C. performed the experiments. H.L. analyzed the data. H.L. drafted the paper. H.L. and J.C. edited the paper.
Peer review
Peer review information
Nature Communications thanks Alex Barnett, Jeremy Manning, Charan Ranganath and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. Peer reviewer reports are available.
Data availability
Source data associated with the figures are provided with this paper. The raw fMRI and behavioral data generated in this study have been deposited in OpenNeuro.org (10.18112/openneuro.ds004042.v1.0.0)97. The region-of-interest labels, activation maps from univariate analysis, movie annotations, and raw behavioural data from the preregistered online experiment are available at GitHub (https://github.com/jchenlab-jhu/filmfest)98. Source data are provided with this paper.
Code availability
The analyses in the current manuscript used code available through MATLAB, R, and Python. Custom scripts that can be used to reproduce the figures from the source data are included in the Source Data file. Other analysis scripts are available upon request to the corresponding author (H.L.).
Competing interests
The authors declare no competing interests.
Footnotes
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
The online version contains supplementary material available at 10.1038/s41467-022-31965-2.
References
- 1.Bird CM. How do we remember events? Curr. Opin. Behav. Sci. 2020;32:120–125. doi: 10.1016/j.cobeha.2020.01.020. [DOI] [Google Scholar]
- 2.Zacks JM. Event perception and memory. Annu. Rev. Psychol. 2020;71:165–191. doi: 10.1146/annurev-psych-010419-051101. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Murdock BB., Jr. The serial position effect of free recall. J. Exp. Psychol. 1962;64:482–488. doi: 10.1037/h0045106. [DOI] [Google Scholar]
- 4.Polyn SM, Natu VS, Cohen JD, Norman KA. Category-specific cortical activity precedes retrieval during memory search. Science. 2005;310:1963–1966. doi: 10.1126/science.1117645. [DOI] [PubMed] [Google Scholar]
- 5.Reiser BJ, Black JB, Abelson RP. Knowledge structures in the organization and retrieval of autobiographical memories. Cogn. Psychol. 1985;17:89–137. doi: 10.1016/0010-0285(85)90005-2. [DOI] [Google Scholar]
- 6.Schank, R. C. & Abelson, R. P. Scripts, Plans, Goals, and Understanding: An Inquiry Into Human Knowledge Structures. (Erlbaum, 1977).
- 7.Lee H, Bellana B, Chen J. What can narratives tell us about the neural bases of human memory? Curr. Opin. Behav. Sci. 2020;32:111–119. doi: 10.1016/j.cobeha.2020.02.007. [DOI] [Google Scholar]
- 8.Willems RM, Nastase SA, Milivojevic B. Narratives for neuroscience. Trends Neurosci. 2020;43:271–273. doi: 10.1016/j.tins.2020.03.003. [DOI] [PubMed] [Google Scholar]
- 9.Clewett D, DuBrow S, Davachi L. Transcending time in the brain: How event memories are constructed from experience. Hippocampus. 2019;29:162–183. doi: 10.1002/hipo.23074. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Shin, Y. S. & DuBrow, S. Structuring Memory Through Inference‐Based Event Segmentation. Top. Cogn. Sci. 10.1111/tops.12505 (2020). [DOI] [PubMed]
- 11.Radvansky GA, Zacks JM. Event perception. WIREs Cogn. Sci. 2011;2:608–620. doi: 10.1002/wcs.133. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Ben-Yakov A, Dudai Y. Constructing realistic engrams: Poststimulus activity of hippocampus and dorsal striatum predicts subsequent episodic memory. J. Neurosci. 2011;31:9032–9042. doi: 10.1523/JNEUROSCI.0702-11.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Ben-Yakov A, Henson RN. The hippocampal film editor: Sensitivity and specificity to event boundaries in continuous experience. J. Neurosci. 2018;38:10057–10068. doi: 10.1523/JNEUROSCI.0524-18.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Buckner RL, DiNicola LM. The brain’s default network: Updated anatomy, physiology and evolving insights. Nat. Rev. Neurosci. 2019;20:593–608. doi: 10.1038/s41583-019-0212-7. [DOI] [PubMed] [Google Scholar]
- 15.Baldassano C, et al. Discovering event structure in continuous narrative perception and memory. Neuron. 2017;95:709–721.e5. doi: 10.1016/j.neuron.2017.06.041. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Ritchey M, Cooper RA. Deconstructing the Posterior Medial Episodic Network. Trends Cogn. Sci. 2020;24:451–465. doi: 10.1016/j.tics.2020.03.006. [DOI] [PubMed] [Google Scholar]
- 17.Chen J, et al. Shared memories reveal shared structure in neural activity across individuals. Nat. Neurosci. 2017;20:115–125. doi: 10.1038/nn.4450. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Zadbood A, Chen J, Leong YC, Norman KA, Hasson U. How we transmit memories to other brains: constructing shared neural representations via communication. Cereb. Cortex. 2017;27:4988–5000. doi: 10.1093/cercor/bhx202. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Schlichting ML, Preston AR. Memory integration: neural mechanisms and implications for behavior. Curr. Opin. Behav. Sci. 2015;1:1–8. doi: 10.1016/j.cobeha.2014.07.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Shohamy D, Wagner AD. Integrating memories in the human brain: Hippocampal-midbrain encoding of overlapping events. Neuron. 2008;60:378–389. doi: 10.1016/j.neuron.2008.09.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Black JB, Bower GH. Story understanding as problem-solving. Poetics. 1980;9:223–250. doi: 10.1016/0304-422X(80)90021-2. [DOI] [Google Scholar]
- 22.Trabasso T, van den Broek P. Causal thinking and the representation of narrative events. J. Mem. Lang. 1985;24:612–630. doi: 10.1016/0749-596X(85)90049-X. [DOI] [Google Scholar]
- 23.Black JB, Bern H. Causal coherence and memory for events in narratives. J. Verbal Learn. Verbal Behav. 1981;20:267–275. doi: 10.1016/S0022-5371(81)90417-5. [DOI] [Google Scholar]
- 24.Trabasso T, Sperry LL. Causal relatedness and importance of story events. J. Mem. Lang. 1985;24:595–611. doi: 10.1016/0749-596X(85)90048-8. [DOI] [Google Scholar]
- 25.Cer, D. et al. Universal Sentence Encoder. arXiv:1803.11175v2 [cs.CL] (2018).
- 26.Heusser AC, Fitzpatrick PC, Manning JR. Geometric models reveal behavioural and neural signatures of transforming experiences into memories. Nat. Hum. Behav. 2021;5:905–919. doi: 10.1038/s41562-021-01051-6. [DOI] [PubMed] [Google Scholar]
- 27.Lydon-Staley DM, Zhou D, Blevins AS, Zurn P, Bassett DS. Hunters, busybodies and the knowledge network building associated with deprivation curiosity. Nat. Hum. Behav. 2021;5:327–336. doi: 10.1038/s41562-020-00985-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Kahana MJ. Computational models of memory search. Annu. Rev. Psychol. 2020;71:107–138. doi: 10.1146/annurev-psych-010418-103358. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Zacks JM, et al. Human brain activity time-locked to perceptual event boundaries. Nat. Neurosci. 2001;4:651–655. doi: 10.1038/88486. [DOI] [PubMed] [Google Scholar]
- 30.Rugg MD, Vilberg KL. Brain networks underlying episodic memory retrieval. Curr. Opin. Neurobiol. 2013;23:255–260. doi: 10.1016/j.conb.2012.11.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Nastase SA, Gazzola V, Hasson U, Keysers C. Measuring shared responses across subjects using intersubject correlation. Soc. Cogn. Affect. Neurosci. 2019;14:667–685. doi: 10.1093/scan/nsz037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Simony E, et al. Dynamic reconfiguration of the default mode network during narrative comprehension. Nat. Commun. 2016;7:12141. doi: 10.1038/ncomms12141. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Schaefer A, et al. Local-global parcellation of the human cerebral cortex from intrinsic functional connectivity MRI. Cereb. Cortex. 2018;28:3095–3114. doi: 10.1093/cercor/bhx179. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Kriegeskorte, N., Mur, M. & Bandettini, P. A. Representational similarity analysis - connecting the branches of systems neuroscience. Front. Syst. Neurosci. 10.3389/neuro.06.004.2008 (2008). [DOI] [PMC free article] [PubMed]
- 35.Ben-Yakov A, Eshel N, Dudai Y. Hippocampal immediate poststimulus activity in the encoding of consecutive naturalistic episodes. J. Exp. Psychol. Gen. 2013;142:1255–1263. doi: 10.1037/a0033558. [DOI] [PubMed] [Google Scholar]
- 36.Reagh ZM, Delarazan AI, Garber A, Ranganath C. Aging alters neural activity at event boundaries in the hippocampus and Posterior Medial network. Nat. Commun. 2020;11:3980. doi: 10.1038/s41467-020-17713-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Chen J, et al. Accessing real-life episodic information from minutes versus hours earlier modulates hippocampal and high-order cortical dynamics. Cereb. Cortex. 2016;26:3428–3441. doi: 10.1093/cercor/bhv155. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Aly M, Chen J, Turk-Browne NB, Hasson U. Learning naturalistic temporal structure in the posterior medial network. J. Cogn. Neurosci. 2018;30:1345–1365. doi: 10.1162/jocn_a_01308. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Ranganath C, Heller A, Cohen MX, Brozinsky CJ, Rissman J. Functional connectivity with the hippocampus during successful memory formation. Hippocampus. 2005;15:997–1005. doi: 10.1002/hipo.20141. [DOI] [PubMed] [Google Scholar]
- 40.Gonzalez-Castillo J, et al. Tracking ongoing cognition in individuals using brief, whole-brain functional connectivity patterns. Proc. Natl Acad. Sci. 2015;112:8762–8767. doi: 10.1073/pnas.1501242112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Zou GY. Toward using confidence intervals to compare correlations. Psychol. Methods. 2007;12:399–413. doi: 10.1037/1082-989X.12.4.399. [DOI] [PubMed] [Google Scholar]
- 42.Chang CHC, Lazaridi C, Yeshurun Y, Norman KA, Hasson U. Relating the past with the present: Information integration and segregation during ongoing narrative processing. J. Cogn. Neurosci. 2021;33:1106–1128. doi: 10.1162/jocn_a_01707. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Yeshurun Y, Nguyen M, Hasson U. The default mode network: where the idiosyncratic self meets the shared social world. Nat. Rev. Neurosci. 2021;22:181–192. doi: 10.1038/s41583-020-00420-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Romney AK, Brewer DD, Batchelder WH. Predicting clustering from semantic structure. Psychol. Sci. 1993;4:28–34. doi: 10.1111/j.1467-9280.1993.tb00552.x. [DOI] [Google Scholar]
- 45.Diamond NB, Levine B. Linking detail to temporal structure in naturalistic-event recall. Psychol. Sci. 2020;31:1557–1572. doi: 10.1177/0956797620958651. [DOI] [PubMed] [Google Scholar]
- 46.Kauttonen J, Hlushchuk Y, Jääskeläinen IP, Tikka P. Brain mechanisms underlying cue-based memorizing during free viewing of movie Memento. NeuroImage. 2018;172:313–325. doi: 10.1016/j.neuroimage.2018.01.068. [DOI] [PubMed] [Google Scholar]
- 47.Cohn-Sheehy BI, et al. The hippocampus constructs narrative memories across distant events. Curr. Biol. 2021;31:4935–4945.e7. doi: 10.1016/j.cub.2021.09.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Ranganath C, Ritchey M. Two cortical systems for memory-guided behaviour. Nat. Rev. Neurosci. 2012;13:713–726. doi: 10.1038/nrn3338. [DOI] [PubMed] [Google Scholar]
- 49.Collin SHP, Milivojevic B, Doeller CF. Memory hierarchies map onto the hippocampal long axis in humans. Nat. Neurosci. 2015;18:1562–1564. doi: 10.1038/nn.4138. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Brandman T, Malach R, Simony E. The surprising role of the default mode network in naturalistic perception. Commun. Biol. 2021;4:1–9. doi: 10.1038/s42003-020-01602-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Price AR, Bonner MF, Peelle JE, Grossman M. Converging evidence for the neuroanatomic basis of combinatorial semantics in the angular gyrus. J. Neurosci. 2015;35:3276–3284. doi: 10.1523/JNEUROSCI.3446-14.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Keidel JL, Oedekoven CSH, Tut AC, Bird CM. Multiscale integration of contextual information during a naturalistic task. Cereb. Cortex. 2018;28:3531–3539. doi: 10.1093/cercor/bhx218. [DOI] [PubMed] [Google Scholar]
- 53.Branzi FM, Pobric G, Jung J, Lambon Ralph MA. The left angular gyrus is causally involved in context-dependent integration and associative encoding during narrative reading. J. Cogn. Neurosci. 2021;33:1082–1095. doi: 10.1162/jocn_a_01698. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Lee H, Samide R, Richter FR, Kuhl BA. Decomposing parietal memory reactivation to predict consequences of remembering. Cereb. Cortex. 2018;29:3305–3318. doi: 10.1093/cercor/bhy200. [DOI] [PubMed] [Google Scholar]
- 55.Finn ES, Corlett PR, Chen G, Bandettini PA, Constable T. Trait paranoia shapes inter-subject synchrony in brain activity during an ambiguous social narrative. Nat. Commun. 2018;9:2043. doi: 10.1038/s41467-018-04387-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Parkinson C, Kleinbaum AM, Wheatley T. Similar neural responses predict friendship. Nat. Commun. 2018;9:332. doi: 10.1038/s41467-017-02722-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Yeshurun Y, et al. Same story, different story: The neural representation of interpretive frameworks. Psychol. Sci. 2017;28:307–319. doi: 10.1177/0956797616682029. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Baldassano C, Hasson U, Norman KA. Representation of real-world event schemas during narrative perception. J. Neurosci. 2018;38:9689–9699. doi: 10.1523/JNEUROSCI.0251-18.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Dixon ML, Fox KCR, Christoff K. A framework for understanding the relationship between externally and internally directed cognition. Neuropsychologia. 2014;62:321–330. doi: 10.1016/j.neuropsychologia.2014.05.024. [DOI] [PubMed] [Google Scholar]
- 60.Honey CJ, Newman EL, Schapiro AC. Switching between internal and external modes: A multiscale learning principle. Netw. Neurosci. 2017;1:339–356. doi: 10.1162/NETN_a_00024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Duncan K, Sadanand A, Davachi L. Memory’s penumbra: Episodic memory decisions induce lingering mnemonic biases. Science. 2012;337:485–487. doi: 10.1126/science.1221936. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Tarder-Stoll H, Jayakumar M, Dimsdale-Zucker HR, Günseli E, Aly M. Dynamic internal states shape memory retrieval. Neuropsychologia. 2020;138:107328. doi: 10.1016/j.neuropsychologia.2019.107328. [DOI] [PubMed] [Google Scholar]
- 63.Song H, Park B, Park H, Shim WM. Cognitive and neural state dynamics of narrative comprehension. J. Neurosci. 2021;41:8972–8990. doi: 10.1523/JNEUROSCI.0037-21.2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Mandler JM, Johnson NS. Remembrance of things parsed: Story structure and recall. Cogn. Psychol. 1977;9:111–151. doi: 10.1016/0010-0285(77)90006-8. [DOI] [Google Scholar]
- 65.Whaley JF. Readers’ expectations for story structures. Read. Res. Q. 1981;17:90–114. doi: 10.2307/747250. [DOI] [Google Scholar]
- 66.Cutting JE. Narrative theory and the dynamics of popular movies. Psychon. Bull. Rev. 2016;23:1713–1743. doi: 10.3758/s13423-016-1051-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Schapiro AC, Rogers TT, Cordova NI, Turk-Browne NB, Botvinick MM. Neural representations of events arise from temporal community structure. Nat. Neurosci. 2013;16:486–492. doi: 10.1038/nn.3331. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Kahn AE, Karuza EA, Vettel JM, Bassett DS. Network constraints on learnability of probabilistic motor sequences. Nat. Hum. Behav. 2018;2:936–947. doi: 10.1038/s41562-018-0463-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Kintsch W. Learning from text. Cogn. Instr. 1986;3:87–108. doi: 10.1207/s1532690xci0302_1. [DOI] [Google Scholar]
- 70.Kintsch W, van Dijk TA. Toward a model of text comprehension and production. Psychol. Rev. 1978;85:363–394. doi: 10.1037/0033-295X.85.5.363. [DOI] [Google Scholar]
- 71.van Dijk, T. & Kintsch, W. Strategies of discourse comprehension. (New York: Academic Press, 1983).
- 72.Zwaan RA, Magliano JP, Graesser AC. Dimensions of situation model construction in narrative comprehension. J. Exp. Psychol. Learn. Mem. Cogn. 1995;21:386–397. doi: 10.1037/0278-7393.21.2.386. [DOI] [Google Scholar]
- 73.Radvansky GA. Across the event horizon. Curr. Dir. Psychol. Sci. 2012;21:269–272. doi: 10.1177/0963721412451274. [DOI] [Google Scholar]
- 74.Gopnik A, et al. A theory of causal learning in children: Causal maps and Bayes nets. Psychol. Rev. 2004;111:3–32. doi: 10.1037/0033-295X.111.1.3. [DOI] [PubMed] [Google Scholar]
- 75.Waldmann MR, Hagmayer Y, Blaisdell AP. Beyond the information given: Causal models in learning and reasoning. Curr. Dir. Psychol. Sci. 2006;15:307–311. doi: 10.1111/j.1467-8721.2006.00458.x. [DOI] [Google Scholar]
- 76.Long NM, Danoff MS, Kahana MJ. Recall dynamics reveal the retrieval of emotional context. Psychon. Bull. Rev. 2015;22:1328–1333. doi: 10.3758/s13423-014-0791-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Tomita TM, Barense MD, Honey CJ. The similarity structure of real-world memories. bioRxiv. 2021 doi: 10.1101/2021.01.28.428278. [DOI] [Google Scholar]
- 78.McAndrews MP, Cohn M, Gold DA. Infusing cognitive neuroscience into the clinical neuropsychology of memory. Curr. Opin. Behav. Sci. 2020;32:94–101. doi: 10.1016/j.cobeha.2020.01.011. [DOI] [Google Scholar]
- 79.Davidesco I, et al. Brain-to-brain synchrony between students and teachers predicts learning outcomes. bioRxiv. 2019 doi: 10.1101/644047. [DOI] [PubMed] [Google Scholar]
- 80.Dikker S, et al. Brain-to-brain synchrony tracks real-world dynamic group Interactions in the classroom. Curr. Biol. 2017;27:1375–1380. doi: 10.1016/j.cub.2017.04.002. [DOI] [PubMed] [Google Scholar]
- 81.Meshulam M, et al. Neural alignment predicts learning outcomes in students taking an introduction to computer science course. Nat. Commun. 2021;12:1922. doi: 10.1038/s41467-021-22202-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.McClelland JL, Hill F, Rudolph M, Baldridge J, Schütze H. Placing language in an integrated understanding system: Next steps toward human-level performance in neural language models. Proc. Natl Acad. Sci. 2020;117:25966–25974. doi: 10.1073/pnas.1910416117. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Lu Q, Hasson U, Norman KA. Learning to use episodic memory for event prediction. bioRxiv. 2020 doi: 10.1101/2020.12.15.422882. [DOI] [Google Scholar]
- 84.Yeo BTT, et al. The organization of the human cerebral cortex estimated by intrinsic functional connectivity. J. Neurophysiol. 2011;106:1125–1165. doi: 10.1152/jn.00338.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.McDonnell, J. et al. psiTurk (Version 2.1.2)[Software]. New York, NY: New York University. Available fromhttps://github.com/NYUCCL/psiTurk. (2014).
- 86.Mrazek, M. D., Phillips, D. T., Franklin, M. S., Broadway, J. M. & Schooler, J. W. Young and restless: validation of the Mind-Wandering Questionnaire (MWQ) reveals disruptive impact of mind-wandering for youth. Front. Psychol. 10.3389/fpsyg.2013.00560 (2013). [DOI] [PMC free article] [PubMed]
- 87.Palombo DJ, Williams LJ, Abdi H, Levine B. The survey of autobiographical memory (SAM): A novel measure of trait mnemonics in everyday life. Cortex. 2013;49:1526–1540. doi: 10.1016/j.cortex.2012.08.023. [DOI] [PubMed] [Google Scholar]
- 88.Andrade J, May J, Deeprose C, Baugh S-J, Ganis G. Assessing vividness of mental imagery: The Plymouth Sensory Imagery Questionnaire. Br. J. Psychol. 2014;105:547–563. doi: 10.1111/bjop.12050. [DOI] [PubMed] [Google Scholar]
- 89.Mitchell SM, Lange S, Brus H. Gendered citation patterns in international relations journals. Int. Stud. Perspect. 2013;14:485–492. doi: 10.1111/insp.12026. [DOI] [Google Scholar]
- 90.Maliniak D, Powers R, Walter BF. The gender citation gap in international relations. Int. Organ. 2013;67:889–922. doi: 10.1017/S0020818313000209. [DOI] [Google Scholar]
- 91.Caplar N, Tacchella S, Birrer S. Quantitative evaluation of gender bias in astronomical publications from citation counts. Nat. Astron. 2017;1:1–5. doi: 10.1038/s41550-016-0001. [DOI] [Google Scholar]
- 92.Dion ML, Sumner JL, Mitchell SM. Gendered citation patterns across political science and social science methodology fields. Polit. Anal. 2018;26:312–327. doi: 10.1017/pan.2018.12. [DOI] [Google Scholar]
- 93.Dworkin JD, et al. The extent and drivers of gender imbalance in neuroscience reference lists. Nat. Neurosci. 2020;23:918–926. doi: 10.1038/s41593-020-0658-y. [DOI] [PubMed] [Google Scholar]
- 94.Zhou, D. et al. Gender diversity statement and code notebook v1.0. (Zenodo, 2020). 10.5281/zenodo.3672110.
- 95.Ambekar, A., Ward, C., Mohammed, J., Male, S. & Skiena, S. Name-ethnicity classification from open sources. in Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining 49–58 (Association for Computing Machinery, 2009). 10.1145/1557019.1557032.
- 96.Sood, G. & Laohaprapanon, S. Predicting race and ethnicity from the sequence of characters in a name. ArXiv180502109 Stat (2018).
- 97.Lee, H., Chen, J., & Hasson, U. FilmFestival. OpenNeuro.10.18112/openneuro.ds004042.v1.0.0 (2022).
- 98.Lee, H., & Chen, J. Predicting memory from the network structure of naturalistic events. Zenodo.10.5281/zenodo.6574792 (2022). [DOI] [PMC free article] [PubMed]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Source data associated with the figures are provided with this paper. The raw fMRI and behavioral data generated in this study have been deposited in OpenNeuro.org (10.18112/openneuro.ds004042.v1.0.0)97. The region-of-interest labels, activation maps from univariate analysis, movie annotations, and raw behavioural data from the preregistered online experiment are available at GitHub (https://github.com/jchenlab-jhu/filmfest)98. Source data are provided with this paper.
The analyses in the current manuscript used code available through MATLAB, R, and Python. Custom scripts that can be used to reproduce the figures from the source data are included in the Source Data file. Other analysis scripts are available upon request to the corresponding author (H.L.).