Abstract
Semantic knowledge includes understanding of objects and their features and also understanding of the characteristics of events. The hub-and-spoke theory holds that these conceptual representations rely on multiple information sources that are integrated in a central hub in the ventral anterior temporal lobes. The dual-hub theory expands this framework with the claim that the ventral anterior temporal lobe hub is specialized for object representation, while a second hub in angular gyrus is specialized for event representation. To test these ideas, we used representational similarity analysis, univariate and psychophysiological interaction analyses of fMRI data collected while participants processed object and event concepts (e.g. “an apple,” “a wedding”) presented as images and written words. Representational similarity analysis showed that angular gyrus encoded event concept similarity more than object similarity, although the left angular gyrus also encoded object similarity. Bilateral ventral anterior temporal lobes encoded both object and event concept structure, and left ventral anterior temporal lobe exhibited stronger coding for events. Psychophysiological interaction analysis revealed greater connectivity between left ventral anterior temporal lobe and right pMTG, and between right angular gyrus and bilateral ITG and middle occipital gyrus, for event concepts compared to object concepts. These findings support the specialization of angular gyrus for event semantics, though with some involvement in object coding, but do not support ventral anterior temporal lobe specialization for object concepts.
Keywords: event semantics, object semantics, anterior temporal lobe, angular gyrus, semantic representation
Introduction
Humans can recognize and reason about single objects, and we can also understand events as coherent conceptual units—complex, context-bound interactions between objects that unfold over time. Object similarity can be captured by shared features, whereas events involve multiple objects’ interactions, temporal sequences, and causal relationships (Altmann and Ekves 2019). A core function of the semantic system is to represent similarities between concepts of various types. For example, apples are more similar to tomatoes than to hammers, and weddings are more similar to parties than to fights. The neural coding of object similarity has been studied in depth (Devereux et al. 2013; Hutchison et al. 2014; Kaneshiro et al. 2015; Bi et al. 2016; Chen et al. 2016; Xu et al. 2018). Event structure coding is also investigated by some studies (Bedny et al. 2014; Baldassano et al. 2017; Morton et al. 2020). However, object representation and event representation are rarely compared directly, meaning that differences in their neural bases remain unclear. Thus, in the present study, we used representational similarity analysis (RSA), univariate fMRI analyses, and functional connectivity analyses to directly compare how the semantic structures of objects and events are represented in the brain.
Vision is crucial for identifying objects and events; thus, specializations for object and event understanding could be driven by the organization of the visual system into dorsal and ventral pathways (Mirman et al. 2017). The dorsal pathway usually refers to the processing stream that lies between early visual cortex and frontal-parietal regions specialized for action, and which courses through temporal-parietal cortex (Mishkin et al. 1983; Kravitz et al. 2013). The dorsal stream is identified as a “where/how” pathway, supporting visually guided action, motion and spatial cognition (Wager and Smith 2003; Husain and Nachev 2007; Andersen and Cui 2009; Buxbaum and Kalénine 2010; Watson and Chatterjee 2011). The dorsal stream may be particularly important for event representation, as this requires processing of objects’ interactions and their spatiotemporal relations. Conversely, the ventral pathway lies between early visual cortex and the ventral anterior temporal lobe (vATL), and courses through the inferior parts of the temporal lobe (Mishkin et al. 1983; Kravitz et al. 2013). This stream is characterized as a “what” pathway, specialized for identifying and categorizing objects. In line with this view, ventral pathway regions are engaged in processing and integrating perceptual features like color, size, and brightness (Baron et al. 2010; Coutanche and Thompson-Schill 2015; Martin et al. 2018). Regions in this pathway show category-selective effects for different object categories like tools, animals and human faces (Hutchison et al. 2014; Bi et al. 2016).
As the junction of the ventral pathway with other processing streams, vATL is thought to act as a transmodal semantic hub that combines visual features with multimodal information sources to generate conceptual representations (for a review, see Lambon Ralph et al. 2017). The ATLs are strongly associated with integrating object features across sensory modalities (Rogers and McClelland 2004; Coutanche and Thompson-Schill 2015), and are engaged in semantic processing irrespective of input modality (e.g. words, pictures and sounds) (Vandenberghe et al. 1996; Marinkovic et al. 2003; Binney et al. 2010; Visser and Lambon Ralph 2011) and across a range of conceptual categories (Hoffman et al. 2015; Rice et al. 2018; Wang et al. 2019; Conca et al. 2021).
Studies using multivariate pattern analysis indicate that ATL regions code semantic relationships between objects (Peelen and Caramazza 2012; Fairhall and Caramazza 2013; Chen et al. 2016; Rogers et al. 2021). For example, in an iEEG study using a picture-naming task, Chen et al. (2016) observed that vATL activity patterns were predicted by semantic similarity between objects, even after controlling for visual and phonological features of the stimuli. The medial part of vATL, the perirhinal cortex, has been implicated specifically in recognizing objects and in differentiating between objects that have many overlapping semantic features (for review, see Clarke and Tyler 2015). Perirhinal cortex activation increases when participants recognize semantically more-confusable objects (Tyler et al. 2013; Clarke and Tyler 2014) and damage to this region results in deficits for naming semantically more-confusable objects (Wright et al. 2015). RSA analyses of fMRI data indicate that more similar objects elicit more similar patterns of activation in the perirhinal cortex (Bruffaerts et al. 2013; Liuzzi et al. 2015; Devereux et al. 2018; Naspi et al. 2021). For example, Liuzzi et al. (2015) presented people with written object names, and found that in left perirhinal cortex, activation pattern similarity was predicted by semantic similarity between objects (measured in terms of their property overlap). However, while it is now well established that regions within vATL code semantic similarity between objects, it remains unclear whether this region also codes semantic similarities between events. Studies of event semantics have instead focused on regions within the temporoparietal cortex (TPC).
An association between TPC and event representation has been suggested by many researchers (for reviews, see Binder and Desai 2011; Mirman et al. 2017). Event representations require frequent processing of interactions or contextual associations (e.g. action, spatial, temporal information). This kind of processing may be well suited to TPC regions, which participate in, and receive inputs from, the dorsal visual stream. TPC regions have been implicated in the semantics of action and in representing thematic relationships between concepts. Posterior temporal lobe is involved in understanding action concepts (Kable et al. 2002; Kable et al. 2005; Bedny et al. 2014) and motion concepts (Noppeney et al. 2005; Gennari et al. 2007; Bedny et al. 2008; Watson et al. 2013; Zhang et al. 2022). The posterior parietal cortex is involved in action planning (for reviews, see Andersen and Cui 2009; Buxbaum and Kalénine 2010). Parietal regions within TPC are also important for integrating spatially distributed objects into a single coherent percept (Huberle and Karnath 2012; Lestou et al. 2014) and for making temporal order judgments (Davis et al. 2009). These roles in supporting the dynamic aspects of semantics make TPC particularly suited to representing interactions between objects. Indeed, based on neuropsychological and neuroimaging evidence, the dual-hub theory of semantic representation proposes that TPC is specialized for coding thematic/event-based semantic relations (e.g. dog-bone) and the ATL for taxonomic/similarity-based semantic relations (e.g. dog-cat) (Schwartz et al. 2011; Mirman et al. 2017). A recent fMRI meta-analysis provided support for this idea by revealing that TPC regions are reliably more activated by thematic than taxonomic relations (Zhang et al. 2023).
Within TPC, the angular gyrus (AG) in particular has been identified as a critical area for multiple functions relevant to event representation: autobiographical memory and episodic memory (Bonnici et al. 2018; Russell et al. 2019), retrieval of multimodal spatiotemporal memories (Yazar et al. 2014; Ben-Zvi et al. 2015; Bonnici et al. 2016; Richter et al. 2016; Yazar et al. 2017), and combinatorial semantics (e.g. computing the meanings of noun+noun and verb+noun phrases) (Boylan et al. 2015; Price et al. 2015). More broadly, AG is a key part of the default mode network (DMN), which is implicated in coding situation models of ongoing events and segmenting experiences into separate events (Zacks et al. 2010; Swallow et al. 2011; Ranganath and Ritchey 2012; Baldassano et al. 2017; Yeshurun et al. 2021; Morales et al. 2022). DMN appears to act as a dynamic network that combines incoming external information with internal information from prior experiences to create detailed, context-specific representations of situations as they develop over time (for reviews, see Ranganath and Ritchey 2012; Yeshurun et al. 2021). In line with these functions, DMN is sensitive to event boundaries in a continuous experience: stronger responses in DMN are observed when participants watch event changes in movies or listen to event changes in narratives (Zacks et al. 2010; Swallow et al. 2011; Baldassano et al. 2017). These various lines of evidence implicate AG in event processing, supporting the idea that this region may act as a semantic hub for event knowledge. If this is the case, it should represent semantic similarities between abstract event concepts (e.g. wedding-party), and it should code event similarities more strongly than object similarities. These predictions have not previously been tested directly.
In summary, vATL has emerged as a representational hub for various aspects of semantic knowledge, and is known to play an important role in coding similarity-based relationships between individual concepts. It is not clear whether this role extends to coding semantic relationships between more complex event concepts. In contrast, AG has been proposed to be a semantic hub that specializes for representing event-based knowledge, by integrating contextual information, interactions, and associations between objects. However, while numerous studies have investigated how this region responds to processing temporally-extended events (e.g. movies or narratives; Zacks et al. 2010; Swallow et al. 2011; Bonnici et al. 2016; Baldassano et al. 2017), it is less clear to what extent this region represents more abstract event concepts, or whether it represents these in preference to object concepts. More generally, the regions involved in representing semantic relations for objects and events have rarely been directly compared.
To address these questions, we used fMRI to scan participants when they were presented with event and object concepts (as written words and still images), then conducted RSA to test whether neural patterns reflected semantic similarity within either set of concepts. We particularly focused on representation similarity effects in vATL and AG, since these have been proposed as core semantic hubs for objects and events respectively. We analyzed left and right vATLs and AGs. Many studies have assumed semantic representations are left-lateralized and have not tested effects in right-hemisphere regions. Here we included both hemispheres, to determine whether effects are specific to the left hemisphere. In addition, univariate analysis was conducted to test general activation differences to event and object concepts. Finally, psychophysiological interaction (PPI) analyses were performed to explore whether, when processing event and object concepts, semantic hubs have different connective patterns with other areas.
Method
Participants
We recruited 43 healthy participants (31 females, 12 males; mean age = 23.07 years, s.d. = 3.23 years, range = 19–32). All participants were right-handed native English speakers, and no-one reported history of dyslexia or other neurological disorders. The study was approved by University of Edinburgh School of Philosophy, Psychology & Language Sciences Research Ethics Committee and all participants gave written informed consent.
Materials
We presented participants with 60 different concepts, each of which was represented by four different pictures (240 pictures in total; see Fig. 1A for examples). 30 of these were event concepts, while the other 30 were object concepts. The list of all concepts can be found in Supplementary Table 1. Object concepts referred to individual entities, and we sampled from a variety of categories: animals (e.g. a dog), food (an apple), manipulable tools (a hammer), vehicles (a car), buildings (a castle), body parts (an arm), and human entities (a woman). Event concepts referred to situations in which multiple people or entities interact, including a range of social (e.g. a party), cultural (an opera), professional (a diagnosis), and everyday events (a picnic). In the experiment, each concept was presented four times, with the concept name shown each time with a different picture. We used images to elicit richer representations of the underlying concepts. In addition, by showing broader contexts and interactions, event pictures encouraged participants to process the situational aspects of these concepts. In contrast, object pictures included no background or interactions, encouraging people to process each object as an individual entity. In RSA analyses, we used the average neural responses across all four presentations of each concept. This ensured that the neural pattern for each concept represented general knowledge of the concept, rather than idiosyncratic features of one particular image.
Fig. 1.
Experimental design. (A) Examples of object and event stimuli. (B) Experimental procedure, showing one trial followed by a catch question. (C) Semantic similarities for event concepts (left) and object concepts (right).
Object and event stimuli differed in several ways, reflecting intrinsic differences between object and event concepts. Compared with object pictures, event pictures were more visually complex because they showed scenes containing multiple people and objects. Object words were more concrete than event words, according to the concreteness norms of Brysbaert et al. (2014) (t(58) = 9.95, P < 0.001), but less social, according to norms of Diveica et al. (2023) (t(30) = 5.31, P < 0.001). The two sets of items did not differ in word frequency values from Van Heuven et al. (2014) (t(58) = 0.2, P = 0.84), or in ratings of emotional valence provided by Warriner et al. (2013) (t(57) = 0.91, P = 0.365).
Given these differences, the main RSA analyses were conducted separately within each of the two sets of concepts. We also present univariate activation contrasts of the two conditions and analyses of functional connectivity but we note that effects in these analyses could arise from differences at various levels of processing (e.g. lower-level visual perceptual processes and demands on social cognition).
For RSA, we constructed four 30 × 30 representational dissimilarity matrices (RDMs) that captured the similarity structures within events and within objects (see Fig. 1C). For each set of concepts, we calculated a semantic RDM and a visual RDM. The semantic RDM was based on vector-based representations of word meaning, generated by training the word2vec neural network with the 100-billion-word Google news corpus (Mikolov et al. 2013). We defined dissimilarity between two concepts as one minus the cosine between their word2vec vectors. Although a number of vector-based models of word meaning are available, we used word2vec because these vectors show the best fit to human semantic relatedness judgments (Pereira et al. 2016). The visual RDM controlled for the low-level visual characteristics of the images we presented. A visual representation of each image was calculated by entering images into the Hmax computational model of vision and extracting the output on the C1 layer of the model, which represents low-level visual attributes (Serre et al. 2007). Visual dissimilarity between images was defined as one minus the Pearson’s correlation between their C1 outputs (for a similar approach, see Naspi et al. 2021). To determine the visual dissimilarity between concepts, we averaged the pairwise dissimilarities between the images representing each concept.
Experimental procedure
Participants viewed the concepts in a single scanning run of approximately 24 minutes, after completing two runs of unrelated tasks described later. The timeline for a single trial is shown in Fig. 1(B). Each trial consisted of a picture presented in the middle of screen for 2.5 s with the concept name shown below. Participants were asked to think about the concept demonstrated by the picture and described by the word. To ensure that participants paid attention to the concepts, in 25% of trials, the concept was followed by a catch question, which asked if the concept is related to another word. For example, for the concept “a diagnosis,” the catch question was “Is it associated with doctor?”. Each concept was followed by a catch question on one of its four presentations. The correct answers for half of these catch questions were “Yes,” and for the other half, they were “No”. All trials were presented with a mean interstimulus interval of 2.5 s, jittered between 1 and 4 s. Trials were presented in four blocks, each containing one instance of each concept. The order of stimuli within each block was randomized separately for each participant, to ensure independence between activation patterns for each concept (Mumford et al. 2014).
Image acquisition and processing
Images were acquired on a 3 T Siemens Prisma scanner with a 32-channel head coil. For the functional images, the multi-echo EPI sequence included 46 slices covering the whole brain with echo time (TE) at 13, 31, and 50 ms, repetition time (TR) = 1.7 s, flip angle = 73, 80 × 80 matrix, reconstructed in-plane resolution = 3 mm × 3 mm, slice thickness = 3.0 mm (no slice gap) and multiband factor = 2. A single run of 858 volumes was acquired. A high-resolution T1-weighted structural image was also acquired for each participant using an MP-RAGE sequence with 1 mm isotropic voxels, TR = 2.5 s, TE = 4.4 ms. To minimize the impact of head movements and signal drop out in the ventral temporal regions (Kundu et al. 2017), the study employed a whole-brain multi-echo acquisition protocol, in which data were simultaneously acquired at three TEs. Data from the three-echo series were weighted and combined, and the resulting time-series were denoised using independent components analysis (ICA).
Images were pre-processed and analyzed using SPM12 and the TE-Dependent Analysis Toolbox (Tedana) (Kundu et al. 2012; Kundu et al. 2013). Estimates of head motion were obtained using the first BOLD echo series. Slice-timing correction was carried out and images were then realigned using the previously obtained motion estimates. Tedana was used to combine the three-echo series into a single-time series and to divide the data into components classified as either BOLD-signal or noise-related based on their patterns of signal decay over increasing TEs (Kundu et al. 2017). Components classified as noise were discarded. After that, images were unwarped with a B0 field-map to correct for irregularities in the scanner’s magnetic field. Finally, functional images were spatially normalized to MNI space using SPM’s DARTEL tool (Ashburner 2007), and were smoothed with a kernel of 8 mm FWHM for univariate and PPI analysis and 4 mm FWHM for RSA analysis. Although multivariate analyses are often performed on unsmoothed images, there is evidence that a small amount of smoothing can slightly improve performance (Gardumi et al. 2016; Hendriks et al. 2017). Data in our study were treated with a high-pass filter with a cut-off of 180 s. Covariates consisted of six motion parameters and their first-order derivatives.
For univariate and PPI analysis, a general linear model (GLM) was used that included 3 regressors for event concepts, object concepts, and catch trials. For RSA, to obtain better estimates of activation patterns of each concept, we used the least squares separate (LSS) approach (Mumford et al. 2012). We ran a separate GLM for each concept, where the 4 trials of that concept were modeled as the regressor of interest and all other trials were combined into a single nuisance regressor (with a further regressor modeling catch questions). This process yielded one activation map for each concept, which were used to compute neural RDMs.
Regions of interest
We defined four regions of interest (ROIs): left ventral anterior temporal lobe (left vATL), left angular gyrus (left AG), right ventral anterior temporal lobe (right vATL), right angular gyrus (right AG). Each ROI was defined as a 10 mm radius sphere centered on specific MNI co-ordinates, which were selected in a two-stage process.
In the first stage, we constructed anatomical masks covering the ATLs and AGs. Masks of ATLs were made in a similar way to Hoffman and Lambon Ralph (2018). We first created masks of the temporal regions: inferior temporal gyrus, fusiform gyrus, superior temporal gyrus, and middle temporal gyrus. These were created by including all voxels with a greater than 50% probability of being located within these areas in the LONI Probabilistic Brain Atlas (LPBA40) (Shattuck et al. 2008). These regions span the full length of the temporal lobe. As there are no anatomical landmarks that demarcate the ATL from the posterior temporal lobe, we had to decide which voxels to include. Following Hoffman and Lambon Ralph (2018), we divided the temporal lobe into six sections of roughly equal length along an anterior-to-posterior axis. These sections were numbered 0–5, with section 0 representing the most anterior section. The divisions were made approximately perpendicular to the long axis of the temporal lobe. Finally, we created left ATL and right ATL masks by combining sections 1 and 2 of temporal regions’ masks in the left hemisphere and right hemisphere, separately. This includes ventral temporal cortex between y ≈ −2 and −28, which is typically the main focus of semantic activation (e.g. Shimotake et al. 2015).
For masks of AGs, we included all voxels with a greater than 30% probability of being located within this particular brain region as defined by the LPBA40 atlas (Shattuck et al. 2008). We used a more lenient voxel inclusion threshold here as the precise boundaries of the AG vary somewhat across individuals (Caspers et al. 2006).
Within these large anatomical masks, we then sought the voxels that were most responsive to semantic processing, using the activation peaks from independent data collected from the same participants (i.e. a functional localizer at the group level). In the scanning runs prior to the object/events task, participants completed a series of tasks which required them to match words based on similarities in color, size, general meaning, and letters (for further details, see Wu and Hoffman 2023). The judgments of color, size and general meaning all required access to semantic knowledge, while the letter similarity task did not. Based on these tasks, we made a semantic > non-semantic contrast at the group level and identified the peak co-ordinates within each anatomical mask.
In the vATLs, the maximal response was in the left and right anterior fusiform region. The maximal AG response was in the ventral part of the AG mask, in the region of the temporoparietal junction. Each ROI was defined as a 10 mm radius sphere centered on the peak semantic > non-semantic co-ordinates within each anatomical mask (see Fig. 3). The center coordinates were as follows: left vATL [−36, −18, −30]; left AG [−51, −54, 15]; right vATL [33, −9, −39]; right AG [66, −45, 15]. These four ROIs were used in univariate, RSA, and PPI analyses.
Fig. 3.

Activation to events and objects in ROIs. Bars show 1 standard error of the mean.
Behavioral analysis
For the behavioral data, we built one linear mixed effect (LME) model to predict accuracy for responses to catch questions of event and object concepts, and another one to predict reaction times. The analyses were conducted with R-4.0.3, and three packages: “lme4,” “effects,” and “afex.” In each LME model, concept type (event/object) was set as a fixed effect, and participant was set as the random effect with intercepts and random slopes for concept type.
Univariate analysis
To compare activation for event concept and object concept conditions, both whole-brain analysis and ROI analyses were conducted with SPM12. The whole-brain analysis was corrected for multiple comparisons (P < 0.05) at the cluster level using SPM’s random field theory, with a cluster-forming threshold of P < 0.005. In ROI analyses, we extracted mean beta values in left vATL, left AG, right vATL, right AG in each condition, which represent activation relative to the implicit baseline (rest). Then a three-way repeated ANOVA analysis was done using R-4.2.2, to examine the effects of concept type (event/object), ROI (AG/vATL), hemisphere (left/right), and their interactions.
Representational similarity analysis
We used RSA to examine which brain areas are sensitive to similarity in event and object concepts’ semantic representations. CoSMoMVPA (Oosterhof et al. 2016) was used for these analyses.
To investigate effects across the brain, we used a searchlight analysis with a spherical searchlight with radius of four voxels. We extracted activation patterns for the 60 concepts, and computed pairwise dis-similarities (1 – Pearson correlation) between activation patterns for the event concepts and separately for the object concepts. Then the partial Spearman correlation between neural RDMs and semantic RDMs, controlling for effects of the visual RDMs, was computed. This process was repeated for all searchlights, resulting in two correlation maps, one for objects and one for events. These showed the degree to which neural similarities between concepts are predicted by their semantic similarity. Correlations were Fisher-z transformed for group-level analysis. We conducted ROI analysis in the same way but using neural patterns from the four spherical ROIs.
To test the significance of the semantic-neural correlations, we used a two-stage method to perform permutation tests (Stelzer et al. 2013). We first computed the correlation maps between semantic RDMs and neural RDMs 100 times for each participant, with random reshuffling of the labels in the semantic and visual RDMs each time. This process provided a distribution of expected correlations under the null hypothesis for each participant. Then we used a Monte Carlo approach to compute a null correlation distribution at the group level (over all participants). To do this, we randomly selected one null correlation map from each participant’s null distribution and averaged these to generate a group mean. This process was repeated 10,000 times to generate a distribution of the expected group correlation under the null hypothesis. In searchlight analyses, we entered the observed and null correlation maps into the Monte Carlo cluster statistics function of CoSMoMVPA to generate a statistical map corrected for multiple comparisons using threshold-free cluster enhancement (Smith and Nichols 2009). These maps were thresholded at corrected P < 0.05. For ROI analyses, we used the position of the observed group correlation in the null distribution to determine the P-value (e.g. if the observed correlation was greater than 95% of correlations in the null distribution, the P-value would be 0.05).
A similar procedure was used to test for regions showing a difference in the strength of the semantic-neural correlation between objects and events. We computed a difference map for each participant by subtracting the object correlation map from the event correlation map. We then used the permutation method to test for regions where the difference was significantly greater or less than zero.
PPI analyses
PPI analysis is a functional connectivity method for investigating task-specific changes in the relationship between different brain regions’ activity (Friston et al. 1997). While functional connectivity analyses often consider the temporal correlations between different brain regions in all conditions (including the resting state), PPI concentrates on connectivity changes caused by experimental manipulations (Gitelman et al. 2003; O’Reilly et al. 2012; Ashburner et al. 2014). For this study, PPI analysis was conducted to examine which brain regions would show increased correlation with our ROIs when representing event concepts relative to object concepts, or vice versa. The PPI analysis for each seed region (left vATL, left AG, right vATL, right AG) was conducted using SPM12 and the gPPI toolbox (McLaren et al. 2012) with the following steps. First, the seed region was defined as described in the ROI section above, and the BOLD signal time-series extracted using the first eigenvariate. Then, gPPI was used to create a GLM with the following regressors:
(i) The signal in the seed region.
(ii) One regressor coding for each experimental effect of interest, including event concepts, object concepts, and catch questions.
(iii) The interaction between the signal in the seed region and each experimental effect.
(iv) Head movement covariates as included in the main univariate analysis.
This model was used for testing differences between PPI regressors (i.e. changes in connectivity driven by concept type) in the whole brain. Results were corrected for multiple comparisons (P < 0.05) at the cluster level using SPM’s random field theory, with a cluster-forming threshold of P < 0.005.
Results
Behavioral data
LME models were used to test whether participants responded differently to catch questions about event and object concepts. There were no significant differences in accuracies between concept types (event M = 97.44%, SD = 0.04, object M = 96.98%, SD = 0.04, z (42) = 21.79, P = 0.29) and overall accuracy was very high, suggesting participants maintained attention through the experiment. Participants responded slightly faster to event questions (event M = 1.26 s, SD = 0.27 s, object M = 1.30 s, SD = 0.26 s, t (1815) = −2.152, P < 0.03).
Univariate fMRI analysis
We began by contrasting activation to events and objects. While these results showed which regions are differentially engaged by the conditions, it is important to note that there were substantial visual differences in the stimuli used in each condition. Thus, these results may reflect both semantic and visual differences between event and object trials. The whole-brain analysis contrasting event and object concepts is displayed in Fig. 2 (see Supplementary Fig. 1 for activation for events and objects relative to rest). Event concepts elicited more activation than objects bilaterally in fusiform gyrus, middle occipital gyrus, and lingual gyrus, as well as anterior and posterior parts of superior and middle temporal gyri, hippocampus and parahippocampal regions, parts of the ventromedial prefrontal cortex and posterior cingulate. Higher activation in visual and scene-processing areas (e.g. parahippocampal gyrus and posterior cingulate) may reflect differences in the images used in the two conditions. Event images were more visually complex, contained a higher number of objects and included contextual elements not present in the object images (see Fig. 1 for examples). In contrast, ventromedial prefrontal cortex, temporoparietal junction, and lateral anterior temporal regions are all frequently engaged by social cognition tasks, so effects here could be due to the relevance of social interactions to event trials (Binney and Ramsey 2020; Diveica et al. 2021; Balgova et al. 2022). Comparatively, object concepts elicited higher activation bilaterally in supramarginal gyrus (SMG), superior parietal cortex, and parts of the dorsolateral prefrontal cortices.
Fig. 2.

Univariate effects of event concepts versus object concepts, FWE corrected (P < 0.05).
Figure 3 shows whether ROIs’ activations were affected by three factors: concept type (event/object), ROI (vATL/AG), hemisphere (left/right). A three-way repeated ANOVA was used to examine these effects. For both event and object concepts, ROIs in left hemisphere showed significantly higher activation (F (1, 42) = 15.88, P < 0.001). Overall, events elicited more activation than objects, and an interaction between concept type and ROI was also found (Concept effect: F (1, 42) = 4.436, P = 0.041; Concept × ROI: F (1, 42) = 5.483, P = 0.024). No other effects were significant. Post hoc tests were performed comparing events vs. objects in each ROI. Left vATL was activated more strongly by events (F (1, 42) = 30.741, P < 0.001), as was right vATL (F (1, 42) = 11.322, P = 0.002). There were no effects of concept type in left AG and right AG. According to dual-hub theory, vATL would be more engaged in processing objects, while AG is more engaged by event representation. The ROI analysis did not show this pattern. However, given the greater complexity of the event images, it is difficult to draw conclusions from these univariate analyses. For example, event images include multiple objects which could drive greater activation in object-specialized regions. To avoid this issue, we next conducted RSA within each concept type.
Representational similarity analyses
The correlation maps, showing regions where neural RDMs were predicted by semantic RDMs, are displayed in Fig. 4(A). Generally, correlation effects were found in a similar set of bilateral regions for both events and objects. Specifically, the strongest effects were found in lateral occipital areas and parts of the ventral visual stream (ventral and medial temporal lobe), extending anteriorly into vATL. We also observed effects spreading into TPC, especially for event concepts. The left inferior frontal area also showed correlations for both events and objects. Thus, neural activation patterns were correlated with semantic relationships not only in canonical semantic regions but also extensively in object and scene processing regions of the visual system. These effects indicate sensitivity to the semantic features of objects and events in these regions, since low-level visual similarity was controlled for in our analyses.
Fig. 4.

(A).Representational similarity maps for each concept type, showing regions where neural similarity is significantly correlated with semantic similarity (corrected P < 0.05); (B). The difference of representational similarities between event and object concepts (corrected P < 0.05). In (A) and (B), low-level visual features are controlled by covarying visual similarities measured with Hmax. Color scale shows the correlation strength.
Figure 4(B) presents regions that showed a significant difference in correlation strength between the event and object analyses. Bilateral primary visual cortex showed stronger correlations for events relative to objects. Conversely, stronger correlations for objects were found in lateral occipital regions, which is consistent with evidence for category-selective responses in this region in object recognition (for a review, see Bi et al. 2016; Carota et al. 2017; Chen et al. 2017; Wu et al. 2020; Wurm and Caramazza 2022). No differences were found in our target regions of vATL and AG, so we turned to more sensitive ROI analyses to investigate effects in these regions.
The correlations between neural and semantic RDMs in the four ROIs are displayed in Fig. 5. Permutation testing indicated that left vATL, right vATL and left AG showed significant correlation between neural RDMs and semantic RDMs for both event and object concepts (all P < 0.0056). Right AG only showed a significant correlation for event concepts (P < 0.001).
Fig. 5.

Representational similarity effects in ROIs. Bars show 1 standard error of the mean.
A three-way repeated ANOVA was conducted to examine whether correlations were affected by three factors: ROI (vATL/AG), hemisphere (left/right), and concept type (event/object). Overall, event concepts’ RDM showed higher correlations with neural RDMs than objects’ (F (1, 42) = 9.467, P = 0.004). No other main effects or interactions were significant at P < 0.05, though there was a suggestion of a weak three-way interaction (F (1, 42) = 3.27, P = 0.078). In post hoc pairwise comparisons of events and objects in each ROI, left vATL and right AG had significantly higher correlations for event concepts (left vATL F (42) = 5.106, P = 0.03; right AG F (42) = 10.951, P = 0.002). The left AG also showed a stronger correlation for event concepts, but this difference was not statistically significant (F (42) = 3.362, P = 0.074). A two-way ANOVA (concept type × hemisphere) conducted on the AG data reported a main effect of concept type (F (1, 42) = 9.379, P = 0.004), but no interaction between concept type and hemisphere (F (1, 42) = 0.509, P = 0.479). This result suggests left AG and right AG had similar effects of concept type.
In a post-hoc two-way ANOVAs in data split by hemisphere (concept type × ROI), both left and right hemispheres showed significantly higher correlations for event concepts (left hemisphere F (1, 42) = 7.112, P = 0.011; right hemisphere F (1, 42) = 4.875, P = 0.033), and only right hemisphere showed interaction between ROI and concept type (F (1, 42) = 6.962, P = 0.012). This result suggests left vATL and left AG had similar effects of concept type, whereas right AG showed a stronger representation al similarity for events than for objects compared to right vATL.
To summarize, stronger correlations for events than objects were found in bilateral AG and in left vATL. The results in AGs are consistent with the dual-hub hypothesis, which proposes that AG is specialized for representing semantic properties of events. However, effects in the vATLs contradict the idea that this region is particularly sensitive to object semantics. Our results instead indicate that right vATL is equally sensitive to events and objects’ semantics, while left vATL is more sensitive to events.
PPI analysis
To investigate how vATL and AG interact with other brain regions in representing concepts, PPI analyses were conducted using left vATL, left AG, right vATL, and right AG as seed regions. Analyses tested for change in connectivity as a function of concept type (event vs. object) and thus could be influenced by the greater visual complexity of the images presented on event trials. When participants processed event concepts, left vATL had stronger connectivity with right posterior MTG (Fig. 6A). Right vATL showed a similar pattern but the effect did not survive cluster correction (see Supplementary Fig. 2). Right AG showed stronger connectivity with bilateral fusiform gyrus and middle occipital gyrus (Fig. 6B). Left AG showed no effects at cluster-corrected significance, though a more lenient uncorrected threshold showed increased connectivity with left fusiform gyrus, left ITG and right IFG for event concepts. Supplementary Figure 2 shows uncorrected events > objects effects for all four seed regions. No effects for objects > events were found at a cluster-corrected threshold and very few significant areas were found at an uncorrected threshold (shown in Supplementary Fig. 3).
Fig. 6.

(A). For events > objects, regions showing increased connectivity with left vATL. (B). For events > objects, regions showing increased connectivity with right AG. Surface render (cluster corrected P < 0.05). Seed regions are shown as colored circles.
Discussion
Both event and object knowledge are critical semantic abilities, but their neural correlates are unclear. Some researchers have suggested that vATL is specialized for object semantics and AG for event semantics (Binder and Desai 2011; Mirman et al. 2017). To test this hypothesis, we used RSA to investigate the neural basis of representing event and object concepts. Left and right AG were found to encode semantic similarity among event concepts more strongly than similarity among object concepts, though left AG also coded objects’ semantic similarity. Left and right vATLs both encoded semantic structure for object and event concepts, and left vATL showed stronger effects for events than objects. Univariate analyses also indicated more engagement of bilateral vATLs for event concepts. These findings support the idea that AG is more specialized for event semantics relative to object semantics. However, vATL specialization for object semantics is not supported by our results, suggesting that this region plays a more global role in semantic representation.
Sensitivity to object and event semantics in the vATLs and AGs
Many previous studies have found that activity patterns in vATL code semantic similarities among object concepts (e.g. Bruffaerts et al. 2013; Tyler et al. 2013; Clarke and Tyler 2014; Liuzzi et al. 2015; Chen et al. 2016; Devereux et al. 2018; Naspi et al. 2021). Our data indicate that the same region is also sensitive to semantic relationships between event concepts.
For vATLs, RSA indicated that their activity patterns reflect the semantic structure of events as well as objects (Fig. 4), and left vATL showed a stronger correlation for events than objects. The simplest explanation for this is that vATL represents not only individual object characteristics, but also objects’ interactions and their context. The RSA finding is consistent with hub-and-spoke models of this region’s function (Patterson et al. 2007; Rice et al. 2015a; Lambon Ralph et al. 2017), which propose that vATL forms conceptual representations by integrating information from a range of neural sources. Our results suggest that, in addition to integrating the features of individual objects, this region may also form representations of more complex event-related concepts. However, an alternative explanation is that vATL is specialized for object representation and that the effects we see are a by-product of processing the objects involved in the depicted event stimuli. If semantically similar events involve semantically-similar objects, then vATL effects for events may reflect the coding for objects involved in those events. For example, picnic and barbeque are semantically similar events but they also contain semantically similar objects (food, plates, knives, etc.).
The univariate analysis showed more vATL activation for event trials (Fig. 2). As we noted in the Results, this univariate effect is difficult to interpret, given that our two sets of stimuli differed in a number of ways. Greater activation might simply stem from the greater number of concepts associated with event processing. According to the hub-and-spoke theory, vATL integrates different modality-specific information sources into a concept, including not only visual features like color or shape, but also objects’ relevant actions or locations (Peelen and Caramazza 2012; Lambon Ralph et al. 2017). Events contain multiple objects and people interacting in a specific environment. Thus, event concepts might require the vATL to encode multiple concepts’ properties before settling on an overall representation of the event concept. The stronger vATL response for event concepts in univariate analysis might be caused by the heavier working load.
PPI analysis indicated that left vATL had stronger connectivity with right pMTG when processing event concepts (Fig. 6A). Right pMTG has been implicated in coding causal relations between objects (Leshinskaya and Thompson-Schill 2020), and in representing action concepts present in videos, still images and in language (Watson et al. 2013; Chen et al. 2020). The increased connectivity between vATL and pMTG may be a result of an enhanced contribution of relational and action-related information when understanding event concepts. This is in line with evidence that the vATL semantic hub alters its connectivity with more specialized spoke regions depending on the type of information that is relevant to the concepts being processed (Coutanche and Thompson-Schill 2015; Chiou and Lambon Ralph 2019).
For AG, RSA showed that activity patterns in both AG were correlated with semantic structure for events more strongly than for objects. Xu et al. (2018) also used RSA and found specialization of TPC for event-based relations among objects relative to category-based relations among the same objects. In contrast, the present study examined a single type of similarity (based on word2vec) and compared different types of concepts (events vs objects). Thus, the two studies provide converging complementary evidence of TPC (more specifically, AG) specialization for event semantics, consistent with this region’s involvement in event representation more generally. AG plays an important role in representing autobiographical and episodic memories of events (Bonnici et al. 2018; Russell et al. 2019), in spatial-temporal feature integration (Yazar et al. 2014; Ben-Zvi et al. 2015; Bonnici et al. 2016; Richter et al. 2016; Yazar et al. 2017), and in combinatorial semantics (Fedorenko et al. 2016; Pylkkänen 2019). In addition, AG may be particularly sensitive to thematic relations because it processes contextual details of events (for a review, see Binder and Desai 2011; Mirman et al. 2017). AG is also part of the broader DMN, which integrates information to form context-specific representations of evolving situations (for a review, see Ranganath and Ritchey 2012; Yeshurun et al. 2021), and is sensitive to event boundaries within a continuous experience (Zacks et al. 2010; Swallow et al. 2011; Baldassano et al. 2017). These functions of AG together suggest that it encodes dynamic and complex combinations of concepts and experiences, where people, objects, and actions are bound together in time and space (for related proposals, see Humphreys and Lambon Ralph 2015; Branzi et al. 2020; Branzi et al. 2021; Humphreys et al. 2021; Branzi and Lambon Ralph 2023).
The univariate analysis did not show significant activation differences in AG between events and objects. This is not consistent with the idea that AG is specialized for event semantics. Bedny et al. (2014) used a similar univariate analysis and found stronger response in TPC (primarily posterior MTG) for event nouns relative to object nouns. A key difference between the two studies is that, in the present study, pictures were presented along with the nouns. It is important to note that there were uncontrolled differences between event and object images, making these results (and differences from the results of Bedny et al.) hard to interpret.
Many previous studies implicating AG in event representation have presented temporally extended stimuli like narratives (e.g. Bonnici et al. 2016) or movies (e.g. Zacks et al. 2010; Swallow et al. 2011; Baldassano et al. 2017), or have required continuous generation of words (e.g. Yazar et al. 2014; Bonnici et al. 2018). In contrast, our study has shown that simple representations of static, abstract events are sufficient to engage AG for semantic processing. Furthermore, while previous language-based studies have focused on the role of left AG in representing thematic/event knowledge, here we found both left AG and right AG code event semantics (Fig. 5). The bilateral effects might be due to our multimodal stimuli: while semantic activations are often left-lateralized for written word processing, more bilateral engagement is common for multimodal and non-verbal stimuli (Rice et al. 2015b). Previous behavioral studies and lesion-symptom mapping studies indicated that left hemisphere injuries impaired verbal knowledge, while right hemisphere damage affected pictorial memory (Grossman and Wilson 1987; Gainotti et al. 1994; Acres et al. 2009; Butler et al. 2009). Neuroimaging investigations further support this view, showing increased involvement of left temporal regions in processing verbal stimuli and right temporal cortex in understanding environmental sounds and images (Thierry et al. 2003; Thierry and Price 2006; Hocking and Price 2009).
In PPI analysis, right AG showed strong connectivity with bilateral ventral visual regions for event concepts (Fig. 6B), which are likely a consequence of this region extracting event-related information from the visual scenes we presented. Images of events were necessarily more complex those of objects: depicting an event requires an image that contains a diverse set of agents and objects interacting in a particular context. To represent the event as a cohesive concept, these individual items must be amalgamated, taking into account their identities, positions, orientations, and interactions. Increased connectivity between right AG and visual regions may reflect the greater quantity of visual information present on event trials and the need for greater visual analysis to encode the relationships between the various elements.
Effects in other regions
In addition to the effects in vATL and AG, our RSA analysis also found that patterns throughout large portions of lateral and ventral occipitotemporal cortex (OTC) were correlated with semantic structure for both objects and events. Within these areas, correlations were stronger for object concepts than event concepts (Fig. 4B). The correlation effects in OTC are consistent with selectivity for specific object categories in these regions (for a review, see Bi et al. 2016). Many studies have reported that when people view pictures or object names, clusters of voxels in OTC are selectively responsive to certain categories of objects, such as faces, bodies, tools, or places (Chao et al. 1999; Ishai et al. 2000; O'Craven and Kanwisher 2000; Goyal et al. 2006; Noppeney et al. 2006; Costantini et al. 2011; Fairhall and Caramazza 2013; Fairhall et al. 2014). In particular, lateral OTC is known to be more strongly activated by small, manipulable objects (such as tools) and by body parts (Chao et al. 1999; Noppeney et al. 2006; Costantini et al. 2011). In ventral OTC, anterior medial regions (parahippocampal and medial fusiform) show preferences for inanimate items broadly related to navigation, including scenes, places, buildings, and large non-manipulable objects (Ishai et al. 2000; O'Craven and Kanwisher 2000; Fairhall and Caramazza 2013; Fairhall et al. 2014), while the posterior fusiform has a preference for animate items including faces and animals (Chao et al. 1999; Ishai et al. 2000; O'Craven and Kanwisher 2000; Goyal et al. 2006). These category-selective responses explain why objects showed stronger semantic correlations with OTC patterns than events: objects from the same category were more semantically related, thus activated similar patches of cortex in OTC. Nevertheless, OTC patterns also showed correlations with event semantics. This could be because pictures of similar events tend to contain objects from similar categories, as discussed earlier.
Event concepts showed stronger correlations than object concepts in primary visual cortex. There are a few possible explanations for this effect. One intriguing possibility is that, when presented with static event images, participants were primed to mentally anticipate the movements of the objects or people depicted in those images. Primary visual cortex (V1) has been associated with motion-inducing illusions and with predicting visual stimuli in many studies (Muckli et al. 2005; Sterzer et al. 2006; Alink et al. 2010; Gavornik and Bear 2014; Kok et al. 2014; Ekman et al. 2017). V1 activation can be modulated by prediction of motion direction or onset (Muckli et al. 2005; Alink et al. 2010) and prior expectation of specific visual stimuli or visual sequences can evoke V1 responses similar to those evoked by viewing the actual stimuli or sequence (Sterzer et al. 2006; Gavornik and Bear 2014; Kok et al. 2014; Ekman et al. 2017). For example, Ekman et al. (2017) found that after familiarizing participants with a spatial sequence, flashing only the starting point of the sequence triggered an activity wave in V1 that resembled the full stimulus sequence. Thus, the observed correlation effects in V1 might indicate the encoding of different predictions about potential motion in event images. This explanation is speculative and verifying it would require more systematic investigation of the motion types present in event images and their correlation with semantic properties. It is also possible that there were other visual properties (e.g. overall image complexity) that covaried with semantic content on event trials and led to the observed effect.
In conclusion, by testing the predictions of dual-hub theory with event and object concepts, our study found AG specialization for coding event semantics, but did not find vATL specialization for object semantics. Left vATL even coded similarity for events more strongly than objects. These findings provide new data on the divisions of labor that exist within the semantic system.
Supplementary Material
Acknowledgments
Imaging was carried out at the Edinburgh Imaging Facility (www.ed.ac.uk/edinburgh-imaging), University of Edinburgh, which is part of the SINAPSE collaboration (www.sinapse.ac.uk). We are grateful to the University of Minnesota Center for Magnetic Resonance Research for sharing their neuroimaging sequences. For the purpose of open access, the author has applied a Creative Commons Attribution (CC BY) license to any Author Accepted Manuscript version arising from this submission.
Contributor Information
Yueyang Zhang, School of Philosophy, Psychology & Language Sciences, University of Edinburgh, Edinburgh EH8 9JZ, United Kingdom.
Wei Wu, School of Philosophy, Psychology & Language Sciences, University of Edinburgh, Edinburgh EH8 9JZ, United Kingdom.
Daniel Mirman, School of Philosophy, Psychology & Language Sciences, University of Edinburgh, Edinburgh EH8 9JZ, United Kingdom.
Paul Hoffman, School of Philosophy, Psychology & Language Sciences, University of Edinburgh, Edinburgh EH8 9JZ, United Kingdom.
CRediT statement
Yueyang Zhang (Conceptualization, Methodology, Investigation, Formal analysis, Writing—original draft. Wei Wu (Methodology, Investigation, Writing—review and editing), Daniel Mirman (Methodology, Writing—review and editing), and Paul Hoffman (Funding acquisition, Conceptualization, Methodology, Formal analysis, Writing—review and editing).
Funding
P.H. was supported by a Biotechnology and Biological Sciences Research Council grant (BB/T004444/1).
Conflict of Interest statement: None declared.
Data availability
Data and code supporting this study are available as follows: Neuroimaging data: https://doi.org/10.7488/ds/7521. Other data and analysis code: https://osf.io/mn8ft/. Group effect maps: https://neurovault.org/collections/NSWUEOPG/
References
- Acres K, Taylor K, Moss H, Stamatakis E, Tyler LK. Complementary hemispheric asymmetries in object naming and recognition: a voxel-based correlational study. Neuropsychologia. 2009:47(8–9):1836–1843. [DOI] [PubMed] [Google Scholar]
- Alink A, Schwiedrzik CM, Kohler A, Singer W, Muckli L. Stimulus predictability reduces responses in primary visual cortex. J Neurosci. 2010:30(8):2960–2966. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Altmann G, Ekves Z. Events as intersecting object histories: a new theory of event representation. Psychol Rev. 2019:126(6):817–840. [DOI] [PubMed] [Google Scholar]
- Andersen RA, Cui H. Intention, action planning, and decision making in parietal-frontal circuits. Neuron. 2009:63(5):568–583. [DOI] [PubMed] [Google Scholar]
- Ashburner J. A fast diffeomorphic image registration algorithm. NeuroImage. 2007:38(1):95–113. [DOI] [PubMed] [Google Scholar]
- Ashburner J, Barnes G, Chen C-C, Daunizeau J, Flandin G, Friston K, Kiebel S, Kilner J, Litvak V, Moran R. SPM12 manual. London: Wellcome Trust Centre for Neuroimaging; 2014. p. 2464. [Google Scholar]
- Baldassano C, Chen J, Zadbood A, Pillow JW, Hasson U, Norman KA. Discovering event structure in continuous narrative perception and memory. Neuron. 2017:95(3):709–721.e5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Balgova E, Diveica V, Walbrin J, Binney RJ. The role of the ventrolateral anterior temporal lobes in social cognition. Hum Brain Mapp. 2022:43(15):4589–4608. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Baron SG, Thompson-Schill SL, Weber M, Osherson D. An early stage of conceptual combination: superimposition of constituent concepts in left anterolateral temporal lobe. Cogn Neurosci. 2010:1(1):44–51. [DOI] [PubMed] [Google Scholar]
- Bedny M, Caramazza A, Grossman E, Pascual-Leone A, Saxe R. Concepts are more than percepts: the case of action verbs. J Neurosci. 2008:28(44):11347–11353. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bedny M, Dravida S, Saxe R. Shindigs, brunches, and rodeos: the neural basis of event words. Cogn Affect Behav Neurosci. 2014:14(3):891–901. [DOI] [PubMed] [Google Scholar]
- Ben-Zvi S, Soroker N, Levy DA. Parietal lesion effects on cued recall following pair associate learning. Neuropsychologia. 2015:73:176–194. [DOI] [PubMed] [Google Scholar]
- Bi Y, Wang X, Caramazza A. Object domain and modality in the ventral visual pathway. Trends Cogn Sci. 2016:20(4):282–290. [DOI] [PubMed] [Google Scholar]
- Binder JR, Desai RH. The neurobiology of semantic memory. Trends Cogn Sci. 2011:15(11):527–536. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Binney RJ, Ramsey R. Social semantics: the role of conceptual knowledge and cognitive control in a neurobiological model of the social brain. Neurosci Biobehav Rev. 2020:112:28–38. [DOI] [PubMed] [Google Scholar]
- Binney RJ, Embleton KV, Jefferies E, Parker GJ, Lambon Ralph MA. The ventral and inferolateral aspects of the anterior temporal lobe are crucial in semantic memory: evidence from a novel direct comparison of distortion-corrected fMRI, rTMS, and semantic dementia. Cereb Cortex. 2010:20(11):2728–2738. [DOI] [PubMed] [Google Scholar]
- Bonnici HM, Richter FR, Yazar Y, Simons JS. Multimodal feature integration in the angular gyrus during episodic and semantic retrieval. J Neurosci. 2016:36(20):5462–5471. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bonnici HM, Cheke LG, Green DA, FitzGerald TH, Simons JS. Specifying a causal role for angular gyrus in autobiographical memory. J Neurosci. 2018:38(49):10438–10443. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Boylan C, Trueswell JC, Thompson-Schill SL. Compositionality and the angular gyrus: a multi-voxel similarity analysis of the semantic composition of nouns and verbs. Neuropsychologia. 2015:78:130–141. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Branzi F, Lambon Ralph M. Semantic-specific and domain-general mechanisms for integration and update of contextual information. Hum Brain Mapp. 2023:44(17):5547–5566. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Branzi FM, Humphreys GF, Hoffman P, Lambon Ralph MA. Revealing the neural networks that extract conceptual gestalts from continuously evolving or changing semantic contexts. NeuroImage. 2020:220:116802. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Branzi FM, Pobric G, Jung J, Lambon Ralph MA. The left angular gyrus is causally involved in context-dependent integration and associative encoding during narrative reading. J Cogn Neurosci. 2021:33(6):1082–1095. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bruffaerts R, Dupont P, Peeters R, De Deyne S, Storms G, Vandenberghe R. Similarity of fMRI activity patterns in left perirhinal cortex reflects semantic similarity between words. J Neurosci. 2013:33(47):18597–18607. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Brysbaert M, Warriner AB, Kuperman V. Concreteness ratings for 40 thousand generally known English word lemmas. Behav Res Methods. 2014:46(3):904–911. [DOI] [PubMed] [Google Scholar]
- Butler CR, Brambati SM, Miller BL, Gorno-Tempini M-L. The neural correlates of verbal and non-verbal semantic processing deficits in neurodegenerative disease. Cogn Behav Neurol. 2009:22(2):73–80. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Buxbaum LJ, Kalénine S. Action knowledge, visuomotor activation, and embodiment in the two action systems. Ann N Y Acad Sci. 2010:1191(1):201–218. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Carota F, Kriegeskorte N, Nili H, Pulvermüller F. Representational similarity mapping of distributional semantics in left inferior frontal, middle temporal, and motor cortex. Cereb Cortex. 2017:27(1):294–309. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Caspers S, Geyer S, Schleicher A, Mohlberg H, Amunts K, Zilles K. The human inferior parietal cortex: cytoarchitectonic parcellation and interindividual variability. NeuroImage. 2006:33(2):430–448. [DOI] [PubMed] [Google Scholar]
- Chao LL, Haxby JV, Martin A. Attribute-based neural substrates in temporal cortex for perceiving and knowing about objects. Nat Neurosci. 1999:2(10):913–919. [DOI] [PubMed] [Google Scholar]
- Chen Y, Shimotake A, Matsumoto R, Kunieda T, Kikuchi T, Miyamoto S, Fukuyama H, Takahashi R, Ikeda A, Lambon Ralph MA. The ‘when’ and ‘where’ of semantic coding in the anterior temporal lobe: temporal representational similarity analysis of electrocorticogram data. Cortex. 2016:79:1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen L, Lambon Ralph MA, Rogers TT. A unified model of human semantic knowledge and its disorders. Nat Hum Behav. 2017:1(3):0039. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen Y, Wang Y, Zhao Q, Wang Y, Lu Y, Zhou C. Watching video of discrete maneuvers yields better action memory and greater activation in the middle temporal gyrus in half-pipe snowboarding athletes. Neurosci Lett. 2020:739:135336. [DOI] [PubMed] [Google Scholar]
- Chiou R, Lambon Ralph MA. Unveiling the dynamic interplay between the hub- and spoke-components of the brain's semantic system and its impact on human behaviour. NeuroImage. 2019:199:114–126. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Clarke A, Tyler LK. Object-specific semantic coding in human perirhinal cortex. J Neurosci. 2014:34(14):4766–4775. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Clarke A, Tyler LK. Understanding what we see: how we derive meaning from vision. Trends Cogn Sci. 2015:19(11):677–687. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Conca F, Catricalà E, Canini M, Petrini A, Vigliocco G, Cappa SF, Della Rosa PA. In search of different categories of abstract concepts: a fMRI adaptation study. Sci Rep. 2021:11(1):22587. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Costantini M, Urgesi C, Galati G, Romani GL, Aglioti SM. Haptic perception and body representation in lateral and medial occipito-temporal cortices. Neuropsychologia. 2011:49(5):821–829. [DOI] [PubMed] [Google Scholar]
- Coutanche MN, Thompson-Schill SL. Creating concepts from converging features in human cortex. Cereb Cortex. 2015:25(9):2584–2593. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Davis B, Christie J, Rorden C. Temporal order judgments activate temporal parietal junction. J Neurosci. 2009:29(10):3182–3188. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Devereux BJ, Clarke A, Marouchos A, Tyler LK. Representational similarity analysis reveals commonalities and differences in the semantic processing of words and objects. J Neurosci. 2013:33(48):18906–18916. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Devereux BJ, Clarke A, Tyler LK. Integrated deep visual and semantic attractor neural networks predict fMRI pattern-information along the ventral object processing pathway. Sci Rep. 2018:8(1):10636. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Diveica V, Koldewyn K, Binney RJ. Establishing a role of the semantic control network in social cognitive processing: a meta-analysis of functional neuroimaging studies. NeuroImage. 2021:245:118702. [DOI] [PubMed] [Google Scholar]
- Diveica V, Pexman PM, Binney RJ. Quantifying social semantics: an inclusive definition of socialness and ratings for 8388 English words. Behav Res Methods. 2023:55(2):461–473. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ekman M, Kok P, de Lange FP. Time-compressed preplay of anticipated events in human primary visual cortex. Nat Commun. 2017:8(1):15276. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fairhall SL, Caramazza A. Brain regions that represent amodal conceptual knowledge. J Neurosci. 2013:33(25):10552–10558. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fairhall SL, Anzellotti S, Ubaldi S, Caramazza A. Person-and place-selective neural substrates for entity-specific semantic access. Cereb Cortex. 2014:24(7):1687–1696. [DOI] [PubMed] [Google Scholar]
- Fedorenko E, Scott TL, Brunner P, Coon WG, Pritchett B, Schalk G, Kanwisher N. Neural correlate of the construction of sentence meaning. Proc Natl Acad Sci. 2016:113(41):E6256–E6262. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Friston K, Buechel C, Fink G, Morris J, Rolls E, Dolan RJ. Psychophysiological and modulatory interactions in neuroimaging. NeuroImage. 1997:6(3):218–229. [DOI] [PubMed] [Google Scholar]
- Gainotti G, Cappa A, Perri R, Silveri MC. Disorders of verbal and pictorial memory in right and left brain-damaged patients. Int J Neurosci. 1994:78(1–2):9–20. [DOI] [PubMed] [Google Scholar]
- Gardumi A, Ivanov D, Hausfeld L, Valente G, Formisano E, Uludağ K. The effect of spatial resolution on decoding accuracy in fMRI multivariate pattern analysis. NeuroImage. 2016:132:32–42. [DOI] [PubMed] [Google Scholar]
- Gavornik JP, Bear MF. Learned spatiotemporal sequence recognition and prediction in primary visual cortex. Nat Neurosci. 2014:17(5):732–737. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gennari SP, MacDonald MC, Postle BR, Seidenberg MS. Context-dependent interpretation of words: evidence for interactive neural processes. NeuroImage. 2007:35(3):1278–1286. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gitelman DR, Penny WD, Ashburner J, Friston KJ. Modeling regional and psychophysiologic interactions in fMRI: the importance of hemodynamic deconvolution. NeuroImage. 2003:19(1):200–207. [DOI] [PubMed] [Google Scholar]
- Goyal MS, Hansen PJ, Blakemore CB. Tactile perception recruits functionally related visual areas in the late-blind. Neuroreport. 2006:17(13):1381–1384. [DOI] [PubMed] [Google Scholar]
- Grossman M, Wilson M. Stimulus categorization by brain-damaged patients. Brain Cogn. 1987:6(1):55–71. [DOI] [PubMed] [Google Scholar]
- Hendriks MH, Daniels N, Pegado F, op de Beeck HP. The effect of spatial smoothing on representational similarity in a simple motor paradigm. Front Neurol. 2017:8:222. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hocking J, Price CJ. Dissociating verbal and nonverbal audiovisual object processing. Brain Lang. 2009:108(2):89–96. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hoffman P, Lambon Ralph MA. From percept to concept in the ventral temporal lobes: graded hemispheric specialisation based on stimulus and task. Cortex. 2018:101:107–118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hoffman P, Binney RJ, Lambon Ralph MA. Differing contributions of inferior prefrontal and anterior temporal cortex to concrete and abstract conceptual knowledge. Cortex. 2015:63:250–266. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Huberle E, Karnath H-O. The role of temporo-parietal junction (TPJ) in global gestalt perception. Brain Struct Funct. 2012:217(3):735–746. [DOI] [PubMed] [Google Scholar]
- Humphreys GF, Lambon Ralph MA. Fusion and fission of cognitive functions in the human parietal cortex. Cereb Cortex. 2015:25(10):3547–3560. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Humphreys GF, Lambon Ralph MA, Simons JS. A unifying account of angular gyrus contributions to episodic and semantic cognition. Trends Neurosci. 2021:44(6):452–463. [DOI] [PubMed] [Google Scholar]
- Husain M, Nachev P. Space and the parietal cortex. Trends Cogn Sci. 2007:11(1):30–36. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hutchison RM, Culham JC, Everling S, Flanagan JR, Gallivan JP. Distinct and distributed functional connectivity patterns across cortex reflect the domain-specific constraints of object, face, scene, body, and tool category-selective modules in the ventral visual pathway. NeuroImage. 2014:96:216–236. [DOI] [PubMed] [Google Scholar]
- Ishai A, Ungerleider LG, Haxby JV. Distributed neural systems for the generation of visual images. Neuron. 2000:28(3):979–990. [DOI] [PubMed] [Google Scholar]
- Kable JW, Lease-Spellmeyer J, Chatterjee A. Neural substrates of action event knowledge. J Cogn Neurosci. 2002:14(5):795–805. [DOI] [PubMed] [Google Scholar]
- Kable JW, Kan IP, Wilson A, Thompson-Schill SL, Chatterjee A. Conceptual representations of action in the lateral temporal cortex. J Cogn Neurosci. 2005:17(12):1855–1870. [DOI] [PubMed] [Google Scholar]
- Kaneshiro B, Perreau Guimaraes M, Kim H-S, Norcia AM, Suppes P. A representational similarity analysis of the dynamics of object processing using single-trial EEG classification. PLoS One. 2015:10(8):e0135697. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kok P, Failing MF, de Lange FP. Prior expectations evoke stimulus templates in the primary visual cortex. J Cogn Neurosci. 2014:26(7):1546–1554. [DOI] [PubMed] [Google Scholar]
- Kravitz DJ, Saleem KS, Baker CI, Ungerleider LG, Mishkin M. The ventral visual pathway: an expanded neural framework for the processing of object quality. Trends Cogn Sci. 2013:17(1):26–49. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kundu P, Inati SJ, Evans JW, Luh W-M, Bandettini PA. Differentiating BOLD and non-BOLD signals in fMRI time series using multi-echo EPI. NeuroImage. 2012:60(3):1759–1770. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kundu P, Brenowitz ND, Voon V, Worbe Y, Vértes PE, Inati SJ, Saad ZS, Bandettini PA, Bullmore ET. Integrated strategy for improving functional connectivity mapping using multiecho fMRI. Proc Natl Acad Sci. 2013:110(40):16187–16192. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kundu P, Voon V, Balchandani P, Lombardo MV, Poser BA, Bandettini PA. Multi-echo fMRI: a review of applications in fMRI denoising and analysis of BOLD signals. NeuroImage. 2017:154:59–80. [DOI] [PubMed] [Google Scholar]
- Lambon Ralph MA, Jefferies E, Patterson K, Rogers TT. The neural and computational bases of semantic cognition. Nat Rev Neurosci. 2017:18(1):42–55. [DOI] [PubMed] [Google Scholar]
- Leshinskaya A, Thompson-Schill SL. Transformation of event representations along middle temporal gyrus. Cereb Cortex. 2020:30(5):3148–3166. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lestou V, Lam JM, Humphreys K, Kourtzi Z, Humphreys GW. A dorsal visual route necessary for global form perception: evidence from neuropsychological fMRI. J Cogn Neurosci. 2014:26(3):621–634. [DOI] [PubMed] [Google Scholar]
- Liuzzi AG, Bruffaerts R, Dupont P, Adamczuk K, Peeters R, De Deyne S, Storms G, Vandenberghe R. Left perirhinal cortex codes for similarity in meaning between written words: comparison with auditory word input. Neuropsychologia. 2015:76:4–16. [DOI] [PubMed] [Google Scholar]
- Marinkovic K, Dhond RP, Dale AM, Glessner M, Carr V, Halgren E. Spatiotemporal dynamics of modality-specific and supramodal word processing. Neuron. 2003:38(3):487–497. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Martin CB, Douglas D, Newsome RN, Man LL, Barense MD. Integrative and distinctive coding of visual and conceptual object features in the ventral visual stream. elife. 2018:7:e31873. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McLaren DG, Ries ML, Xu G, Johnson SC. A generalized form of context-dependent psychophysiological interactions (gPPI): a comparison to standard approaches. NeuroImage. 2012:61(4):1277–1286. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mikolov T, Chen K, Corrado G, Dean J. Efficient estimation of word representations in vector space. 2013: arXiv preprint arXiv:13013781. 10.48550/arXiv.1301.3781. [DOI]
- Mirman D, Landrigan J-F, Britt AE. Taxonomic and thematic semantic systems. Psychol Bull. 2017:143(5):499–520. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mishkin M, Ungerleider LG, Macko KA. Object vision and spatial vision: two cortical pathways. Trends Neurosci. 1983:6:414–417. [Google Scholar]
- Morales M, Patel T, Tamm A, Pickering MJ, Hoffman P. Similar neural networks respond to coherence during comprehension and production of discourse. Cereb Cortex. 2022:32(19):4317–4330. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Morton NW, Schlichting ML, Preston AR. Representations of common event structure in medial temporal lobe and frontoparietal cortex support efficient inference. Proc Natl Acad Sci. 2020:117(47):29338–29345. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Muckli L, Kohler A, Kriegeskorte N, Singer W. Primary visual cortex activity along the apparent-motion trace reflects illusory perception. PLoS Biol. 2005:3(8):e265. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mumford JA, Turner BO, Ashby FG, Poldrack RA. Deconvolving BOLD activation in event-related designs for multivoxel pattern classification analyses. NeuroImage. 2012:59(3):2636–2643. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mumford JA, Davis T, Poldrack RA. The impact of study design on pattern estimation for single-trial multivariate pattern analysis. NeuroImage. 2014:103:130–138. [DOI] [PubMed] [Google Scholar]
- Naspi L, Hoffman P, Devereux B, Morcom AM. Perceptual and semantic representations at encoding contribute to true and false recognition of objects. J Neurosci. 2021:41(40):8375–8389. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Noppeney U, Josephs O, Kiebel S, Friston KJ, Price CJ. Action selectivity in parietal and temporal cortex. Cogn Brain Res. 2005:25(3):641–649. [DOI] [PubMed] [Google Scholar]
- Noppeney U, Price CJ, Penny WD, Friston KJ. Two distinct neural mechanisms for category-selective responses. Cereb Cortex. 2006:16(3):437–445. [DOI] [PubMed] [Google Scholar]
- O’Reilly JX, Woolrich MW, Behrens TE, Smith SM, Johansen-Berg H. Tools of the trade: psychophysiological interactions and functional connectivity. Soc Cogn Affect Neurosci. 2012:7(5):604–609. [DOI] [PMC free article] [PubMed] [Google Scholar]
- O'Craven KM, Kanwisher N. Mental imagery of faces and places activates corresponding stimulus-specific brain regions. J Cogn Neurosci. 2000:12(6):1013–1023. [DOI] [PubMed] [Google Scholar]
- Oosterhof NN, Connolly AC, Haxby JV. CoSMoMVPA: multi-modal multivariate pattern analysis of neuroimaging data in Matlab/GNU Octave. Front Neuroinform. 2016:10:27. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Patterson K, Nestor PJ, Rogers TT. Where do you know what you know? The representation of semantic knowledge in the human brain. Nat Rev Neurosci. 2007:8(12):976–987. [DOI] [PubMed] [Google Scholar]
- Peelen MV, Caramazza A. Conceptual object representations in human anterior temporal cortex. J Neurosci. 2012:32(45):15728–15736. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pereira F, Gershman S, Ritter S, Botvinick M. A comparative evaluation of off-the-shelf distributed semantic representations for modelling behavioural data. Cogn Neuropsychol. 2016:33(3–4):175–190. [DOI] [PubMed] [Google Scholar]
- Price AR, Bonner MF, Peelle JE, Grossman M. Converging evidence for the neuroanatomic basis of combinatorial semantics in the angular gyrus. J Neurosci. 2015:35(7):3276–3284. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pylkkänen L. The neural basis of combinatory syntax and semantics. Science. 2019:366(6461):62–66. [DOI] [PubMed] [Google Scholar]
- Ranganath C, Ritchey M. Two cortical systems for memory-guided behaviour. Nat Rev Neurosci. 2012:13(10):713–726. [DOI] [PubMed] [Google Scholar]
- Rice GE, Hoffman P, Lambon Ralph MA. Graded specialization within and between the anterior temporal lobes. Ann N Y Acad Sci. 2015a:1359(1):84–97. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rice GE, Lambon Ralph MA, Hoffman P. The roles of left versus right anterior temporal lobes in conceptual knowledge: an ALE meta-analysis of 97 functional neuroimaging studies. Cereb Cortex. 2015b:25(11):4374–4391. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rice GE, Hoffman P, Binney RJ, Lambon Ralph MA. Concrete versus abstract forms of social concept: an fMRI comparison of knowledge about people versus social terms. Phil Trans R Soc B: Biol Sci. 2018:373(1752):20170136. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Richter FR, Cooper RA, Bays PM, Simons JS. Distinct neural mechanisms underlie the success, precision, and vividness of episodic memory. Elife. 2016:5:e18260. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rogers TT, McClelland JL. Semantic cognition: a parallel distributed processing approach. Cambridge, Massachusetts: MIT press; 2004. [DOI] [PubMed] [Google Scholar]
- Rogers TT, Cox CR, Lu Q, Shimotake A, Kikuchi T, Kunieda T, Miyamoto S, Takahashi R, Ikeda A, Matsumoto R, et al. Evidence for a deep, distributed and dynamic code for animacy in human ventral anterior temporal cortex. elife. 2021:10:e66276. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Russell C, Davies S, Li K, Musil A-S, Malhotra PA, Williams AL. Self-perspective in episodic memory after parietal damage and in healthy ageing. Neuropsychologia. 2019:124:171–181. [DOI] [PubMed] [Google Scholar]
- Schwartz MF, Kimberg DY, Walker GM, Brecher A, Faseyitan OK, Dell GS, Mirman D, Coslett HB. Neuroanatomical dissociation for taxonomic and thematic knowledge in the human brain. Proc Natl Acad Sci. 2011:108(20):8520–8524. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Serre T, Wolf L, Bileschi S, Riesenhuber M, Poggio T. Robust object recognition with cortex-like mechanisms. IEEE Trans Pattern Anal Mach Intell. 2007:29(3):411–426. [DOI] [PubMed] [Google Scholar]
- Shattuck DW, Mirza M, Adisetiyo V, Hojatkashani C, Salamon G, Narr KL, Poldrack RA, Bilder RM, Toga AW. Construction of a 3D probabilistic atlas of human cortical structures. NeuroImage. 2008:39(3):1064–1080. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shimotake A, Matsumoto R, Ueno T, Kunieda T, Saito S, Hoffman P, Kikuchi T, Fukuyama H, Miyamoto S, Takahashi R, et al. Direct exploration of the role of the ventral anterior temporal lobe in semantic memory: cortical stimulation and local field potential evidence from subdural grid electrodes. Cereb Cortex. 2015:25(10):3802–3817. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Smith SM, Nichols TE. Threshold-free cluster enhancement: addressing problems of smoothing, threshold dependence and localisation in cluster inference. NeuroImage. 2009:44(1):83–98. [DOI] [PubMed] [Google Scholar]
- Stelzer J, Chen Y, Turner R. Statistical inference and multiple testing correction in classification-based multi-voxel pattern analysis (MVPA): random permutations and cluster size control. NeuroImage. 2013:65:69–82. [DOI] [PubMed] [Google Scholar]
- Sterzer P, Haynes J-D, Rees G. Primary visual cortex activation on the path of apparent motion is mediated by feedback from hMT+/V5. NeuroImage. 2006:32(3):1308–1316. [DOI] [PubMed] [Google Scholar]
- Swallow KM, Barch DM, Head D, Maley CJ, Holder D, Zacks JM. Changes in events alter how people remember recent information. J Cogn Neurosci. 2011:23(5):1052–1064. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thierry G, Price CJ. Dissociating verbal and nonverbal conceptual processing in the human brain. J Cogn Neurosci. 2006:18(6):1018–1028. [DOI] [PubMed] [Google Scholar]
- Thierry G, Giraud A-L, Price C. Hemispheric dissociation in access to the human semantic system. Neuron. 2003:38(3):499–506. [DOI] [PubMed] [Google Scholar]
- Tyler LK, Chiu S, Zhuang J, Randall B, Devereux BJ, Wright P, Clarke A, Taylor KI. Objects and categories: feature statistics and object processing in the ventral stream. J Cogn Neurosci. 2013:25(10):1723–1735. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Vandenberghe R, Price C, Wise R, Josephs O, Frackowiak RS. Functional anatomy of a common semantic system for words and pictures. Nature. 1996:383(6597):254–256. [DOI] [PubMed] [Google Scholar]
- van Heuven WJB, Mandera P, Keuleers E, Brysbaert M. SUBTLEX-UK: a new and improved word frequency database for British English. Q J Exp Psychol. 2014:67(6):1176–1190. [DOI] [PubMed] [Google Scholar]
- Visser M, Lambon Ralph M. Differential contributions of bilateral ventral anterior temporal lobe and left anterior superior temporal gyrus to semantic processes. J Cogn Neurosci. 2011:23(10):3121–3131. [DOI] [PubMed] [Google Scholar]
- Wager TD, Smith EE. Neuroimaging studies of working memory. Cogn Affect Behav Neurosci. 2003:3(4):255–274. [DOI] [PubMed] [Google Scholar]
- Wang X, Wang B, Bi Y. Close yet independent: dissociation of social from valence and abstract semantic dimensions in the left anterior temporal lobe. Hum Brain Mapp. 2019:40(16):4759–4776. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Warriner AB, Kuperman V, Brysbaert M. Norms of valence, arousal, and dominance for 13,915 English lemmas. Behav Res Methods. 2013:45(4):1191–1207. [DOI] [PubMed] [Google Scholar]
- Watson CE, Chatterjee A. The functional neuroanatomy of actions. Neurology. 2011:76(16):1428–1434. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Watson CE, Cardillo ER, Ianni GR, Chatterjee A. Action concepts in the brain: an activation likelihood estimation meta-analysis. J Cogn Neurosci. 2013:25(8):1191–1205. [DOI] [PubMed] [Google Scholar]
- Wright P, Randall B, Clarke A, Tyler LK. The perirhinal cortex and conceptual processing: effects of feature-based statistics following damage to the anterior temporal lobes. Neuropsychologia. 2015:76:192–207. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wu W, Hoffman P. Age differences in the neural processing of semantics, within and beyond the core semantic network. Neurobiol Aging. 2023:131:88–105. [DOI] [PubMed] [Google Scholar]
- Wu W, Wang X, Wei T, He C, Bi Y. Object parsing in the left lateral occipitotemporal cortex: whole shape, part shape, and graspability. Neuropsychologia. 2020:138:107340. [DOI] [PubMed] [Google Scholar]
- Wurm MF, Caramazza A. Two ‘what’ pathways for action and object recognition. Trends Cogn Sci. 2022:26(2):103–116. [DOI] [PubMed] [Google Scholar]
- Xu Y, Wang X, Wang X, Men W, Gao J-H, Bi Y. Doctor, teacher, and stethoscope: neural representation of different types of semantic relations. J Neurosci. 2018:38(13):3303–3317. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yazar Y, Bergström ZM, Simons JS. Continuous theta burst stimulation of angular gyrus reduces subjective recollection. PLoS One. 2014:9(10):e110414. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yazar Y, Bergström ZM, Simons JS. Reduced multimodal integration of memory features following continuous theta burst stimulation of angular gyrus. Brain Stimul. 2017:10(3):624–629. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yeshurun Y, Nguyen M, Hasson U. The default mode network: where the idiosyncratic self meets the shared social world. Nat Rev Neurosci. 2021:22(3):181–192. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zacks JM, Speer NK, Swallow KM, Maley CJ. The brain’s cutting-room floor: segmentation of narrative cinema. Front Hum Neurosci. 2010:4:168. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang Y, Lemarchand R, Asyraff A, Hoffman P. Representation of motion concepts in occipitotemporal cortex: fMRI activation, decoding and connectivity analyses. NeuroImage. 2022:259:119450. [DOI] [PubMed] [Google Scholar]
- Zhang Y, Mirman D, Hoffman P. Taxonomic and thematic relations rely on different types of semantic features: evidence from an fMRI meta-analysis and a semantic priming study. Brain Lang. 2023:242:105287. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Data and code supporting this study are available as follows: Neuroimaging data: https://doi.org/10.7488/ds/7521. Other data and analysis code: https://osf.io/mn8ft/. Group effect maps: https://neurovault.org/collections/NSWUEOPG/

