Significance
Knowledge about other people is critical for group survival and may have unique cognitive processing demands. Here, we investigate how person knowledge is represented, organized, and retrieved in the brain. We show that the anterior temporal lobe (ATL) stores abstract person identity representation that is commonly embedded in multiple sources (e.g. face, name, scene, and personal object). We also found the ATL serves as a “neural switchboard,” coordinating with a network of other brain regions in a rapid and need-specific way to retrieve different aspects of biographical information (e.g., occupation and personality traits). Our findings endorse the ATL as a central hub for representing and retrieving person knowledge.
Keywords: person knowledge, anterior temporal lobe, person identity node, semantic memory, social neuroscience
Abstract
Social behavior is often shaped by the rich storehouse of biographical information that we hold for other people. In our daily life, we rapidly and flexibly retrieve a host of biographical details about individuals in our social network, which often guide our decisions as we navigate complex social interactions. Even abstract traits associated with an individual, such as their political affiliation, can cue a rich cascade of person-specific knowledge. Here, we asked whether the anterior temporal lobe (ATL) serves as a hub for a distributed neural circuit that represents person knowledge. Fifty participants across two studies learned biographical information about fictitious people in a 2-d training paradigm. On day 3, they retrieved this biographical information while undergoing an fMRI scan. A series of multivariate and connectivity analyses suggest that the ATL stores abstract person identity representations. Moreover, this region coordinates interactions with a distributed network to support the flexible retrieval of person attributes. Together, our results suggest that the ATL is a central hub for representing and retrieving person knowledge.
As social creatures, it is essential that we develop a rich storehouse of knowledge about other members of our social network, such as who they are, how they look and sound, where they live, and what they do for a living. However, little is known about how and where such “person knowledge” is represented, stored, and retrieved in the brain. This inquiry is challenging because person knowledge is highly multimodal and multifaceted, being linked to both abstract features such as personality and social status as well as more concrete features such as eye color; in addition, familiar individuals are associated with detailed episodic and semantic memories (e.g., memories of shared experiences and biographic information) (1, 2). The neural circuit for person knowledge must therefore have the ability to combine multiple sources of information into an abstract representation accessible from multiplicative cues.
An influential theory by Burton and Bruce (3) proposes that person recognition is achieved through a hierarchical process that begins with the activation of modality-specific recognition units that selectively respond to the presence of a known face, name, or voice. This information is then sent to an amodal person identity node (PIN) that integrates information from the modality-specific recognition units into a multimodal representation for that individual. Excitation of the PIN ultimately allows the retrieval of person-specific semantic information independently of stimulus modality (4, 5). A similar design is embedded in the “hub-and-spoke” theory of semantic knowledge, which proposes that different features of a concept (such as its color or taste) are distributed throughout the brain (the “spokes”) and that a centralized “hub” integrates these features into a coherent, modality-invariant concept (6–8).
Here, we tested the hypothesis that a region in the anterior temporal lobe (ATL) has a function akin to a person identity node (1, 9, 10), subserving access to abstract person identity representations that can be retrieved from multiple cues, such as when you see a photo of Graceland and it cues your knowledge of Elvis (study 1). Next, we asked whether the ATL acts as a neural switchboard, performing in concert with other brain regions to enable the retrieval of different facets of person knowledge in a flexible and context-appropriate manner (study 2). We focus on the ATL because multiple lines of evidence from neuropsychology, electrophysiology, and neuroimaging have documented the critical role of the ATL in person identification (4, 5, 11–16), person-related learning (10, 17–21), semantic memory (6–8), and abstract social knowledge (1, 22–33). Individuals with ATL damage due to resection or stroke have multimodal person recognition deficits (34), lose access to stored knowledge about familiar people (35, 36), and have difficulties learning information about new people (4, 22, 37, 38). A subregion of the ATL contains a face-sensitive patch, first identified in monkeys, and more recently in humans (9).
Across two fMRI studies, 50 participants learned biographical information about a group of fictitious people for 2 d and then completed a person memory test in the MRI scanner on day 3. In study 1, we used stimuli from different categories commonly associated with familiar people (e.g., faces, names, homes, and objects) to cue memories for specific individuals (Fig. 1A). We compared the similarity of response patterns elicited by stimuli from different categories but associated with the same individual to test the abstract person representation properties of the ATL. In study 2, we asked participants to recollect specific content of fictitious people’s biographies (Fig. 1B) and examined whether the ATL coordinates the retrieval of different aspects of person knowledge by recruiting the activation of different brain regions depending on task requirements.
In study 1, findings from a series of multivariate pattern analyses (MVPAs) consistently suggested that portions of the ATL represent person knowledge in an abstract form that is divorced from the ground state of a person’s face or name. Specifically, the ATL contains multivoxel patterns for facial identity that are similarly sensitive to identity accessed through a person’s name, an image of their home, or even an object associated with that person’s occupation. In study 2, we found that the ATL person identity node is embedded in a neural circuit that is consistently engaged during person memory tasks. Multivariate analyses suggested that different content areas of person knowledge—social status, personality traits, and identity—were represented in discrete nodes within this distributed person identification circuit. Connectivity analyses further revealed that the ATL may serve as a “neural switchboard” and is capable of coordinating the flow of person-specific information between sensory brain regions that encode incoming cues and other nodes of this circuit that are engaged when retrieving specific person knowledge content.
Results
Study 1: The ATL Is a Person Identity Node.
In study 1, we used two-way MVPA cross-category (CC) classification (39–42) to identify regions in which abstract and category-invariant person information is represented. The classifier was first trained to discriminate individual identities using one category of cues (e.g., that person’s face) and subsequently tested on a different category of cues (e.g., that person’s name), and vice versa (Materials and Methods). In this way, only the abstract conceptual information that was general to both stimulus categories was informative to the classifier. Three different decoding approaches were performed: regions of interest (ROI-based), combinatorial (43, 44), and whole-brain searchlight (11).
In the ROI-based analyses, we tested whether identity-specific information in face-selective regions can generalize across different categories of cues that have been linked to a specific person (e.g., Bill’s face ↔ Bill’s name, Bill’s face ↔ Bill’s house, Bill’s face ↔ Bill’s object). A separate functional localizer was used to identify face-selective regions in both hemispheres including the ATL, occipital face area (OFA), fusiform face area (FFA), amygdala (AMY), and orbitofrontal cortex face patch (OFC), as well as a place-selective region [i.e., parahippocampal place area (PPA)] and a control region in early visual cortex (V1) (Fig. S1A). In the current paradigm, the face-selective ATL region was the only ROI that showed above-chance CC classification accuracy for “face ↔ name” [left ATL: t(23) = 3.114, P = 0.030; right ATL: t(23) = 3.941, P = 0.006], “face ↔ house” [left ATL: t(23) = 3.952, P = 0.006; right ATL: t(23) = 4.346, P = 0.001], and “face ↔ object” [left ATL: t(23) = 3.555, P = 0.012] identity representation (Fig. 2). None of the face-selective ROIs were able to cross-classify individual identities based on “nonface” stimulus pairs (i.e., “name ↔ house,” “name ↔ object,” “house ↔ object”), likely due to the face-selective nature of these ROIs (Fig. S2).
Next, we used a combinatorial MVPA analysis to validate the unique role of the face-selective ATL region (among all face-selective ROIs) for the category-invariant representation of person identity. Combinatorial analyses allow the estimation of the unique information carried within an ROI (43, 44). For every ROI in the present analysis, we calculated the average increase (or decrease) in CC classification accuracy when paired with other ROIs, referred to as the unique combinatorial performance (UCP). A positive UCP means that an ROI carries information that contributes to classification beyond what is obtained from other ROIs. A negative UCP means that information within an ROI is largely redundant with that of other ROIs. As expected, the results suggested that only the ATL displayed significantly positive UCP for face ↔ name [UCP = 1.027, t(5) = 2.879, P = 0.018], face ↔ house [UCP = 2.083, t(5) = 4.471, P = 0.004], and face ↔ object [UCP = 1.273, t(5) = 2.063, P = 0.047] identity representation (Fig. 2).
Finally, we conducted searchlight analyses to explore whether any regions in the brain beyond our face-selective ROIs can represent person identity across stimulus categories. Consistent with the previous two approaches, significant clusters in the ATL for face ↔ name, face ↔ house, and face ↔ object CC classification (Fig. 2) were revealed in this analysis. Although we did not find any regions that represented identity across name ↔ house, name ↔ object, and house ↔ object at the standard threshold, further exploratory analyses suggest that the ATL was the only subthreshold region showing CC properties for all nonface CC pairs when using at a lenient threshold (Fig. S2 and Table S1).
Table S1.
Region | Hemisphere | Voxels | T | PFWE-corr | x | y | z |
Face ↔ name cross-category classification | |||||||
Cerebellum | Midline | 6 | 7.55 | 0.002 | 6 | −85 | −32 |
Anterior temporal lobe | R | 3 | 7.14 | 0.015 | 57 | −19 | −23 |
Face ↔ house cross-category classification | |||||||
Anterior temporal lobe | L | 19 | 8.76 | 0.001 | −57 | −16 | −29 |
Face ↔ object cross-category classification | |||||||
Anterior temporal lobe | L | 2 | 6.62 | 0.038 | −45 | 8 | −29 |
Name ↔ house cross-category classification* | |||||||
Inferior temporal sulcus | R | 21 | 6.34 | 0.066 | 48 | −58 | −8 |
Inferior parietal lobe | R | 21 | 5.91 | 0.145 | 51 | −52 | 43 |
Anterior temporal lobe | R | 12 | 5.36 | 0.360 | 57 | 8 | −26 |
Name ↔ object cross-category classification* | |||||||
Anterior temporal lobe | L | 22 | 6.49 | 0.053 | −36 | 14 | −23 |
House ↔ object cross-category classification* | |||||||
Anterior temporal lobe | L | 14 | 5.43 | 0.311 | −60 | −10 | −26 |
For exploratory purposes, three cross-classification conditions (marked with an asterisk) used a lenient voxel-level threshold of P < 0.0001 (uncorrected).
Taken together, the CC MVPA analyses in study 1 suggest that the ATL may be the only brain region that holds an abstract representation of person identity that can be accessed via multiple stimulus categories. Previous studies using face or voice stimuli have found amodal representations of person identity in unimodal and multimodal face-processing regions, including the FFA, superior temporal sulcus, and the ATL (5, 16, 42, 45). In the present paradigm, the only region consistently demonstrating classification of identity across all stimulus types was the ATL. In addition, we found that the bilateral ATLs engage in cross-modal representation, whereas hemispheric asymmetry has been previously reported in the ATLs for unimodal stimuli (1, 31, 46, 47).
Study 2: The ATL Is the Hub of Flexible Person Memory Retrieval.
Univariate analyses of the neural circuit for person memory.
In study 2, we first used univariate analyses to identify brain regions that were strongly engaged when retrieving information associated with an individual (Fig. S3). The contrast of “all memory conditions > baseline” revealed a set of brain regions that included the ATL, inferior parietal lobe (IPL), posterior cingulate cortex (PCC), and hippocampus (HIPP). A similar set of brain regions was found for the contrasts of “status memory > baseline” and “trait memory > baseline.” When comparing memory conditions cued by two different categories (face vs. name), we found that the right fusiform gyrus (i.e., FFA) was activated more during face-cued memory, whereas left fusiform gyrus [also known as the visual word form area (VWFA)] was activated more during name-cued memory.
MVPA analyses for person knowledge representations.
The univariate analyses delineated a network of brain regions (ROIs) important for person knowledge retrieval (i.e., left ATL, right FFA, left VWFA, left IPL, PCC, and left HIPP) (Fig. S1B). Next, we used MVPA analyses to examine whether distinct regions within this network represent different aspects of knowledge associated with a person (e.g., that person’s status, personality traits, or identity). As knowledge content associated with an individual per se should be invariant to changes in the stimulus category used to cue that person, we collapsed our MVPA analysis of memory condition across the two cue categories (i.e., face and name). Here, we used the three decoding approaches used in study 1 and a “leave-one-run-out” cross-validation (CV) classification scheme to identify brain regions sensitive to different aspects of person knowledge.
For status and personality-trait knowledge, ROI-based analyses suggested that the IPL was the only region within the person-knowledge network that accurately represented learned people’s occupational status [i.e., manager vs. janitor; t(25)=4.500, P < 0.001], whereas the PCC was the only region that accurately represented trait information [i.e., whether a person was extroverted vs. introverted; t(25) = 3.192, P = 0.012] (Fig. 3). Combinatorial analyses confirmed these findings by showing the only positive UCP in IPL for status information [UCP = 2.372, t(4) = 3.767, P = 0.01] and the highest positive UCP in PCC for personality trait knowledge [UCP = 2.628, t(4) = 3.982, P = 0.008]. Searchlight analyses further identified significant clusters in IPL for status classification and PCC for trait classification (Table S2). These consistent results are in line with previous neuroimaging studies suggesting a critical role of the IPL in representing social status (48, 49) and of the PCC in representing personality traits (50–52).
Table S2.
Region | Hemisphere | Voxels | T | PFWE-corr | x | y | z |
Social status (i.e., manager vs. janitor) | |||||||
Cuneus | Midline | 101 | 10.49 | <0.001 | −21 | −70 | 10 |
169 | 8.81 | <0.001 | −3 | −91 | 19 | ||
16 | 7.19 | 0.010 | −18 | −97 | 22 | ||
Postcentral gyrus | L | 112 | 9.08 | <0.001 | −48 | −22 | 61 |
Inferior parietal lobe | L | 97 | 9.04 | <0.001 | −42 | −46 | 37 |
7.73 | <0.001 | −30 | −64 | 43 | |||
Inferior temporal gyrus | L | 21 | 8.32 | 0.001 | −48 | −70 | −2 |
Personality trait (i.e., extrovert vs. introvert) | |||||||
Cuneus | Midline | 432 | 11.08 | <0.001 | 21 | −94 | 4 |
Precentral gyrus | L | 206 | 10.80 | <0.001 | −42 | −10 | 46 |
Posterior cingulate cortex | Midline | 18 | 9.24 | <0.001 | 12 | −31 | 46 |
Person identity (eight fictitious people) | |||||||
Cuneus | Midline | 661 | 12.26 | <0.001 | 21 | −94 | 4 |
Anterior temporal lobe | L | 46 | 8.92 | <0.001 | −54 | −16 | −29 |
Inferior frontal gyrus | L | 32 | 8.54 | <0.001 | −54 | −4 | 25 |
Temporoparietal junction | L | 225 | 8.46 | <0.001 | −45 | −73 | 34 |
R | 22 | 8.38 | <0.001 | 45 | −73 | 40 |
For identity knowledge, we observed that the ATL was the only region that could accurately distinguish the eight learned people in ROI-based analyses [t(25) = 3.166, P = 0.012], combinatorial analyses [UCP = 1.811, t(4) = 5.232, P = 0.003], and searchlight analyses (Fig. 3 and Table S2). Along with the results of study 1, our MVPA analyses provide convergent evidence suggesting that the ATL is a critical region for person identity representation, regardless of whether the decoding was performed across categories (e.g., face/name/scene/object in study 1) or with categories collapsed together (e.g., face/name in study 2).
Overall, the MVPA analyses in study 2 suggest that a distributed network represents specific content of person knowledge: status information was represented in the IPL, trait information in the PCC, and, consistent with the results of study 1, person identity was represented in the ATL.
Functional connectivity analyses for neural dynamics of person memory.
To elucidate the underlying neural dynamics supporting person knowledge retrieval, we performed a psychophysiological interaction (PPI) analysis to examine functional connectivity between nodes of the person-memory network during the retrieval of person-specific knowledge. PPI measures task-dependent interactions between different brain regions and by examining the similarity of activity patterns (“connectivity”) between a seed and other brain areas as a function of specific task demands (53). We took a comprehensive approach to defining the seed regions for our PPI analysis using regions identified by our univariate and MVPA analyses as being important for the distributed representation of person-related knowledge (Fig. S1B). We defined three types of seeds for the PPI analyses: sensory cue regions (FFA, VWFA), knowledge representation regions (IPL, PCC), and the ATL (Fig. S4).
Because information generally flows from posterior to more anterior regions during person perception (54), we first used sensory regions (FFA or VWFA) as seeds to explore where in the brain the sensory input is forwarded to, and then used regions in association cortex that may be involved in knowledge representation (IPL or PCC) as seeds to examine which brain area(s) are engaged in retrieving knowledge content. The results of our PPI analysis suggest that sensory regions had stronger coupling with the IPL and ATL during the retrieval of a person’s status, but with the PCC and ATL during the retrieval of a person’s personality traits (Fig. S4A). Knowledge representation regions showed enhanced coupling with the FFA and ATL when cued by faces, but with the VWFA and ATL when cued by names (Fig. S4B). The hippocampus also showed increased coupling with these two groups of seed regions, but not in all conditions (Fig. 4A and Tables S3 and S4).
Table S3.
Region | Hemisphere | Voxels | T | PFWE-corr | x | y | z |
FFA as seed region: Face status > face baseline | |||||||
Precuneus | Midline | 449 | 10.03 | <0.001 | 3 | −61 | 19 |
8.26 | <0.001 | 6 | −43 | 31 | |||
Middle occipital gyrus | L | 215 | 8.82 | <0.001 | −42 | −73 | −11 |
R | 130 | 8.53 | <0.001 | 48 | −67 | −14 | |
19 | 6.50 | 0.007 | 39 | −85 | 1 | ||
Fusiform gyrus | L | 215 | 7.32 | <0.001 | −51 | −55 | −14 |
R | 130 | 7.01 | <0.001 | 45 | −52 | −17 | |
Amygdala | L | 427 | 8.28 | <0.001 | −30 | 5 | −11 |
R | 427 | 6.78 | <0.001 | 24 | 2 | −11 | |
Posterior cingulate cortex | Midline | 67 | 7.60 | 0.001 | 9 | −7 | 40 |
Anterior temporal lobe | R | 40 | 6.81 | 0.004 | 60 | −7 | −20 |
L | 45 | 6.61 | 0.006 | −51 | −10 | −29 | |
Precentral gyrus | R | 109 | 6.69 | 0.005 | 36 | −16 | 43 |
Inferior frontal gyrus | L | 32 | 6.67 | 0.005 | −48 | 29 | 13 |
R | 12 | 6.45 | 0.008 | 48 | 5 | 34 | |
Inferior parietal lobe | L | 83 | 6.65 | 0.005 | −30 | −49 | 52 |
6.58 | 0.006 | −21 | −61 | 49 | |||
FFA as seed region: Face trait > face baseline | |||||||
Inferior frontal gyrus | L | 88 | 13.95 | <0.001 | −45 | 23 | 28 |
R | 112 | 13.91 | <0.001 | 36 | 8 | 25 | |
Hippocampus | R | 50 | 11.00 | <0.001 | 27 | −19 | −17 |
L | 10 | 9.44 | <0.001 | −21 | −10 | −26 | |
Precuneus and posterior cingulate cortex | Midline | 404 | 11.71 | <0.001 | −3 | −52 | 28 |
10.39 | <0.001 | −6 | −34 | 43 | |||
Middle occipital gyrus | R | 19 | 11.55 | <0.001 | 42 | −73 | 7 |
35 | 10.54 | <0.001 | 45 | −61 | −11 | ||
L | 102 | 10.40 | <0.001 | −36 | −88 | −8 | |
Superior temporal gyrus | R | 19 | 11.46 | <0.001 | 48 | −34 | 10 |
Precentral gyrus | L | 195 | 11.41 | <0.001 | −30 | −28 | 70 |
R | 153 | 10.10 | <0.001 | 27 | −31 | 70 | |
Fusiform gyrus | R | 117 | 11.16 | <0.001 | 39 | −46 | −23 |
Amygdala | L | 37 | 10.51 | <0.001 | −18 | −4 | −11 |
Inferior parietal lobe | R | 31 | 10.43 | <0.001 | 30 | −58 | 43 |
L | 68 | 9.61 | <0.001 | −27 | −67 | 40 | |
Anterior temporal lobe | L | 6 | 10.32 | <0.001 | −54 | 5 | −14 |
8 | 9.81 | <0.001 | −63 | −13 | −8 | ||
Insula | L | 17 | 9.79 | <0.001 | −27 | 23 | −2 |
Orbitofrontal cortex | R | 156 | 6.17 | <0.001 | 33 | 23 | −11 |
Anterior cingulate cortex | Midline | 70 | 5.52 | 0.020 | 12 | 20 | 28 |
VWFA as seed region: Name status > name baseline | |||||||
Middle occipital gyrus | L | 48 | 8.44 | <0.001 | −42 | −73 | −11 |
R | 12 | 6.41 | 0.007 | 42 | −70 | −17 | |
Inferior parietal lobe | L | 127 | 7.27 | 0.001 | −51 | −40 | 40 |
6.83 | 0.003 | −36 | −61 | 40 | |||
Inferior frontal gyrus | L | 74 | 6.78 | 0.003 | −48 | 14 | 22 |
R | 12 | 6.74 | 0.004 | 51 | 26 | 22 | |
Anterior temporal lobe | L | 15 | 6.89 | 0.003 | −48 | 5 | −38 |
Cuneus | Midline | 18 | 6.71 | 0.004 | −12 | −103 | 4 |
Middle temporal gyrus | L | 77 | 6.66 | 0.004 | −63 | −43 | −8 |
R | 10 | 6.30 | 0.009 | 63 | −49 | −8 | |
Precentral gyrus | L | 17 | 6.24 | 0.010 | −45 | −16 | 40 |
VWFA as seed region: Name trait > name baseline | |||||||
Insula | R | 452 | 11.72 | <0.001 | 33 | −22 | 19 |
L | 10 | 7.86 | <0.001 | −36 | −25 | 22 | |
Inferior parietal lobe | L | 593 | 11.71 | <0.001 | −36 | −55 | 43 |
R | 160 | 9.15 | <0.001 | 45 | −49 | 31 | |
Posterior cingulate cortex, extended to precentral gyrus and precuneus | Midline | 1,300 | 11.42 | <0.001 | 27 | −25 | 55 |
11.29 | <0.001 | −24 | −31 | 52 | |||
11.20 | <0.001 | 12 | −25 | 52 | |||
9.99 | <0.001 | −6 | −55 | 40 | |||
Superior temporal gyrus | L | 593 | 9.77 | <0.001 | −57 | −49 | 16 |
9.39 | <0.001 | −57 | −31 | −2 | |||
R | 75 | 7.91 | <0.001 | 57 | −31 | 10 | |
Anterior temporal lobe | R | 109 | 9.28 | <0.001 | 60 | −10 | −23 |
7.08 | 0.002 | 51 | 8 | −29 | |||
L | 14 | 8.74 | <0.001 | −51 | 5 | −35 | |
Fusiform gyrus | L | 12 | 7.85 | <0.001 | −27 | −46 | −26 |
Cuneus | Midline | 37 | 7.80 | 0.001 | 24 | −91 | 19 |
Cerebellum | Midline | 34 | 7.55 | 0.001 | −6 | −73 | −41 |
Inferior occipital gyrus | L | 10 | 7.46 | 0.001 | −36 | −79 | −11 |
Precentral gyrus | L | 18 | 7.38 | 0.001 | −42 | −16 | 31 |
Table S4.
Region | Hemisphere | Voxels | T | PFWE-corr | x | y | z |
IPL as seed region: Face status > face baseline | |||||||
Middle occipital gyrus | R | 389 | 10.24 | <0.001 | 33 | −85 | 16 |
L | 137 | 8.88 | <0.001 | −27 | −85 | 16 | |
Inferior parietal lobe | L | 154 | 9.62 | <0.001 | −36 | −49 | 43 |
R | 82 | 8.50 | <0.001 | 33 | −46 | 52 | |
Anterior temporal lobe | L | 17 | 9.61 | <0.001 | −42 | 11 | −41 |
R | 15 | 7.12 | 0.003 | 54 | 2 | −29 | |
Orbitofrontal cortex | R | 144 | 9.39 | <0.001 | 33 | 35 | −14 |
L | 15 | 6.79 | 0.005 | −24 | 26 | −17 | |
Midline | 54 | 9.04 | <0.001 | 15 | 14 | −14 | |
Fusiform gyrus | R | 120 | 9.29 | <0.001 | 45 | −52 | −29 |
L | 115 | 6.90 | 0.004 | −36 | −61 | −26 | |
Cerebellum | R | 28 | 9.11 | <0.001 | 45 | −73 | −35 |
Inferior frontal gyrus | R | 65 | 9.01 | <0.001 | 45 | 8 | 37 |
Precuneus | Midline | 130 | 8.46 | <0.001 | −6 | −58 | 25 |
Medial prefrontal cortex | Midline | 17 | 8.22 | <0.001 | −6 | 38 | 43 |
22 | 6.79 | 0.005 | −12 | 53 | −2 | ||
Middle temporal gyrus | L | 57 | 8.08 | <0.001 | −60 | −31 | −17 |
Cuneus | Midline | 48 | 6.98 | 0.004 | −9 | −94 | 1 |
IPL as seed region: Name status > name baseline | |||||||
Inferior frontal gyrus | L | 71 | 9.95 | <0.001 | −42 | 17 | 40 |
226 | 8.31 | <0.001 | −48 | 23 | 4 | ||
R | 37 | 7.24 | 0.002 | 42 | 32 | 28 | |
64 | 7.09 | 0.002 | 39 | 41 | −8 | ||
Inferior parietal lobe | L | 381 | 8.92 | <0.001 | −42 | −55 | 46 |
R | 112 | 8.77 | <0.001 | 42 | −52 | 46 | |
Cerebellum | R | 114 | 8.63 | <0.001 | 33 | −76 | −35 |
L | 104 | 8.59 | <0.001 | −33 | −76 | −35 | |
Anterior temporal lobe and middle temporal gyrus | L | 171 | 8.42 | <0.001 | −60 | −25 | −20 |
7.35 | 0.001 | −60 | −40 | −8 | |||
6.00 | 0.021 | −57 | −16 | −29 | |||
R | 36 | 6.35 | 0.010 | 57 | −37 | −8 | |
12 | 6.26 | 0.012 | 60 | −13 | −26 | ||
Medial frontal gyrus | Midline | 45 | 8.06 | <0.001 | −9 | 26 | 49 |
Precentral gyrus | L | 161 | 7.82 | 0.001 | −30 | −16 | 61 |
Precuneus | Midline | 246 | 7.68 | 0.001 | −9 | −46 | 28 |
Medial orbitofrontal cortex | Midline | 13 | 6.34 | 0.010 | −3 | 41 | −11 |
Fusiform gyrus | L | 104 | 6.07 | 0.018 | −36 | −64 | −20 |
PCC as seed region: Face trait > face baseline | |||||||
Hippocampus | L | 31 | 10.76 | <0.001 | −21 | −19 | −23 |
Anterior temporal lobe and superior temporal gyrus | R | 67 | 10.01 | <0.001 | 45 | −10 | −11 |
6.70 | 0.004 | 45 | −7 | −20 | |||
L | 119 | 9.56 | <0.001 | −45 | −13 | −11 | |
7.79 | <0.001 | −54 | −1 | −11 | |||
Posterior cingulate cortex | Midline | 145 | 9.59 | <0.001 | −9 | −28 | 40 |
61 | 8.42 | <0.001 | 9 | −46 | 43 | ||
Parahippocampal gyrus | L | 210 | 9.35 | <0.001 | −24 | −43 | −14 |
Fusiform gyrus | R | 230 | 8.89 | <0.001 | 36 | −40 | −20 |
Precentral gyrus | L | 87 | 8.99 | <0.001 | −24 | −28 | 58 |
Supramarginal gyrus | R | 101 | 8.90 | <0.001 | 60 | −28 | 22 |
L | 49 | 8.60 | <0.001 | −48 | −28 | 22 | |
Superior parietal lobule | R | 17 | 8.44 | <0.001 | 21 | −55 | 58 |
L | 23 | 8.09 | <0.001 | −21 | −49 | 58 | |
PCC as seed region: Name trait > name baseline | |||||||
Insula | R | 445 | 9.77 | <0.001 | 27 | 11 | −2 |
9.42 | <0.001 | 36 | −19 | −2 | |||
L | 110 | 8.72 | <0.001 | −51 | −16 | 4 | |
Anterior temporal lobe and superior temporal gyrus | R | 400 | 9.57 | <0.001 | 54 | −13 | −17 |
6.25 | 0.009 | 57 | −4 | −23 | |||
L | 98 | 8.72 | <0.001 | −51 | −16 | 4 | |
6.29 | 0.009 | −51 | −13 | −2 | |||
Fusiform gyrus | L | 261 | 8.98 | <0.001 | −30 | −40 | −11 |
7.57 | 0.001 | −39 | −55 | −11 | |||
R | 59 | 7.63 | 0.001 | 30 | −46 | −14 | |
Thalamus | Midline | 147 | 8.60 | <0.001 | −6 | −13 | 1 |
Posterior cingulate cortex | Midline | 155 | 7.75 | <0.001 | −12 | −34 | 58 |
151 | 7.39 | 0.001 | 15 | −37 | 55 | ||
Cuneus | Midline | 56 | 7.23 | 0.001 | 18 | −76 | −5 |
Hippocampus | R | 24 | 7.52 | 0.001 | 24 | −16 | −26 |
Middle occipital gyrus | R | 23 | 6.62 | 0.004 | 48 | −58 | 4 |
We also used the ATL as a seed region to more specifically test whether its functional connectivity with other nodes of the person-memory network changed as a function of task demands. The ATL showed increased connectivity with the IPL during the retrieval of a person’s status, but with the PCC during the retrieval of a person’s personality traits. The functional connectivity of the ATL with the FFA increased during memory retrieval when cued by faces, but with the VWFA when cued by names (Fig. S4C and Table S5). The functional connectivity between the ATL and the hippocampus did not change in any conditions.
Table S5.
Region | Hemisphere | Voxels | T | PFWE-corr | x | y | z |
Status memory > baseline: (face–status, name–status) > (face–baseline, name–baseline) | |||||||
Inferior parietal lobe | L | 317 | 9.82 | <0.001 | −42 | −40 | 37 |
109 | 6.69 | 0.007 | −36 | −58 | 43 | ||
Precuneus | Midline | 120 | 8.95 | <0.001 | −9 | −67 | 37 |
Supplementary motor area | Midline | 35 | 8.39 | <0.001 | 0 | −4 | 61 |
Middle temporal gyrus | L | 41 | 7.71 | 0.001 | −51 | −22 | −14 |
Superior temporal gyrus | R | 27 | 7.67 | 0.001 | 54 | −37 | 22 |
L | 11 | 6.74 | 0.006 | −57 | −52 | 22 | |
Fusiform gyrus | L | 46 | 7.63 | 0.001 | −33 | −52 | −20 |
R | 10 | 7.01 | 0.003 | 30 | −46 | −17 | |
Thalamus | Midline | 24 | 7.60 | 0.001 | −15 | −16 | 7 |
Cuneus | Midline | 10 | 6.71 | 0.006 | 12 | −91 | 22 |
Trait memory > baseline: (face–trait, name–trait) > (face–baseline, name–baseline) | |||||||
Fusiform gyrus | R | 341 | 9.27 | <0.001 | 27 | −79 | −5 |
7.88 | <0.001 | 30 | −46 | −20 | |||
L | 377 | 9.27 | <0.001 | −30 | −82 | −11 | |
8.02 | <0.001 | −30 | −55 | −23 | |||
Middle temporal gyrus, extended to superior temporal gyrus and inferior frontal gyrus | L | 1,254 | 8.88 | <0.001 | −42 | −25 | −8 |
8.80 | <0.001 | −48 | 5 | 7 | |||
R | 51 | 8.85 | <0.001 | 57 | −22 | −11 | |
Temporoparietal junction | R | 415 | 8.03 | 0.001 | 54 | −40 | 19 |
7.29 | 0.002 | 63 | −52 | 16 | |||
Anterior temporal lobe | L | 14 | 7.15 | 0.003 | −48 | 8 | −38 |
Posterior cingulate cortex | Midline | 99 | 6.96 | 0.004 | −6 | −43 | 58 |
6.26 | 0.017 | 6 | −40 | 52 | |||
20 | 6.92 | 0.005 | 0 | −19 | 49 | ||
Postcentral gyrus | L | 25 | 6.81 | 0.006 | −42 | −28 | 61 |
Inferior frontal gyrus | R | 36 | 6.68 | 0.007 | 51 | 29 | 7 |
Precuneus | Midline | 19 | 6.61 | 0.009 | −9 | −73 | 40 |
16 | 6.48 | 0.011 | −27 | −76 | 31 | ||
Face memory > face baseline | |||||||
Middle temporal gyrus, extended to superior temporal gyrus, postcentral gyrus and inferior parietal lobe | L | 1,113 | 9.76 | <0.001 | −48 | −19 | −11 |
9.59 | <0.001 | −60 | −13 | 22 | |||
8.92 | <0.001 | −54 | −19 | 1 | |||
8.30 | <0.001 | −36 | −34 | 55 | |||
7.75 | 0.001 | −51 | −49 | 25 | |||
61 | 8.42 | <0.001 | 9 | −46 | 43 | ||
Superior temporal gyrus | R | 535 | 9.49 | <0.001 | 54 | −40 | 22 |
Cuneus | R | 273 | 9.26 | <0.001 | 12 | −85 | −2 |
Fusiform gyrus | R | 154 | 8.91 | <0.001 | 42 | −46 | −23 |
Precuneus | Midline | 66 | 8.09 | 0.001 | −9 | −70 | 37 |
Supplementary motor area and posterior cingulate cortex | Midline | 129 | 7.82 | 0.001 | −3 | −7 | 61 |
6.61 | 0.011 | 0 | −19 | 37 | |||
Name memory > name baseline | |||||||
Fusiform gyrus, extended to inferior parietal lobe and precuneus | L | 1,192 | 9.25 | <0.001 | −27 | −46 | −20 |
7.68 | 0.001 | −36 | −67 | 31 | |||
7.19 | 0.002 | −36 | −52 | 34 | |||
6.95 | 0.003 | −42 | −73 | −11 | |||
6.57 | 0.007 | −6 | −73 | 43 | |||
R | 86 | 7.61 | 0.001 | 27 | −43 | −20 | |
32 | 6.61 | 0.006 | 30 | −70 | −5 | ||
Anterior temporal lobe | L | 35 | 8.82 | <0.001 | −42 | 8 | −41 |
Precentral gyrus, extended to posterior cingulate cortex | L | 264 | 7.74 | 0.001 | −42 | −22 | 61 |
6.51 | 0.008 | −9 | −40 | 55 | |||
5.80 | 0.033 | −3 | −49 | 37 | |||
22 | 6.09 | 0.019 | −3 | −19 | 55 | ||
Inferior frontal gyrus | L | 234 | 7.66 | 0.001 | −51 | 11 | 25 |
Middle temporal gyrus | L | 210 | 7.48 | 0.001 | −48 | −25 | −8 |
Thalamus | Midline | 37 | 7.24 | 0.002 | −18 | −19 | 7 |
Insula | L | 163 | 7.17 | 0.002 | −33 | −13 | 4 |
Supplementary motor area | Midline | 29 | 6.13 | 0.017 | −3 | −4 | 67 |
In sum, the results of our PPI analyses suggest that functional activity in secondary sensory regions (i.e., FFA, VWFA), knowledge representation regions in association cortex (i.e., IPL, PCC), and the ATL is tightly coupled during the retrieval of person-specific information, and that different regions within this circuit are differentially coupled depending on the content of the knowledge being retrieved. In contrast, the hippocampus, a region that has an undisputed role in the formation of episodic memories (55), does not appear to interact with other regions during the retrieval of person knowledge (Fig. 4A).
Dynamic causal modeling to test the “ATL-hub” theory.
PPI analysis does not provide information about the direction of causal influences between source and target regions, nor whether the connectivity is mediated by other regions. To gain traction on these issues, we performed dynamic causal modeling (DCM) to explore detailed information processing dynamics within the aforementioned network. We tested three models that might support the implementation of person knowledge retrieval. In model 1, retrieval is implemented by a direct link between sensory regions and knowledge representation regions, which could be formed by associative learning during the 2-d training session that preceded scanning. Alternatively, in model 2, the ATL acts as a “switchboard-like” hub to flexibly coordinate the flow of information when retrieving different content of person knowledge. In model 3, the hippocampus, rather than the ATL, acts as a domain-general memory hub (Materials and Methods and Fig. S5 for details regarding the model specifications).
Bayesian model selection clearly selected model 2 as the optimal model (Fig. 4B): it has the highest relative group log evidence in fixed-effects analysis (71.29) as well as the highest exceedance probability in random-effects analysis (59.23%). An analysis of absolute model fits confirmed that model 2 provides the highest explained variance among the three models, accounting for 24 ± 6% (mean ± SD) of the observed variance. Therefore, our DCM support a neural mechanism of person memory in which the ATL serves as the central hub coordinating the person-information retrieval from discrete sources.
Discussion
The results of the two studies reported here suggest that the ATL plays a critical role in representing and retrieving person knowledge. Using MVPA CC classification, study 1 provided strong evidence that the ATL face patch has a function akin to a person identity node, representing abstract conceptually invariant person identity information. Research in nonhuman primates has shown that cells in anterior–ventral temporal cortex are highly sensitive to particular facial identities as well as to facial familiarity (56, 57). Previous studies in humans using intracranial recording (16) or fMRI analyses (15, 42, 45) have suggested that the ATL can distinguish between different people using their faces, voices, or names. Our study extends these findings by using sophisticated multivariate analyses and a wider range of stimulus categories. Based on our results, we suggest that the ATL may be the only region that can merge such a wide variety of multicategorical person information.
The results of study 2 shed light onto the neural architecture of person knowledge. Study 2 provides strong support for the idea that person knowledge is organized in a hub-and-spoke manner such that specific features of person knowledge are neurally distributed and portions of the ATL serve as a convergence zone or hub (6–8). MVPA analyses suggested that status and personality trait information are stored separately in portions of the IPL and PCC, respectively, and PPI analyses showed that the ATL acts in concert with these regions to flexibly retrieve different aspects of person knowledge. Moreover, when a particular person memory was cued by a face, there was enhanced coupling between the FFA and ATL, whereas if the same memory was cued by a name, there was enhanced coupling between the visual word form area and the ATL. Critically, we performed DCM to directly test the hub-and-spoke accounts (i.e., model 2 and 3) against an alternative “distributed-only” account (5) (i.e., model 1), and the ATL-hub model was dominantly favored in the comparison.
Functional coupling as measured by PPI and DCM does not necessitate a direct anatomical connection between regions (53). However, the ATL is viewed as a convergence zone precisely because of its unusual pattern of white matter connectivity (58). It receives direct input from the ventral visual stream, which includes the FFA and the VWFA, via the inferior longitudinal fasciculus. Portions of the ATL are also directly connected to the PCC via a limbic pathway, the cingulum bundle, whereas superior aspects of the ATL are structurally interconnected with the IPL by a pathway most closely associated with language, the middle longitudinal fasciculus (59). Thus, the results of the PPI and DCM analyses rest on a verified neurostructural ground truth.
The hub-and-spoke model is a popular account of general semantic knowledge (6–8), and our DCM findings generally support that model. However, one aspect of our findings is challenging for this model. The hub-and-spoke model predicts that, within the ATL, there should be a single person identity node that links all features of all stimuli associated with a particular person. Across ROI-based and searchlight analyses in study 1, we did find that the ATL contains convergence zones (60) that bind different categories of cues related to an identity; however, we were unable to identify a single site—a “master node”—within the ATL that linked all four categories (faces, houses, names, and objects) and survived all six CC pairings. These findings are consistent with neuropsychological evidence that patients with ATL lesions often have person identification deficits from one or two modalities (46, 61), but no patient has been shown to have impairments in all (three or more) modalities at the same time (4). Moreover, our ATL ROI seemed to be biased toward faces: we found CC classifications only worked well for face-related pairings (i.e., face ↔ name, face ↔ house, and face ↔ object; Fig. 2), but not for nonface pairing (i.e., name ↔ house, name ↔ object, and house ↔ object; Fig. S2). One possibility is that known faces serve as singular categories (62), grouping and linking the bits of biographical knowledge and descriptors that define a person, but in a manner that is inherently asymmetrical. Another possibility is that our training regime created a person representation biased toward faces.
It is important to consider that biographical knowledge bears all of the signatures of semantic memory with little to none of the signatures of episodic memory, such as spatiotemporal context. It has been proposed that the learning of semantic associations can bypass the hippocampus entirely and, instead, rely on the ATL (63). Discrete hippocampal damage slows the learning rate of semantic associations but does not impair it completely (64). Indeed, individuals with hippocampal damage can learn new semantic associations through a type of incidental learning called “fast mapping,” whereas patients with ATL lesions cannot (65). Consistent with this small literature, our results suggest that the hippocampus plays a nonsignificant role in the retrieval of biographical information. The results of our univariate analysis showed that the hippocampus was active across all memory conditions so it was included in our network analyses. However, the functional connectivity analysis suggested that it had only minimal interactions with other brain regions during person knowledge retrieval (Fig. 4A), and DCM ruled out any critical role of the hippocampus in person knowledge retrieval (Fig. 4B). We speculate that the hippocampus may be important during the initial acquisition of semantic associations, but within a short amount of time, it is no longer needed and, instead, the ATL becomes the nexus for retrieval (21). Thus, the ATL, which is geographically proximal to the hippocampus, and structurally interconnected with the hippocampus, may play a vital role in orchestrating the cascade of neural events that ultimately result in the retrieval of person knowledge.
There are several limitations of the current investigation that should be noted. First, across the two studies reported here, we tested different groups of subjects and defined “ATL” ROIs in separate ways (i.e., “faces > houses” or “memory > baseline”). We are, therefore, unable to examine whether these ROIs in the separate groups correspond to the same anatomical region. Visual inspection of the each individual’s ROIs in both groups suggest that these regions fall in roughly the same portion of the ATLs (Fig. S1); however, whether the same population of neurons in the ATL of a given individual has both amodal and hub-like properties remains an open question. Second, we only tested the ATL-hub model for two person attributes (i.e., status and trait). It will be interesting for future researchers to assess whether the ATL similarly coordinates the retrieval of other attributes, such as physical (i.e., eye color, body shape) or evaluative characteristics (i.e., likeability, reputation). If our model is generalizable, one would expect to see increased connectivity between the ATL and area V4 during the retrieval of eye color, extrastriate body area during the retrieval of body characteristics, and orbitofrontal cortex during the retrieval of social-evaluative information. Third, the present study only focused on a subset of social knowledge—the biographical information about others. Although biographical information is important, several other types of social knowledge contribute to social behavior, including knowledge about other people’s mental states and goals, semantic knowledge about people that is not specific to particular individuals (e.g., people need to eat, people tend to like chocolate), and knowledge about social etiquette and norms (66). Future research is needed to elucidate the neural architectures supporting each knowledge type and how they relate to one another.
Materials and Methods
Subjects.
A total of 24 subjects (9 males; Mage = 23.21) participated in study 1, and a new cohort of 26 subjects (21 males; Mage = 20.38) participated in study 2. All subjects were native English speakers, right-handed, and had no history of psychological or neurological disorders. They were financially compensated and gave informed consent in a manner approved by the Institutional Review Board of Temple University. The sample size (n ∼ 25) was chosen based on previous studies using similar training paradigms and analytic approaches (10, 41, 67, 68).
Materials.
Study 1.
Biographies of four fictitious males were provided during training sessions (Fig. 1A). Each male’s face was professionally photographed (i.e., full-frontal, neutral expressions, uniform lighting/background) and was linked to six pieces of biographical information: a name, age, marital status, occupation, city of residence, and a house. The first five pieces of biographical information were presented as words and house was presented as a color image, procured from publically available sources on the Internet. Proper names were taken from the social security database of popular names (https://www.ssa.gov/cgi-bin/popularnames.cgi) and each had two syllables. A critical manipulation was made purposely on biographies: each male’s face, name, house, and occupation were distinctive (“unique bios”), whereas their age, marital status, and city of residence could be shared (“common bios”) with one other person. Biographical information was randomly assigned to each face such that each participant learned different associations for any given face.
In the fMRI session, a new set of stimuli was constructed for each fictitious male (Fig. 1B). Specifically, the learned faces/houses in training sessions were rephotographed from five different camera angles (0°, ±45°, and ±90°), and the learned names were presented in five different fonts/colors. In object runs, participants viewed daily-used objects that were representative of each male’s occupation from five different viewpoints (e.g., if the fictitious male was a doctor, a stethoscope was shown).
Study 2.
Biographies of eight fictitious males were provided in training sessions (Fig. 1C). Participants learned four pieces of information about each male: their face, name, occupation, and one personality trait. Eight distinct faces were generated by FaceGen Modeler 3.5, which allowed us to control low-level visual features such as color, brightness, and illumination. Eight artificial names were adapted from a previous study (69). The attractiveness and likeability of these artificial faces and names were carefully controlled by asking a different group of 20 participants to assign ratings. Social status and personality traits were manipulated in a 2 × 2 factorial design: each fictitious male was associated with an occupation title and salary that suggested high or low status (e.g., “Maur is a manager earning more than $9,000 per month” or “Lorc is a janitor earning less than $1,000 per month”) and with personality trait descriptions that denoted high or low sociability (e.g., “Jora is an introverted, quiet bookworm” or “Gris is an extroverted, chatty party animal”). In real life, there are infinite attributes associated with an individual (e.g., attractiveness, reputation, temper, athleticism, etc.); here, we only manipulated “social status” and “sociability traits” because they are the two principal dimensions of person concepts and impression formation (70). Again, biographical information was randomly combined such that each participant learned different occupation/personality characteristics of any given face/name.
In the fMRI session, the eight learned faces or names were presented during the “memory” conditions (Fig. 1D). In addition, we generated eight novel faces/names for the “nonmemory” baseline conditions. These faces/names matched the learned faces/names in regard of low-level visual features, phonemic properties (e.g., start with the same consonant), and subjective ratings of attractiveness/likeability. Participants could easily identify them as unfamiliar during the practice session right before the fMRI session.
Procedure.
Behavioral training sessions.
Participants were told to learn biographical information about four males in study 1 (Fig. 1A) and eight males in study 2 (Fig. 1C). The training protocol was adapted from the literature (10, 71). Training was conducted over 2 d, with the first day session lasting ∼30 min, and the second day session lasting 20 min. During each training session, participants first completed “show” trials in which they viewed slides containing a face image, a name, a house (only in study 1), and a short paragraph containing other biographical information. Each slide was presented three times. There was no time limit on slide presentation; participants pressed a button when they wished to move to the next screen. Next, participants completed “naming” trials in which they viewed a previously learned face and were asked to type that person’s name. After responding, participants were told whether they were correct or incorrect, and the correct biographical information for that individual was presented again on the screen. For naming trials, each male was presented six times in a random order. Finally, participants completed “matching” trials in which they were presented either with learned faces or names and were asked to select the corresponding biographical information (e.g., occupation) from all options presented below (e.g., pilot/mailman/photographer/doctor in study 1; manager/janitor in study 2). The matching phase consisted of blocks of 40 trials in study 1 (or 32 trials in study 2) and participants received correctness feedback after each trial as well as an accumulative accuracy at the end of each block. Participants were considered fully trained if they performed a matching block with accuracy over 95% in day 1 and 100% in day 2. If participants did not reach this level of accuracy, they had to do extra matching blocks until they reached the appropriate level (e.g., overlearning ensured). On day 3, all participants were retested using a paper and pencil test immediately before the scan, to ensure that everyone could accurately recall the learned biographical information.
fMRI session in study 1.
The fMRI session was scheduled on day 3 and consisted of two parts: two runs of a functional localizer and eight runs of the “person memory” task. Both tasks used a block design. In the localizer task, participants were instructed to pay attention to the images and respond whenever the same image was presented twice in a row (one-back task). Each run consisted of 30 blocks, evenly divided between three alternating stimulus types: “faces,” “places,” or “fixation cross.” Each stimulus was presented for 800 ms followed by a 200-ms interstimulus interval. All faces and places stimuli were adapted from a previous study in our laboratory (10).
In the person memory task, participants were asked to retrieve knowledge about the four males they had learned about during training (Fig. 1B). Participants first observed a series of stimuli related to a unique aspect of the “target” male’s identity (i.e., their face, name, house, or occupation) and then responded to a question regarding a common aspect of that male’s biography (i.e., their age, city of residence, or marital status). For instance, after viewing five images of Tyler’s house (presented at different vantage points), participants had to respond to a question regarding Tyler’s age: “True or False: the person associated with this house is 28”?
The person memory task consisted of eight runs (i.e., two runs for each category). Only one category (face/name/house/object) was consistently used to cue target males across the whole run. Each run consisted of 16 blocks of stimulus presentation (4 males × 4 repetitions) and each block contained five cue stimuli related to the same individual (3 s each) and one true/false question about that person (9 s). Note that face/house stimuli were only displayed in full-frontal view during training, whereas during the fMRI session the face/house cue stimuli were presented from five different vantage points. Similarly, the name cue stimuli here were presented in five different fonts and colors. These manipulations aimed to vary the low-level visual properties of each male’s cue stimuli while holding the high-level person identity constant, and also avoid visual habituation (56). The object cue stimuli were related to each male’s occupation. Five pictures depicting a daily-used object associated with each male’s occupation were presented. Note that these objects were not shown during training, because occupation information was presented in written form during training. They were used here to ensure participants thoroughly learned each male and their associated biographical information. The cue stimuli order (within a block), the target male order (within a run), and the category order (across runs) alternated and were randomized across participants.
fMRI session in study 2.
Participants completed six runs of a person memory task in the scanner on day 3 (Fig. 1D). In each run, we adopted a 3 × 2 factorial block-design where the factors were “task type” (status memory, trait memory, or nonmemory baseline) and “cue category” (name or face of the target male). In status/trait memory blocks, participants viewed faces or names of the eight learned males and were asked to indicate their social status (manager/janitor) or personality trait (extroverted/introverted). In the “nonmemory baseline” blocks, participants viewed eight novel faces/names they never learned before and were asked to indicate the spatial location of the cue stimuli appearing on the screen (left/right). At the beginning of each block, participants were informed of the upcoming task type (presented for 3 s). Each run consisted of 18 blocks (i.e., 3 task types × 2 cue types × 3 repetitions), with eight trials in each block to cover the whole set of faces/names. To make the memory and nonmemory conditions similar in terms of cognitive demand, participants had to respond in 4 s for status/trait trials but in 2 s for baseline trials. The target male order (within a block) and the task order (within a run) alternated and were randomized across participants.
Data Acquisition and Preprocessing.
The fMRI session was conducted at the Temple University Hospital on a 3-T Siemens Verio scanner, equipped with a 12-channel head coil. In study 1, functional images were acquired using a gradient-echo echo-planar imaging (EPI) sequence [repetition time (TR) = 3,000 ms; echo time (TE) = 20 ms; field of view = 240 × 240; matrix size = 80 × 80; flip angle = 90°]. Sixty-one interleaved slices (3 × 3 × 2.5-mm voxels) were acquired aligned to 30° tilted from the anterior commissure–posterior commissure line, with full brain coverage. These imaging parameters (i.e., short TE, tilted slices) were optimized for mitigating susceptibility artifacts around ATL and OFC (72), and were validated by pilot scans as well as previous studies in the laboratory (10, 28). In study 2, we adopted the same pulse sequence except the TR (2,000 ms) and 40 interleaved slices (3 × 3 × 3.5-mm voxels). In study 1 the visual stimulus was delivered by E-Prime software; in study 2, stimuli were delivered by Cogent toolbox running under Matlab R2014b. Responses were recorded using a four-button fiber optic response pad system.
To remove sources of noise and artifact, functional data were corrected for slice timing, realigned, unwarped, normalized to the EPI template [Montreal Neurological Institute (MNI) space, resampled to 3 × 3 × 3 mm], spatially smoothed (8 mm), high-pass filtered at 128 s, and prewhitened by means of an autoregressive model AR (1) using SPM8 software.
Univariate General Linear Model Analyses.
Subject-specific parameter estimates (β weights) for each condition were derived through a general linear model (GLM). For each subject, the data were best-fitted at every voxel using a combination of effects of interest. These were delta functions representing the onset of each of the experiment conditions, convolved with the SPM8 hemodynamic response function. The six motion regressors and memory error trials were also included as nuisance regressors. Next, subject-specific β weights were entered into a group-level random-effect GLM to allow statistical inference. Statistics maps were generated using a voxel-level familywise error (FWE)-corrected threshold of P < 0.05. Stereotaxic coordinates are reported in MNI space and regional labels were derived using the automated anatomical labeling (AAL) atlas in xjView.
In study 1, we set up one GLM for each of the four categories (face, name, house, and object) across two runs using unsmoothed data. This was prepared for subsequent MVPA CC classification. β maps for each target male (i.e., Tyler, Justin, Aaron, and Cameron) were extracted from each block of a run. Each male had four β maps per run and eight β maps in total per category.
In study 2, four different GLMs were set up across six runs, based on different analytic purposes. One GLM was set up for a univariate analysis of the main effect of task types. β maps for each task type regressor (i.e., face–status, face–trait, face–baseline, name–status, name–trait, and name–baseline) were extracted and contrasts of interest were conducted upon them (i.e., all memory > baseline, status memory > baseline, trait memory > baseline, face memory > name memory, and name memory > face memory) (Fig. S3 and Table S6). This GLM was also used for DCM analysis. For MVPA decoding of status, personality traits, and identity representations, three GLMs were set up separately using unsmoothed data. β maps for each status regressor (i.e., manager or janitor), each trait regressor (i.e., extrovert or introvert), and each learned male regressor (i.e., eight males) were extracted from status memory blocks, trait memory blocks, and all memory blocks, respectively.
Table S6.
Region | Hemisphere | Voxels | T | P FWE-corr | x | y | z |
All-memory > baseline: (face–status, face–trait, name–status, name–trait) > (face–baseline, name–baseline) | |||||||
Insula | L | 218 | 10.12 | <0.001 | −30 | 26 | 7 |
R | 84 | 6.86 | 0.003 | 33 | 17 | 1 | |
Posterior cingulate cortex | Midline | 132 | 8.97 | <0.001 | 15 | −28 | 52 |
Precuneus | Midline | 52 | 8.47 | <0.001 | −9 | −73 | 37 |
Cerebellum | Midline | 75 | 8.08 | <0.001 | −12 | −46 | −17 |
Anterior cingulate cortex | Midline | 75 | 7.75 | 0.001 | −6 | 23 | 31 |
Inferior frontal gyrus | L | 218 | 7.13 | 0.002 | −36 | 23 | 16 |
Anterior temporal lobe | L | 12 | 6.75 | 0.004 | −51 | 2 | −14 |
Inferior parietal lobe | L | 11 | 6.56 | 0.006 | −27 | −61 | 40 |
Hippocampus | L | 11 | 6.47 | 0.007 | −36 | −34 | −5 |
Status memory > baseline: (face–status, name–status) > (face–baseline, name–baseline) | |||||||
Insula | L | 157 | 9.48 | <0.001 | −30 | 26 | 7 |
R | 40 | 6.07 | 0.016 | 33 | 20 | −2 | |
Posterior cingulate cortex | Midline | 129 | 8.81 | <0.001 | 15 | −28 | 52 |
Precuneus | Midline | 41 | 7.83 | <0.001 | −9 | −73 | 37 |
Cerebellum | Midline | 48 | 7.38 | 0.001 | −12 | −46 | −17 |
Inferior parietal lobe | L | 18 | 7.33 | 0.001 | −27 | −61 | 40 |
Anterior cingulate cortex | Midline | 36 | 7.31 | 0.001 | −6 | 23 | 31 |
Superior temporal gyrus | R | 11 | 6.92 | 0.003 | 45 | −16 | −5 |
Inferior frontal gyrus | L | 157 | 6.82 | 0.003 | −42 | 20 | 28 |
Anterior temporal lobe | R | 10 | 6.68 | 0.005 | 51 | −4 | −11 |
L | 6 | 6.37 | 0.009 | −51 | 2 | −14 | |
Hippocampus | L | 5 | 5.99 | 0.019 | −33 | −34 | −5 |
Trait memory > baseline: (face–trait, name–trait) > (face–baseline, name–baseline) | |||||||
Insula | L | 228 | 9.67 | <0.001 | −30 | 26 | 7 |
R | 117 | 7.62 | <0.001 | 33 | 17 | 1 | |
Precuneus | Midline | 57 | 8.97 | <0.001 | −9 | −73 | 37 |
Anterior cingulate cortex | Midline | 91 | 7.94 | <0.001 | −9 | 20 | 43 |
Posterior cingulate cortex | Midline | 125 | 8.33 | <0.001 | 15 | −28 | 52 |
Cerebellum | Midline | 82 | 8.21 | <0.001 | −12 | −46 | −17 |
Inferior frontal gyrus | L | 228 | 7.24 | 0.002 | −36 | 23 | 16 |
Superior temporal gyrus | L | 11 | 6.80 | 0.004 | −60 | −34 | 10 |
Hippocampus | L | 18 | 6.76 | 0.004 | −36 | −34 | −5 |
Anterior temporal lobe | L | 7 | 6.67 | 0.005 | −51 | 2 | −14 |
Face memory > name memory: (face–status, face–trait) > (name–status, name–trait) | |||||||
Fusiform gyrus | R | 119 | 9.95 | <0.001 | 39 | −58 | −14 |
8.09 | <0.001 | 39 | −79 | −11 | |||
Cuneus | Midline | 70 | 7.96 | <0.001 | −12 | −97 | −5 |
Inferior occipital gyrus | R | 108 | 7.33 | 0.001 | 27 | −91 | −5 |
Name memory > face memory: (name–status, name–trait) > (face–status, face–trait) | |||||||
Cuneus | Midline | 219 | 11.73 | <0.001 | 6 | −79 | 28 |
205 | 8.15 | <0.001 | −6 | −82 | 19 | ||
Middle cingulate cortex | Midline | 109 | 9.58 | <0.001 | 6 | −13 | 40 |
Superior temporal gyrus | R | 67 | 8.12 | <0.001 | 48 | −22 | 22 |
12 | 8.00 | <0.001 | 60 | −61 | 16 | ||
L | 52 | 7.14 | 0.002 | −57 | −43 | 28 | |
Orbitofrontal cortex | Midline | 12 | 7.31 | 0.001 | −12 | 23 | −17 |
Medial prefrontal cortex | Midline | 47 | 7.24 | 0.002 | −12 | 53 | 28 |
Fusiform gyrus and inferior temporal gyrus | L | 154 | 7.23 | 0.002 | −57 | −61 | −8 |
6.77 | 0.004 | −54 | −46 | 1 | |||
Precuneus | Midline | 9 | 6.57 | 0.006 | 12 | −55 | 58 |
ROIs Localization.
Study 1 used an established functional localizer (10) to localize subject-specific ROIs. Face-selective regions (i.e., OFA, FFA, ATL, AMY, and OFC) were defined by individuating the peaks showing greater activity for faces than for places (“faces > places”; P < 0.05, uncorrected). A place-selective region (i.e., PPA) was defined by the opposite contrast (“places > faces”). To rule out effects driven by the low-level perceptual features of our stimuli, we additionally defined a control V1 ROI in early visual cortex around the voxel showing the greatest activation for all visual stimuli (“faces + places” > “fixation”) (Fig. S1A).
In study 2, subject-specific ROIs were functionally localized using a series of contrasts between task types in univariate analysis: ATL (“all memory > baseline”), FFA (“face memory > name memory”), VWFA (“name memory > face memory”), IPL (“status memory > baseline”), PCC (“trait memory > baseline”), and the hippocampus (“all memory > baseline”) (Fig. S1B). We stress that, even though ROI definition and subsequent MVPA analysis were based on the same data, our analyses did not suffer from circular logic (73), because the voxel selection criteria were based on “task types” in univariate analyses where all eight learned males were modeled by the same regressor whereas MVPA analysis aimed to classify each learned male. In other words, the voxel selection procedure used no information about the specific biographies of each learned male (74, 75). Similar rationale also applies to our PPI (32, 76) and DCM analysis (i.e., DCM inferences were made via model comparison, which were conditioned on prespecified regions but not biased by the selection of the regions per se) (77, 78).
A 6-mm spherical mask was generated for all ROIs (68, 79) in study 1 and study 2, centered on MNI coordinates with the highest activation within each peak (Table S7), and confirmed by anatomical AAL atlas. In study 1, we localized bilateral areas for each ROI (except midline areas such as OFC and V1) for ROI-based MVPA analyses; then we merged both hemispheres into one unified mask to represent each ROI for combinatorial MVPA analyses. In study 2, we generated a unilateral mask for each ROI because only one hemisphere survived in univariate analysis. In study 2, the same ROIs were used in subsequent MVPA, PPI, and DCM analysis.
Table S7.
Region | Hemisphere | x | y | z |
Study 1 | ||||
Early vison cortex (V1) | Midline | 0 | −89 | 4 |
Occipital face area (OFA) | R | 44 | −77 | −8 |
L | −42 | −80 | −7 | |
Fusiform face area (FFA) | R | 42 | −52 | −20 |
L | −42 | −52 | −22 | |
Anterior temporal lobe (ATL) | R | 51 | −4 | −27 |
L | −50 | −2 | −26 | |
Amygdala (AMY) | R | 21 | −6 | −18 |
L | −19 | −6 | −17 | |
Orbitofrontal cortex face patch (OFC) | Midline | −1 | 38 | −15 |
Parahippocampal place area (PPA) | R | 29 | −47 | −11 |
L | −27 | −48 | −11 | |
Study 2 | ||||
Anterior temporal lobe (ATL) | L | −50 | −2 | −16 |
Fusiform face area (FFA) | R | 39 | −55 | −18 |
Visual word form area (VWFA) | L | −45 | −56 | −12 |
Inferior parietal lobe (IPL) | L | −30 | −58 | 45 |
Posterior cingulate cortex (PCC) | Midline | 10 | −31 | 46 |
Hippocampus (HIPP) | L | −33 | −30 | −8 |
MVPA Analyses.
In both studies, we adopted three MVPA decoding approaches: ROI-based, combinatorial, and searchlight analysis. All were implemented by the Decoding Toolbox (80) using a support vector machine as a classifier. Combinatorial analysis was executed in the same way as the ROI-based approach, except that conjunction masks were created between all possible combinations of two ROIs. Searchlight analysis was executed with a radius of 4 voxels (i.e., 12 mm) across the whole brain.
In study 1, a two-way CC classification scheme was used for each subject. The classifier was first trained on one category (e.g., face runs) and subsequently tested on the other category (e.g., name runs); the reverse decoding direction was also performed (e.g., first trained on name runs and then tested on face runs). The average two-way CC classification accuracy was then calculated for each ROI for each CC pair and compared with chance performance (identity of four males = 0.25). In total, six CC classification analyses were performed in a pairwise fashion between any two of the four categories (i.e., face–name, face–house, face–object, name–house, name–object, house–object).
Study 2 used a leave-one-run-out CV scheme in which the classifier was trained on five runs of data and tested on the remaining untrained run. This procedure was repeated six times, each time using a different test run, and the average CV accuracy was calculated for each ROI and compared with chance performance (status = 0.5, personality trait = 0.5, identity of eight fictitious males = 0.125).
For group-level inference, ROI-based analysis used one-tailed t tests with a Bonferroni correction of P < 0.05. For combinatorial analysis, we used one-tailed t tests of P < 0.05 to test whether the average change of CC/CV accuracy (by combining a particular ROI with each of the rest ROIs) was significantly above 0% (i.e., increase the prediction) (43). For searchlight analysis, a voxel-level FWE-corrected threshold of P < 0.05 was used. If null results were found, a lenient voxel-level threshold (P < 0.0001, uncorrected) was then used to reveal subthreshold results for exploratory purposes (Fig. S2 and Table S1).
PPI Analyses.
We carried out PPI analysis in SPM8 (53) to examine functional coupling between the six ROIs identified using univariate analyses during the person memory task. Seed regions included sensory input ROIs (VWFA/FFA) or knowledge representations ROIs (IPL/PPC) or the ATL (identity representation); and four experimental conditions were examined (i.e., face/name × status/personality trait) vs. baseline. For instance, in one PPI analysis, we used the FFA as the seed region and explored its “coupling” brain regions during the retrieval of social status; in another PPI analysis, we set the IPL as a seed region and explored its coupling regions during face–status conditions. In total, 12 PPI analyses were performed for each subject (Fig. S4 and Tables S3–S5).
For each PPI analysis, we created a new GLM with (i) a “physiological” regressor in which the seed region’s time course (i.e., first eigenvariate) was deconvolved to estimate the underlying neural activity; (ii) a “psychological” regressor in which task type trials (vs. baseline trials) were convolved with the canonical hemodynamic response function; and (iii) a “PPI interaction” regressor in which the psychological regressor was multiplied by the physiological regressor. We used this interaction regressor to identify voxels in which functional activity covaried in a task-dependent manner with the seed region. Subject-level PPI analyses were run to generate SPM contrast images similar to a subject-level GLM model, and these contrast images were entered into a group-level random-effects GLM and thresholded at P < 0.05 (voxel-level FWE corrected).
DCM.
We performed DCM analysis using DCM10 in SPM8 (81). Three distinct, but not mutually exclusive, models were tested (Fig. 4B and Fig. S5): (i) Model 1 hypothesizes that person knowledge is retrieved via a direct link from sensory input areas (FFA/VWFA) to knowledge representation areas (IPL/PCC), with the ATL being mainly implemented in identity recognition and therefore not involved in specific knowledge content retrieval. In model 1, one can expect two separate information flows during face–status memory: “FFA→ IPL” (for status knowledge retrieval) and “FFA→ ATL” (for implicit person identification). This model is conceptually similar to the distributed-only account in semantic memory research. (ii) Model 2 assumes that the ATL serves as a hub that coordinates activity within the network during person memory retrieval: a person’s identity is first recognized in the ATL, which then directs information flow to knowledge representation areas (IPL/PCC). In model 2, the information flows for face–status memory and name–trait memory can be expected as “FFA → ATL → IPL” and “VWFA → ATL → PCC,” respectively. Note that we did not set ATL’s modulation to hippocampus in this model because MVPA analyses suggest that the hippocampus does not contain any person-knowledge representations and PPI analyses found no functional connectivity between the ATL and hippocampus during person memory retrieval. (iii) In model 3, the hippocampus is the central hub: all person knowledge (including identity) is coordinated by this domain-general hub. The information flows during face–status memory would be “FFA → hippocampus → IPL (for status knowledge)” and “FFA → hippocampus → ATL (for implicit person identification).” For each model, the six ROIs were all set to be bidirectionally connected (Fig. S5A), and their time courses (i.e., first eigenvariate) were extracted for each subject individually.
To prevent poor model fit, we calculated the mean explained variance of each model across subjects by using the SPM function “spm_dcm_fmri_check.” To determine the optimal model, fixed-effects (FFX) and random-effects (RFX) group analyses were implemented by Bayesian models selection (82). In the FFX case, one assumes that the optimal model is identical across the population. It uses group log evidence to quantify the relative goodness of models, which is the exponentiated sum of the log model evidence of each subject-specific model. Usually, a difference in a group log evidence of 3 is taken as statistically strong evidence. Thus, if the group log evidence of one model is bigger than the other models’ by 3 or more, that model would be considered by FFX analysis to be the optimal model (78).
Because the FFX analysis is vulnerable to outlier subjects, we also implemented an RFX analysis, which accounts for the heterogeneity of the model structure across subjects. It uses hierarchical Bayesian modeling that estimates the parameters of a Dirichlet distribution over the probabilities of all models considered. These probabilities define a multinomial distribution over model space, enabling the computation of the posterior probability of each model given the data of all subjects and the models considered. The results of RFX analysis are reported in terms of the exceedance probability that one model is more likely than any other model. The optimal model in RFX analysis would be considered to be the one with the largest exceedance probability.
Acknowledgments
We thank Ashley Unger for assistance in data collection, David V. Smith for helpful comments, and two anonymous reviewers for their time and effort in reviewing this manuscript. This work was supported by NIH Grant R01 MH091113 (to I.R.O.).
Footnotes
The authors declare no conflict of interest.
This article is a PNAS Direct Submission. R.S. is a Guest Editor invited by the Editorial Board.
See Commentary on page 4042.
This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1621234114/-/DCSupplemental.
References
- 1.Olson IR, McCoy D, Klobusicky E, Ross LA. Social cognition and the anterior temporal lobes: A review and theoretical framework. Soc Cogn Affect Neurosci. 2013;8(2):123–133. doi: 10.1093/scan/nss119. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Barton JJS, Corrow SL. Recognizing and identifying people: A neuropsychological review. Cortex. 2016;75:132–150. doi: 10.1016/j.cortex.2015.11.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Burton AM, Bruce V. I recognize your face but I can’t remember your name: A simple explanation? Br J Psychol. 1992;83(Pt 1):45–60. doi: 10.1111/j.2044-8295.1992.tb02424.x. [DOI] [PubMed] [Google Scholar]
- 4.Blank H, Wieland N, von Kriegstein K. Person recognition and the brain: Merging evidence from patients and healthy individuals. Neurosci Biobehav Rev. 2014;47:717–734. doi: 10.1016/j.neubiorev.2014.10.022. [DOI] [PubMed] [Google Scholar]
- 5.Perrodin C, Kayser C, Abel TJ, Logothetis NK, Petkov CI. Who is that? Brain networks and mechanisms for identifying individuals. Trends Cogn Sci. 2015;19(12):783–796. doi: 10.1016/j.tics.2015.09.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Patterson K, Nestor PJ, Rogers TT. Where do you know what you know? The representation of semantic knowledge in the human brain. Nat Rev Neurosci. 2007;8(12):976–987. doi: 10.1038/nrn2277. [DOI] [PubMed] [Google Scholar]
- 7.Martin A. The representation of object concepts in the brain. Annu Rev Psychol. 2007;58(1):25–45. doi: 10.1146/annurev.psych.57.102904.190143. [DOI] [PubMed] [Google Scholar]
- 8.Lambon Ralph MA. Neurocognitive insights on conceptual knowledge and its breakdown. Philos Trans R Soc Lond B Biol Sci. 2013;369(1634):20120392. doi: 10.1098/rstb.2012.0392. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Collins JA, Olson IR. Beyond the FFA: The role of the ventral anterior temporal lobes in face processing. Neuropsychologia. 2014;61(1):65–79. doi: 10.1016/j.neuropsychologia.2014.06.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Collins JA, Koski JE, Olson IR. More than meets the eye: The merging of perceptual and conceptual knowledge in the anterior temporal face area. Front Hum Neurosci. 2016;10(May):189. doi: 10.3389/fnhum.2016.00189. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Kriegeskorte N, Formisano E, Sorger B, Goebel R. Individual faces elicit distinct response patterns in human anterior temporal cortex. Proc Natl Acad Sci USA. 2007;104(51):20600–20605. doi: 10.1073/pnas.0705654104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Nestor A, Plaut DC, Behrmann M. Unraveling the distributed neural code of facial identity through spatiotemporal pattern analysis. Proc Natl Acad Sci USA. 2011;108(24):9998–10003. doi: 10.1073/pnas.1102433108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Von Der Heide RJ, Skipper LM, Olson IR. Anterior temporal face patches: A meta-analysis and empirical study. Front Hum Neurosci. 2013;7(February):17. doi: 10.3389/fnhum.2013.00017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Anzellotti S, Caramazza A. The neural mechanisms for the recognition of face identity in humans. Front Psychol. 2014;5(Jun):672. doi: 10.3389/fpsyg.2014.00672. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Blank H, Kiebel SJ, von Kriegstein K. How the human brain exchanges information across sensory modalities to recognize other people. Hum Brain Mapp. 2015;36(1):324–339. doi: 10.1002/hbm.22631. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Abel TJ, et al. Direct physiologic evidence of a heteromodal convergence region for proper naming in human left anterior temporal lobe. J Neurosci. 2015;35(4):1513–1520. doi: 10.1523/JNEUROSCI.3387-14.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.von Kriegstein K, Giraud AL. Implicit multisensory associations influence voice recognition. PLoS Biol. 2006;4(10):e326. doi: 10.1371/journal.pbio.0040326. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Tsukiura T, Suzuki C, Shigemune Y, Mochizuki-Kawai H. Differential contributions of the anterior temporal and medial temporal lobe to the retrieval of memory for person identity information. Hum Brain Mapp. 2008;29(12):1343–1354. doi: 10.1002/hbm.20469. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Tsukiura T, et al. Dissociable roles of the anterior temporal regions in successful encoding of memory for person identity information. J Cogn Neurosci. 2010;22(10):2226–2237. doi: 10.1162/jocn.2009.21349. [DOI] [PubMed] [Google Scholar]
- 20.Skipper LM, Ross LA, Olson IR. Sensory and semantic category subdivisions within the anterior temporal lobes. Neuropsychologia. 2011;49(12):3419–3429. doi: 10.1016/j.neuropsychologia.2011.07.033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Nieuwenhuis ILC, et al. The neocortical network representing associative memory reorganizes with time in a process engaging the anterior temporal lobe. Cereb Cortex. 2012;22(11):2622–2633. doi: 10.1093/cercor/bhr338. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Olson IR, Plotzker A, Ezzyat Y. The enigmatic temporal pole: A review of findings on social and emotional processing. Brain. 2007;130(Pt 7):1718–1731. doi: 10.1093/brain/awm052. [DOI] [PubMed] [Google Scholar]
- 23.Schurz M, Radua J, Aichhorn M, Richlan F, Perner J. Fractionating theory of mind: A meta-analysis of functional brain imaging studies. Neurosci Biobehav Rev. 2014;42:9–34. doi: 10.1016/j.neubiorev.2014.01.009. [DOI] [PubMed] [Google Scholar]
- 24.Mitchell JP, Heatherton TF, Macrae CN. Distinct neural systems subserve person and object knowledge. Proc Natl Acad Sci USA. 2002;99(23):15238–15243. doi: 10.1073/pnas.232395699. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Frith U, Frith CD. Development and neurophysiology of mentalizing. Philos Trans R Soc Lond B Biol Sci. 2003;358(1431):459–473. doi: 10.1098/rstb.2002.1218. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Zahn R, et al. Social concepts are represented in the superior anterior temporal cortex. Proc Natl Acad Sci USA. 2007;104(15):6430–6435. doi: 10.1073/pnas.0607061104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Zahn R, et al. The neural basis of human social values: Evidence from functional MRI. Cereb Cortex. 2009;19(2):276–283. doi: 10.1093/cercor/bhn080. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Ross LA, Olson IR. Social cognition and the anterior temporal lobes. Neuroimage. 2010;49(4):3452–3462. doi: 10.1016/j.neuroimage.2009.11.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Simmons WK, Reddish M, Bellgowan PSF, Martin A. The selectivity and functional connectivity of the anterior temporal lobes. Cereb Cortex. 2010;20(4):813–825. doi: 10.1093/cercor/bhp149. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Gilbert SJ, Swencionis JK, Amodio DM. Evaluative vs. trait representation in intergroup social judgments: Distinct roles of anterior temporal lobe and prefrontal cortex. Neuropsychologia. 2012;50(14):3600–3611. doi: 10.1016/j.neuropsychologia.2012.09.002. [DOI] [PubMed] [Google Scholar]
- 31.Wong C, Gallate J. The function of the anterior temporal lobe: A review of the empirical evidence. Brain Res. 2012;1449:94–116. doi: 10.1016/j.brainres.2012.02.017. [DOI] [PubMed] [Google Scholar]
- 32.Greven IM, Downing PE, Ramsey R. Linking person perception and person knowledge in the human brain. Soc Cogn Affect Neurosci. 2016;11(4):641–651. doi: 10.1093/scan/nsv148. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Pobric G, Lambon Ralph MA, Zahn R. Hemispheric specialization within the superior anterior temporal cortex for social and nonsocial concepts. J Cogn Neurosci. 2016;28(3):351–360. doi: 10.1162/jocn_a_00902. [DOI] [PubMed] [Google Scholar]
- 34.Gainotti G. Different patterns of famous people recognition disorders in patients with right and left anterior temporal lesions: A systematic review. Neuropsychologia. 2007;45(8):1591–1607. doi: 10.1016/j.neuropsychologia.2006.12.013. [DOI] [PubMed] [Google Scholar]
- 35.Ellis AW, Young AW, Critchley EMR. Loss of memory for people following temporal lobe damage. Brain. 1989;112(Pt 6):1469–1483. doi: 10.1093/brain/112.6.1469. [DOI] [PubMed] [Google Scholar]
- 36.Gainotti G, Marra C. Differential contribution of right and left temporo-occipital and anterior temporal lesions to face recognition disorders. Front Hum Neurosci. 2011;5(June):55. doi: 10.3389/fnhum.2011.00055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Tsukiura T, et al. Neural basis of the retrieval of people’s names: Evidence from brain-damaged patients and fMRI. J Cogn Neurosci. 2002;14(6):922–937. doi: 10.1162/089892902760191144. [DOI] [PubMed] [Google Scholar]
- 38.Todorov A, Olson IR. Robust learning of affective trait associations with faces when the hippocampus is damaged, but not when the amygdala and temporal pole are damaged. Soc Cogn Affect Neurosci. 2008;3(3):195–203. doi: 10.1093/scan/nsn013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Peelen MV, Atkinson AP, Vuilleumier P. Supramodal representations of perceived emotions in the human brain. J Neurosci. 2010;30(30):10127–10134. doi: 10.1523/JNEUROSCI.2161-10.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Oosterhof NN, Wiggett AJ, Diedrichsen J, Tipper SP, Downing PE. Surface-based information mapping reveals crossmodal vision-action representations in human parietal and occipitotemporal cortex. J Neurophysiol. 2010;104(2):1077–1089. doi: 10.1152/jn.00326.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Fairhall SL, Caramazza A. Brain regions that represent amodal conceptual knowledge. J Neurosci. 2013;33(25):10552–10558. doi: 10.1523/JNEUROSCI.0051-13.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Awwad Shiekh Hasan B, Valdes-Sosa M, Gross J, Belin P. “Hearing faces and seeing voices”: Amodal coding of person identity in the human brain. Sci Rep. 2016;6:37494. doi: 10.1038/srep37494. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Clithero JA, Carter RM, Huettel SA. Local pattern classification differentiates processes of economic valuation. Neuroimage. 2009;45(4):1329–1338. doi: 10.1016/j.neuroimage.2008.12.074. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Smith DV, Clithero JA, Rorden C, Karnath HO. Decoding the anatomical network of spatial attention. Proc Natl Acad Sci USA. 2013;110(4):1518–1523. doi: 10.1073/pnas.1210126110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Watson R, Latinus M, Charest I, Crabbe F, Belin P. People-selectivity, audiovisual integration and heteromodality in the superior temporal sulcus. Cortex. 2014;50:125–136. doi: 10.1016/j.cortex.2013.07.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Liu RR, et al. Voice recognition in face-blind patients. Cereb Cortex. 2016;26(4):1473–1487. doi: 10.1093/cercor/bhu240. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Gainotti G. Cognitive models of familiar people recognition and hemispheric asymmetries. Front Biosci (Elite Ed) 2014;6(1):148–158. doi: 10.2741/e698. [DOI] [PubMed] [Google Scholar]
- 48.Chiao JY. Neural basis of social status hierarchy across species. Curr Opin Neurobiol. 2010;20(6):803–809. doi: 10.1016/j.conb.2010.08.006. [DOI] [PubMed] [Google Scholar]
- 49.Koski JE, Xie H, Olson IR. Understanding social hierarchies: The neural and psychological foundations of status perception. Soc Neurosci. 2015;10(5):527–550. doi: 10.1080/17470919.2015.1013223. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Hassabis D, et al. Imagine all the people: How the brain creates and uses personality models to predict behavior. Cereb Cortex. 2014;24(8):1979–1987. doi: 10.1093/cercor/bht042. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Van Overwalle F, Ma N, Baetens K. Nice or nerdy? The neural representation of social and competence traits. Soc Neurosci. 2015;11(6):567–78. doi: 10.1080/17470919.2015.1120239. [DOI] [PubMed] [Google Scholar]
- 52.Heleven E, Van Overwalle F. The person within: Memory codes for persons and traits using fMRI repetition suppression. Soc Cogn Affect Neurosci. 2016;11(1):159–171. doi: 10.1093/scan/nsv100. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Friston KJ, et al. Psychophysiological and modulatory interactions in neuroimaging. Neuroimage. 1997;6(3):218–229. doi: 10.1006/nimg.1997.0291. [DOI] [PubMed] [Google Scholar]
- 54.Barbeau EJ, et al. Spatio temporal dynamics of face recognition. Cereb Cortex. 2008;18(5):997–1009. doi: 10.1093/cercor/bhm140. [DOI] [PubMed] [Google Scholar]
- 55.Eichenbaum H, Cohen NJ. Can we reconcile the declarative memory and spatial navigation views on hippocampal function? Neuron. 2014;83(4):764–770. doi: 10.1016/j.neuron.2014.07.032. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Eifuku S, De Souza WC, Nakata R, Ono T, Tamura R. Neural representations of personally familiar and unfamiliar faces in the anterior inferior temporal cortex of monkeys. PLoS One. 2011;6(4):e18913. doi: 10.1371/journal.pone.0018913. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Freiwald WA, Tsao DY. Functional compartmentalization and viewpoint generalization within the macaque face-processing system. Science. 2010;330(6005):845–851. doi: 10.1126/science.1194908. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Bajada CJ, et al. The tract terminations in the temporal lobe: Their location and associated functions. Cortex. April 1, 2016 doi: 10.1016/j.cortex.2016.03.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Makris N, et al. Human middle longitudinal fascicle: Segregation and behavioral-clinical implications of two distinct fiber connections linking temporal pole and superior temporal gyrus with the angular gyrus or superior parietal lobule using multi-tensor tractography. Brain Imaging Behav. 2013;7(3):335–352. doi: 10.1007/s11682-013-9235-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Tranel D. Impaired naming of unique landmarks is associated with left temporal polar damage. Neuropsychology. 2006;20(1):1–10. doi: 10.1037/0894-4105.20.1.1. [DOI] [PubMed] [Google Scholar]
- 61.Gainotti G. Is the right anterior temporal variant of prosopagnosia a form of “associative prosopagnosia” or a form of “multimodal person recognition disorder”? Neuropsychol Rev. 2013;23(2):99–110. doi: 10.1007/s11065-013-9232-7. [DOI] [PubMed] [Google Scholar]
- 62.Tanaka JW. The entry point of face recognition: Evidence for face expertise. J Exp Psychol Gen. 2001;130(3):534–543. doi: 10.1037//0096-3445.130.3.534. [DOI] [PubMed] [Google Scholar]
- 63.Merhav M, Karni A, Gilboa A. Not all declarative memories are created equal: Fast mapping as a direct route to cortical declarative representations. Neuroimage. 2015;117:80–92. doi: 10.1016/j.neuroimage.2015.05.027. [DOI] [PubMed] [Google Scholar]
- 64.Gardiner JM, Brandt KR, Baddeley AD, Vargha-Khadem F, Mishkin M. Charting the acquisition of semantic knowledge in a case of developmental amnesia. Neuropsychologia. 2008;46(11):2865–2868. doi: 10.1016/j.neuropsychologia.2008.05.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Sharon T, Moscovitch M, Gilboa A. Rapid neocortical acquisition of long-term arbitrary associations independent of the hippocampus. Proc Natl Acad Sci USA. 2011;108(3):1146–1151. doi: 10.1073/pnas.1005238108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Adolphs R. The social brain: Neural basis of social knowledge. Annu Rev Psychol. 2009;60:693–716. doi: 10.1146/annurev.psych.60.110707.163514. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Peelen MV, Caramazza A. Conceptual object representations in human anterior temporal cortex. J Neurosci. 2012;32(45):15728–15736. doi: 10.1523/JNEUROSCI.1953-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Axelrod V, Yovel G. Successful decoding of famous faces in the fusiform face area. PLoS One. 2015;10(2):e0117126. doi: 10.1371/journal.pone.0117126. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Rezlescu C, Barton JJS, Pitcher D, Duchaine B. Normal acquisition of expertise with greebles in two cases of acquired prosopagnosia. Proc Natl Acad Sci USA. 2014;111(14):5123–5128. doi: 10.1073/pnas.1317125111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Fiske ST, Cuddy AJC, Glick P. Universal dimensions of social cognition: Warmth and competence. Trends Cogn Sci. 2007;11(2):77–83. doi: 10.1016/j.tics.2006.11.005. [DOI] [PubMed] [Google Scholar]
- 71.Goesaert E, Op de Beeck HP. Representations of facial identity information in the ventral visual stream investigated with multivoxel pattern analyses. J Neurosci. 2013;33(19):8549–8558. doi: 10.1523/JNEUROSCI.1829-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Visser M, Jefferies E, Lambon Ralph MA. Semantic processing in the anterior temporal lobes: A meta-analysis of the functional neuroimaging literature. J Cogn Neurosci. 2010;22(6):1083–1094. doi: 10.1162/jocn.2009.21309. [DOI] [PubMed] [Google Scholar]
- 73.Kriegeskorte N, Simmons WK, Bellgowan PSF, Baker CI. Circular analysis in systems neuroscience: The dangers of double dipping. Nat Neurosci. 2009;12(5):535–540. doi: 10.1038/nn.2303. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Oosterhof NN, Tipper SP, Downing PE. Viewpoint (in)dependence of action representations: An MVPA study. J Cogn Neurosci. 2012;24(4):975–989. doi: 10.1162/jocn_a_00195. [DOI] [PubMed] [Google Scholar]
- 75.Kim NY, Lee SM, Erlendsdottir MC, McCarthy G. Discriminable spatial patterns of activation for faces and bodies in the fusiform gyrus. Front Hum Neurosci. 2014;8(August):632. doi: 10.3389/fnhum.2014.00632. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.O’Reilly JX, Woolrich MW, Behrens TEJ, Smith SM, Johansen-Berg H. Tools of the trade: Psychophysiological interactions and functional connectivity. Soc Cogn Affect Neurosci. 2012;7(5):604–609. doi: 10.1093/scan/nss055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Wang Y, Ramsey R, Hamilton AF. The control of mimicry by eye contact is mediated by medial prefrontal cortex. J Neurosci. 2011;31(33):12001–12010. doi: 10.1523/JNEUROSCI.0845-11.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78.Stephan KE, et al. Ten simple rules for dynamic causal modeling. Neuroimage. 2010;49(4):3099–3109. doi: 10.1016/j.neuroimage.2009.11.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Anzellotti S, Fairhall SL, Caramazza A. Decoding representations of face identity that are tolerant to rotation. Cereb Cortex. 2014;24(8):1988–1995. doi: 10.1093/cercor/bht046. [DOI] [PubMed] [Google Scholar]
- 80.Hebart MN, Görgen K, Haynes J-D. The Decoding Toolbox (TDT): A versatile software package for multivariate analyses of functional imaging data. Front Neuroinform. 2015;8(January):88. doi: 10.3389/fninf.2014.00088. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Friston KJ, Harrison L, Penny W. Dynamic causal modelling. Neuroimage. 2003;19(4):1273–1302. doi: 10.1016/s1053-8119(03)00202-7. [DOI] [PubMed] [Google Scholar]
- 82.Stephan KE, Penny WD, Daunizeau J, Moran RJ, Friston KJ. Bayesian model selection for group studies. Neuroimage. 2009;46(4):1004–1017. doi: 10.1016/j.neuroimage.2009.03.025. [DOI] [PMC free article] [PubMed] [Google Scholar]