Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2019 Jan 5.
Published in final edited form as: Personal Neurosci. 2018 Jul 5;1:e6. doi: 10.1017/pen.2018.8

Resting-state functional brain connectivity best predicts the personality dimension of openness to experience

Julien Dubois 1,4,*, Paola Galdi 5,*, Yanting Han 2, Lynn K Paul 1, Ralph Adolphs 1,2,3
PMCID: PMC6138449  NIHMSID: NIHMS948443  PMID: 30225394

Abstract

Personality neuroscience aims to find associations between brain measures and personality traits. Findings to date have been severely limited by a number of factors, including small sample size and omission of out-of-sample prediction. We capitalized on the recent availability of a large database, together with the emergence of specific criteria for best practices in neuroimaging studies of individual differences. We analyzed resting-state functional magnetic resonance imaging data from 884 young healthy adults in the Human Connectome Project (HCP) database. We attempted to predict personality traits from the “Big Five”, as assessed with the NEO-FFI test, using individual functional connectivity matrices. After regressing out potential confounds (such as age, sex, handedness and fluid intelligence), we used a cross-validated framework, together with test-retest replication (across two sessions of resting-state fMRI for each subject), to quantify how well the neuroimaging data could predict each of the five personality factors. We tested three different (published) denoising strategies for the fMRI data, two inter-subject alignment and brain parcellation schemes, and three different linear models for prediction. As measurement noise is known to moderate statistical relationships, we performed final prediction analyses using average connectivity across both imaging sessions (1 h of data), with the analysis pipeline that yielded the highest predictability overall. Across all results (test/retest; 3 denoising strategies; 2 alignment schemes; 3 models), Openness to experience emerged as the only reliably predicted personality factor. Using the full hour of resting-state data and the best pipeline, we could predict Openness to experience (NEOFAC_O: r=0.24, R2=0.024) almost as well as we could predict the score on a 24-item intelligence test (PMAT24_A_CR: r=0.26, R2=0.044). Other factors (Extraversion, Neuroticism, Agreeableness and Conscientiousness) yielded weaker predictions across results that were not statistically significant under permutation testing. We also derived two superordinate personality factors (“α” and “β”) from a principal components analysis of the NEO-FFI factor scores, thereby reducing noise and enhancing the precision of these measures of personality. We could account for 5% of the variance in the β superordinate factor (r=0.27, R2=0.050), which loads highly on Openness to experience. We conclude with a discussion of the potential for predicting personality from neuroimaging data and make specific recommendations for the field.

Introduction

Personality refers to the relatively stable disposition of an individual that influences long-term behavioral style (Back, Schmukle, & Egloff, 2009; Furr, 2009; Hong, Paunonen, & Slade, 2008; Jaccard, 1974). It is especially conspicuous in social interactions, and in emotional expression. It is what we pick up on when we observe a person for an extended time, and what leads us to make predictions about general tendencies in behaviors and interactions in the future. Often, these predictions are inaccurate stereotypes, and they can be evoked even by very fleeting impressions, such as merely looking at photographs of people (Todorov, 2017). Yet there is also good reliability (Viswesvaran & Ones, 2000) and consistency (Roberts & DelVecchio, 2000) for many personality traits currently used in psychology, which can predict real-life outcomes (Roberts, Kuncel, Shiner, Caspi, & Goldberg, 2007).

While human personality traits are typically inferred from questionnaires, viewed as latent variables they could plausibly be derived also from other measures. In fact, there are good reasons to think that biological measures other than self-reported questionnaires can be used to estimate personality traits. Many of the personality traits similar to those used to describe human dispositions can be applied to animal behavior as well, and again they make some predictions about real-life outcomes (Gosling & John, 1999; Gosling & Vazire, 2002). For instance, anxious temperament has been a major topic of study in monkeys, as a model of human mood disorders. Hyenas show neuroticism in their behavior, and also show sex differences in this trait as would be expected from human data (in humans, females tend to be more neurotic than males; in hyenas, the females are socially dominant and the males are more neurotic). Personality traits are also highly heritable. Anxious temperament in monkeys is heritable and its neurobiological basis is being intensively investigated (Oler et al., 2010). Twin studies in humans typically report heritability estimates for each trait between .4 and .6 (Bouchard & McGue, 2003; Jang, Livesley, & Vernon, 1996; Verweij et al., 2010), even though no individual genes account for much variance (studies using common single-nucleotide polymorphisms report estimates between 0 and 0.2; see Power & Pluess, 2015; Vinkhuyzen et al., 2012).

Just as gene-environment interactions constitute the distal causes of our phenotype, the proximal cause of personality must come from brain-environment interactions, since these are the basis for all behavioral patterns. Some aspects of personality have been linked to specific neural systems -- for instance, behavioral inhibition and anxious temperament have been linked to a system involving the medial temporal lobe and the prefrontal cortex (Birn et al., 2014). Although there is now universal agreement that personality is generated through brain function in a given context, it is much less clear what type of brain measure might be the best predictor of personality. Neurotransmitters, cortical thickness or volume of certain regions, and functional measures have all been explored with respect to their correlation with personality traits (for reviews see Canli, 2006; Yarkoni, 2015). We briefly summarize this literature next and refer the interested reader to review articles and primary literature for the details.

The search for neurobiological substrates of personality traits

Since personality traits are relatively stable over time (unlike state variables, such as emotions), one might expect that brain measures that are similarly stable over time are the most promising candidates for predicting such traits. The first types of measures to look at might thus be structural, connectional and neurochemical; indeed a number of such studies have reported correlations with personality differences. Here we briefly review studies using structural and functional magnetic resonance imaging (MRI) of humans, but leave aside research on neurotransmission. Although a number of different personality traits have been investigated, we emphasize those most similar to the “Big Five”, since they are the topic of the present paper (see below).

Structural MRI studies

Many structural MRI studies of personality to date have used voxel-based morphometry (VBM) (Blankstein, Chen, Mincic, McGrath, & Davis, 2009; Coutinho, Sampaio, Ferreira, Soares, & Gonçalves, 2013; DeYoung et al., 2010; Hu et al., 2011; Kapogiannis, Sutin, Davatzikos, Costa, & Resnick, 2013; Liu et al., 2013; Lu et al., 2014; Omura, Constable, & Canli, 2005; Taki et al., 2013). Results have been quite variable, sometimes even contradictory (e.g., the volume of the posterior cingulate cortex has been found to be both positively and negatively correlated with agreeableness; see Coutinho et al., 2013; DeYoung et al., 2010). Methodologically, this is in part due to the rather small sample sizes (typically less than 100; 116 in DeYoung et al., 2010; 52 in Coutinho et al., 2013) which undermine replicability (Button et al., 2013); studies with larger sample sizes (Liu et al., 2013) typically fail to replicate previous results.

More recently, surface-based morphometry (SBM) has emerged as a promising measure to study structural brain correlates of personality (Bjørnebekk et al., 2013; Holmes et al., 2012; Rauch et al., 2005; Riccelli, Toschi, Nigro, Terracciano, & Passamonti, 2017; Wright et al., 2006). It has the advantage of disentangling several geometric aspects of brain structure which may contribute to differences detected in VBM, such as cortical thickness (Hutton, Draganski, Ashburner, & Weiskopf, 2009), cortical volume, and folding. Although many studies using SBM are once again limited by small sample sizes, one recent study (Riccelli et al., 2017) used 507 subjects to investigate personality, although it had other limitations (e.g., using a correlational, rather than a predictive framework; see Dubois & Adolphs, 2016; Woo, Chang, Lindquist, & Wager, 2017; Yarkoni & Westfall, 2017).

There is much room for improvement in structural MRI studies of personality traits. The limitation of small sample sizes can now be overcome, since all MRI studies regularly collect structural scans, and recent consortia and data sharing efforts have lead to the accumulation of large publicly available datasets (Job et al., 2017; Miller et al., 2016; Van Essen et al., 2013). One could imagine a mechanism by which personality assessments, if not available already within these datasets, are collected later (Mar, Spreng, & Deyoung, 2013), yielding large samples for relating structural MRI to personality. Lack of out-of-sample generalizability, a limitation of almost all studies that we raised above, can be overcome using cross-validation techniques, or by setting aside a replication sample. In short: despite a considerable historical literature that has investigated the association between personality traits and structural MRI measures, there are as yet no very compelling findings because prior studies have been unable to surmount this list of limitations.

Diffusion MRI studies

Several studies have looked for a relationship between white-matter integrity as assessed by DTI and personality factors (Cohen, Schoene-Bake, Elger, & Weber, 2009; Kim & Whalen, 2009; Westlye, Bjørnebekk, Grydeland, Fjell, & Walhovd, 2011; Xu & Potenza, 2012). As with structural MRI studies, extant focal findings often fail to replicate with larger samples of subjects, which tend to find more widespread differences linked to personality traits (Bjørnebekk et al., 2013). The same concerns mentioned in the previous section, in particular the lack of a predictive framework (e.g., using cross-validation), plague this literature; similar recommendations can be made to increase the reproducibility of this line of research, in particular aggregating data (Miller et al., 2016; Van Essen et al., 2013) and using out-of-sample prediction (Yarkoni & Westfall, 2017).

Functional MRI studies

Functional MRI (fMRI) measures local changes in blood flow and blood oxygenation as a surrogate of the metabolic demands due to neuronal activity (Logothetis & Wandell, 2004). There are two main paradigms that have been used to relate fMRI data to personality traits: task-based fMRI and resting-state fMRI.

Task-based fMRI studies are based on the assumption that differences in personality may affect information-processing in specific tasks (Yarkoni, 2015). Personality variables are hypothesized to influence cognitive mechanisms, whose neural correlates can be studied with fMRI. For example, differences in neuroticism may materialize as differences in emotional reactivity, which can then be mapped onto the brain ( Canli et al., 2001). There is a very large literature on task-fMRI substrates of personality, which is beyond the scope of this overview. In general, some of the same concerns we raised above also apply to task-fMRI studies, which typically have even smaller sample sizes (Yarkoni, 2009), greatly limiting power to detect individual differences (in personality or any other behavioral measures). Several additional concerns on the validity of fMRI-based individual differences research apply (Dubois & Adolphs, 2016) and a new challenge arises as well: whether the task used has construct validity for a personality trait.

The other paradigm, resting-state fMRI, offers a solution to the sample size problem, as resting-state data is often collected alongside other data, and can easily be aggregated in large online databases (Biswal et al., 2010; Eickhoff, Nichols, Van Horn, & Turner, 2016; Poldrack & Gorgolewski, 2017; Van Horn & Gazzaniga, 2013). It is the type of data we used in the present paper. Resting-state data does not explicitly engage cognitive processes that are thought to be related to personality traits. Instead, it is used to study correlated self-generated activity between brain areas while a subject is at rest. These correlations, which can be highly reliable given enough data (Finn et al., 2015; Laumann et al., 2015; Noble et al., 2017), are thought to reflect stable aspects of brain organization (Shen et al., 2017; Smith et al., 2013). There is a large ongoing effort to link individual variations in functional connectivity (FC) assessed with resting-state fMRI to individual traits and psychiatric diagnosis (for reviews see Dubois & Adolphs, 2016; Orrù, Pettersson-Yeo, Marquand, Sartori, & Mechelli, 2012; Smith et al., 2013; Woo et al., 2017).

A number of recent studies have investigated functional connectivity markers from resting-state fMRI and their association with personality traits (Adelstein et al., 2011; Aghajani et al., 2014; Baeken et al., 2014; Beaty et al., 2014, 2016; Gao et al., 2013; Jiao et al., 2017; Lei, Zhao, & Chen, 2013; Pang et al., 2016; Ryan, Sheu, & Gianaros, 2011; Takeuchi et al., 2012; Wu, Li, Yuan, & Tian, 2016). Somewhat surprisingly, these resting-state fMRI studies typically also suffer from low sample sizes (typically less than 100 subjects, usually about 40), and the lack of a predictive framework to assess effect size out-of-sample. One of the best extant datasets, the Human Connectome Project (HCP) has only in the past year reached its full sample of over 1000 subjects, now making large sample sizes readily available. To date, only the exploratory “MegaTrawl” (Smith et al., 2016) has investigated personality in this database; we believe that ours is the first comprehensive study of personality on the full HCP dataset, offering very substantial improvements over all prior work.

Measuring Personality

Although there are a number of different schemes and theories for quantifying personality traits, by far the most common and well validated one, and also the only one available for the Human Connectome Project dataset, is the five-factor solution of personality (aka “The Big Five”). This was originally identified through systematic examination of the adjectives in English language that are used to describe human traits. Based on the hypothesis that all important aspects of human personality are reflected in language, Raymond Cattell applied factor analysis to peer ratings of personality and identified 16 common personality factors (Cattell, 1945). Over the next 3 decades, multiple attempts to replicate Cattell’s study using a variety of methods (e.g. self-description and description of others with adjective lists and behavioral descriptions) agreed that the taxonomy of personality could be robustly described through a five-factor solution (Borgatta, 1964; Fiske, 1949; Norman, 1963; Smith, 1967; Tupes & Christal, 1961). Since the 1980s, the Big Five has emerged as the leading psychometric model in the field of personality psychology (Goldberg, 1981; McCrae & John, 1992). The five factors are commonly termed “openness to experience,” “conscientiousness,” “extraversion,” “agreeableness,” and “neuroticism.”

While the Big Five personality dimensions are not based on an independent theory of personality, and in particular have no basis in neuroscience theories of personality, proponents of the Big Five maintain that they provide the best empirically-based integration of the dominant theories of personality, encompassing the alternative theories of Cattell, Guilford and Eysenck (Amelang & Borkenau, 1982). Self-report questionnaires, such as the NEO-FFI (McCrae & Costa, 2004), can be used to reliably assess an individual with respect to these five factors. Even though there remain critiques of the Big Five (Block, 1995; Uher, 2015), its proponents argue that its five factors “are both necessary and reasonably sufficient for describing at a global level the major features of personality” (McCrae & Costa, 1986).

The present study

As we emphasized above, personality neuroscience based on MRI data confronts two major challenges. First, nearly all studies to date have been severely underpowered due to small sample sizes (Button et al., 2013; Schönbrodt & Perugini, 2013; Yarkoni, 2009). Second, most studies have failed to use a predictive or replication framework (but see Deris, Montag, Reuter, Weber, & Markett, 2017), making their generalizability unclear -- a well-recognized problem in neuroscience studies of individual differences (Dubois & Adolphs, 2016; Gabrieli, Ghosh, & Whitfield-Gabrieli, 2015; Yarkoni & Westfall, 2017). The present paper takes these two challenges seriously by applying a predictive framework, together with a built-in replication, to a large, homogeneous resting-state fMRI dataset. We chose to focus on resting-state fMRI data to predict personality, because this is a predictor that could have better mechanistic interpretation than structural MRI measures (since ultimately it is brain function, not structure, that generates the behavior on the basis of which we can infer personality).

Our dataset, the Human Connectome Project resting-state fMRI data (HCP rs-fMRI) makes available over 1000 well assessed healthy adults. With respect to our study, it provided three types of relevant data: (1) substantial high-quality resting-state fMRI (2 sessions per subject on separate days, each consisting of two 15 minute runs, for 1 hour total); (2) personality assessment for each subject (using the NEO-FFI 2); (3) additional basic cognitive assessment (including fluid intelligence and others), as well as demographic information, which can be assessed as potential confounds.

Our primary question was straightforward: given the challenges noted above, is it possible to find evidence that any personality trait can be reliably predicted from fMRI data, using the best available resting-state fMRI dataset together with the best generally used current analysis methods? If the answer to this question is negative, this might suggest that studies to date that have claimed to find associations between resting-state fMRI and personality are false positives (but of course it would still leave open future positive findings, if more sensitive measures are available). If the answer is positive, it would provide an estimate of the effect size that can be expected in future studies; it would provide initial recommendations for data preprocessing, modeling, and statistical treatment; and it would provide a basis for hypothesis-driven investigations that could focus on particular traits and brain regions. As a secondary aim, we wanted to explore the sensitivity of the results to the details of the analysis used and gain some reassurance that any positive findings would be relatively robust with respect to the details of the analysis; we therefore used a few (well established) combinations of inter-subject alignment, preprocessing, and learning models. This is not intended as a systematic, exhaustive foray into all choices that could be made; such an investigation would be extremely valuable, yet is outside the scope of this work.

Methods

Dataset

We used data from a public repository, the 1200 subjects release of the Human Connectome Project (HCP) (Van Essen et al., 2013). The HCP provides MRI data and extensive behavioral assessment from almost 1200 subjects. Acquisition parameters and “minimal” preprocessing of the resting-state fMRI data is described in the original publication (Glasser et al., 2013). Briefly, each subject underwent two sessions of resting-state fMRI on separate days, each session with two separate 15 minute acquisitions generating 1200 volumes (customized Siemens Skyra 3 Tesla MRI scanner, TR = 720 ms, TE = 33 ms, flip angle= 52°, voxel size = 2 mm isotropic, 72 slices, matrix = 104 × 90, FOV = 208 mm × 180 mm, multiband acceleration factor = 8). The two runs acquired on the same day differed in the phase encoding direction, left-right and right-left (which leads to differential signal intensity especially in ventral temporal and frontal structures). The HCP data was downloaded in its minimally preprocessed form, i.e. after motion correction, B0 distortion correction, coregistration to T1-weighted images and normalization to MNI space (the T1w image is registered to MNI space with a FLIRT 12 DOF affine and then a FNIRT nonlinear registration, producing the final nonlinear volume transformation from the subject’s native volume space to MNI space).

Personality assessment, and personality factors

The 60 item version of the Costa and McCrae Neuroticism/Extraversion/Openness Five Factor Inventory (NEO-FFI), which has shown excellent reliability and validity (McCrae & Costa, 2004), was administered to HCP subjects. This measure was collected as part of the Penn Computerized Cognitive Battery (Gur et al., 2001; Gur et al., 2010). Note that the NEO-FFI was recently updated (NEO-FFI-3, 2010), but the test administered to the HCP subjects is the older version (NEO-FFI-2, 2004).

The NEO-FFI is a self-report questionnaire -- the abbreviated version of the 240-item Neuroticism/Extraversion/Openness Personality Inventory Revised (NEO-PI-R; Costa & McCrae, 1992). For each item, participants reported their level of agreement on a 5-point Likert scale, from strongly disagree to strongly agree.

The Openness, Conscientiousness, Extraversion, Agreeableness and Neuroticism scores are derived by coding each item’s answer (strongly disagree = 0; disagree = 1; neither agree nor disagree = 2; agree = 3; strongly agree = 4) and then reverse coding appropriate items and summing into subscales. As the item scores are available in the database, we recomputed the Big Five scores with the following item coding published in the NEO-FFI 2 manual, where * denotes reverse coding :

  • Openness: (3*, 8*, 13, 18*, 23*, 28, 33*, 38*, 43, 48*, 53, 58)

  • Conscientiousness: (5, 10, 15*, 20, 25, 30*, 35, 40, 45*, 50, 55*, 60)

  • Extraversion: (2, 7, 12*, 17, 22, 27*, 32, 37, 42*, 47, 52, 57*)

  • Agreeableness: (4, 9*, 14*, 19, 24*, 29*, 34, 39*, 44*, 49, 54*, 59*)

  • Neuroticism: (1*, 6, 11, 16*, 21, 26, 31*, 36, 41, 46*, 51, 56)

We note that the Agreeableness factor score that we calculated was slightly discrepant with the score in the HCP database due to an error in the HCP database in not reverse-coding item 59 at that time (downloaded 06/07/2017). This issue was reported on the HCP listserv (Gray, 2017).

To test the internal consistency of each of the Big Five personality traits in our sample, Cronbach’s alpha was calculated.

Each of the Big Five personality traits can be decomposed into further facets (Costa & McCrae, 1995), but we did not attempt to predict these facets from our data. Not only does each facet rely on fewer items and thus constitute a noisier measure, which necessarily reduces predictability from neural data (Gignac & Bates, 2017); also, trying to predict many traits leads to a multiple comparison problem which then needs to be accounted for (for an extreme example, see the HCP “MegaTrawl” Smith et al., 2016).

Despite their theoretical orthogonality, the Big Five are often found to be correlated with one another in typical subject samples. Some authors have suggested that these inter-correlations suggest a higher-order structure, and two superordinate factors have been described in the literature, often referred to as {α/socialization/stability} and {β/personal growth/plasticity} (Blackburn, Renwick, Donnelly, & Logan, 2004; DeYoung, 2006; Digman, 1997). The theoretical basis for the existence of these superordinate factors is highly debated (McCrae et al., 2008), and it is not our intention to enter this debate. However, these superordinate factors are less noisy (have lower associated measurement error) than the Big 5, as they are derived from a larger number of test items; this may improve predictability (Gignac & Bates, 2017). Hence, we performed a principal component analysis on the 5 factor scores to extract 2 orthogonal superordinate components, and tested the predictability of these from the HCP rs-fMRI data, in addition to the original five factors.

While we used resting-state fMRI data from two separate sessions (typically collected on consecutive days), there was only a single set of behavioral data available; the NEO-FFI was typically administered on the same day as the second session of resting-state fMRI (Van Essen et al., 2013).

Fluid intelligence assessment

An estimate of fluid intelligence is available as the PMAT24_A_CR measure in the HCP dataset. This proxy for fluid intelligence is based on a short version of Raven’s progressive matrices (24 items) (Bilker et al., 2012); scores are integers indicating number of correct items. We used this fluid intelligence score for two purposes: i) as a benchmark comparison in our predictive analyses, since others have previously reported that this measure of fluid intelligence could be predicted from resting-state fMRI in the HCP dataset (Finn et al., 2015; Noble et al., 2017); ii) as a de-confounding variable (see “Assessment and removal of potential confounds” below). Note that we recently performed a factor analysis of the scores on all cognitive tasks in the HCP to derive a more reliable measure of intelligence; this g-factor could be predicted better than the 24-item score from resting-state data (Dubois, Galdi, Paul, & Adolphs, 2018).

Subject selection

The total number of subjects in the 1200-subject release of the HCP dataset is N=1206. We applied the following criteria to include/exclude subjects from our analyses (listing in parentheses the HCP database field codes). i) Complete neuropsychological datasets. Subjects must have completed all relevant neuropsychological testing (PMAT_Compl=True, NEO-FFI_Compl=True, Non-TB_Compl=True, VisProc_Compl=True, SCPT_Compl=True, IWRD_Compl=True, VSPLOT_Compl=True) and the Mini Mental Status Exam (MMSE_Compl=True). Any subjects with missing values in any of the tests or test items were discarded. This left us with N = 1183 subjects. ii) Cognitive compromise. We excluded subjects with a score of 26 or below on the MMSE, which could indicate marked cognitive impairment in this highly educated sample of adults under age 40 (Crum, Anthony, Bassett, & Folstein, 1993). This left us with N = 1181 subjects (638 females, 28.8 +/− 3.7 y.o., range 22–37 y.o). Furthermore, iii) subjects must have completed all resting-state fMRI scans (3T_RS-fMRI_PctCompl=100), which leaves us with N = 988 subjects. Finally, iv) we further excluded subjects with a root-mean-squared frame-to-frame head motion estimate (Movement_Relative_RMS.txt) exceeding 0.15mm in any of the 4 resting-state runs (threshold similar to Finn et al., 2015). This left us with the final sample of N = 884 subjects (Table S1; 475 females, 28.6 +/− 3.7 y.o., range 22–36 y.o.) for predictive analyses based on resting-state data.

Assessment and removal of potential confounds

We computed the correlation of each of the personality factors with gender (Gender), age (Age_in_Yrs, restricted), handedness (Handedness, restricted) and fluid intelligence (PMAT24_A_CR). We also looked for differences in personality in our subject sample with other variables that are likely to affect FC matrices, such as brain size (we used FS_BrainSeg_Vol), motion (we computed the sum of framewise displacement in each run), and the multiband reconstruction algorithm which changed in the third quarter of HCP data collection (fMRI_3T_ReconVrs). Correlations are shown in Figure 2a below. We then used multiple linear regression to regress these variables from each of the personality scores and remove their confounding effects.

Figure 2. Structure of Personality factors in our subject sample (N=884).

Figure 2

a. The five personality factors were not orthogonal in our sample. Neuroticism was anticorrelated with Conscientiousness, Extraversion and Agreeableness, and the latter three were positively correlated with each other. Openness correlated more weakly with other factors. There were highly significant correlations with other behavioral and demographic variables, which we accounted for in our subsequent analyses by regressing them out of the personality scores (see next section). b. Distributions of the five personality scores in our sample. Each of the five personality scores was approximately normally distributed by visual inspection. c. Two-dimensional principal component projection; the value for each personality factor in this projection is represented by the color of the dots. The weights for each personality factor are shown at the bottom.

Note that we do not control for differences in cortical thickness and other morphometric features, which have been reported to be correlated with personality factors (e.g. Riccelli et al., 2017). These likely interact with FC measures and should eventually be accounted for in a full model, yet this was deemed outside the scope of the present study.

The five personality factors are intercorrelated to some degree (see Results, Figure 2a). We did not orthogonalize them-- consequently predictability would be expected also to correlate slightly among personality factors.

It could be argued that controlling for variables such as gender and fluid intelligence risks producing a conservative, but perhaps overly pessimistic picture. Indeed, there are well established gender differences in personality (Feingold, 1994; Schmitt, Realo, Voracek, & Allik, 2008), which might well be based on gender differences in functional connectivity (similar arguments can be made with respect to age (Allemand, Zimprich, & Hendriks, 2008; Soto, John, Gosling, & Potter, 2011) and fluid intelligence (Chamorro-Premuzic & Furnham, 2004; Rammstedt, Danner, & Martin, 2016)). Since the causal primacy of these variables with respect to personality is unknown, it is possible that regressing out sex and age could regress out substantial meaningful information about personality. We therefore also report supplemental results with a less conservative de-confounding procedure -- only regressing out obvious confounds which are not plausibly related to personality, but which would plausibly influence functional connectivity data: image reconstruction algorithm, framewise displacement, and brain size measures.

Data preprocessing

Resting-state data must be preprocessed beyond “minimal preprocessing”, due to the presence of multiple noise components, such as subject motion and physiological fluctuations. Several approaches have been proposed to remove these noise components and clean the data, however the community has not yet reached a consensus on the “best” denoising pipeline for resting-state fMRI data (Caballero-Gaudes & Reynolds, 2017; Ciric et al., 2017; Murphy & Fox, 2017; Siegel et al., 2017). Most of the steps taken to denoise resting-state data have limitations, and it is unlikely that there is a set of denoising steps that can completely remove noise without also discarding some of the signal of interest. Categories of denoising operations that have been proposed comprise tissue regression, motion regression, noise component regression, temporal filtering and volume censoring. Each of these categories may be implemented in several ways. There exist several excellent reviews of the pros and cons of various denoising steps (Caballero-Gaudes & Reynolds, 2017; Liu, 2016; Murphy, Birn, & Bandettini, 2013; Power et al., 2014).

Here, instead of picking a single denoising strategy combining steps used in the previous literature, we set out to explore three reasonable alternatives, which we refer to as A, B, and C (Figure 1c). To easily apply these preprocessing strategies in a single framework, using input data that is either volumetric or surface-based, we developed an in-house, Python (v2.7.14)-based pipeline, mostly based on open source libraries and frameworks for scientific computing including SciPy (v0.19.0), Numpy (v1.11.3), NiLearn (v0.2.6), NiBabel (v2.1.0), Scikit-learn (v0.18.1) (Abraham et al., 2014; Gorgolewski et al., 2011; Gorgolewski et al., 2017; Pedregosa et al., 2011; Walt, Colbert, & Varoquaux, 2011), implementing the most common denoising steps described in previous literature.

Figure 1. Overview of our approach.

Figure 1

In total, we separately analyzed 36 different sets of results: 2 data sessions × 2 alignment/brain parcellation schemes × 3 preprocessing pipelines × 3 predictive models (univariate positive, univariate negative, and multivariate). a. The data from each selected HCP subject (Nsubjects=884) and each run (REST1_LR, REST1_RL, REST2_LR, REST2_RL) was downloaded after minimal preprocessing, both in MNI space, and in MSM-All space. The _LR and _RL runs within each session were averaged, producing 2 datasets that we call REST1 and REST2 henceforth. Data for REST1 and REST2, and for both spaces (MNI, MSM-All) were analyzed separately. We applied three alternate denoising pipelines to remove typical confounds found in resting-state fMRI data (see c). We then parcellated the data (see d) and built a functional connectivity matrix separately for each alternative. This yielded 6 FC matrices per run and per subject. In red: alternatives taken and presented in this paper. b. For each of the 6 alternatives, an average FC matrix was computed for REST1 (from REST1_LR and REST1_RL), for REST2 (from REST2_LR and REST2_RL), and for all runs together, REST12. For a given session, we built a (Nsubjects × Nedges) matrix, stacking the upper triangular part of all subjects’ FC matrices (the lower triangular part is discarded, because FC matrices are diagonally symmetric). Each column thus corresponds to a single entry in the upper triangle of the FC matrix (a pairwise correlation between two brain parcels, or edge) across all 884 subjects. There are a total of Nparcels(Nparcels−1)/2 edges (thus: 35778 edges for the MNI parcellation, 64620 edges for the MSM-All parcellation). This was the data from which we then predicted individual differences in each of the personality factors. We used two different linear models (see text), and a leave-one-family-out cross validation scheme. The final result is a predicted score for each subject, against which we correlate the observed score for statistical assessment of the prediction. Permutations are used to assess statistical significance. c. Detail of the three denoising alternatives. These are common denoising strategies for resting-state fMRI. The steps are color-coded to indicate the category of operation they correspond to (legend at the bottom). See text for details. d. The parcellations used for the MNI-space and MSM-All space, respectively. Parcels are randomly colored for visualization. Note that the parcellation used for MSM-All space does not include subcortical structures, while the parcellation used for MNI space does.

Pipeline A reproduces as closely as possible the strategy described in (Finn et al., 2015) and consists of seven consecutive steps: 1) the signal at each voxel is z-score normalized; 2) using tissue masks, temporal drifts from cerebrospinal fluid (CSF) and white matter (WM) are removed with third-degree Legendre polynomial regressors; 3) the mean signals of CSF and WM are computed and regressed from gray matter voxels; 4) translational and rotational realignment parameters and their temporal derivatives are used as explanatory variables in motion regression; 5) signals are low-pass filtered with a Gaussian kernel with a standard deviation of 1 TR, i.e. 720ms in the HCP dataset; 6) the temporal drift from gray matter signal is removed using a third-degree Legendre polynomial regressor; 7) global signal regression is performed.

Pipeline B, described in (Ciric et al., 2017; Satterthwaite, Wolf, et al., 2013), is composed of four steps in our implementation: 1) normalization at voxel-level is performed by subtracting the mean from each voxel’s time series; 2) linear and quadratic trends are removed with polynomial regressors; 3) temporal filtering is performed with a first order Butterworth filter with a passband between 0.01 and 0.08 Hz (after linearly interpolating volumes to be censored, cf. step 4); 4) tissue regression (CSF and WM signals with their derivatives and quadratic terms), motion regression (realignment parameters with their derivatives, quadratic terms and square of derivatives), global signal regression (whole brain signal with derivative and quadratic term), and censoring of volumes with a root-mean-squared (RMS) displacement that exceeded 0.25 mm are combined in a single regression model.

Pipeline C, inspired by (Siegel et al., 2017), is implemented as follows: 1) an automated independent component (IC)-based denoising was performed with ICA-FIX (Salimi-Khorshidi et al., 2014). Instead of running ICA-FIX ourselves, we downloaded the FIX-denoised data which is available from the HCP database; 2) voxel signals were demeaned and 3) detrended with a first degree polynomial; 4) CompCor, a PCA-based method proposed by (Behzadi, Restom, Liau, & Liu, 2007) was applied to derive 5 components from CSF and WM signals; these were regressed out of the data, together with gray matter and whole-brain mean signals; volumes with a framewise displacement greater than 0.25 mm or a variance of differentiated signal (DVARS) greater than 105% of the run median DVARS were discarded as well; 5) temporal filtering was performed with a first-order Butterworth band-pass filter between 0.01 and 0.08 Hz, after linearly interpolating censored volumes.

Inter-subject alignment, parcellation, and functional connectivity matrix generation

An important choice in processing fMRI data is how to align subjects in the first place. The most common approach is to warp individual brains to a common volumetric template, typically MNI152. However, cortex is a 2D structure; hence, surface-based algorithms that rely on cortical folding to map individual brains to a template may be a better approach. Yet another improvement in aligning subjects may come from using functional information alongside anatomical information - this is what the multimodal surface matching (MSM) framework achieves (Robinson et al., 2014). MSM-All aligned data, in which intersubject registration uses individual cortical folding, myelin maps, and resting-state fMRI correlation data, is available for download from the HCP database.

Our prediction analyses below are based on functional connectivity matrices. While voxel- (or vertex-) wise functional connectivity matrices can be derived, their dimensionality is too high compared to the number of examples in the context of a machine-learning based predictive approach. PCA or other dimensionality reduction applied to the voxelwise data can be used, but this often comes at the cost of losing neuroanatomical specificity. Hence, we work with the most common type of data: parcellated data, in which data from many voxels (or vertices) is aggregated anatomically and the signal within a parcel is averaged over its constituent voxels. Choosing a parcellation scheme is the first step in a network analysis of the brain (Sporns, 2013), yet once again there is no consensus on the “best” parcellation. There are two main approaches to defining network nodes in the brain: nodes may be a set of overlapping, weighted masks, e.g. obtained using independent component analysis (ICA) of BOLD fMRI data (Smith et al., 2013); or a set of discrete, non-overlapping binary masks, also known as a hard parcellation (Glasser, Coalson, et al., 2016; Gordon et al., 2016). We chose to work with a hard parcellation, which we find easier to interpret.

Here we present results based on a classical volumetric alignment, together with a volumetric parcellation of the brain into 268 nodes (Finn et al., 2015; Shen, Tokoglu, Papademetris, & Constable, 2013); and, for comparison, results based on MSM-All data, together with a parcellation that was specifically derived from this data (Glasser, Coalson, et al., 2016) (Figure 1d).

Timeseries extraction simply consisted in averaging data from voxels (or grayordinates) within each parcel, and matrix generation in pairwise correlating parcel time series (Pearson correlation coefficient). FC matrices were averaged across runs (all averaging used Fisher-z transforms) acquired with left-right and right-left phase encoding in each session, i.e. we derived two FC matrices per subject, one for REST1 (from REST1_LR and REST1_RL) and one for REST2 (from REST2_LR and REST2_RL); we also derived a FC matrix averaged across all runs (REST12).

Test-retest comparisons

We applied all three denoising pipelines to the data of all subjects. We then compared the functional connectivity (FC) matrices produced by each of these strategies, using several metrics. One metric that we used follows from the connectome fingerprinting work of (Finn et al., 2015), and was recently labeled the identification success rate (ISR) (Noble et al., 2017). Identification of subject S is successful if, out of all subjects’ FC matrices derived from REST2, subject S’s is the most highly correlated with subject S’s FC matrix from REST1 (identification can also be performed from REST2 to REST1; results are very similar). The ISR gives an estimate of the reliability and specificity of the entire FC matrix at the individual subject level, and is influenced both by within-subject test-retest reliability as well as by discriminability amongst all subjects in the sample. Relatedly, it is desirable to have similarities (and differences) between all subjects be relatively stable across repeated testing sessions. Following an approach introduced in (Geerligs, Rubinov, Cam-Can, & Henson, 2015), we computed the pairwise similarity between subjects separately for session 1 and session 2, constructing a Nsubjects×Nsubjects matrix for each session. We then compared these matrices using a simple Pearson correlation. Finally, we used a metric targeted at behavioral utility, and inspired by (Geerligs, Rubinov, et al., 2015): for each edge (the correlation value between a given pair of brain parcels) in the FC matrix, we computed its correlation with a stable trait across subjects, and built a matrix representing the relationship of each edge to this trait, separately for session 1 and session 2. We then compared these matrices using a simple Pearson correlation. The more edges reliably correlate with the stable trait, the higher the correlation between session 1 and session 2 matrices. It should be noted that trait stability is an untested assumption with this approach, because in fact only a single trait score was available in the HCP, collected at the time of session 2. We performed this analysis for the measure of fluid intelligence available in the HCP (PMAT24_A_CR) as well as all Big Five personality factors.

Prediction models

There is no obvious “best” model available to predict individual behavioral measures from functional connectivity data (Abraham et al., 2017). So far, most attempts have relied on linear machine learning approaches. This is partly related to the well-known “curse of dimensionality”: despite the relatively large sample size that is available to us (N=884 subjects), it is still about an order of magnitude smaller than the typical number of features included in the predictive model. In such situations, fitting relatively simple linear models is less prone to overfitting than fitting complex nonlinear models.

There are several choices of linear prediction models. Here, we present the results of two methods that have been used in the literature for similar purposes: 1) a simple, “univariate” regression model as used in (Finn et al., 2015), and further advocated by (Shen et al., 2017), preceded by feature selection; and 2) a regularized linear regression approach, based on elastic-net penalization (Zou & Hastie, 2005). We describe each of these in more detail next.

Model (1) is the simplest model, and the one proposed by (Finn et al., 2015), consisting in a univariate regressor where the dependent variable is the score to be predicted and the explanatory variable is a scalar value that summarizes the functional connectivity network strength (i.e., the sum of edge weights). A filtering approach is used to select features (edges in the FC correlation matrix) that are correlated with the behavioral score on the training set: edges that correlate with the behavioral score with a p-value less than 0.01 are kept. Two distinct models are built using edges of the network that are positively and negatively correlated with the score, respectively. This method has the advantage of being extremely fast to compute, but some main limitations are that i) it condenses all the information contained in the connectivity network into a single measure and does not account for any interactions between edges and ii) it arbitrarily builds two separate models (one for positively correlated edges, one for negatively correlated edges; they are referred to as the positive and the negative models (Finn et al., 2015)) and does not offer a way to integrate them. We report results from both the positive and negative models for completeness.

To address the limitations of the univariate model(s), we also included a multivariate model. Model (2) kept the same filtering approach as for the univariate model (discard edges for which the p-value of the correlation with the behavioral score is greater than 0.01); this choice allows for a better comparison of the multivariate and univariate models, and for faster computation. Elastic Net is a regularized regression method that linearly combines L1- (lasso) and L2- (ridge) penalties to shrink some of the regressor coefficients toward zero, thus retaining just a subset of features. The lasso model performs continuous shrinkage and automatic variable selection simultaneously, but in the presence of a group of highly correlated features, it tends to arbitrarily select one feature from the group. With high-dimensional data and few examples, the ridge model has been shown to outperform lasso; yet it cannot produce a sparse model since all the predictors are retained. Combining the two approaches, elastic net is able to do variable selection and coefficient shrinkage while retaining groups of correlated variables. Here however, based on preliminary experiments and on the fact that it is unlikely that just a few edges contribute to prediction, we fixed the L1 ratio (which weights the L1- and L2- regularizations) to 0.01, which amounts to almost pure ridge regression. We used 3-fold nested cross-validation (with balanced “classes”, based on a partitioning of the training data into quartiles) to choose the alpha parameter (among 50 possible values) that weighs the penalty term.

Cross-validation scheme

In the HCP dataset, several subjects are genetically related (in our final subject sample, there were 410 unique families). To avoid biasing the results due to this family structure (e.g., perhaps having a sibling in the training set would facilitate prediction for a test subject), we implemented a leave-one-family-out cross-validation scheme for all predictive analyses.

Statistical assessment of predictions

Several measures can be used to assess the quality of prediction. A typical approach is to plot observed vs. predicted values (rather than predicted vs. observed (Piñeiro, Perelman, Guerschman, & Paruelo, 2008)). The Pearson correlation coefficient between observed scores and predicted scores is often reported as a measure of prediction (e.g. Finn et al., 2015), given its clear graphical interpretation. However, in the context of cross-validation, it is incorrect to square this correlation coefficient to obtain the coefficient of determination R2, which is often taken to reflect the proportion of variance explained by the model (Alexander, Tropsha, & Winkler, 2015); instead, the coefficient of determination R2 should be calculated as:

R2=1-i=1n(yi-yι^)2i=1n(yi-y¯)2 (1)

where n is the number of observations (subjects), y is the observed response variable, ȳ is its mean, and ŷ is the corresponding predicted value. Equation (1) therefore measures the size of the residuals from the model compared with the size of the residuals for a null model where all of the predictions are the same, i.e., the mean value ȳ. In a cross-validated prediction context, R2 can actually take negative values (in cases when the denominator is larger than the numerator, i.e. when the sum of squared errors is larger than that of the null model)! Yet another, related statistic to evaluate prediction outcome is the Root Mean Square Deviation (RMSD), defined in (Piñeiro et al., 2008) as:

RMSD=1n-1i=1n(yi-yι^)2 (2)

RMSD as defined in (2) represents the standard deviation of the residuals. To facilitate interpretation, it can be normalized by dividing it by the standard deviation of the observed values:

nRMSD=1n-1i=1n(yi-yι^)21n-1i=1n(yi-y¯)2=i=1n(yi-yι^)2i=1n(yi-y¯)2=1-R2 (3)

nRMSD thus has a very direct link to R2 (3); it is interpretable as the average deviation of each predicted value to the corresponding observed value, and is expressed as a fraction of the standard deviation of the observed values.

In a cross-validation scheme, the folds are not independent of each other. This means that statistical assessment of the cross-validated performance using parametric statistical tests is problematic (Combrisson & Jerbi, 2015; Noirhomme et al., 2014). Proper statistical assessment should thus be done using permutation testing on the actual data. To establish the empirical distribution of chance, we ran our final predictive analyses using 1000 random permutations of the scores (shuffling scores randomly between subjects, keeping everything else exactly the same, including the family structure).

Results

Characterization of behavioral measures

Internal consistency, distribution, and inter-correlations of personality traits

In our final subject sample (N=884), there was good internal consistency for each personality trait, as measured with Cronbach’s α. We found: Openness, α = 0.76; Conscientiousness α = 0.81; Extraversion, α = 0.78; Agreeableness, α = 0.76; and Neuroticism, α = 0.85. These compare well with the values reported by (McCrae & Costa, 2004).

Scores on all factors were nearly normally distributed by visual inspection, although the null hypothesis of a normal distribution was rejected for all but Agreeableness (using D’Agostino and Pearson’s (1973) normality test as implemented in SciPy) (Figure 2b).

Although in theory the Big Five personality traits should be orthogonal, their estimation from the particular item scoring of versions of the NEO in practice deviates considerably from orthogonality. This intercorrelation amongst the five factors has been reported for the NEO-PI-R (Block, 1995; Saucier, 2002), the NEO-FFI (Block, 1995; Egan, Deary, & Austin, 2000), and alternate instruments (DeYoung, 2006) (but, see McCrae et al., 2008). Indeed, in our subject sample, we found that the five personality factors were correlated with one another (Figure 2a). For example, Neuroticism was anticorrelated with Conscientiousness (r = −0.41, p <10−37), Extraversion (r = −0.34, p < 10−25), and Agreeableness (r = −0.28, p <10−16), while these latter three factors were positively correlated with one another (all r>0.21). Though the theoretical interpretation of these inter-correlations in terms of higher-order factors of personality remains a topic of debate (DeYoung, 2006; Digman, 1997; McCrae et al., 2008), we derived two orthogonal higher-order personality dimensions using a principal component analysis of the Big 5 factor scores; we labeled the two derived dimensions α and β, following (Digman, 1997). The first component [α] accounted for 40.3% of the variance, and the second [β] for 21.6% (total variance explained by the two-dimensional PC solution was thus 61.9%). Figure 2c shows how the Big Five project on this two-dimensional solution, and the PC loadings.

Confounding variables

There are known effects of gender (Ruigrok et al., 2014; Trabzuni et al., 2013), age (Dosenbach et al., 2010; Geerligs, Renken, Saliasi, Maurits, & Lorist, 2015), handedness (Pool, Rehme, Eickhoff, Fink, & Grefkes, 2015), in-scanner motion (Power, Barnes, Snyder, Schlaggar, & Petersen, 2012; Satterthwaite, Elliott, et al., 2013; Tyszka, Kennedy, Paul, & Adolphs, 2014), brain size (Hänggi, Fövenyi, Liem, Meyer, & Jäncke, 2014) and fluid intelligence (Cole, Yarkoni, Repovs, Anticevic, & Braver, 2012; Finn et al., 2015; Noble et al., 2017) on the functional connectivity patterns measured in the resting-state with fMRI. It is thus necessary to control for these variables: indeed, if a personality factor is correlated with gender, one would be able to predict some of the variance in that personality factor solely from functional connections that are related to gender. The easiest way (though perhaps not the best way, see Westfall & Yarkoni, 2016) to control for these confounds is by regressing the confounding variables on the score of interest in our sample of subjects.

We characterized the relationship between each of the personality factors and each of the confounding variables listed above in our subject sample (Figure 2a). All personality factors but Extraversion were correlated with gender: women scored higher on Conscientiousness, Agreeableness and Neuroticism, while men scored higher on Openness. In previous literature, women have been reliably found to score higher on Neuroticism and Agreeableness, which we replicated here, while other gender differences are generally inconsistent at the level of the factors (Costa, Terracciano, & McCrae, 2001; Feingold, 1994; Weisberg, DeYoung, & Hirsh, 2011). Agreeableness and Openness were significantly correlated with age in our sample, despite our limited age range (22–36 y.o.): younger subjects scored higher on Openness, while older subjects scored higher on Agreeableness. The finding for Openness does not match previous reports (Allemand et al., 2008; Soto et al., 2011), but this may be confounded by other factors such as gender, as our analyses here do not use partial correlations. Motion, quantified as the sum of frame-to-frame displacement over the course of a run (and averaged separately for REST1 and REST2) was correlated with Openness: subjects scoring lower on Openness moved more during the resting-state. Note that motion in REST1 was highly correlated (r=0.72, p<10−143) with motion in REST2, indicating that motion itself may be a stable trait, and correlated with other traits. Brain size, obtained from Freesurfer during the minimal preprocessing pipelines, was found to be significantly correlated with all personality factors but Extraversion. Fluid intelligence was positively correlated with Openness, and negatively correlated with Conscientiousness, Extraversion, and Neuroticism, consistent with other reports (Bartels et al., 2012; Chamorro-Premuzic & Furnham, 2004). While the interpretation of these complex relationships would require further work outside the scope of this study, we felt that it was critical to remove shared variance between each personality score and the primary confounding variables before proceeding further. This ensures that our model is trained specifically to predict personality, rather than confounds that covary with personality, although it may also reduce power by removing shared variance (thus providing a conservative result).

Another possible confound, specific to the HCP dataset, is a difference in the image reconstruction algorithm between subjects collected prior to and after April 2013. The reconstruction version leaves a notable signature on the data that can make a large difference in the final analyses produced (Elam, 2015). We found a significant correlation with the Openness factor in our sample. This indicates that the sample of subjects who were scanned with the earlier reconstruction version happened to score slightly less high for the Openness factor than the sample of subjects who were scanned with the later reconstruction version (purely by sampling chance); this of course is meaningless, and a simple consequence of working with finite samples. Therefore, we also included the reconstruction factor as a confound variable.

Importantly, the multiple linear regression used for removing the variance shared with confounds was performed on training data only (in each cross-validation fold during the prediction analysis), and then the fitted weights were applied to both the training and test data. This is critical to avoid any leakage of information, however negligible, from the test data into the training data.

Authors of the HCP-MegaTrawl have used transformed variables (Age2) and interaction terms (Gender × Age, Gender × Age2) as further confounds (Smith et al., 2016). After accounting for the confounds described above, we did not find sizeable correlations with these additional terms (all correlations <0.008), and thus we did not use these additional terms in our confound regression.

Preprocessing affects test-retest reliability of FC matrices

As we were interested in relatively stable traits (which are unlikely to change much between sessions REST1 and REST2), one clear goal for the denoising steps applied to the minimally preprocessed data was to yield functional connectivity matrices that are as “similar” as possible across the two sessions. We computed several metrics (see Methods) to assess this similarity for each of our three denoising strategies (A, B, and C; cf Figure 1c). Of course, no denoising strategy would achieve perfect test-retest reliability of FC matrices since, in addition to inevitable measurement error, the two resting-state sessions for each subject likely feature somewhat different levels of states such as arousal and emotion.

In general, differences in test-retest reliability across metrics were small when comparing the three denoising strategies. Considering the entire FC matrix, the Identification Success Rate (ISR) (Finn et al., 2015; Noble et al., 2017) was high for all strategies, and highest for pipeline B (Figure 3a). The multivariate pairwise distances between subjects were also best reproduced across sessions by pipeline B (Figure 3b). In terms of behavioral utility, i.e. reproducing the pattern of correlations of the different edges with a behavioral score, pipeline A outperformed the others (Figure 3c). All three strategies appear to be reasonable choices, and we would thus expect a similar predictive accuracy under each of them, if there is information about a given score in the functional connectivity matrix.

Figure 3. Test-retest comparisons between spaces and denoising strategies.

Figure 3

a. Identification success rate, and other statistics related to connectome fingerprinting (Finn et al., 2015; Noble et al., 2017). All pipelines had a success rate superior to 87% for identifying the functional connectivity matrix of a subject in REST2 (out of N=884 choices) based on their functional connectivity matrix in REST1. Pipeline B slightly outperformed the others. b. Test-retest of the pairwise similarities (based on Pearson correlation) between all subjects (Geerligs, Rubinov, et al., 2015). Overall, for the same session, the three pipelines gave similar pairwise similarities between subjects. About 25% of the variance in pairwise distances was reproduced in REST2, with pipeline B emerging as the winner (0.542=29%). c. Test-retest reliability of behavioral utility, quantified as the pattern of correlations between each edge and a behavioral score of interest (Geerligs, Rubinov, et al., 2015). Shown are fluid intelligence, Openness to experience, and Neuroticism (all de-confounded, see main text). Pipeline A gave slightly better test-retest reliability for all behavioral scores. MSM-All outperformed MNI alignment. Neuroticism showed lower test-retest reliability than fluid intelligence or Openness to experience.

We note here already that Neuroticism stands out as having lower test-retest reliability in terms of its relationship to edge values across subjects (Figure 3c). This may be a hint that the FC matrices do not carry information about Neuroticism.

Prediction of fluid intelligence (PMAT24_A_CR)

It has been reported that a measure of fluid intelligence, the raw score on a 24-item version of the Raven’s Progressive Matrices (PMAT24_A_CR), could be predicted from FC matrices in previous releases of the HCP dataset (Finn et al., 2015; Noble et al., 2017). We generally replicated this result qualitatively for the de-confounded fluid intelligence score (removing variance shared with gender, age, handedness, brain size, motion, and reconstruction version), using a leave-one-family-out cross-validation approach. We found positive correlations across all 36 of our result datasets: 2 sessions × 3 denoising pipelines (A, B & C) × 2 parcellation schemes (in volumetric space and in MSM-All space) × 3 models (univariate positive, univariate negative, and multivariate learning models) (Figure 4a, Table 1). We note however that, using MNI space and denoising strategy A as in (Finn et al., 2015), the prediction score was very low (REST1: r=0.04; REST2: r=0.03). One difference is that the previous study did not use de-confounding, hence some variance from confounds may have been used in the predictions; also the sample size was much smaller in Finn et al. (2015) (N=118; but N=606 in Noble et al., 2017), and family structure was not accounted for in the cross-validation. We generally found that prediction performance was better in MSM-All space (Figure 4a, Table 1).

Figure 4. Prediction results for de-confounded fluid intelligence (PMAT24_A_CR).

Figure 4

a. All predictions were assessed using the correlation between the observed scores (the actual scores of the subjects) and the predicted scores. This correlation obtained using the REST2 dataset was plotted against the correlation from the REST1 dataset, to assess test-retest reliability of the prediction outcome. Results in MSM-All space outperformed results in MNI space. The multivariate model slightly outperformed the univariate models (positive and negative). Our results generally showed good test-retest reliability across sessions, although REST1 tended to produce slightly better predictions than REST2. Pearson correlation scores for the predictions are listed in Table 1. Supplementary Figure 1 shows prediction scores with minimal deconfounding. b. We ran a final prediction using combined data from all resting-state runs (REST12), in MSM-All space with denoising strategy A (results are shown as vertical red lines). We randomly shuffled the PMAT24_A_CR scores 1000 times while keeping everything else the same, for the univariate model (positive, top) and the multivariate model (bottom). The distribution of prediction scores (Pearson r, and R2) under the null hypothesis is shown (black histograms). Note that the empirical 99% confidence interval (shaded gray area) is wider than the parametric CI (shown for reference, magenta dotted lines), and features a heavy tail on the left side for the univariate model. This demonstrates that parametric statistics are not appropriate in the context of cross-validation. Such permutation testing may be computationally prohibitive for more complex models, yet since the chance distribution is model-dependent, it must be performed for statistical assessment.

Table 1. Test-retest prediction results using deconfounded scores.

Listed are Pearson correlation coefficients between predicted and observed individual scores, for all behavioral scores and analytical alternatives (the two columns for each score correspond to the two resting-state sessions). See Supplementary Figure 1 for results with minimal deconfounding.

graphic file with name nihms948443f7.jpg

To generate a final prediction, we combined data from all four resting-state runs (REST12). We chose to use pipeline A and MSM-All space, which we had found to yield the best test-retest reliability in terms of behavioral utility (Figure 3c). We obtained r=0.22 (R2=0.007, nRMSD=0.997) for the univariate positive model, r=0.18 (R2=-0.023, nRMSD=1.012) for the univariate negative model, and r=0.26 (R2=0.044, nRMSD=0.978) for the multivariate model. Interestingly, these performances on combined data outperformed performance on REST1 or REST2 alone, suggesting that decreasing noise in the neural data boosts prediction performance. For statistical assessment of predictions, we estimated the distribution of chance for the prediction score under both the univariate positive and the multivariate models, using 1000 random permutations of the subjects’ fluid intelligence scores (Figure 4b). For reference we also show parametric statistics thresholds for the correlation coefficients; we found that parametric statistics underestimate the confidence interval for the null hypothesis, hence overestimate significance. Interestingly, the null distributions differed between the univariate and the multivariate models: while the distribution under the multivariate model was roughly symmetric about 0, the distribution under the univariate model was asymmetric with a long tail on the left. The empirical, one-tailed p-values for REST12 MSM-All space data denoised with strategy A and using the univariate positive model, and using the multivariate model, both achieved p<0.001 (none of the 1000 random permutations resulted in a higher prediction score).

Prediction of the Big Five

We established that our approach reproduces and improves on the previous finding that fluid intelligence can be predicted from resting-state functional connectivity (Finn et al., 2015; Noble et al., 2017). We next turned to predicting each of the Big Five personality factors using the same approach (including de-confounding, which in this case removes variance shared with gender, age, handedness, brain size, motion, reconstruction version and, importantly, fluid intelligence).

Test-retest results across analytical choices are shown in Figure 5a, and in Table 1. Predictability was lower than for fluid intelligence (PMAT24_A_CR) for all Big Five personality factors derived from the NEO-FFI. Openness to experience showed the highest predictability overall, and also the most reproducible across sessions; prediction of Extraversion was moderately reproducible; in contrast, the predictability of the other three personality factors (Agreeableness and Neuroticism, and Conscientiousness) was low and lacked reproducibility.

Figure 5. Prediction results for the Big Five personality factors.

Figure 5

a. Test-retest prediction results for each of the Big Five. Representation is the same as in Figure 4a. The only factor that showed consistency across parcellation schemes, denoising strategies, models and sessions was Openness (NEOFAC_O), although Extraversion (NEOFAC_E) also showed substantial positive correlations. See also Table 1. b. Prediction results for each of the (demeaned and deconfounded) Big Five, from REST12 FC matrices, using MSM-All inter-subject alignment, denoising strategy A, and the multivariate prediction model. The blue line shows the best fit to the cloud of points (its slope should be close to 1 (dotted line) for good predictions, see Piñeiro et al., 2008). The variance of predicted values is noticeably smaller than the variance of observed values.

It is worth noting that the NEO-FFI test was administered closer in time to REST2 than to REST1 on average; hence one might expect REST2 to yield slightly better results, if the NEO-FFI factor scores reflect a state component. We found that REST2 produced better predictability than REST1 for Extraversion (results fall mostly to the left of the diagonal line of reproducibility), while REST1 produced better results for Openness, hence the data does not reflect our expectation of state effects on predictability.

Although we conducted 18 different analyses for each session with the intent to present all of them in a fairly unbiased manner, it is notable that certain combinations produced the best predictions across different personality scores - some of the same combinations that yielded the best predictability for fluid intelligence (Figure 4, above). While the findings strongly encourage the exploration of additional processing alternatives (see Discussion), some of which may produce results yet superior to those here, we can provisionally recommend MSM-All alignment and the associated multimodal brain parcellation (Glasser, Coalson, et al., 2016), together with a multivariate learning model such as elastic net regression.

Finally, results for REST12 (all resting-state runs combined), using MSM-All alignment and denoising strategy A, and the multivariate learning model, are shown in Figure 5b together with statistical assessment using 1000 permutations. Only Openness to experience could be predicted above chance, albeit with a very small effect size (r=0.24, R2=0.024).

Predicting higher-order dimensions of personality (α and β)

In previous sections, we qualitatively observed that decreasing noise in individual FC matrices by averaging data over all available resting state runs (REST12, 1 h of data) leads to improvements in prediction performance compared to session-wise predictions (REST1 and REST2, 30 min of data each). We can also decrease noise in the behavioral data, by deriving composite scores that pool over a larger number of test items than the Big Five factor scores (each factor relies solely on 12 items in the NEO-FFI). The principal component analysis presented in Figure 2c is a way to achieve such pooling. We therefore next attempted to predict these two principal component scores, which we refer to as α and β, from REST12 FC matrices, using denoising A and MSM-all inter-subject alignment.

α was not predicted above chance, which was somewhat expected because it loads most highly on Neuroticism, which we could not predict well in the previous section.

β was predicted above chance (p1000<0.002), which we also expected because it loads most highly on Openness to experience (which had r = 0.24, R2=0.024; Figure 5b). Since β effectively combines variance from Openness with that from other factors (Conscientiousness, Extraversion, and Agreeableness; see Figure 2c) this leads to a slight improvement in predictability, and a doubling of the explained variance (β: r = 0.27, R2 = 0.050). This result strongly suggests that improving the reliability of scores on the behavioral side helps boost predictability (Gignac & Bates, 2017), just as improving the reliability of FC matrices by combining REST1 and REST2 improved predictability.

Discussion

Summary of results

Connectome-based predictive modeling (Dubois & Adolphs, 2016; Shen et al., 2017) has been an active field of research in the past years: it consists in using functional connectivity as measured from resting-state fMRI data to predict individual differences in demographics, behavior, psychological profile, or psychiatric diagnosis. Here, we applied this approach and attempted to predict the Big Five personality factors (McCrae & Costa, 1987) from resting-state data in a large public dataset, the Human Connectome Project (N=884 after exclusion criteria). We can summarize our findings as follows.

  1. We found that personality traits were not only intercorrelated with one another, but were also correlated with fluid intelligence, age, sex, handedness and other measures. We therefore regressed these possible confounds out, producing a residualized set of personality trait measures (that were, however, still intercorrelated amongst themselves).

  2. Comparing different processing pipelines and data from different fMRI sessions showed generally good stability of functional connectivity across time, a prerequisite for attempting to predict a personality trait that is also stable across time.

  3. We qualitatively replicated and extended a previously published finding, the prediction of a measure of fluid intelligence (Finn et al., 2015; Noble et al., 2017) from functional connectivity patterns, providing reassurance that our approach is able to predict individual differences when possible.

  4. We then carried out a total of 36 different analyses for each of the five personality factors. The 36 different analyses resulted from separately analysing data from 2 sessions (establishing test-retest reliability), each with 3 different preprocessing pipelines (exploring sensitivity to how the fMRI data are processed), 2 different alignment and hard parcellation schemes (providing initial results whether multimodal surface-based alignment improves on classical volumetric alignment), and 3 different predictive models (univariate positive, univariate negative, and multivariate). Across all of these alternatives, we generally found that the MSM-All multimodal alignment together with the parcellation scheme of Glasser et al. (2016) was associated with the greatest predictability; and likewise for the multivariate model (elastic net).

  5. Among the personality measures, Openness to experience showed the most reliable prediction between the two fMRI sessions, followed by Extraversion; for all other factors, predictions were often highly unstable, showing large variation depending on small changes in preprocessing, or across sessions.

  6. Combining data from both fMRI sessions improved predictions. Likewise, combining behavioral data through principal component analysis improved predictions. At both the neural and behavioral ends, improving the quality of our measurements could improve predictions.

  7. We best predicted the β superordinate factor, with r=0.27 and R2=0.05. This is highly significant as per permutation testing (though, in interpreting the statistical significance of any single finding, we note that one would have to correct for all the multiple analysis pipelines that we tested; future replications or extensions of this work would benefit from a pre-registered single approach to reduce the degrees of freedom in the analysis).

Though some of our findings achieve statistical significance in the large sample of subjects provided by the HCP, resting-state functional connectivity still only explains at most 5% of the variance in any personality score. We are thus still far from understanding the neurobiological substrates of personality (Yarkoni, 2015) (and, for that matter, of fluid intelligence which we predicted at a similar, slightly lower level; but, see Dubois et al., 2018). Indeed, based on this finding, it seems unlikely that findings from predictive approaches using whole-brain resting-state fMRI will inform hypotheses about specific neural systems that provide a causal mechanistic explanation of how personality is expressed in behavior.

Taken together, our approach provides important general guidelines for personality neuroscience studies using resting-state fMRI data: i) operations that are sometimes taken for granted, such as resting-state fMRI denoising (Abraham et al., 2017), make a difference to the outcome of connectome-based predictions and their test-retest reliability; ii) new inter-subject alignment procedures, such as multimodal surface matching (Robinson et al., 2014), improve performance and test-retest reliability; iii) a simple multivariate linear model may be a good alternative to the separate univariate models proposed by (Finn et al., 2015), yielding improved performance.

Our approach also draws attention to the tremendous analytical flexibility that is available in principle (Carp, 2012), and to the all-too-common practice of keeping such explorations “behind the scenes” and only reporting the “best” strategy, leading to an inflation of positive findings reported in the literature (Neuroskeptic, 2012; Simonsohn, Nelson, & Simmons, 2014). At a certain level, if all analyses conducted make sense (i.e., would pass a careful expert reviewer’s scrutiny), they should all give a similar answer to the final question (conceptually equivalent to inter-rater reliability; see Dubois & Adolphs, 2016).

Effect of subject alignment

The recently proposed multimodal surface matching framework uses a combination of anatomical and functional features to best align subject cortices. It improves functional inter-subject alignment over the classical approach of warping brains volumetrically (Dubois & Adolphs, 2016). For the scores that can be predicted from functional connectivity, alignment in the MSM-All space outperformed alignment in the MNI space. However, more work needs to be done to further establish the superiority of the MSM-All approach. Indeed, the parcellations used in this study differed between the MNI and MSM-All space: the parcellation in MSM-All space had more nodes (360, vs. 268) and no subcortical structures were included. Also, it is unclear how the use of resting-state data during the alignment process in the MSM-All framework interacts with resting-state based predictions, since the same data used for predictions has already been used to align subjects. Finally, it has recently been shown that the precise anatomy of each person’s brain, even after the best alignment, introduces variability that interacts with functional connectivity (Bijsterbosch et al., 2018). The complete description of brain variability at both structural and functional levels will need to be incorporated into future studies of individual differences.

Effect of preprocessing

We applied three separate, reasonable denoising strategies, inspired from published work (Ciric et al., 2017; Finn et al., 2015; Satterthwaite, Elliott, et al., 2013; Siegel et al., 2017) and our current understanding of resting-state fMRI confounds (Caballero-Gaudes & Reynolds, 2017; Murphy et al., 2013). The differences between the three denoising strategies in terms of the resulting test-retest reliability, based on several metrics, were not very large - yet, there were differences. Pipeline A appeared to yield the best reliability in terms of behavioral utility, while Pipeline B was best at conserving differences across subjects. Pipeline C performed worst on these metrics in our hands, despite its use of the automated artifact removal tool ICA-FIX (Salimi-Khorshidi et al., 2014); it is possible that performing CompCor and censoring are in fact detrimental after ICA-FIX (see also Muschelli et al., 2014). Finally, in terms of the final predictive score, all three strategies demonstrated acceptable test-retest reliability for scores that were successfully predicted.

The particular choices of pipelines that we made were intended to provide an initial survey of some commonly used schemes, but substantial future work will be needed to explore the space of possibilities more comprehensively. For instance, global signal regression - which was a part of all three chosen strategies - remains a somewhat controversial denoising step, and could be omitted if computing partial correlations, or replaced with a novel temporal ICA decomposition approach (Glasser et al., 2017). The bandpass filtering used in all our denoising approaches to reduce high frequency noise could also be replaced with alternatives such as PCA decomposition combined with “Wishart rolloff” (Glasser, Smith, et al., 2016). All of these choices impact the amount and quality of information in principle available, and how that information can be used to build a predictive model.

Effect of predictive algorithm

Our exploration of a multivariate model was motivated by the seemingly arbitrary decision to weight all edges equally in the univariate models (positive and negative) proposed by Finn et al. (2015). However, we also recognize the need for simple models, given the paucity of data compared to the number of features (curse of dimensionality). We thus explored a regularized regression model that would combine information from negative and positive edges optimally, after performing the same feature-filtering step as in the univariate models. The multivariate model performed best on the scores that were predicted most reliably, yet it also seemed to have lower test-retest reliability. More work remains to be done on this front to find the best simple model that optimally combines information from all edges and can be trained in a situation with limited data.

Statistical significance

It is inappropriate to assess statistical significance using parametric statistics in the case of a cross-validation analysis (Figure 4b). However, for complex analyses, it is often the preferred option, due to the prohibitive computational resources needed to run permutation tests. Here we showed the empirical distribution of chance prediction scores for both the univariate (positive)- and multivariate-model predictions of fluid intelligence (PMAT24_A_CR) using denoising pipeline A in MSM-All space (Figure 4b). As expected, the permutation distribution is wider than the parametric estimate; it also differs significantly between the univariate and the multivariate models. This finding stresses that one needs to calculate permutation statistics for the specific analysis that one runs. The calculation of permutation statistics should be feasible given the rapid increase and ready availability of computing clusters with multiple processors. We show permutation statistics for all our key findings, but we did not correct for the multiple comparisons (5 personality factors, multiple processing pipelines). Future studies should ideally provide analyses that are pre-registered to reduce the degrees of freedom available and aid interpretation of statistical reliability.

Will our findings reproduce?

It is common practice in machine learning competitions to set aside a portion of your data and not look at it at all until a final analysis has been decided, and only then to run that single final analysis on the held-out data to establish out-of-sample replication. We decided not to split our dataset in that way due to its already limited sample size, and instead used a careful cross-validation framework, assessed test-retest reliability across data from different sessions, and refrained from adaptively changing parameters upon examining the final results. The current paper should now serve as the basis of a pre-registered replication, to be performed on an independent dataset (a good candidate would be the NKI enhanced dataset (Nooner et al., 2012), which also contains assessment of the Big Five).

On the relationship between brain and personality

The best neural predictor of personality may be distinct, wholly or in part, from the actual neural mechanisms by which personality expresses itself on any given occasion. Personality may stem from a disjunctive and heterogeneous set of biological constraints that in turn influence brain function in complex ways (Yarkoni, 2015); neural predictors may simply be conceived of as “markers” of personality: any correlated measures that a machine learning algorithm could use as information, on the basis of which it could be trained in a supervised fashion to discriminate among personality traits. Our goal in this study was to find such predictions, not a causal explanation (see Yarkoni & Westfall, 2017). It may well someday be possible to predict personality differences from fMRI data with much greater accuracy than what we found here. However, we think it likely that, in general, such an approach will still fall short of uncovering the neural mechanisms behind personality, in the sense of explaining the proximal causal processes whereby personality is expressed in behavior on specific occasions.

Subjective and objective measures of personality

As noted already in the introduction, it is worth keeping in mind the history of the Big Five: they derive from factor analyses of words, of the vocabularies that we use to describe people. As such, they fundamentally reflect our folk psychology, and our social inferences (“theory of mind”) about other people. This factor structure was then used to design a self-report instrument, in which participants are asked about themselves (the NEO or variations thereof). Unlike some other self-report indices (such as the MMPI), the NEO-FFI does not assess test-taking approach (e.g. consistency across items or tendency toward a particular response set), and thus, offers no insight regarding validity of any individual’s responses. This is a notable limitation, as there is substantial evidence that NEO-FFI scores may be intentionally manipulated by the subject’s response set (Furnham, 1997; Topping & O’Gorman, 1997). Even in the absence of intentional ‘faking’, NEO outcomes are likely to be influenced by an individual’s insight, impression management, and reference group effects. However, these limitations may be addressed by applying the same analysis to multiple personality measures with varying degrees of face-validity and objectivity, as well as measures that include indices of response bias. This might include ratings provided by a familiar informant, implicit-association tests (e.g. Schnabel, Asendorpf, & Greenwald, 2008), and spontaneous behavior (e.g. Mehl, Gosling, & Pennebaker, 2006). Future development of behavioral measures of personality that provide better convergent validity and discriminative specificity will be an important component of personality neuroscience.

Limitations and Future Directions

There are several limitations of the present study that could be improved upon or extended in future work. In addition to the obvious issue of simply needing more, and/or better quality, data, there is the important issue of obtaining a better estimate of variability within a single subject. This is especially pertinent for personality traits, which are supposed to be relatively stable within an individual. Thus, collecting multiple fMRI datasets, perhaps over weeks or even years, could help to find those features in the data with the best cross-temporal stability. Indeed several such dense datasets across multiple sessions in a few subjects have already been collected, and may help guide the intelligent selection of features with the greatest temporal stability (Gordon et al., 2017; Noble et al., 2017; Poldrack et al., 2015). Against expectations, initial analyses seem to indicate that the most reliable edges in FC from such studies are not necessarily the most predictive edges (for fluid intelligence; see Noble et al., 2017), yet more work needs to be done to further test this hypothesis. It is also possible that shorter timescale fluctuations in resting-state fMRI provide additional information (if these are stable over longer times), and it might thus be fruitful to explore dynamic FC, as some work has done (Calhoun, Miller, Pearlson, & Adalı, 2014; Jia, Hu, & Deshpande, 2014; Vidaurre, Smith, & Woolrich, 2017).

No less important would be improvements on the behavioral end, as we alluded to in the previous section. Developing additional tests of personality to provide convergent validity to the personality dimension constructs would help provide a more accurate estimate of these latent variables. Just as with the fMRI data, collecting personality scores across time should help to prioritize those items that have the greatest temporal stability and reduce measurement error.

Another limitation is signal-to-noise. It may be worth exploring fMRI data obtained while watching a movie that drives relevant brain function, rather than during rest, in order to maximize the signal variance in the fMRI signal. Similarly, it could be beneficial to include participants with a greater range of personality scores, perhaps even including those with a personality disorder. A greater range of signal both on the fMRI end and on the behavioral end would help provide greater power to detect associations.

One particularly relevant aspect of our approach is that the models we used, like most in the literature, were linear. Nonlinear models may be more appropriate, yet the difficulty in using such models is that they would require a much larger number of training samples relative to the number of features in the dataset. This could be accomplished both by accruing ever larger databases of rs-fMRI data, and by further reducing the dimensionality of the data, for instance through PCA or coarser parcellations. Alternatively, one could form a hypothesis about the shape of the function that might best predict personality scores and explicitly include this in a model.

A final important but complex issue concerns the correlation between most behavioral measures. In our analyses, we regressed out fluid intelligence, age, and sex, amongst other variables. But there are many more that are likely to be correlated with personality at some level. If one regressed out all possible measures, one would likely end up removing what one is interested in, since eventually the residual of personality would shrink to a very small range. An alternative approach is to use the raw personality scores (without any removal of confounds at all), and then selectively regress out fluid intelligence, memory task performance, mood, etc., and make comparisons between the results obtained (we provide such minimally deconfounded results in Supplementary Figure 2). This could yield insights into which other variables are driving the predictability of a personality trait. It could also suggest specific new variables to investigate in their own right. Finally, multiple regression may not be the best approach to addressing these confounds, due to noise in the measurements. Specifying confounds within a structural equation model may be a better approach (Westfall & Yarkoni, 2016).

Recommendations for Personality Neuroscience

There are well known challenges to the reliability and reproducibility of findings in personality neuroscience, which we have already mentioned. The field shares these with any other attempt to link neuroscience data with individual differences (Dubois & Adolphs, 2016). We conclude with some specific recommendations for the field going forward, focusing on the use of resting-state fMRI data.

  1. Given the effect sizes that we report here (which are by no means a robust estimate, yet do provide a basis on which to build), we think it would be fair to recommend a minimum sample size of 500 or so subjects (Schönbrodt & Perugini, 2013) for connectome-based predictions. If other metrics are used, a careful estimate of effect size that adjusts for bias in the literature should be undertaken for the specific brain measure of interest (cf. Anderson, Kelley, & Maxwell, 2017).

  2. A predictive framework is essential (Dubois & Adolphs, 2016; Yarkoni & Westfall, 2017), as it ensures out-of-sample reliability. Personality neuroscience studies should use proper cross-validation (in the case of the HCP, taking family structure into account), with permutation statistics. Even better, studies should include a replication sample which is held out and not examined at all until the final model has been decided from the discovery sample (advanced methods may help implement this in a more practical manner; e.g. Dwork et al., 2015).

  3. Data sharing. If new data are collected by individual labs, it would be very important to make these available, in order to eventually accrue the largest possible sample size in a database. It has been suggested that contact information about the participants would also be valuable, so that additional measures (or retest reliability) could be collected (Mar et al., 2013). Some of these data could be collected over the internet.

  4. Complete transparency and documentation of all analyses, including sharing of all analysis scripts, so that the methods of published studies can be reproduced. Several papers give more detailed recommendations for using and reporting fMRI data, see Dubois & Adolphs, 2016; Nichols et al., 2016; Poldrack et al., 2008). Our paper makes specific recommendation about detailed parcellation, processing and modeling pipelines; however, this is a continuously evolving field and these recommendations will likely change with future work. For personality in particular, detailed assessment for all participants, and justified exclusionary and inclusionary criteria should be provided. As suggested above, authors should consider pre-registering their study, on the Open Science Framework or a similar platform.

  5. Ensure reliable and uniform behavioral estimates of personality. This is perhaps one of the largest unsolved challenges. Compared with the huge ongoing effort and continuous development of the processing and analysis of fMRI data, the measures for personality are mostly stagnant and face many problems of validity. For the time being, a simple recommendation would be to use a consistent instrument and stick with the Big Five, so as not to mix apples and oranges by using very different instruments. That said, it will be important to explore other personality measures and structures. As we noted above, there is in principle a large range of more subjective, or more objective, measures of personality. It would be a boon to the field if these were more systematically collected, explored, and possibly combined to obtain the best estimate of the latent variable of personality they are thought to measure.

  6. Last but not least, we should consider methods in addition to fMRI and species in addition to humans. To the extent that a human personality dimension appears to have a valid correlate in an animal model, it might be possible to collect large datasets, and to complement fMRI with optical imaging or other modalities. Studies in animals may also yield the most powerful tools to examine specific neural circuits, a level of causal mechanism that, as we argued above, may largely elude analyses using resting-state fMRI.

Supplementary Material

Figure 6. Prediction results for superordinate factors/principal components α and β, using REST12 data (1h of resting-state fMRI per subject).

Figure 6

These results use MSM-All inter-subject alignment, denoising strategy A, and the multivariate prediction model. As in Figure 5b, the range of predicted scores is much narrower than the range of observed scores. a. The first PC, α, is not predicted better than chance. α loads mostly on Neuroticism (see Figure 2c), which was itself not predicted well (cf. Figure 5). b. We can predict about 5% of the variance in the score on the second PC, β. This is better than chance, as established by permutation statistics (p1000<0.002). β loads mostly on Openness to Experience (see Figure 2c), which showed good predictability in the previous section.

Acknowledgments

This work was supported by NIMH grant 2P50MH094258 (RA), the Carver Mead Seed Fund (RA), and a NARSAD Young Investigator Grant from the Brain and Behavior Research Foundation (JD).

Footnotes

Author contributions

J. Dubois and P. Galdi developed the overall general analysis framework and conducted some of the initial analyses for the paper. J. Dubois conducted all final analyses and produced all figures. Y. Han helped with literature search and analysis of behavioral data. L. Paul helped with literature search, analysis of behavioral data, and interpretation of the results. J. Dubois and R. Adolphs wrote the initial manuscript and all authors contributed to the final manuscript. All authors contributed to planning and discussion on this project. The authors declare no conflict of interest.

Data Sharing

The Young Adult HCP dataset is publicly available at https://www.humanconnectome.org/study/hcp-young-adult. Analysis scripts are available in the following public repository: https://github.com/adolphslab/HCP_MRI-behavior.

References

  1. Abraham A, Milham MP, Di Martino A, Craddock RC, Samaras D, Thirion B, Varoquaux G. Deriving reproducible biomarkers from multi-site resting-state data: An Autism-based example. NeuroImage. 2017;147:736–745. doi: 10.1016/j.neuroimage.2016.10.045. [DOI] [PubMed] [Google Scholar]
  2. Abraham A, Pedregosa F, Eickenberg M, Gervais P, Mueller A, Kossaifi J, … Varoquaux G. Machine learning for neuroimaging with scikit-learn. Frontiers in Neuroinformatics. 2014;8:14. doi: 10.3389/fninf.2014.00014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Adelstein JS, Shehzad Z, Mennes M, Deyoung CG, Zuo XN, Kelly C, … Milham MP. Personality is reflected in the brain’s intrinsic functional architecture. PloS One. 2011;6(11):e27633. doi: 10.1371/journal.pone.0027633. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Aghajani M, Veer IM, van Tol MJ, Aleman A, van Buchem MA, Veltman DJ, … van der Wee NJ. Neuroticism and extraversion are associated with amygdala resting-state functional connectivity. Cognitive, Affective & Behavioral Neuroscience. 2014;14(2):836–848. doi: 10.3758/s13415-013-0224-0. [DOI] [PubMed] [Google Scholar]
  5. Alexander DLJ, Tropsha A, Winkler DA. Beware of R(2): Simple, Unambiguous Assessment of the Prediction Accuracy of QSAR and QSPR Models. Journal of Chemical Information and Modeling. 2015;55(7):1316–1322. doi: 10.1021/acs.jcim.5b00206. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Allemand M, Zimprich D, Hendriks AAJ. Age differences in five personality domains across the life span. Developmental Psychology. 2008;44(3):758–770. doi: 10.1037/0012-1649.44.3.758. [DOI] [PubMed] [Google Scholar]
  7. Amelang M, Borkenau Über die faktorielle Struktur und externe Validität einiger Fragebogen-Skalen zur Erfassung von Dimensionen der Extraversion und emotionalen Labilität. Zeitschrift für Differentielle und Diagnostische Psychologie. 1982;3(2):119–145. Retrieved from http://psycnet.apa.org/psycinfo/1983-22406-001. [Google Scholar]
  8. Anderson SF, Kelley K, Maxwell SE. Sample-Size Planning for More Accurate Statistical Power: A Method Adjusting Sample Effect Sizes for Publication Bias and Uncertainty. Psychological Science. 2017;28(11):1547–1562. doi: 10.1177/0956797617723724. [DOI] [PubMed] [Google Scholar]
  9. Back MD, Schmukle SC, Egloff B. Predicting actual behavior from the explicit and implicit self-concept of personality. Journal of Personality and Social Psychology. 2009;97(3):533–548. doi: 10.1037/a0016229. [DOI] [PubMed] [Google Scholar]
  10. Baeken C, Marinazzo D, Van Schuerbeek P, Wu GR, De Mey J, Luypaert R, De Raedt R. Left and right amygdala - mediofrontal cortical functional connectivity is differentially modulated by harm avoidance. PloS One. 2014;9(4):e95740. doi: 10.1371/journal.pone.0095740. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bartels M, van Weegen FI, van Beijsterveldt CEM, Carlier M, Polderman TJC, Hoekstra RA, Boomsma DI. The five factor model of personality and intelligence: A twin study on the relationship between the two constructs. Personality and Individual Differences. 2012;53(4):368–373. doi: 10.1016/j.paid.2012.02.007. [DOI] [Google Scholar]
  12. Beaty RE, Benedek M, Wilkins RW, Jauk E, Fink A, Silvia PJ, … Neubauer AC. Creativity and the default network: A functional connectivity analysis of the creative brain at rest. Neuropsychologia. 2014;64:92–98. doi: 10.1016/j.neuropsychologia.2014.09.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Beaty RE, Kaufman SB, Benedek M, Jung RE, Kenett YN, Jauk E, … Silvia PJ. Personality and complex brain networks: The role of openness to experience in default network efficiency. Human Brain Mapping. 2016;37(2):773–779. doi: 10.1002/hbm.23065. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Behzadi Y, Restom K, Liau J, Liu TT. A component based noise correction method (CompCor) for BOLD and perfusion based fMRI. NeuroImage. 2007;37(1):90–101. doi: 10.1016/j.neuroimage.2007.04.042. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Bijsterbosch JD, Woolrich MW, Glasser MF, Robinson EC, Beckmann CF, Van Essen DC, … Smith SM. The relationship between spatial configuration and functional connectivity of brain regions. eLife. 2018;7:e32992. doi: 10.7554/eLife.32992. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Bilker WB, Hansen JA, Brensinger CM, Richard J, Gur RE, Gur RC. Development of abbreviated nine-item forms of the Raven’s standard progressive matrices test. Assessment. 2012;19(3):354–369. doi: 10.1177/1073191112446655. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Birn RM, Shackman AJ, Oler JA, Williams LE, McFarlin DR, Rogers GM, … Kalin NH. Evolutionarily conserved prefrontal-amygdalar dysfunction in early-life anxiety. Molecular Psychiatry. 2014;19(8):915–922. doi: 10.1038/mp.2014.46. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Biswal BB, Mennes M, Zuo XN, Gohel S, Kelly C, Smith SM, … Milham MP. Toward discovery science of human brain function. Proceedings of the National Academy of Sciences of the United States of America. 2010;107(10):4734–4739. doi: 10.1073/pnas.0911855107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Bjørnebekk A, Fjell AM, Walhovd KB, Grydeland H, Torgersen S, Westlye LT. Neuronal correlates of the five factor model (FFM) of human personality: Multimodal imaging in a large healthy sample. NeuroImage. 2013;65(Supplement C):194–208. doi: 10.1016/j.neuroimage.2012.10.009. [DOI] [PubMed] [Google Scholar]
  20. Blackburn R, Renwick SJD, Donnelly JP, Logan C. Big Five or Big Two? Superordinate factors in the NEO Five Factor Inventory and the Antisocial Personality Questionnaire. Personality and Individual Differences. 2004;37(5):957–970. doi: 10.1016/j.paid.2003.10.017. [DOI] [Google Scholar]
  21. Blankstein U, Chen JYW, Mincic AM, McGrath PA, Davis KD. The complex minds of teenagers: neuroanatomy of personality differs between sexes. Neuropsychologia. 2009;47(2):599–603. doi: 10.1016/j.neuropsychologia.2008.10.014. [DOI] [PubMed] [Google Scholar]
  22. Block J. A contrarian view of the five-factor approach to personality description. Psychological Bulletin. 1995;117(2):187–215. doi: 10.1037/0033-2909.117.2.187. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/7724687. [DOI] [PubMed] [Google Scholar]
  23. Borgatta EF. THE STRUCTURE OF PERSONALITY CHARACTERISTICS. Behavioral Science. 1964;61(1):8–17. doi: 10.1002/bs.3830090103. [DOI] [PubMed] [Google Scholar]
  24. Bouchard TJ, Jr, McGue M. Genetic and environmental influences on human psychological differences. Journal of Neurobiology. 2003;54(1):4–45. doi: 10.1002/neu.10160. [DOI] [PubMed] [Google Scholar]
  25. Button KS, Ioannidis JPA, Mokrysz C, Nosek BA, Flint J, Robinson ESJ, Munafò MR. Power failure: why small sample size undermines the reliability of neuroscience. Nature Reviews. Neuroscience. 2013;14(5):365–376. doi: 10.1038/nrn3475. [DOI] [PubMed] [Google Scholar]
  26. Caballero-Gaudes C, Reynolds RC. Methods for cleaning the BOLD fMRI signal. NeuroImage. 2017;154:128–149. doi: 10.1016/j.neuroimage.2016.12.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Calhoun VD, Miller R, Pearlson G, Adalı T. The Chronnectome: Time-Varying Connectivity Networks as the Next Frontier in fMRI Data Discovery. Neuron. 2014;84(2):262–274. doi: 10.1016/j.neuron.2014.10.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Canli T. Biology of Personality and Individual Differences. New York: Guilford Press; 2006. Retrieved from https://market.android.com/details?id=book-DH-IWaVdRHIC. [Google Scholar]
  29. Canli T, Zhao Z, Desmond JE, Kang E, Gross J, Gabrieli JD. An fMRI study of personality influences on brain reactivity to emotional stimuli. Behavioral Neuroscience. 2001;115(1):33–42. doi: 10.1037/0735-7044.115.1.33. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/11256451. [DOI] [PubMed] [Google Scholar]
  30. Carp J. On the plurality of (methodological) worlds: estimating the analytic flexibility of FMRI experiments. Frontiers in Neuroscience. 2012;6:149. doi: 10.3389/fnins.2012.00149. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Cattell RB. The Description of Personality: Principles and Findings in a Factor Analysis. The American Journal of Psychology. 1945;58(1):69–90. doi: 10.2307/1417576. [DOI] [Google Scholar]
  32. Chamorro-Premuzic T, Furnham A. A possible model for understanding the personality--intelligence interface. British Journal of Psychology. 2004;95(Pt 2):249–264. doi: 10.1348/000712604773952458. [DOI] [PubMed] [Google Scholar]
  33. Ciric R, Wolf DH, Power JD, Roalf DR, Baum GL, Ruparel K, … Satterthwaite TD. Benchmarking of participant-level confound regression strategies for the control of motion artifact in studies of functional connectivity. NeuroImage. 2017;154:174–187. doi: 10.1016/j.neuroimage.2017.03.020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Cohen MX, Schoene-Bake JC, Elger CE, Weber B. Connectivity-based segregation of the human striatum predicts personality characteristics. Nature Neuroscience. 2009;12(1):32–34. doi: 10.1038/nn.2228. [DOI] [PubMed] [Google Scholar]
  35. Cole MW, Yarkoni T, Repovs G, Anticevic A, Braver TS. Global connectivity of prefrontal cortex predicts cognitive control and intelligence. The Journal of Neuroscience: The Official Journal of the Society for Neuroscience. 2012;32(26):8988–8999. doi: 10.1523/JNEUROSCI.0536-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Combrisson E, Jerbi K. Exceeding chance level by chance: The caveat of theoretical chance levels in brain signal classification and statistical assessment of decoding accuracy. Journal of Neuroscience Methods. 2015;250:126–136. doi: 10.1016/j.jneumeth.2015.01.010. [DOI] [PubMed] [Google Scholar]
  37. Costa PT, Jr, McCrae RR. Domains and facets: hierarchical personality assessment using the revised NEO personality inventory. Journal of Personality Assessment. 1995;64(1):21–50. doi: 10.1207/s15327752jpa6401_2. [DOI] [PubMed] [Google Scholar]
  38. Costa PT, McCrae RR. Neo PI-R professional manual. Vol. 396. Odessa, FL: Psychological Assessment Resources; 1992. pp. 653–665. [Google Scholar]
  39. Costa PT, Terracciano A, McCrae RR. Gender differences in personality traits across cultures: robust and surprising findings. Journal of Personality and Social Psychology. 2001;81(2):322–331. doi: 10.1037/0022-3514.81.2.322. [DOI] [PubMed] [Google Scholar]
  40. Coutinho JF, Sampaio A, Ferreira M, Soares JM, Gonçalves OF. Brain correlates of pro-social personality traits: a voxel-based morphometry study. Brain Imaging and Behavior. 2013;7(3):293–299. doi: 10.1007/s11682-013-9227-2. [DOI] [PubMed] [Google Scholar]
  41. Crum RM, Anthony JC, Bassett SS, Folstein MF. Population-based norms for the Mini-Mental State Examination by age and educational level. JAMA: The Journal of the American Medical Association. 1993;269(18):2386–2391. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/8479064. [PubMed] [Google Scholar]
  42. D’Agostino R, Pearson ES. Tests for Departure from Normality. Empirical Results for the Distributions of b2 and √ b1. Biometrika. 1973;60(3):613–622. doi: 10.2307/2335012. [DOI] [Google Scholar]
  43. Deris N, Montag C, Reuter M, Weber B, Markett S. Functional connectivity in the resting brain as biological correlate of the Affective Neuroscience Personality Scales. NeuroImage. 2017;147(Supplement C):423–431. doi: 10.1016/j.neuroimage.2016.11.063. [DOI] [PubMed] [Google Scholar]
  44. DeYoung CG. Higher-order factors of the Big Five in a multi-informant sample. Journal of Personality and Social Psychology. 2006;91(6):1138–1151. doi: 10.1037/0022-3514.91.6.1138. [DOI] [PubMed] [Google Scholar]
  45. DeYoung CG, Hirsh JB, Shane MS, Papademetris X, Rajeevan N, Gray JR. Testing predictions from personality neuroscience. Brain structure and the big five. Psychological Science. 2010;21(6):820–828. doi: 10.1177/0956797610370159. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Digman JM. Higher-order factors of the Big Five. Journal of Personality and Social Psychology. 1997;73(6):1246–1256. doi: 10.1037//0022-3514.73.6.1246. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/9418278. [DOI] [PubMed] [Google Scholar]
  47. Dosenbach NUF, Nardos B, Cohen AL, Fair DA, Power JD, Church JA, … Schlaggar BL. Prediction of individual brain maturity using fMRI. Science. 2010;329(5997):1358–1361. doi: 10.1126/science.1194144. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Dubois J, Adolphs R. Building a Science of Individual Differences from fMRI. Trends in Cognitive Sciences. 2016;20(6):425–443. doi: 10.1016/j.tics.2016.03.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Dubois J, Galdi P, Paul LK, Adolphs R. A distributed brain network predicts general intelligence from resting-state human neuroimaging data. bioRxiv. 2018 Jan 31; doi: 10.1101/257865. [DOI] [PMC free article] [PubMed]
  50. Dwork C, Feldman V, Hardt M, Pitassi T, Reingold O, Roth A. The reusable holdout: Preserving validity in adaptive data analysis. Science. 2015;349(6248):636–638. doi: 10.1126/science.aaa9375. [DOI] [PubMed] [Google Scholar]
  51. Egan V, Deary I, Austin E. The NEO-FFI: emerging British norms and an item-level analysis suggest N, A and C are more reliable than O and E. Personality and Individual Differences. 2000;29(5):907–920. doi: 10.1016/S0191-8869(99)00242-1. [DOI] [Google Scholar]
  52. Eickhoff S, Nichols TE, Van Horn JD, Turner JA. Sharing the wealth: Neuroimaging data repositories. NeuroImage. 2016;124(Pt B):1065–1068. doi: 10.1016/j.neuroimage.2015.10.079. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Elam J. Ramifications of Image Reconstruction Version Differences. 2015 Mar 16; Retrieved from https://wiki.humanconnectome.org/display/PublicData/Ramifications+of+Image+Reconstruction+Version+Differences.
  54. Feingold A. Gender differences in personality: a meta-analysis. Psychological Bulletin. 1994;116(3):429. doi: 10.1037/0033-2909.116.3.429. Retrieved from http://psycnet.apa.org/psycinfo/1995-09434-001. [DOI] [PubMed] [Google Scholar]
  55. Finn ES, Shen X, Scheinost D, Rosenberg MD, Huang J, Chun MM, … Constable RT. Functional connectome fingerprinting: identifying individuals using patterns of brain connectivity. Nature Neuroscience. 2015;18(11):1664–1671. doi: 10.1038/nn.4135. [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Fiske DW. Consistency of the factorial structures of personality ratings from different sour sources. Journal of Abnormal Psychology. 1949;44(3):329–344. doi: 10.1037/h0057198. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/18146776. [DOI] [PubMed] [Google Scholar]
  57. Furnham AF. Knowing and Faking One’s Five-Factor Personality Score. Journal of Personality Assessment. 1997;69(1):229–243. doi: 10.1207/s15327752jpa6901_14. [DOI] [Google Scholar]
  58. Furr RM. Personality psychology as a truly behavioural science. European Journal of Personality. 2009;23(5):369–401. doi: 10.1002/per.724. [DOI] [Google Scholar]
  59. Gabrieli JDE, Ghosh SS, Whitfield-Gabrieli S. Prediction as a humanitarian and pragmatic contribution from human cognitive neuroscience. Neuron. 2015;85(1):11–26. doi: 10.1016/j.neuron.2014.10.047. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Gao Q, Xu Q, Duan X, Liao W, Ding J, Zhang Z, … Chen H. Extraversion and neuroticism relate to topological properties of resting-state brain networks. Frontiers in Human Neuroscience. 2013;7:257. doi: 10.3389/fnhum.2013.00257. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Geerligs L, Renken RJ, Saliasi E, Maurits NM, Lorist MM. A Brain-Wide Study of Age-Related Changes in Functional Connectivity. Cerebral Cortex. 2015;25(7):1987–1999. doi: 10.1093/cercor/bhu012. [DOI] [PubMed] [Google Scholar]
  62. Geerligs L, Rubinov M, Cam-Can &, Henson RN. State and Trait Components of Functional Connectivity: Individual Differences Vary with Mental State. The Journal of Neuroscience: The Official Journal of the Society for Neuroscience. 2015;35(41):13949–13961. doi: 10.1523/JNEUROSCI.1324-15.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Gignac GE, Bates TC. Brain volume and intelligence: The moderating role of intelligence measurement quality. Intelligence. 2017;64(Supplement C):18–29. doi: 10.1016/j.intell.2017.06.004. [DOI] [Google Scholar]
  64. Glasser MF, Coalson TS, Bijsterbosch JD, Harrison SJ, Harms MP, Anticevic A, … Smith SM. Using Temporal ICA to Selectively Remove Global Noise While Preserving Global Signal in Functional MRI Data. bioRxiv. 2017 Dec 13; doi: 10.1101/193862. [DOI] [PMC free article] [PubMed]
  65. Glasser MF, Coalson TS, Robinson EC, Hacker CD, Harwell J, Yacoub E, … Van Essen DC. A multi-modal parcellation of human cerebral cortex. Nature. 2016;536(7615):171–178. doi: 10.1038/nature18933. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Glasser MF, Smith SM, Marcus DS, Andersson JLR, Auerbach EJ, Behrens TEJ, … Van Essen DC. The Human Connectome Project’s neuroimaging approach. Nature Neuroscience. 2016;19(9):1175–1187. doi: 10.1038/nn.4361. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Glasser MF, Sotiropoulos SN, Wilson JA, Coalson TS, Fischl B, Andersson JL … WU-Minn HCP Consortium. The minimal preprocessing pipelines for the Human Connectome Project. NeuroImage. 2013;80:105–124. doi: 10.1016/j.neuroimage.2013.04.127. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Goldberg LR. Language and individual differences: The search for universals in personality lexicons. Personality and Social Psychology Review: An Official Journal of the Society for Personality and Social Psychology, Inc. 1981;2(1):141–165. Retrieved from http://projects.ori.org/lrg/PDFs_papers/universals.lexicon.81.pdf. [Google Scholar]
  69. Gordon EM, Laumann TO, Adeyemo B, Huckins JF, Kelley WM, Petersen SE. Generation and Evaluation of a Cortical Area Parcellation from Resting-State Correlations. Cerebral Cortex. 2016;26(1):288–303. doi: 10.1093/cercor/bhu239. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Gordon EM, Laumann TO, Gilmore AW, Newbold DJ, Greene DJ, Berg JJ, … Dosenbach NUF. Precision Functional Mapping of Individual Human Brains. Neuron. 2017;95(4):791–807e7. doi: 10.1016/j.neuron.2017.07.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Gorgolewski K, Burns CD, Madison C, Clark D, Halchenko YO, Waskom ML, Ghosh SS. Nipype: a flexible, lightweight and extensible neuroimaging data processing framework in python. Frontiers in Neuroinformatics. 2011;5:13. doi: 10.3389/fninf.2011.00013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Gorgolewski KJ, Esteban O, Ellis DG, Notter MP, Ziegler E, Johnson H, … Ghosh S. nipy/nipype: Release 0.13.1. 2017 May; doi: 10.5281/zenodo.581704. [DOI]
  73. Gosling SD, John OP. Personality Dimensions in Nonhuman Animals: A Cross-Species Review. Current Directions in Psychological Science. 1999;8(3):69–75. doi: 10.1111/1467-8721.00017. [DOI] [Google Scholar]
  74. Gosling SD, Vazire S. Are we barking up the right tree? Evaluating a comparative approach to personality. Journal of Research in Personality. 2002;36(6):607–614. doi: 10.1016/S0092-6566(02)00511-1. [DOI] [Google Scholar]
  75. Gray JC. NEO-FFI Agreeableness scoring. 2017 Oct 4; Retrieved from https://www.mail-archive.com/hcp-users@humanconnectome.org/msg05266.html.
  76. Gur RC, Ragland JD, Moberg PJ, Turner TH, Bilker WB, Kohler C, … Gur RE. Computerized neurocognitive scanning: I. Methodology and validation in healthy people. Neuropsychopharmacology: Official Publication of the American College of Neuropsychopharmacology. 2001;25(5):766–776. doi: 10.1016/S0893-133X(01)00278-0. [DOI] [PubMed] [Google Scholar]
  77. Gur RC, Richard J, Hughett P, Calkins ME, Macy L, Bilker WB, … Gur RE. A cognitive neuroscience-based computerized battery for efficient measurement of individual differences: standardization and initial construct validation. Journal of Neuroscience Methods. 2010;187(2):254–262. doi: 10.1016/j.jneumeth.2009.11.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Hänggi J, Fövenyi L, Liem F, Meyer M, Jäncke L. The hypothesis of neuronal interconnectivity as a function of brain size-a general organization principle of the human connectome. Frontiers in Human Neuroscience. 2014;8:915. doi: 10.3389/fnhum.2014.00915. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Holmes AJ, Lee PH, Hollinshead MO, Bakst L, Roffman JL, Smoller JW, Buckner RL. Individual differences in amygdala-medial prefrontal anatomy link negative affect, impaired social functioning, and polygenic depression risk. The Journal of Neuroscience: The Official Journal of the Society for Neuroscience. 2012;32(50):18087–18100. doi: 10.1523/JNEUROSCI.2531-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Hong RY, Paunonen SV, Slade HP. Big Five personality factors and the prediction of behavior: A multitrait–multimethod approach. Personality and Individual Differences. 2008;45(2):160–166. doi: 10.1016/j.paid.2008.03.015. [DOI] [Google Scholar]
  81. Hutton C, Draganski B, Ashburner J, Weiskopf N. A comparison between voxel-based cortical thickness and voxel-based morphometry in normal aging. NeuroImage. 2009;48(2):371–380. doi: 10.1016/j.neuroimage.2009.06.043. [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Hu X, Erb M, Ackermann H, Martin JA, Grodd W, Reiterer SM. Voxel-based morphometry studies of personality: issue of statistical model specification--effect of nuisance covariates. NeuroImage. 2011;54(3):1994–2005. doi: 10.1016/j.neuroimage.2010.10.024. [DOI] [PubMed] [Google Scholar]
  83. Jaccard JJ. Predicting social behavior from personality traits. Journal of Research in Personality. 1974;7(4):358–367. doi: 10.1016/0092-6566(74)90057-9. [DOI] [Google Scholar]
  84. Jang KL, Livesley WJ, Vernon PA. Heritability of the big five personality dimensions and their facets: a twin study. Journal of Personality. 1996;64(3):577–591. doi: 10.1111/j.1467-6494.1996.tb00522.x. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/8776880. [DOI] [PubMed] [Google Scholar]
  85. Jia H, Hu X, Deshpande G. Behavioral relevance of the dynamics of the functional brain connectome. Brain Connectivity. 2014;4(9):741–759. doi: 10.1089/brain.2014.0300. [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Jiao B, Zhang D, Liang A, Liang B, Wang Z, Li J, … Liu M. Association between resting-state brain network topological organization and creative ability: Evidence from a multiple linear regression model. Biological Psychology. 2017;129:165–177. doi: 10.1016/j.biopsycho.2017.09.003. [DOI] [PubMed] [Google Scholar]
  87. Job DE, Dickie DA, Rodriguez D, Robson A, Danso S, Pernet C, … Wardlaw JM. A brain imaging repository of normal structural MRI across the life course: Brain Images of Normal Subjects (BRAINS) NeuroImage. 2017;144(Pt B):299–304. doi: 10.1016/j.neuroimage.2016.01.027. [DOI] [PubMed] [Google Scholar]
  88. Kapogiannis D, Sutin A, Davatzikos C, Costa P, Jr, Resnick S. The five factors of personality and regional cortical variability in the Baltimore longitudinal study of aging. Human Brain Mapping. 2013;34(11):2829–2840. doi: 10.1002/hbm.22108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  89. Kim MJ, Whalen PJ. The structural integrity of an amygdala-prefrontal pathway predicts trait anxiety. The Journal of Neuroscience: The Official Journal of the Society for Neuroscience. 2009;29(37):11614–11618. doi: 10.1523/JNEUROSCI.2335-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  90. Laumann TO, Gordon EM, Adeyemo B, Snyder AZ, Joo SJ, Chen MY, … Petersen SE. Functional System and Areal Organization of a Highly Sampled Individual Human Brain. Neuron. 2015;87(3):657–670. doi: 10.1016/j.neuron.2015.06.037. [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Lei X, Zhao Z, Chen H. Extraversion is encoded by scale-free dynamics of default mode network. NeuroImage. 2013;74:52–57. doi: 10.1016/j.neuroimage.2013.02.020. [DOI] [PubMed] [Google Scholar]
  92. Liu TT. Noise contributions to the fMRI signal: An overview. NeuroImage. 2016;143:141–151. doi: 10.1016/j.neuroimage.2016.09.008. [DOI] [PubMed] [Google Scholar]
  93. Liu WY, Weber B, Reuter M, Markett S, Chu WC, Montag C. The Big Five of Personality and structural imaging revisited: a VBM - DARTEL study. Neuroreport. 2013;24(7):375–380. doi: 10.1097/WNR.0b013e328360dad7. [DOI] [PubMed] [Google Scholar]
  94. Logothetis NK, Wandell BA. Interpreting the BOLD signal. Annual Review of Physiology. 2004;66:735–769. doi: 10.1146/annurev.physiol.66.082602.092845. [DOI] [PubMed] [Google Scholar]
  95. Lu F, Huo Y, Li M, Chen H, Liu F, Wang Y, … Chen H. Relationship between personality and gray matter volume in healthy young adults: a voxel-based morphometric study. PloS One. 2014;9(2):e88763. doi: 10.1371/journal.pone.0088763. [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Mar RA, Spreng RN, Deyoung CG. How to produce personality neuroscience research with high statistical power and low additional cost. Cognitive, Affective & Behavioral Neuroscience. 2013;13(3):674–685. doi: 10.3758/s13415-013-0202-6. [DOI] [PubMed] [Google Scholar]
  97. McCrae RR, Costa PT. Clinical assessment can benefit from recent advances in personality psychology. The American Psychologist. 1986;41(9):1001–1003. doi: 10.1037/0003-066X.41.9.1001. [DOI] [Google Scholar]
  98. McCrae RR, Costa PT. A contemplated revision of the NEO Five-Factor Inventory. Personality and Individual Differences. 2004;36(3):587–596. doi: 10.1016/S0191-8869(03)00118-1. [DOI] [Google Scholar]
  99. McCrae RR, Costa PT., Jr Validation of the five-factor model of personality across instruments and observers. Journal of Personality and Social Psychology. 1987;52(1):81–90. doi: 10.1037//0022-3514.52.1.81. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/3820081. [DOI] [PubMed] [Google Scholar]
  100. McCrae RR, John OP. An Introduction to the Five-Factor Model and Its Applications. Journal of Personality. 1992;60(2):175–215. doi: 10.1111/j.1467-6494.1992.tb00970.x. [DOI] [PubMed] [Google Scholar]
  101. McCrae RR, Yamagata S, Jang KL, Riemann R, Ando J, Ono Y, … Spinath FM. Substance and artifact in the higher-order factors of the Big Five. Journal of Personality and Social Psychology. 2008;95(2):442–455. doi: 10.1037/0022-3514.95.2.442. [DOI] [PMC free article] [PubMed] [Google Scholar]
  102. Mehl MR, Gosling SD, Pennebaker JW. Personality in its natural habitat: manifestations and implicit folk theories of personality in daily life. Journal of Personality and Social Psychology. 2006;90(5):862–877. doi: 10.1037/0022-3514.90.5.862. [DOI] [PubMed] [Google Scholar]
  103. Miller KL, Alfaro-Almagro F, Bangerter NK, Thomas DL, Yacoub E, Xu J, … Smith SM. Multimodal population brain imaging in the UK Biobank prospective epidemiological study. Nature Neuroscience. 2016;19(11):1523–1536. doi: 10.1038/nn.4393. [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Murphy K, Birn RM, Bandettini PA. Resting-state fMRI confounds and cleanup. NeuroImage. 2013;80:349–359. doi: 10.1016/j.neuroimage.2013.04.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Murphy K, Fox MD. Towards a consensus regarding global signal regression for resting state functional connectivity MRI. NeuroImage. 2017;154:169–173. doi: 10.1016/j.neuroimage.2016.11.052. [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Muschelli J, Nebel MB, Caffo BS, Barber AD, Pekar JJ, Mostofsky SH. Reduction of motion-related artifacts in resting state fMRI using aCompCor. NeuroImage. 2014;96:22–35. doi: 10.1016/j.neuroimage.2014.03.028. [DOI] [PMC free article] [PubMed] [Google Scholar]
  107. Neuroskeptic. The Nine Circles of Scientific Hell. Perspectives on Psychological Science: A Journal of the Association for Psychological Science. 2012;7(6):643–644. doi: 10.1177/1745691612459519. [DOI] [PubMed] [Google Scholar]
  108. Nichols TE, Das S, Eickhoff SB, Evans AC, Glatard T, Hanke M, … Thomas Yeo BT. Best Practices in Data Analysis and Sharing in Neuroimaging using MRI. bioRxiv. 2016 May 20; doi: 10.1101/054262. [DOI] [PMC free article] [PubMed]
  109. Noble S, Spann MN, Tokoglu F, Shen X, Constable RT, Scheinost D. Influences on the Test-Retest Reliability of Functional Connectivity MRI and its Relationship with Behavioral Utility. Cerebral Cortex. 2017;27(11):5415–5429. doi: 10.1093/cercor/bhx230. [DOI] [PMC free article] [PubMed] [Google Scholar]
  110. Noirhomme Q, Lesenfants D, Gomez F, Soddu A, Schrouff J, Garraux G, … Laureys S. Biased binomial assessment of cross-validated estimation of classification accuracies illustrated in diagnosis predictions. NeuroImage. Clinical. 2014;4:687–694. doi: 10.1016/j.nicl.2014.04.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Nooner KB, Colcombe SJ, Tobe RH, Mennes M, Benedict MM, Moreno AL, … Milham MP. The NKI-Rockland Sample: A Model for Accelerating the Pace of Discovery Science in Psychiatry. Frontiers in Neuroscience. 2012;6:152. doi: 10.3389/fnins.2012.00152. [DOI] [PMC free article] [PubMed] [Google Scholar]
  112. Norman WT. Toward an adequate taxonomy of personality attributes: replicated factors structure in peer nomination personality ratings. Journal of Abnormal and Social Psychology. 1963;66:574–583. doi: 10.1037/h0040291. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/13938947. [DOI] [PubMed] [Google Scholar]
  113. Oler JA, Fox AS, Shelton SE, Rogers J, Dyer TD, Davidson RJ, … Kalin NH. Amygdalar and hippocampal substrates of anxious temperament differ in their heritability. Nature. 2010;466(7308):864–868. doi: 10.1038/nature09282. [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Omura K, Todd Constable R, Canli T. Amygdala gray matter concentration is associated with extraversion and neuroticism. Neuroreport. 2005;16(17):1905–1908. doi: 10.1097/01.wnr.0000186596.64458.76. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/16272876. [DOI] [PubMed] [Google Scholar]
  115. Orrù G, Pettersson-Yeo W, Marquand AF, Sartori G, Mechelli A. Using Support Vector Machine to identify imaging biomarkers of neurological and psychiatric disease: a critical review. Neuroscience and Biobehavioral Reviews. 2012;36(4):1140–1152. doi: 10.1016/j.neubiorev.2012.01.004. [DOI] [PubMed] [Google Scholar]
  116. Pang Y, Cui Q, Wang Y, Chen Y, Wang X, Han S, … Chen H. Extraversion and neuroticism related to the resting-state effective connectivity of amygdala. Scientific Reports. 2016;6:35484. doi: 10.1038/srep35484. [DOI] [PMC free article] [PubMed] [Google Scholar]
  117. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, … Duchesnay É. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research: JMLR. 2011;12(Oct):2825–2830. Retrieved from http://www.jmlr.org/papers/volume12/pedregosa11a/pedregosa11a.pdf. [Google Scholar]
  118. Piñeiro G, Perelman S, Guerschman JP, Paruelo JM. How to evaluate models: Observed vs. predicted or predicted vs. observed? Ecological Modelling. 2008;216(3):316–322. doi: 10.1016/j.ecolmodel.2008.05.006. [DOI] [Google Scholar]
  119. Poldrack RA, Fletcher PC, Henson RN, Worsley KJ, Brett M, Nichols TE. Guidelines for reporting an fMRI study. NeuroImage. 2008;40(2):409–414. doi: 10.1016/j.neuroimage.2007.11.048. [DOI] [PMC free article] [PubMed] [Google Scholar]
  120. Poldrack RA, Gorgolewski KJ. OpenfMRI: Open sharing of task fMRI data. NeuroImage. 2017;144(Pt B):259–261. doi: 10.1016/j.neuroimage.2015.05.073. [DOI] [PMC free article] [PubMed] [Google Scholar]
  121. Poldrack RA, Laumann TO, Koyejo O, Gregory B, Hover A, Chen MY, … Mumford JA. Long-term neural and physiological phenotyping of a single human. Nature Communications. 2015;6:8885. doi: 10.1038/ncomms9885. [DOI] [PMC free article] [PubMed] [Google Scholar]
  122. Pool EM, Rehme AK, Eickhoff SB, Fink GR, Grefkes C. Functional resting-state connectivity of the human motor network: differences between right- and left-handers. NeuroImage. 2015;109:298–306. doi: 10.1016/j.neuroimage.2015.01.034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  123. Power JD, Barnes KA, Snyder AZ, Schlaggar BL, Petersen SE. Spurious but systematic correlations in functional connectivity MRI networks arise from subject motion. NeuroImage. 2012;59(3):2142–2154. doi: 10.1016/j.neuroimage.2011.10.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  124. Power JD, Mitra A, Laumann TO, Snyder AZ, Schlaggar BL, Petersen SE. Methods to detect, characterize, and remove motion artifact in resting state fMRI. NeuroImage. 2014;84:320–341. doi: 10.1016/j.neuroimage.2013.08.048. [DOI] [PMC free article] [PubMed] [Google Scholar]
  125. Power RA, Pluess M. Heritability estimates of the Big Five personality traits based on common genetic variants. Translational Psychiatry. 2015;5:e604. doi: 10.1038/tp.2015.96. [DOI] [PMC free article] [PubMed] [Google Scholar]
  126. Rammstedt B, Danner D, Martin S. The association between personality and cognitive ability: Going beyond simple effects. Journal of Research in Personality. 2016;62:39–44. doi: 10.1016/j.jrp.2016.03.005. [DOI] [Google Scholar]
  127. Rauch SL, Milad MR, Orr SP, Quinn BT, Fischl B, Pitman RK. Orbitofrontal thickness, retention of fear extinction, and extraversion. Neuroreport. 2005;16(17):1909–1912. doi: 10.1097/01.wnr.0000186599.66243.50. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/16272877. [DOI] [PubMed] [Google Scholar]
  128. Riccelli R, Toschi N, Nigro S, Terracciano A, Passamonti L. Surface-based morphometry reveals the neuroanatomical basis of the five-factor model of personality. Social Cognitive and Affective Neuroscience. 2017;12(4):671–684. doi: 10.1093/scan/nsw175. [DOI] [PMC free article] [PubMed] [Google Scholar]
  129. Roberts BW, DelVecchio WF. The rank-order consistency of personality traits from childhood to old age: a quantitative review of longitudinal studies. Psychological Bulletin. 2000;126(1):3–25. doi: 10.1037/0033-2909.126.1.3. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/10668348. [DOI] [PubMed] [Google Scholar]
  130. Roberts BW, Kuncel NR, Shiner R, Caspi A, Goldberg LR. The Power of Personality: The Comparative Validity of Personality Traits, Socioeconomic Status, and Cognitive Ability for Predicting Important Life Outcomes. Perspectives on Psychological Science: A Journal of the Association for Psychological Science. 2007;2(4):313–345. doi: 10.1111/j.1745-6916.2007.00047.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  131. Robinson EC, Jbabdi S, Glasser MF, Andersson J, Burgess GC, Harms MP, … Jenkinson M. MSM: a new flexible framework for Multimodal Surface Matching. NeuroImage. 2014;100:414–426. doi: 10.1016/j.neuroimage.2014.05.069. [DOI] [PMC free article] [PubMed] [Google Scholar]
  132. Ruigrok ANV, Salimi-Khorshidi G, Lai MC, Baron-Cohen S, Lombardo MV, Tait RJ, Suckling J. A meta-analysis of sex differences in human brain structure. Neuroscience and Biobehavioral Reviews. 2014;39:34–50. doi: 10.1016/j.neubiorev.2013.12.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  133. Ryan JP, Sheu LK, Gianaros PJ. Resting state functional connectivity within the cingulate cortex jointly predicts agreeableness and stressor-evoked cardiovascular reactivity. NeuroImage. 2011;55(1):363–370. doi: 10.1016/j.neuroimage.2010.11.064. [DOI] [PMC free article] [PubMed] [Google Scholar]
  134. Salimi-Khorshidi G, Douaud G, Beckmann CF, Glasser MF, Griffanti L, Smith SM. Automatic denoising of functional MRI data: combining independent component analysis and hierarchical fusion of classifiers. NeuroImage. 2014;90:449–468. doi: 10.1016/j.neuroimage.2013.11.046. [DOI] [PMC free article] [PubMed] [Google Scholar]
  135. Satterthwaite TD, Elliott MA, Gerraty RT, Ruparel K, Loughead J, Calkins ME, … Wolf DH. An improved framework for confound regression and filtering for control of motion artifact in the preprocessing of resting-state functional connectivity data. NeuroImage. 2013;64:240–256. doi: 10.1016/j.neuroimage.2012.08.052. [DOI] [PMC free article] [PubMed] [Google Scholar]
  136. Satterthwaite TD, Wolf DH, Ruparel K, Erus G, Elliott MA, Eickhoff SB, … Gur RC. Heterogeneous impact of motion on fundamental patterns of developmental changes in functional connectivity during youth. NeuroImage. 2013;83:45–57. doi: 10.1016/j.neuroimage.2013.06.045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  137. Saucier G. Orthogonal Markers for Orthogonal Factors: The Case of the Big Five. Journal of Research in Personality. 2002;36(1):1–31. doi: 10.1006/jrpe.2001.2335. [DOI] [Google Scholar]
  138. Schmitt DP, Realo A, Voracek M, Allik J. Why can’t a man be more like a woman? Sex differences in Big Five personality traits across 55 cultures. Journal of Personality and Social Psychology. 2008;94(1):168–182. doi: 10.1037/0022-3514.94.1.168. [DOI] [PubMed] [Google Scholar]
  139. Schnabel K, Asendorpf JB, Greenwald AG. Understanding and using the implicit association test: V. measuring semantic aspects of trait self-concepts. European Journal of Personality. 2008;22(8):695–706. doi: 10.1002/per.697. [DOI] [Google Scholar]
  140. Schönbrodt FD, Perugini M. At what sample size do correlations stabilize? Journal of Research in Personality. 2013;47(5):609–612. doi: 10.1016/j.jrp.2013.05.009. [DOI] [Google Scholar]
  141. Shen X, Finn ES, Scheinost D, Rosenberg MD, Chun MM, Papademetris X, Constable RT. Using connectome-based predictive modeling to predict individual behavior from brain connectivity. Nature Protocols. 2017;12(3):506–518. doi: 10.1038/nprot.2016.178. [DOI] [PMC free article] [PubMed] [Google Scholar]
  142. Shen X, Tokoglu F, Papademetris X, Constable RT. Groupwise whole-brain parcellation from resting-state fMRI data for network node identification. NeuroImage. 2013;82:403–415. doi: 10.1016/j.neuroimage.2013.05.081. [DOI] [PMC free article] [PubMed] [Google Scholar]
  143. Siegel JS, Mitra A, Laumann TO, Seitzman BA, Raichle M, Corbetta M, Snyder AZ. Data Quality Influences Observed Links Between Functional Connectivity and Behavior. Cerebral Cortex. 2017;27(9):4492–4502. doi: 10.1093/cercor/bhw253. [DOI] [PMC free article] [PubMed] [Google Scholar]
  144. Simonsohn U, Nelson LD, Simmons JP. P-curve: a key to the file-drawer. Journal of Experimental Psychology. General. 2014;143(2):534–547. doi: 10.1037/a0033242. [DOI] [PubMed] [Google Scholar]
  145. Smith GM. Usefulness of Peer Ratings of Personality in Educational Research. Educational and Psychological Measurement. 1967;27(4):967–984. doi: 10.1177/001316446702700445. [DOI] [Google Scholar]
  146. Smith SM, Vidaurre D, Beckmann CF, Glasser MF, Jenkinson M, Miller KL, … Van Essen DC. Functional connectomics from resting-state fMRI. Trends in Cognitive Sciences. 2013;17(12):666–682. doi: 10.1016/j.tics.2013.09.016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  147. Smith S, Vidaurre D, Glasser M, Winkler A, McCarthy P, Robinson E, … Van Essen D. Second beta-release of the HCP Functional Connectivity MegaTrawl. 2016 Apr; Retrieved from https://db.humanconnectome.org/megatrawl/HCP820_MegaTrawl_April2016.pdf.
  148. Soto CJ, John OP, Gosling SD, Potter J. Age differences in personality traits from 10 to 65: Big Five domains and facets in a large cross-sectional sample. Journal of Personality and Social Psychology. 2011;100(2):330–348. doi: 10.1037/a0021717. [DOI] [PubMed] [Google Scholar]
  149. Sporns O. The human connectome: origins and challenges. NeuroImage. 2013;80:53–61. doi: 10.1016/j.neuroimage.2013.03.023. [DOI] [PubMed] [Google Scholar]
  150. Takeuchi H, Taki Y, Hashizume H, Sassa Y, Nagase T, Nouchi R, Kawashima R. The association between resting functional connectivity and creativity. Cerebral Cortex. 2012;22(12):2921–2929. doi: 10.1093/cercor/bhr371. [DOI] [PubMed] [Google Scholar]
  151. Taki Y, Thyreau B, Kinomura S, Sato K, Goto R, Wu K, … Fukuda H. A longitudinal study of the relationship between personality traits and the annual rate of volume changes in regional gray matter in healthy adults. Human Brain Mapping. 2013;34(12):3347–3353. doi: 10.1002/hbm.22145. [DOI] [PMC free article] [PubMed] [Google Scholar]
  152. Todorov A. Face Value: The Irresistible Influence of First Impressions. Princeton University Press; 2017. Retrieved from https://market.android.com/details?id=book-mH8rDgAAQBAJ. [Google Scholar]
  153. Topping GD, O’Gorman JG. Effects of faking set on validity of the NEO-FFI. Personality and Individual Differences. 1997;23(1):117–124. doi: 10.1016/S0191-8869(97)00006-8. [DOI] [Google Scholar]
  154. Trabzuni D, Ramasamy A, Imran S, Walker R, Smith C, Weale ME … North American Brain Expression Consortium. Widespread sex differences in gene expression and splicing in the adult human brain. Nature Communications. 2013;4:2771. doi: 10.1038/ncomms3771. [DOI] [PMC free article] [PubMed] [Google Scholar]
  155. Tupes EC, Christal RE. Recurrent personality factors based on trait ratings. Personnel Research Lab; Lackland AFB, TX: 1961. (No. Technical Report ASD-TR-61-97) Retrieved from http://www.dtic.mil/dtic/tr/fulltext/u2/267778.pdf. [DOI] [PubMed] [Google Scholar]
  156. Tyszka JM, Kennedy DP, Paul LK, Adolphs R. Largely typical patterns of resting-state functional connectivity in high-functioning adults with autism. Cerebral Cortex. 2014;24(7):1894–1905. doi: 10.1093/cercor/bht040. [DOI] [PMC free article] [PubMed] [Google Scholar]
  157. Uher J. Developing “Personality” Taxonomies: Metatheoretical and Methodological Rationales Underlying Selection Approaches, Methods of Data Generation and Reduction Principles. Integrative Psychological & Behavioral Science. 2015;49(4):531–589. doi: 10.1007/s12124-014-9280-4. [DOI] [PubMed] [Google Scholar]
  158. Van Essen DC, Smith SM, Barch DM, Behrens TEJ, Yacoub E, Ugurbil K WU-Minn HCP Consortium. The WU-Minn Human Connectome Project: an overview. NeuroImage. 2013;80:62–79. doi: 10.1016/j.neuroimage.2013.05.041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  159. Van Horn JD, Gazzaniga MS. Why share data? Lessons learned from the fMRIDC. NeuroImage. 2013;82:677–682. doi: 10.1016/j.neuroimage.2012.11.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  160. Verweij KJH, Zietsch BP, Medland SE, Gordon SD, Benyamin B, Nyholt DR, … Wray NR. A genome-wide association study of Cloninger’s temperament scales: implications for the evolutionary genetics of personality. Biological Psychology. 2010;85(2):306–317. doi: 10.1016/j.biopsycho.2010.07.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  161. Vidaurre D, Smith SM, Woolrich MW. Brain network dynamics are hierarchically organized in time. Proceedings of the National Academy of Sciences of the United States of America. 2017;114(48):12827–12832. doi: 10.1073/pnas.1705120114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  162. Vinkhuyzen AAE, Pedersen NL, Yang J, Lee SH, Magnusson PKE, Iacono WG, … Wray NR. Common SNPs explain some of the variation in the personality dimensions of neuroticism and extraversion. Translational Psychiatry. 2012;2:e102. doi: 10.1038/tp.2012.27. [DOI] [PMC free article] [PubMed] [Google Scholar]
  163. Viswesvaran C, Ones DS. Measurement Error in “Big Five Factors” Personality Assessment: Reliability Generalization across Studies and Measures. Educational and Psychological Measurement. 2000;60(2):224–235. doi: 10.1177/00131640021970475. [DOI] [Google Scholar]
  164. van der Walt S, Colbert SC, Varoquaux G. The NumPy Array: A Structure for Efficient Numerical Computation. Computing in Science & Engineering. 2011;13(2):22–30. doi: 10.1109/MCSE.2011.37. [DOI] [Google Scholar]
  165. Weisberg YJ, Deyoung CG, Hirsh JB. Gender Differences in Personality across the Ten Aspects of the Big Five. Frontiers in Psychology. 2011;2:178. doi: 10.3389/fpsyg.2011.00178. [DOI] [PMC free article] [PubMed] [Google Scholar]
  166. Westfall J, Yarkoni T. Statistically Controlling for Confounding Constructs Is Harder than You Think. PloS One. 2016;11(3):e0152719. doi: 10.1371/journal.pone.0152719. [DOI] [PMC free article] [PubMed] [Google Scholar]
  167. Westlye LT, Bjørnebekk A, Grydeland H, Fjell AM, Walhovd KB. Linking an anxiety-related personality trait to brain white matter microstructure: diffusion tensor imaging and harm avoidance. Archives of General Psychiatry. 2011;68(4):369–377. doi: 10.1001/archgenpsychiatry.2011.24. [DOI] [PubMed] [Google Scholar]
  168. Woo CW, Chang LJ, Lindquist MA, Wager TD. Building better biomarkers: brain models in translational neuroimaging. Nature Neuroscience. 2017;20(3):365–377. doi: 10.1038/nn.4478. [DOI] [PMC free article] [PubMed] [Google Scholar]
  169. Wright CI, Williams D, Feczko E, Barrett LF, Dickerson BC, Schwartz CE, Wedig MM. Neuroanatomical correlates of extraversion and neuroticism. Cerebral Cortex. 2006;16(12):1809–1819. doi: 10.1093/cercor/bhj118. [DOI] [PubMed] [Google Scholar]
  170. Wu Y, Li L, Yuan B, Tian X. Individual differences in resting-state functional connectivity predict procrastination. Personality and Individual Differences. 2016;95(Supplement C):62–67. doi: 10.1016/j.paid.2016.02.016. [DOI] [Google Scholar]
  171. Xu J, Potenza MN. White matter integrity and five-factor personality measures in healthy adults. NeuroImage. 2012;59(1):800–807. doi: 10.1016/j.neuroimage.2011.07.040. [DOI] [PMC free article] [PubMed] [Google Scholar]
  172. Yarkoni T. Big Correlations in Little Studies Inflated fMRI Correlations Reflect Low Statistical Power—Commentary on Vul et al. (2009) Perspectives on Psychological Science: A Journal of the Association for Psychological Science. 2009;4:294–298. doi: 10.1111/j.1745-6924.2009.01127.x. [DOI] [PubMed] [Google Scholar]
  173. Yarkoni T. Neurobiological substrates of personality: A critical overview. In: Mikulincer Mario Shaver, Cooper Phillip R, Lynne Larsen M, Randy J., editors. Personality processes and individual differences. Vol. 4. American Psychological Association; 2015. [Google Scholar]
  174. Yarkoni T, Westfall J. Choosing Prediction Over Explanation in Psychology: Lessons From Machine Learning. Perspectives on Psychological Science: A Journal of the Association for Psychological Science. 2017;12(6):1100–1122. doi: 10.1177/1745691617693393. [DOI] [PMC free article] [PubMed] [Google Scholar]
  175. Zou H, Hastie T. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society. Series B, Statistical Methodology. 2005;67(2):301–320. doi: 10.1111/j.1467-9868.2005.00503.x. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

RESOURCES