Skip to main content
Cerebral Cortex (New York, NY) logoLink to Cerebral Cortex (New York, NY)
. 2024 Feb 22;34(2):bhad506. doi: 10.1093/cercor/bhad506

Measures of resting-state brain network segregation and integration vary in relation to data quantity: implications for within and between subject comparisons of functional brain network organization

Liang Han 1, Micaela Y Chan 2, Phillip F Agres 3, Ezra Winter-Nelson 4, Ziwei Zhang 5, Gagan S Wig 6,7,
PMCID: PMC10883417  PMID: 38385891

Abstract

Measures of functional brain network segregation and integration vary with an individual’s age, cognitive ability, and health status. Based on these relationships, these measures are frequently examined to study and quantify large-scale patterns of network organization in both basic and applied research settings. However, there is limited information on the stability and reliability of the network measures as applied to functional time-series; these measurement properties are critical to understand if the measures are to be used for individualized characterization of brain networks. We examine measurement reliability using several human datasets (Midnight Scan Club and Human Connectome Project [both Young Adult and Aging]). These datasets include participants with multiple scanning sessions, and collectively include individuals spanning a broad age range of the adult lifespan. The measurement and reliability of measures of resting-state network segregation and integration vary in relation to data quantity for a given participant’s scan session; notably, both properties asymptote when estimated using adequate amounts of clean data. We demonstrate how this source of variability can systematically bias interpretation of differences and changes in brain network organization if appropriate safeguards are not included. These observations have important implications for cross-sectional, longitudinal, and interventional comparisons of functional brain network organization.

Keywords: resting-state functional correlations, brain networks, reliability, system segregation, modularity, integration

Introduction

Measures of network segregation are widely used to describe the topological organization of large-scale brain networks (Sporns and Betzel 2016; Wig 2017). Network segregation refers to the degree of separation between the sub-networks (also referred to as modules or communities) of a network. Network segregation can be quantified directly by calculating the relative strength of relationships within versus between sub-networks (system segregation; Chan et al. 2014) or inferred indirectly using the measure of modularity (Q) obtained via the modularity maximization algorithm (Newman 2004). In the context of brain networks, both measures are calculated from adjacency matrices created by evaluating pairwise structural or functional relationships between brain network nodes.

A large body of research has revealed that functional brain network segregation varies in relation to age, task performance, cognitive ability, and health status (for a review, see Wig 2017). In healthy individuals, segregation of resting-state functional correlation (RSFC) brain networks increases in youth and adolescence (Chen and Deem 2015; Gu et al. 2015) and declines with increasing adult age (Betzel et al. 2014; Chan et al. 2014; Geerligs et al. 2015; Han et al. 2018; Chong et al. 2019; Cassady et al. 2020; Kong et al. 2020; Madden et al. 2020; Manza et al. 2020; Stumme et al. 2020; Pedersen et al. 2021; Setton et al. 2022). Lower resting-state system segregation is associated with reduced activation selectivity across visual and frontal–parietal control areas during performance of tasks supported by these brain systems (Chan et al. 2017), worse cognitive and motor function (Chan et al. 2014; King et al. 2018; Varangis et al. 2019), and socioeconomic disadvantage (Chan et al. 2018; Chan et al. 2021). Measures of brain network segregation also have relevance toward characterizing patient populations. For example, baseline brain network modularity is predicative of rehabilitation outcomes in aphasic patients (Duncan and Small 2016), and brain network modularity changes (increases) during recovery from stroke (Siegel et al. 2018). Among Alzheimer’s disease (AD) patients, individuals diagnosed with dementia have lower modularity compared to healthy “controls” (Brier et al. 2014), increasing AD dementia severity Is associated with lower system segregation, independent of age and amyloid burden (Zhang et al. 2023), and system segregation moderates the effect of tau deposition on cognitive impairment (Ewers et al. 2021). Finally, longitudinal measurement of healthy adults indicates that declining system segregation is prognostic of dementia severity, independent of atrophy, AD genetic risk, and baseline biomarkers of AD-related brain pathology (Chan et al. 2021).

As highlighted above, the measurement of network segregation has become a popular target for researchers investigating properties of brain network organization in healthy and diseased individuals, in part because the approach offers a straightforward and objective method of effectively summarizing complex patterns of meso-scale brain network topology using a single measurement value. This value is often treated as a modifiable brain network “trait” for an individual, which allows for cross-sectional comparisons (e.g. across age, cognitive ability, or psycho-social characteristics), evaluation of longitudinal changes within individuals (e.g. development and aging), and determination of the impacts of an intervention on an individual’s brain network organization (e.g. following brain stimulation, pharmaceutics, cognitive training).

A central assumption throughout all these applications is that measures of network segregation (and other measures of network organization more generally) are a reliable index of an individual’s network topology. This requirement is in turn partly predicated on obtaining reliable RSFC relationships between distributed brain locations, which has been previously examined using specific seed-based correlation maps (Van Dijk et al. 2010; Shah et al. 2016; O'Connor et al. 2017) or specific ICA patterns (Zuo et al. 2010; Zuo and Xing 2014). These reports have demonstrated that reliability of RSFC can vary across different brain regions and sets of relationships between brain regions. In keeping with this, measures quantifying whole-brain RSFC patterns (Shehzad et al. 2009; Birn et al. 2013; Noble et al. 2017), multivariate analysis (Zuo et al. 2019; Noble et al. 2021), or summary measures of RSFC network topologies (e.g. small-worldness [Braun et al. 2012], modularity and clustering coefficient [Cao et al. 2014], and degree centrality [Chen et al. 2015]) have been shown to be generally more reliable than single pair-wise correlations or correlation maps. However, even among these studies there is a considerable range of measurement reliability, some of which relates to differences in data acquisition (e.g. eyes open vs eyes closed, awake vs. asleep), test-retest intervals, data pre-processing decisions, and topological characteristics of the examined RSFC relationships (for reviews, see Noble et al. 2019; Zuo and Xing 2014).

One source of variability in these preceding observations likely relates to the amount of data used to measure an individual’s RSFCs (Birn et al. 2013). Measures of RSFC network topography and topology can vary in relation to data quantity used to calculate the measures when resting-state data are combined across multiple sessions (Laumann et al. 2015; Gordon et al. 2017). These important observations indicate that the inter-session reliability of RSFC network measures may also vary as a consequence of data amount. Critically, the amount of data used to calculate an individual’s functional network organization can vary across studies, as well as across participants within a study, either due to variability in data acquisition or variability in the effectiveness of preprocessing methods used to minimize known sources of non-neuronal variance.

Why would variable amounts of data contribute toward obtaining valid and reliable estimates of measures of network organization for an individual? There exist multiple time-varying sources of information that contribute variance to resting-state signals. RSFC has been shown to vary over time, both when evaluating different scans collected in close proximity (Honey et al. 2009; Liu et al. 2009; Shehzad et al. 2009; Van Dijk et al. 2010), and also on much shorter timescales within a single scan (from seconds to minutes) (Chang and Glover 2010; Sakoglu et al. 2010; Kiviniemi et al. 2011; Handwerker et al. 2012; Jones et al. 2012). It has been proposed that the temporal fluctuation of resting-state signals relates to changes in ongoing cognitive processing (e.g. Esposito et al. 2006; Fransson 2006; Sun et al. 2007; Fornito et al. 2012; Gratton et al. 2018), outcomes of learning (Albert et al. 2009; Lewis et al. 2009; Stevens et al. 2010; Tambini et al. 2010; Bassett et al. 2011), and transitions between states related to sleep (Horovitz et al. 2008; Horovitz et al. 2009) or level of arousal (Tagliazucchi et al. 2013; Becker et al. 2016). However, a considerable portion of both the moment-to-moment and day-to-day RSFC variability can be attributed to data sampling error, head motion, and fluctuating drowsiness (Laumann et al. 2017). Importantly, efforts to alleviate the described sources of non-neuronal variance include truncation of time series (e.g. “frame scrubbing” [Power et al. 2020]), or inclusion of regressors that tag moments which can vary in their number across individuals (e.g. “spike regression” [Satterthwaite et al. 2013]). These processing steps are necessary for mitigation of sources of variance of non-interest (Ciric et al. 2017), but they typically result in both less data and variable amounts of “clean” data across scan sessions within and/or between participants.

Altogether then, it is clear that (i) measures of RSFC network organization can vary at shorter timescales, and (ii) that the efforts to “clean” functional data can result in a variable amount of data used to estimate measures of functional relationships across individuals, or within individuals when comparing different scan sessions. This poses independent but related challenges for using measures of RSFC network segregation as an individual’s brain network trait and for assessing the reliability of the measures. Do measures of brain network segregation vary relative to the amount of data used to derive them? If so, data cleaning techniques can introduce potential sources of variance related to data quantity. Further, it is critical to determine the extent to which within-individual variability is small enough such that the network estimates are deemed reliable across independent scans/sessions.

In the current report, we examine the extent to which measures of brain network segregation vary in relation to data quantity used to derive the adjacency matrices from a given scanning session for an individual, and whether the measures are reliable across independent scan sessions for that individual. Based on these observations, we highlight several practical issues for the analysis of cross-sectional, longitudinal, and interventional comparison, and offer means of addressing them. To achieve these aims, we use both cross-sectional and multi-session datasets comprising multiple independent cohorts of individuals that range in age from younger to older adulthood (21–100 y).

We primarily focus on measures of brain network segregation throughout this report (i.e. system segregation [Chan et al. 2014; Wig 2017] and modularity [Newman 2004; Rubinov and Sporns 2010]). However, given their close relationship to measures of network segregation, we also include related measures of brain network integration to evaluate their sensitivity to data amount (specifically, mean clustering coefficient and mean participation coefficient [Guimera et al. 2005; Watts and Strogatz 1998]). While our concentration is on patterns of resting-state correlations obtained using blood oxygen level dependent (BOLD) imaging, the conclusions are also relevant for quantifying other types of functional relationships (e.g. during task performance or passive viewing of dynamic stimuli such as movies) and other modalities of data acquisition that involve estimating time-series relationships [e.g. electroencephalography (EEG), magnetoencephalography (MEG)].

Materials and methods

Datasets

Three datasets were analyzed in this report: data from the Midnight Scan Club (MSC; Gordon et al. 2017), Human Connectome Project Young Adult (HCP-YA; Van Essen et al. 2013), and the Lifespan Human Connectome Project Aging (HCP-A; Bookheimer et al. 2019). The first dataset was used for the main analyses and the latter two were used for the purpose of demonstrating the implications of the findings.

Midnight Scan Club

The MSC dataset comprises data from 10 participants (5 females; age range: 24–34 y, M = 29.1 y, SD = 3.3). These data are publicly available (https://openneuro.org/datasets/ds000224/). One participant was excluded from this study due to excessive sleep and head motion (Gordon et al. 2017). The scanning protocol was approved by the Washington University Institutional Review Board and School of Medicine Human Studies Committee. All participants provided written informed consent.

Human Connectome Project Young Adult

The dataset includes data collected from 1,206 participants (age range: 21–35 y; HCP s1200 release dataset; Glasser et al. 2013; Marcus et al. 2013). This release has been made publicly available (http://www.humanconnectome.org). The scanning protocol was approved by the Washington University in St. Louis’s Human Research Protection Office and all participants provided written informed consent. Participants with at least 20 min of clean resting-state data per session (see RSFC Preprocessing) were included in the final sample [n = 730 (nfemale = 410), age range: 22–35 y, M = 28.6 y, SD = 3.6].

Lifespan Human Connectome Project Aging

The dataset includes data collected from 689 participants (age range: 35–100 y; HCP-A release 1.0 dataset [https://nda.nih.gov]). The scanning protocol was approved by the Washington University in St. Louis’s Human Research Protection Office and all participants provided written informed consent. Participants with at least 20 min of clean resting-state data total (see RSFC Preprocessing) were included in the final sample [n = 535 (nfemale = 314), age range: 36–100 y, M = 57.9 y, SD = 14.8].

Imaging data acquisition

Midnight Scan Club

Each participant’s brain images were collected on a Siemens TRIO 3 T MRI scanner from 12 sessions on separate days.

Anatomical images

Across two separate days, four T1-weighted images (TR = 2,400 ms, TE = 3.74 ms, TI = 1,000 ms, flip angle = 8°, 224 slices) and four T2-weighted images (TR = 3,200 ms, TE = 479 ms, resolution = 0.8 × 0.8 × 0.8 mm3, 224 slices) were obtained for each participant.

Functional images

Resting-state functional MRI images were acquired across 10 subsequent days (one session per day). In each session, 30 min of resting state fMRI data were collected, in which participants visually fixated on a white crosshair presented against a black background. All functional imaging was performed using a gradient-echo EPI sequence (TR = 2,200 ms, TE = 27 ms, flip angle = 90°, resolution = 4 × 4 × 4 mm3, 36 slices).

Human Connectome Project

Participants were scanned on a Siemens 3 T Prisma whole-body scanner (Siemens, Erlangen, Germany) with a Siemens 32-channel head coil at one of 4 different sites (Massachusetts General Hospital, University of California-Los Angeles, University of Minnesota, and Washington University in St. Louis). Each participant completed two scanning sessions on two separate days.

Anatomical images

Each participant has two T1-weighted magnetization-prepared rapid acquisition gradient echo (MP-RAGE) structural scans (TR = 2,500 ms, TE = 2.1 ms, TI = 1,000 ms, resolution = 0.7 × 0.7 ×  0.7 mm3, flip angle = 8°) and two T2-weighted structural scans (TR = 3,200 ms, TE = 563 ms, resolution = 0.7 × 0.7 × 0.7 mm3).

Functional images

Resting-state functional MRI images were acquired while the participants fixated on a white crosshair on a black background using a gradient-echo EPI sequence (multiband factor = 8, TR = [HCP-YA: 720 ms; HCP-A: 800 ms], TE = [HCP-YA: 33.1 ms; HCP-A: 37 ms], flip angle = 52°, 104 × 90 matrix size, 72 slices, 2 mm isotropic voxels, and 1200 time points [14.4 min] per scan for HCP-YA, and 688 time points [6.5 min] per scan for HCP-A). There were two resting-state scans in each scan session, with different phase-encoding directions (HCP-YA: RL and LR; HCP-A: AP and PA) in each scan.

Processing of anatomical MRI image and cortical surface

Midnight Scan Club

For each participant, their mean T1- and T2-weighted images were derived by averaging multiple structural images of the same modality, then underwent distortion correction using the mean field map (Laumann et al. 2015; Gordon et al. 2017). The following steps were taken to generate cortical surfaces: (i) generating anatomical surfaces from the participant’s average T1-weighted image in native volumetric space using FreeSurfer’s (version 5.3) default recon-all processing pipeline; (ii) registering the fsaverage-registered surfaces from FreeSurfer pipeline using deformation maps from a landmark-based registration of fsaverage surfaces to a hybrid atlas surfaces (fs_LR; Van Essen et al. 2012a) and resampling to a resolution of 163,842 vertices (164 k fs_LR) using Caret tools (Van Essen et al. 2001); and (iii) down-sampling each participant’s 164 k fs_LR surfaces to a 32,492 vertex surfaces (fs_LR 32 k). The deformations from the original surfaces to the fs_LR 32 k surfaces were composed into a single deformation map allowing for a one-step resampling.

Human Connectome Project

Anatomical MRI images were processed using HCP structural pipelines that consist of 3 parts (PreFreeSurfer, FreeSurfer and PostFreeSurfer). The PreFreeSurfer pipeline is used to produce an undistorted native structural volume space for each participant, align the T1- and T2-weighted images, perform a B1 (bias field) correction, and register the participant’s native structural volume space to MNI space. The FreeSurfer pipeline (HCP-YA: version 5.3; HCP-A: version 6.0) segments the volume into predefined structures, reconstructs white and pial cortical surfaces, and performs FreeSurfer’s standard folding-based surface registration to their surface atlas (fsaverage). The final structural pipeline (PostFreeSurfer) produces all of the NIFTI volume and GIFTI surface files, along with applying the surface registration to the Conte69 surface template (Van Essen et al. 2012a), down-sampling registered surfaces, creating the final brain mask, and creating myelin maps.

Basic fMRI preprocessing

Midnight Scan Club

The MSC dataset was preprocessed with slice-timing correction (odd vs. even slice due to interleaved acquisition), intensity normalization to a whole brain mode value of 1,000 (Miezin et al. 2000), and realignment within-run to correct for head movement. BOLD data were transformed to Talairach atlas space by registering the mean intensity image from a single BOLD session to the atlas (Talairach and Tournoux 1988) using the average high-resolution T1- and T2-weighted images. All other BOLD sessions were linearly registered to the single atlas registered BOLD session. The atlas transformation, mean field distortion correction, and resampling to 3-mm isotropic atlas space were combined and performed via a single interpolation.

Human Connectome Project

Both HCP datasets employed the HCP fMRI volume pipeline to correct gradient-nonlinearity-induced distortion, realign the time series to correct for head motion, perform EPI fMRI image distortion correction due to phase encoding directions, and combine all the transforms for each registration and distortion correction steps into a single nonlinear transformation that can be applied in a single resampling step. After transforming the fMRI volumes, the intensity across runs was normalized to a whole brain mode value of 1,000 (Miezin et al. 2000).

RSFC preprocessing

For all three datasets, additional preprocessing steps were taken to reduce spurious variance unlikely to reflect neuronal activity in RSFC data (Power et al. 2014): (i) demeaning and detrending; (ii) multiple regression of the BOLD data to remove variance related to the whole brain gray matter signal (global signal regression; defined by each participant’s own anatomy), ventricular signal, white matter signal, six detrended head realignment parameters obtained by rigid-body head motion correction, and the first-order derivative terms for all aforementioned nuisance variables. While the use of global signal regression in RSFC processing has attracted varying opinions, it is objectively effective for minimizing motion-related artifacts (Satterthwaite et al. 2013; Power et al. 2017), and the recommended method to reliably remove global cardiac- and respiration-related artifacts when direct estimates of these signals are unavailable (Power et al. 2018). Because older adults are more prone to head movement (Van Dijk et al. 2012; Savalia et al. 2017) that lead to altered RSFC profiles [Satterthwaite et al. 2013; Power et al. 2014], it is critical to minimize this source of bias that may contribute to erroneous estimation of RSFC. (iii) To reduce the effect of motion artifacts on RSFC, data were processed following a “scrubbing” procedure (Power et al. 2014). Motion-contaminated volumes were identified by frame-by-frame displacement (FD) that was calculated as the sum of absolute values of the differentials of the three translational motion parameters and three rotational motion parameters (Power et al. 2014). Volumes with excessive head motion (FD ≥ 0.2 mm for MSC) were flagged, based on previous demonstrations of the effectiveness of this FD cutoff for this dataset (Gordon et al. 2017). Recent studies have demonstrated that high-frequency respiratory artifact can confound estimates of FD, particularly for scans using multiband sequences with short TRs (e.g. Fair et al. 2020). As such, the motion parameters for HCP datasets were filtered to remove high-frequency components prior to the estimate of FD and more stringent cutoffs for FD were used for HCP-YA (0.04 mm; Gordon et al. 2017) and HCP-A datasets (0.08 mm; as has been used for analyzing datasets with wider age-range and shorter total scan duration; Gratton et al. 2020). In addition, for all datasets, data between two motion-contaminated frames that were fewer than five frames apart were also flagged. These flagged motion-contaminated frames were removed and interpolated for the subsequent processing. (iv) Band-pass filtering (0.009 < f < 0.08 Hz). (v) Removing the interpolated frames that were used to preserve the time series during regression and bandpass filtering.

CIFTI generation

Connectivity Informatics Technology Initiative (CIFTI) format files were generated to integrate information from all possible brain ordinates, including the cerebral cortex, subcortical structures, and the cerebellum. To do so, the time series data of the cortical surface were derived by resampling functional volumes to 32 k mesh surface using single deformation maps derived from surface data processing and smoothed with a Gaussian kernel (MSC: 6 mm full width half-maximum [FWHM]; HCP: 2 mm FWHM). The subcortical data were smoothed in volumetric space with a Gaussian kernel (MSC: 6 mm FWHM; HCP: 2 mm FWHM). Finally, the time-series data of the cortical surface and volumetric time series of FreeSurfer-labeled subcortical structures and the cerebellum were combined to create CIFTI files (Glasser et al. 2013).

Data sampling

To estimate the impact of including different amounts of data on estimating brain networks, expanding windows were used to sample frames from the entire functional time series in each session of a participant. Specifically, we started by sampling from the 1st frame to the nth frame needed to obtain 1 minute of data to form the first window (Window 1 = 28 frames of data for MSC dataset [TR = 2.2 s]), then used this data window to compute the participant's correlation matrix and corresponding network measures. This size of this data window was incremented by one and the process was repeated until all frames were sampled (i.e. 1st–nth frames, 1st–nth + 1 frames, …, 1st–t frames; t = total number of frames in each session; Fig. 1).

Fig. 1.

Fig. 1

To examine impacts of data quantity, individuals’ resting-state data were sampled using expanding windows for each of their scan sessions. (A) Within a given session for a participant, we began by sampling only the 1st–nth frames needed to obtain 1 minute of data (window 1; e.g. for the MSC dataset this is the 1st–28th frame given a 2.2 s TR), followed by the 1st to n + 1 frames (window 2) and so on, successively adding a single additional frame until all frames were included. The participant’s correlation matrices and resting-state network measures were calculated separately for each window, thereby estimating the participant’s brain network organization across an increasing range of data amounts. (B) The process in (A) was repeated for each of the participant’s independent scan sessions.

Brain network construction and estimation

Brain network nodes

For the analysis using MSC data, individual parcellations were used to define nodes for the construction of RSFC correlation matrix (Gordon et al. 2017), given that each participant had a substantial amount of data available, making individual parcellations more reliable (Laumann et al. 2015). In the validation analyses using HCP datasets, we used a previously published atlas node set derived from an independent group of participants (Gordon et al. 2016). The reasoning for using a parcellation atlas in these cases was that the parcellation from a large pool of participants with high-quality data can capture important features of brain network organization in alternate datasets (Power et al. 2011; Yeo et al. 2011; Wig et al. 2014), which would also ensure generalization of the findings.

Constructing brain networks

For each data-sampling window, the functional time series were averaged across all the vertices within each node, resulting in an n × t matrix (where n is the number of nodes, and t is the total number of frames in this window). The time series of each node was correlated with the time series of every other node using Pearson’s correlations, and Fisher’s r-to-z transformed to generate a node-to-node correlation matrix (an n × n network graph matrix). Following previous reports (Chan et al. 2014; Han et al. 2018; Chan et al. 2021), negative correlations were excluded in the brain network graph matrix (negative values were set to 0). Network measures were calculated from these correlation matrices.

Community assignments

In the primary analysis using the MSC dataset, network communities of the node-wise graphs were identified using the Infomap algorithm (Rosvall and Bergstrom 2008). Specifically, each participant’s network graph matrix in a data-sampling window was (i) thresholded at each edge density ranging from 0.3 to 5% in steps of 0.1% (i.e. correlations above each threshold were retained, and correlations below the threshold were set to 0); (ii) filtered to exclude any correlations between nodes whose centroids were less than 30 mm apart (these edges were set to 0); and (iii) entered into the Infomap algorithm for the estimation of community assignments at each edge density. For the analysis in HCP datasets, we used a previously published atlas node set and community assignments (Gordon et al. 2016).

Calculation of network measures

Two measures of brain network segregation were evaluated: (i) system segregation and (ii) modularity. In addition, two widely studied measures of brain network integration were also evaluated. Network integration reflects the degree to which information from distributed brain network nodes is communicated and combined (Rubinov and Sporns 2010); measures of network integration have typically relied on measurement of the (i) mean clustering coefficient (CC; reflecting the degree to which nodes in a network tend to cluster together) and (ii) participation coefficient (PC; reflecting intermodular connectivity between brain network nodes.) across nodes of the network.

To quantify the changes/differences of measures attributable to different amounts of data being used, network measures were computed from matrices defined for each data-sampling window.

System segregation summarizes values of within-system correlations in relation to between-system correlations, reflecting the degree to which the functional brain network is organized into distinct subnetworks (Chan et al. 2014; Chan et al. 2021). Following previous reports (Chan et al. 2014; Chan et al. 2021), negative correlations were excluded in the brain network graph matrix (negative values were set to 0). Without thresholding the correlation coefficients, system segregation takes the difference in mean within-system and mean between-system correlation as a proportion of mean within-system correlation, as noted in the following formula (Chan et al. 2021):

graphic file with name DmEquation1.gif

where Inline graphic represents the connectivity among nodes (Fisher z-transformed correlation coefficients) within the same system, Inline graphic denotes the connectivity between nodes of one system and nodes of other systems, Inline graphic is the total number of within-system edges across all functional systems, and Inline graphic is the total number of between-system edges across all functional systems. In each data-sampling window, system segregation was computed based on the community assignments at each edge density (top 0.3–5%) and then averaged across all edge thresholds for MSC dataset. In the HCP datasets, system segregation was calculated using a previously published set of community labels (Gordon et al. 2016).

Modularity quantifies the degree to which the entire brain network could be divided into separate functional systems (Newman 2004):

graphic file with name DmEquation2.gif

where Inline graphic is the connection between Inline graphic and Inline graphic, Inline graphic is the number of connections, Inline graphic is the degree of node Inline graphic, Inline graphic is the module containing node Inline graphic, and Inline graphic = 1 if Inline graphic = Inline graphic and 0 otherwise. In each expanding window, modularity was computed using thresholded RSFC graph matrix at each edge density (top 0.3–5%) and then averaged across all edge thresholds.

CC averages the fraction of triangles around each node in the network, which reflects the prevalence of clustered connectivity around individual nodes (Watts and Strogatz 1998):

graphic file with name DmEquation3.gif

where Inline graphic is the degree of node Inline graphic, and Inline graphic is the number of triangles around node Inline graphic. In each expanding window, CC was computed using thresholded RSFC graph matrix at each edge density (top 0.3–5%), and then averaged across all edge thresholds. For each participant, CC values were averaged across all nodes, to compute a mean CC for that participant’s brain network.

PC quantifies the diversity of intermodular interconnections of individual nodes (Guimera et al. 2005). A higher PC indicates that the node has connections with multiple communities in a network. PC is formally expressed by

graphic file with name DmEquation4.gif

where Inline graphic is a set of nonoverlapping modules in the network, Inline graphic is the number of links between Inline graphic and all nodes in module Inline graphic, and Inline graphic is the degree of node Inline graphic. In each expanding window, PC was computed based on the community assignments at each edge density (top 0.3–5%), averaged across all edge thresholds for MSC data. For HCP data, PC was calculated using a previously published set of community labels (Gordon et al. 2016). For each participant, PC values were averaged across all nodes, to compute a mean PC for that participant’s brain network.

Evaluation of reliability

The reliability of network measures was evaluated by quantifying within- and between-session variability using the MSC data. To do so, we first calculated network measures as a function of data quantity (i.e. for each expanding window) for each session, resulting in an s × w matrix for each participant (where s is the number of sessions, and w is the number of expanding windows).

Within-session variability

The within-session variability of a network measure was assessed by calculating the first derivative of each session’s curve, to quantify the change of network measures in relation to the amount of data used to calculate the measure. Derivative values were averaged across all available sessions to produce a 1 × w vector for each participant, which summarized the within-session variability of the network measures in relation to data quantity, for that participant.

Between-session variability

The between-session variability of a network measure was assessed by calculating the standard deviation (SD) of the measure across all sessions, for a given participant. This calculation was repeated for varying amounts of data (i.e. each expanding window size, w), resulting in a 1 × w vector of SD values for each participant, which represents the cross-session variability of a network measure as a function of the amount of data used to calculate the measure.

Intra-class correlation of network measures

Existing studies have assessed the reliability of RSFC networks using the intra-class correlation (ICC; Shrout and Fleiss 1979). ICC takes both within- and between-subject variability into account. ICC quantifies the degree to which the data in their own class are similar to one another. ICC values range from 0 to 1, indicating different degrees of reliability. A high ICC value denotes that each participant’s network measure estimates at different time points are more similar to one another than the estimates in other participants. ICC is computed with the following formula (Shrout and Fleiss 1979):

graphic file with name DmEquation5.gif

where Inline graphic is the between-subject mean squared error, Inline graphic is the within-subject mean squared error, and Inline graphic is number of time points. In this report, we computed one ICC value in each expanding window for a network measure, resulting in a row vector of ICC values quantifying the changes of ICC as a function of amount of data being used.

Application to within and between subject comparisons

Within-participant comparisons

We evaluated how including unequal amounts of data from scanning sessions collected close in time can impact the estimate of within-participant “changes” in brain system segregation. This analysis utilized the HCP-YA dataset, which consists of two independent scanning sessions per individual. The 730 HCP-YA participants were first randomly assigned into 6 non-overlapping groups of 120 participants. For every individual within a given group, brain system segregation was calculated for one scanning session using 5 min of data and compared to brain system segregation calculated from that individual’s alternate scanning session using either 5, 6, 7.5, 10, 15, or 20 min of data. The order of scanning sessions (i.e. which scan session was treated as the alternate scanning session) was randomized for each individual and each comparison. The random sampling of participants into groups was repeated 10,000 times. Comparisons of system segregation across scan sessions were evaluated using paired-samples t-tests.

Between-participant comparisons

We evaluated how including unequal amounts of data across participants will impact between-participant system segregation comparisons. Using data from the HCP-A dataset, two sets of analyses were conducted to evaluate the correlation of age and brain system segregation (12 subjects were excluded from this analysis due to their outlier system segregation [< 3 SD from the mean] values calculated using 20 min of data, resulting in a final sample of 523 subjects with an age range of 36–100 y). In the first analysis, the amount of data included in the brain system segregation calculation was equivalent across participants (range = 2.4 s–20 min). This process was repeated 10,000 times to build a distribution of correlation values. In the second analysis, the amount of data per participant was set to randomly vary across participants (range = 2.4 s–20 min), before calculating the relationship between age and brain system segregation. Given that older adults exhibit more data loss due to greater head motion (Savalia et al. 2017), this analysis was conducted such that frames were removed in a pseudo-random manner that was based on participant’s age. To do so, we first calculated distributions of number of movement-censored frames defined by the participant’s age. Using the HCP-A dataset, three age cohorts were created (earlier middle-age: 35–49 y, older middle-age: 50–64 y, older-age: 65–100 y). For each age cohort, the mean and SD of available frames after motion processing were estimated across participants. As expected, frame loss was significantly correlated with age (r = 0.195, P < 0.001), indicating older adults had a greater degree of frame loss. These parameters (mean and SD) were used to construct three age-cohort specific normal distributions of frames, which were subsequently used to sample data amounts for each participant. Specifically, data were randomly sampled for each participant but sampling was constrained by the cohort-specific distribution of frames available based on their age. Using the sampled data, segregation was computed for each participant and correlated with age across all the participants. The sampling process was repeated 10,000 times, resulting in 10,000 correlation coefficients between age and system segregation.

Results

Measures of brain network segregation and integration vary in relation to the amount of data used to calculate brain network edges

The primary analysis of the current report was performed using MSC dataset that comprised data from nine participants (age range, 24–34 y). Each participant underwent 10 sessions of resting-state functional MRI images acquisitions, with each session occurring on a separate day (all sessions were within 7 weeks), enabling examination of cross-session (day) reliability of brain network measures for each individual.

Each of the brain network measures were computed on an adjacency matrix that was constructed with varying amount of data (in frames) used to calculate the edges of the matrix. Across data amount comparisons, frames were always sampled from the same scan session of a participant to avoid mixing within- and between-session variance (see Methods for details). We first investigated how estimates of an individual participant’s network measures varied as a function of the amount of resting-state data included. In Fig. 2(A), it is evident that both brain system segregation and modularity increase with the inclusion of more data, whereas mean CC and mean PC decrease with increasing data. The range of values can vary substantially in relation to the amount of data used to calculate the measures for a given participant. For example, system segregation values can vary up to 21.4% when comparing the measure calculated from 2.5 min of clean data to 25 min of clean data (mean difference when comparing values calculated from 2.5 min vs. 25 min, across participants and sessions: 9.12%). This observation has particularly important implications for comparisons of a participant’s brain network organization across sessions (e.g. during an interventional study examining pre- vs. post-intervention changes, or in longitudinal studies of development/aging that examine changes in brain network organization over extended durations), a point we will expand on further below.

Fig. 2.

Fig. 2

Participants’ functional network measures vary in relation to the amount of data used to construct their brain networks. Network measures were computed by sampling an increasing amount of data from a single session. (A) The resultant curves depict differences in network measure estimates as a function of the amount of data used to calculate the measure, from a given session. Curves from all sessions for a given participant were averaged to generate the curves depicted in the figure; error bars reflect within participant standard deviation across sessions for a given amount of data. Within a session, segregation and modularity increase with increasing amounts of data, whereas clustering coefficient and participation coefficient decrease. In all cases the values appear to plateau (asymptote) with larger quantities of data, and the patterns are consistent across participants. Data points are displayed in 50-frame increments (i.e. ~ 1.8 min) to facilitate visualization. (B) Participant’s functional network measures asymptote when a greater quantity of data is used to construct their brain network. The derivative of network measure curves in each session (upper panel) was calculated to quantify the change of network measures in relation to the amount of data used. Derivative values were averaged across sessions and participants. Each panel depicts the mean derivative of a network measure as a function of the amount of data used in its calculation, and the shaded area depicts the standard deviation of derivatives across all sessions of all participants. Vertical lines mark the positions at which derivative curves first reach zero and stabilize thereafter, indicating the network measures eventually asymptote with enough data. Most network measures need large amounts of data to become asymptotic (system segregation: 24 min, modularity: 20 min, mean clustering coefficient: 25 min, mean participation coefficient: 13 min).

A participant’s network measures stabilize within a session, with sufficient amounts of data

In Fig. 2(A), it appears that each of the measures may plateau (asymptote) with larger quantities of data. Asymptotes were confirmed empirically by determining whether there exists a point at which the first derivative of the data curves reaches zero and remains stable thereafter. Figure 2(B) demonstrates that all examined network measures asymptote with sufficient amounts of data, although the asymptote necessitates a considerable amount of data and varies across measures (system segregation: 24 min, modularity: 20 min, mean CC: 25 min, mean PC: 13 min).

A perfect asymptote may be too conservative to serve as a general criterion, and it is possible that the measures get “close enough” to their stable values with considerably less data (e.g. within 2% of their observed asymptote; Fig. 3A). Most network measures still require a considerable quantity of data to settle toward their identified asymptotic values (Msegregation = 18 ± 5 min, Mmodularity = 13 ± 4 min, MCC = 19 ± 5 min, MPC = 12 ± 2 min; Fig. 3B). Naturally, relaxing the error tolerance relative to the stable values results in less data required for the network measures to settle.

Fig. 3.

Fig. 3

Network measures stabilize within each session for a given participant with sufficient data. (A) In an example session of an MSC participant, the settling time for system segregation is 18 min, from which the values stay steadily within 2% (error tolerance) of the final value using maximum data points. (B) The mean settling time of each network measure across all sessions and all participants using different thresholds of error tolerance (1–10%). The settling time is generally shorter if a more relaxed error tolerance is used.

Stable estimates of network measures are reliable across sessions collected close in time

We have established that measures of functional brain network organization vary in relation to the amount of data used to calculate the edges of the matrices, within a scan session. These observations parallel the conclusions made in a number of previous reports (Laumann et al. 2015; Gordon et al. 2017), although, in those cases, frames were sampled randomly across multiple sessions that were aggregated across days, limiting inferences regarding within-session bias and variability.

Given that the value of a brain network measure can differ based on how much data is used to calculate it, it is also important to determine whether the between-session reliability of brain network measures vary in relation to data quantity. The MSC data allow for effective evaluation of measurement reliability, whereby we can test how reliable the brain network measures are across scanning sessions which have been collected close in time (within days to weeks). The between-session reliability of each network measures was quantified by assessing the variability of measures across scans, within participants. Variability was calculated as the ratio of the standard deviation (SD) across sessions relative to the mean measured value. Figure 4 demonstrates that between-session variability quickly decreases with increasing amounts of data for all network measures (critically, equivalent amounts of data were used for each session). Brain system segregation and modularity have lower between-session variability in general, relative to mean CC and mean PC.

Fig. 4.

Fig. 4

Network measures are reliable between sessions, especially when larger quantities of clean data are used in their calculation. (A) Between-session reliability was quantified by assessing the variability of measures across scans, within participants. This was calculated as the ratio of the standard deviation (SD) across sessions relative to the mean measured value. For each participant and brain network measure, the between-session variability decreases with increasing amounts of data used to calculate the brain network measure. (B) For each brain network measure, the between-session variability asymptotes with increasing amounts of data. The derivatives were calculated for each between-session variability curve in panel (A), and then averaged (black line in each panel), with the shaded area representing the standard deviation of derivatives across all sessions of all participants. Vertical lines mark the positions at which derivative curves reach zero, indicating that the between-session reliability of network measures asymptotes relatively quickly, and is high even with lesser data quantities.

It is evident that while the value of network measures takes a greater amount of data (time) to reach an asymptotic value (Fig. 2), considerably less data are necessary to observe high between-session reliability (Fig. 4A). This was formally evaluated by identifying the asymptotic point of the curves in Fig. 4(A). As is evident in Fig. 4(B), the between-session variability of all examined network measures asymptotes relatively quickly, although the asymptote point varies across measures (system segregation: 5 min, modularity: 11 min, clustering coefficient: 9 min, participation coefficient: 7 min).

Another common way of assessing reliability of a given measure is to examine the ICC (e.g. Birn et al. 2013; Braun et al. 2012; Choe et al. 2017; Noble et al. 2017; Xu et al. 2016; Zuo and Xing 2014). ICC is a descriptive measure that takes both within- and between- subject variability into account. ICC was calculated as a function of the amount of data used for each session. The ICC values of most network measures (except modularity) increase with increasing amounts of data included per session (see Fig. 5 and Supplemental Text S1 and Fig. S1 for related observations). These values appear to plateau with about 12–15 min of data (see System Segregation and PC). ICC values for system segregation are consistently the highest, while ICC values for modularity are lowest, irrespective of the amount of data used.

Fig. 5.

Fig. 5

ICC of network measures varies with increasing amount of data. ICC compares between-subject variability relative to within-subject variability of each network measure. Higher ICC values indicate greater differences between participants and lesser differences within each participant across sessions. Network measures and ICC values were computed using increasing amounts of data in each session. The ICC values of system segregation and mean PC are relatively higher and continue to increase with the inclusion of greater amounts of data. ICC values of modularity and mean CC are relatively lower, and do not substantially change with greater amounts of data used to calculate the measures.

Collectively, these findings reveal that measures of functional brain network segregation and integration vary with increasing amounts of data. However, both within- and between-session variability of network measures stabilize with sufficient data quantities, suggesting that reliable estimates of network segregation and integration are obtainable when enough clean data is included per participant. These results lead to important applications toward understanding the impact of data quantity on estimating network measures in cross-sectional, longitudinal, and interventional comparisons.

Using unequal quantities of data across sessions results in artifactual estimation of brain network changes within a participant and distorts between-participant comparisons

Artifactual estimation of brain network changes within participants

The preceding results have demonstrated that measures of large-scale functional brain network organization vary as a function of the amount of time-series data being used in each session. This has important implications both for (i) studies that aim to compare how an individual’s network measure changes over multiple sessions (e.g. longitudinal studies of development and aging, pre/post interventional studies) and (ii) studies that aim to compare network measures across participants (e.g. in relation to age, cognitive ability, individual phenotypes, disease status, etc.). Here we demonstrate how comparing sessions with unequal amounts of data can result in misestimation of changes or differences, within or across participants, respectively.

To demonstrate the first of these problems more explicitly, we first examined a scenario where changes in an individual’s brain network organization are not expected. Specifically, we examined the HCP-YA dataset (n = 730, age range: 22–35 y), comparing participants’ brain system segregation across the two scanning sessions. Given that the two scanning sessions were collected in close proximity (for most participants, on two consecutive days, but at maximum a week apart, as per descriptions of the HCP-YA protocol; Van Essen et al. 2012b), and no systematic interventions were administered to participants in between the scan sessions, we expected that their brain system segregation values should remain relatively consistent across sessions. The absence of change is confirmed when the amount of data is equated across each of the participant’s scanning sessions (Fig. 6). However, Fig. 6 additionally illustrates that comparing two sessions with unequal amounts of data results in systematic inflation of the magnitude of calculated “change.” The observed “change” in brain system segregation is an artifact of failure to equate the amount of data across scan sessions and the degree of this artifact varies in proportion to the magnitude of data quantity differences across sessions. This artifact is evident even when the differences in amount of data used to calculate the brain network measure is not that large (i.e. 5 min vs. 6 min).

Fig. 6.

Fig. 6

Using unequal amounts of data across scanning sessions results in systematic distortion of a participant’s change in brain network organization. HCP-YA data were collected on two consecutive days without experimental interventions, which makes this dataset a suitable candidate to examine the impact of including different amounts of data for estimating between-session differences within a participant. When the amount of data is equated across the two scan sessions, system segregation in session 1 does not differ from system segregation in session 2 (paired-samples t-test: t(1458) = −1.25, P = 0.212; far left panel). However, slightly varying participant’s data quantity in session 2 (6 min) leads to significantly higher system segregation values when compared to participant’s system segregation from session 1 (5 min) (t(1458) = 6.64, P < 0.001; second panel). As more data are used in session 2 relative to session 1, this between-session “change” of system segregation becomes greater (ts > 12.48, Ps < 0.001). These results demonstrate that using unequal amounts of data leads to systematically biased estimates of between-session “change” of network measures within a participant. Each dot in the figure represents system segregation value of a participant; the line between each two dots shows between-session “change” within the same participant. In each plot the thick line represents the mean group change, “ns” = non-significant, “*” = P < 0.05.

Misestimation of brain network differences across participants

The sensitivity of brain network measures to data quantity also indicates that between participant comparisons can be misestimated if appropriate steps are not taken to minimize this source of bias.

To illustrate this impact directly, we focus on the well-established relationship between increasing adult age and decreasing brain system segregation (e.g. Chan et al. 2014; for a review, see Wig 2017). To further highlight the robustness of this result, we analyzed the HCP-A dataset. Based on the preceding observations demonstrating that the measure of brain system segregation varies as a function of data quantity within participants, the number of clean frames was first equated across participants within the dataset (included HCP-A participants had 20 min of data). As expected, increasing adult age is associated with decreasing brain system segregation (r = −0.407, P < 0.001; Fig. 7A).

Fig. 7.

Fig. 7

Cross-sectional relationships between system segregation and a variable of interest (in this case, age) can be misestimated using unequal quantities of data across participants. (A) System segregation computed using the maximally available amounts of equated data (20 min) across all HCP-A participants. System segregation decreases with age (r = −0.407, P < 0.001), indicating less segregated functional systems in older adults. (B) This correlation coefficient observed in (A) serves as a “standard” reference (vertical dotted line in B) for comparisons of correlation distributions. The rightmost distribution (requal) depicts correlation coefficients between age and system segregation calculated using equal amounts of data across participants (ranging from 3 to 1500 frames). In general, the distribution is tightly distributed around the 20-min reference line. However, using unequal amounts of data tends to misestimate the relationship between segregation and age. The leftmost distribution (runequal) depicts correlation coefficients between age and segregation calculated using unequal amounts of data across participants, whereby frame sampling was constrained in a manner that mimics the relationship between age and frame loss (see the text for details). The range of calculated correlation coefficients between age and segregation is wider, and the calculated relationships are systematically overestimated (stronger negative correlation values) than both the maximum data reference and the distribution of correlations observed when data amounts are equated across participants. The bias is still evident even when the number of frames is included as a covariate in the statistical model (the wide distribution in the middle, labeled Partial runequal). The correlation distributions are systematically left-shifted relative to the 20-min reference line and the distribution of r-values when data are equated, demonstrating that this statistical maneuver does not correct misestimation of relationships between age and system segregation.

In practice however, data-cleaning strategies (e.g. scrubbing, frame regression) can result in an unequal number of “clean” frames that are available across participants (e.g. individuals will vary in the extent to which they move, which would result in different amounts of data remaining following scrubbing). This effect is more prominent in older adults with greater data loss due to excessive head motion (Savalia et al. 2017). This aspect of data acquisition and processing often leads researchers to include varying amounts of clean data across participants, based on the assumption that more data are better. However, as we have seen in earlier sections, data quantity can influence the estimates of network measures, so differences in data quantity across participants is potentially problematic.

To illustrate the impact of relaxing the requirement of equating frames across participants on cross-sectional comparisons, we computed correlation coefficients between age and system segregation estimated by (i) sampling a random number of frames for each participant in the HCP-A dataset (ranging from three frames [2.4 s] to 1,500 frames [20 min]), (ii) calculating system segregation for that participant, and (iii) calculating the correlation between age and system segregation across participants. Random frame sampling was constrained to mimic the relationship between age and data loss (see methods for details). This process was repeated 10,000 times, resulting in 10,000 correlation coefficients. This distribution of correlation coefficients was compared to the correlation coefficients using an equated number of frames across all participants (again ranging from 3 to 1,500 frames).

Using equal amounts of data produces greater correlation coefficients that are relatively tightly distributed (Mean = −0.409, Median = −0.348, SD = 0.014; red distribution [i.e. the rightmost distribution] in Fig. 7B) centered closely around the correlation observed when maximum data are included (red dotted line in Fig. 7B, corresponding to 20 min of data per participant as observed in Fig. 7A).

In contrast, using unequal amounts of data across participants (randomly sampled), there exists a broad spread of the correlation values (darker blue distribution [i.e. the leftmost distribution] in Fig. 7B), which are also systematically stronger than the correlation values obtained when data amounts are equated and when maximum data are used, resulting in inflated effect sizes (96.8% of the correlations are stronger than the correlation using the maximum amount of data). This observation underscores how using unequal amounts of data could significantly misestimate the relationship between segregation and age.

In some instances, investigators have opted to include number of frames as a covariate when estimating relationships between variables of interest, in order to control for potential confounds resulting from inclusion of unequal amounts of data across participants. However, Fig. 7(B) also demonstrates the inadequacy with this approach. For unequal sampling, the majority (77.8%) of partial correlations between age and segregation (in lighter blue [i.e. distribution in the middle]) are still stronger than the correlation using the maximum number of equated frames. These results demonstrate that using unequal amounts of data across participants often significantly misestimates the relationship between variables of interest in cross-sectional comparisons, even after statically controlling for the number of frames across participants.

Finally, it is important to point out that the above scenarios related to within and between participant comparisons are not rare. Unequal data comparisons for a participant and group of participants can easily be introduced from changes in imaging protocols (e.g. increasing or decreasing the number of frames or scans of resting-state data on a follow-up session), or as a consequence of data cleaning techniques that result in different amounts of clean data being used across imaging sessions. Critically, these biases are not restricted to frame scrubbing but are also evident following alternate data processing and cleaning techniques which retain all frames (see Supplemental Text S2 and Figs. S2–S5). The impacts of this source of variance are important to consider, as a systematic relationship between a given variable of interest (e.g. age of participants, patients vs. controls, time) and data quantity/quality (e.g. due to head movement changes/differences) can result in biased estimation of change that gets mis-attributed to the variable of interest.

Discussion

The described findings demonstrate that: (i) measures of functional brain network segregation and integration are sensitive to the amount of RSFC data used to calculate each measure, although the measures asymptote with greater data quantities, (ii) obtaining reliable estimates of RSFC brain network measures within a scanning session is achievable, and (iii) the comparison of network measures within and between participants can be biased when the amount of data is not equated across scanning sessions.

Reliable network measures are necessary for both basic science and clinical applications

Several previous studies have evaluated the reliability of RSFC maps (Van Dijk et al. 2010; Birn et al. 2013; Noble et al. 2017) and measures of RSFC networks (Cao et al. 2014; Chen et al. 2015). Reliability has been assessed with respect to various factors, including the impact of eyes open vs. eyes closed and sleep vs. awake scanning (Patriat et al. 2013), spatial differences (Van Dijk et al. 2010; Zuo and Xing 2014; Shah et al. 2016), and direct comparisons between network measures (Braun et al. 2012; Cao et al. 2014; Chen et al. 2015). While the bulk of this work has focused on data collection efforts that were more limited in terms of absolute scan length (e.g. ≤ 5 min) and number of sessions across time, they revealed weaker to modest estimates of reliability. More recent studies examining RSFC reliability have included data with greater scan lengths and sessions (e.g. O'Connor et al. 2017; Noble et al. 2019). These latter studies align with the observations presented here, demonstrating that more reliable measures of RSFC patterns are achievable with sufficient data quantity. The availability of data collection efforts that include longer lengths of resting-state data acquisition and multiple scan sessions offers the opportunity to evaluate whether, and how, measures of functional network organization vary as a function of data quantity and what the resultant reliability of these measures is. In keeping with this, we demonstrated that the reliability of network measures varies in relation to within-session data quantity, but between-session variability quickly decreases especially when data amounts are equated across sessions. This latter point underscores that in addition to considering the amount of data necessary to achieve a low degree of variability, the comparison of measures of brain network segregation and integration estimated with unequal amounts of data across sessions could bias cross-sectional, longitudinal, and interventional comparisons in relation to variables of interest. This influence of varying data amounts persists even after controlling for amount of data in the models using participant-level covariates.

Although we have focused our efforts on network measures that are defined by graph analysis, the same impacts of data equating likely apply to the estimations of node-to-node correlation values and correlation maps, given that they are used as inputs to the network measures. Likewise, the implications extend to alternate analytic approaches that examine and compare functional correlation patterns across individuals and groups of individuals (e.g. finger-printing [Finn et al. 2015; Gratton et al. 2018; Mars et al. 2018], machine-learning based prediction [Dosenbach et al. 2010; Khosla et al. 2019], gradients of connectivity patterns [Margulies et al. 2016; Huntenburg et al. 2018], and parcellation of cerebral cortex [Wig et al. 2014; Eickhoff et al. 2018]). It is likely that equating frames of data during the formation of the network matrix within and across participants is important for accurately estimating and evaluating brain network organization in general.

RSFC system segregation varies in relation to age and cognitive performance, but has also been used to differentiate unhealthy populations from control participants, which has seeded the idea of incorporating this measure in clinical applications (for a review, see Wig 2017). For example, differences in system segregation (and other related measures) are evident in adult individuals with AD (Ewers et al. 2021; Zhang et al. 2023), epilepsy (Pedersen et al. 2015), schizophrenia (Yang et al. 2016), and younger individuals with ADHD (Mills et al. 2018), suggesting that these disorders may be associated with altered “meso-scale” organization of large-scale resting-state brain networks. Among older individuals classified as “healthier,” declining system segregation has been shown to be prognostic of dementia up to 10 y prior to cognitive impairment, and independent of AD-related genetic risk, the presence of baseline AD-associated pathology and longitudinal structural changes (Chan et al. 2021). Altogether, these observations highlight the potential of applying system segregation as a biomarker of disease diagnosis and/or incipient cognitive decline. At a basic level, this application requires that the measure is a reliable descriptor of an individual’s brain network organization. Our report takes steps toward establishing this criterion by evaluating both the measurement values and the between session measurement reliability as a function of data amount. First, with sufficient data quantity in a scanning session, the system segregation values asymptote. This observation is in line with previous reports that have examined several related network measures (although not specifically system segregation; Laumann et al. 2015; Gordon et al. 2017). In addition, in the present report, system segregation was evaluated within individual scan sessions, which allowed interpretation of the impact of data quantity on measurement values as it relates to within-session estimates, an approach which is pragmatically more likely and feasible for application purposes. Second, despite the influence of data amount on measurement values, system segregation was found to be reliable across measurements obtained on multiple days as long as data amounts were equated across sessions, suggesting that the measure can be used to reliably describe the organization of large-scale functional brain networks within individuals. One notable feature was that system segregation was found to exhibit greater participant discernability than modularity, with the latter achieving notably lower ICC values irrespective of data amount (Fig. 5), indicating that system segregation may be a better individual indicator of the community-level organization of RSFC networks.

Exploring the sources of diminished reliability and methods to alleviate them

The current study indicates that RSFC network measures vary with varying amounts of data, which may be explained from the perspective of sampling error. Using insufficient data could lead to greater sampling error that misestimates functional networks. Conversely, when more clean data are used, the degree of sampling error is minimized, which results in better estimation of the network measures. To test this hypothesis, we conducted an additional analysis where noise was directly added to the resting-state data to simulate the scenario of varying sampling error on functional network estimation; the result demonstrates that system segregation decreases with increasing noise (greater sampling error) in a manner that parallels the differences in data quantity observed throughout this report (Supplemental Text S3 and Figs. S6S7).

Toward the previous point, in order to improve accurate estimation of resting-state functional relationships, it is critical to include data cleaning techniques that remove variance related to head motion, cardiac-signals, breathing, and other non-neuronal signals (Power et al. 2020). There exist multiple approaches that are reasonably effective toward achieving this aim. Data censoring (“scrubbing”) accompanied with global signal regression, has been argued to be an important step toward collectively minimizing distance-dependent covariance that are caused by multiple sources of noise that cannot be removed by other data cleaning techniques (Ciric et al. 2017; Power et al. 2020). The present results may lead one to conclude that data censoring is less appealing due to the introduction of variable data quantity across participants/scans. However, additional supplementary analyses indicate that other data processing techniques, including those relying on frame regression (“spike regression”) (Satterthwaite et al. 2013; Yan et al. 2013) and ICA-based regression (Beckmann et al. 2009; Nickerson et al. 2017) are not immune to the observed impacts of including a variable number of “clean frames” across scans (Supplemental Text S2 and Figs. S2S5). As participants, scans, and scan sessions often differ in their number of clean frames present, this will naturally introduce the biases noted throughout the present report which must be dealt with, irrespective of the data cleaning technique.

Practical considerations for data acquisition

Older adults (Mowinckel et al. 2012; Van Dijk et al. 2012; Savalia et al. 2017) and children (Power et al. 2014) exhibit greater amounts of head movement relative to younger adults, and many patient groups exhibit greater head movement relative to age-matched control participants (Haller et al. 2014; Pardoe et al. 2016). This means that less clean data are available for these cohorts following preprocessing if a fixed amount of data is collected (even in cases where data are not removed, as with outlier frame regression techniques). Importantly, supplementary analyses indicate that it is the total scan time as opposed to number of frames that drive the reliability effects described in this report (Supplemental Text S4 and Fig. S8). This observation is in line with the properties of resting-state time series that by definition comprise lower-frequency signals (< 0.1 Hz). Using longer scan length can capture more variance from lower-frequency domains and result in better estimation of functional networks. This conclusion is similar to that reached by a previous study, which focused on ICC values of specific connections (Birn et al. 2013). As a result, in order to obtain a sufficient and equivalent amount of data that will allow reliable measurement of RSFC brain network segregation or integration across children, older individuals, and patient populations, data collection efforts in these cohorts will generally need to be longer relative to younger adult cohorts (either via longer duration scans protocols, and/or via real-time monitoring of head motion to increase the likelihood that enough data is collected to obtain reliable estimates of function brain networks; Dosenbach et al. 2017). While newer generations of data acquisition techniques (e.g. multi-echo fMRI; Lynch et al. 2020) or data cleaning methods (e.g. Gratton et al. 2018; Kaplan et al. 2022) may reduce the amount of data needed to achieve high-quality signal, the necessity of equating clean frames across sessions remains.

With very large quantities of data, equating data may be less necessary given that measures eventually asymptote [i.e. the confounding effect is particularly insidious when less data are available for participants (e.g. <=20 min)]. However, many data acquisition protocols and existing datasets (including several large-scale initiatives including UK Biobank, Adolescent Brain Cognitive Development, Alzheimer’s Disease Neuroimaging Initiative, etc.) do not have the luxury of collecting the identified scan lengths given the constraints introduced by scanning special populations (in terms of age and clinical status) and limited amount of scanning time available per participant. As such, while the present results support the value of acquiring greater amounts of data, in many practical situations, it is critical to recognize the importance of equating data quantity toward achieving consistency in brain network measurement.

Conclusion

Given the increased focus on measures of large-scale brain network organization, it is critical that researchers understand the consistency and reliability of RSFC-based measures of brain network segregation and integration. We have demonstrated that measures of brain system segregation, modularity, mean CC, and mean PC vary with increasing amounts of data for a given individual. With sufficient data, low inter-session variability of network measures can be achieved, indicating that obtaining reliable estimates of these measures is feasible within an individual. Notably, obtaining accurate and reliable measures of segregation and integration necessitates equating the quantity of data included in their calculation across scan sessions (between and/or within individuals). The failure to equate data quantity results in misestimation of cross-sectional differences and within individual changes of brain network changes, which can impact the interpretation of results. In most circumstances then, it is reasonable to assert that while more data is better, equal data is necessary.

Supplementary Material

Han_SegregationReliability_SI_20240115_bhad506

Acknowledgments

The authors thank Babatunde Adeyemo and Steven E. Petersen for sharing their pre-processed HCP-YA data. The authors are also thankful to the reviewers for their insightful comments on an earlier version of this article.

Contributor Information

Liang Han, Center for Vital Longevity and School of Behavioral and Brain Sciences, University of Texas at Dallas, Dallas, TX 75235, United States.

Micaela Y Chan, Center for Vital Longevity and School of Behavioral and Brain Sciences, University of Texas at Dallas, Dallas, TX 75235, United States.

Phillip F Agres, Center for Vital Longevity and School of Behavioral and Brain Sciences, University of Texas at Dallas, Dallas, TX 75235, United States.

Ezra Winter-Nelson, Center for Vital Longevity and School of Behavioral and Brain Sciences, University of Texas at Dallas, Dallas, TX 75235, United States.

Ziwei Zhang, Center for Vital Longevity and School of Behavioral and Brain Sciences, University of Texas at Dallas, Dallas, TX 75235, United States.

Gagan S Wig, Center for Vital Longevity and School of Behavioral and Brain Sciences, University of Texas at Dallas, Dallas, TX 75235, United States; Department of Psychiatry, University of Texas Southwestern Medical Center, Dallas, TX 75390, United States.

Author contributions

Gagan S. Wig (Conceptualization, Formal analysis, Funding acquisition, Methodology, Project administration, Resources, Supervision, Writing—original draft, Writing—review and editing), Liang Han (Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Visualization, Writing—original draft, Writing—review and editing), Micaela Y. Chan (Data curation, Methodology, Software, Validation, Visualization, Writing—review and editing), Phillip F. Agres (Writing—review and editing), Ezra Winter-Nelson (Data curation, Visualization, Writing—review and editing), Ziwei Zhang (Visualization, Writing—review and editing).

Funding

This work was supported by an Understanding Human Cognition Scholar Award from the James S. McDonnell Foundation (to G.S.W.) and National Institutes of Health (Grant R01 AG063930 to G.S.W.). HCP-YA data were provided by the Human Connectome Project, WU-Minn Consortium (Principal Investigators: David Van Essen and Kamil Ugurbil; 1U54MH091657) funded by the 16 NIH Institutes and Centers that support the NIH Blueprint for Neuroscience Research; and by the McDonnell Center for Systems Neuroscience at Washington University. HCP-A data were supported by the National Institute on Aging of the National Institutes of Health under Award Number U01AG052564 and by funds provided by the McDonnell Center for Systems Neuroscience at Washington University in St. Louis.

Conflict of interest statement: None declared.

Data availability

The MSC dataset is available via OpenNeuro data repository (https://openneuro.org/datasets/ds000224). The HCP-YA dataset is available via the Human Connectome Project with ConnectomeDB (https://www.humanconnectome.org). The HCP-A dataset is available on the NIMH Data Archive (https://nda.nih.gov/).

References

  1. Albert NB, Robertson EM, Miall RC. The resting human brain and motor learning. Curr Biol. 2009:19(12):1023–1027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Bassett DS, Wymbs NF, Porter MA, Mucha PJ, Carlson JM, Grafton ST. Dynamic reconfiguration of human brain networks during learning. Proc Natl Acad Sci U S A. 2011:108(18):7641–7646. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Becker R, Braun U, Schwarz AJ, Gass N, Schweiger JI, Weber-Fahr W, Schenker E, Spedding M, Clemm von Hohenberg C, Risterucci C, et al. Species-conserved reconfigurations of brain network topology induced by ketamine. Transl Psychiatry. 2016:6(4):7, e786. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Beckmann CF, Mackay CE, Filippini N, Smith SM. Group comparison of resting-state FMRI data using multi-subject ICA and dual regression. NeuroImage. 2009:47:S148. [Google Scholar]
  5. Betzel RF, Byrge L, He Y, Goni J, Zuo XN, Sporns O. Changes in structural and functional connectivity among resting-state networks across the human lifespan. NeuroImage. 2014:102:345–357. [DOI] [PubMed] [Google Scholar]
  6. Birn RM, Molloy EK, Patriat R, Parker T, Meier TB, Kirk GR, Nair VA, Meyerand ME, Prabhakaran V. The effect of scan length on the reliability of resting-state fMRI connectivity estimates. NeuroImage. 2013:83:550–558. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bookheimer SY, Salat DH, Terpstra M, Ances BM, Barch DM, Buckner RL, Burgess GC, Curtiss SW, Diaz-Santos M, Elam JS, et al. The Lifespan Human Connectome Project in aging: an overview. NeuroImage. 2019:185:335–348. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Braun U, Plichta MM, Esslinger C, Sauer C, Haddad L, Grimm O, Mier D, Mohnke S, Heinz A, Erk S, et al. Test-retest reliability of resting-state connectivity network characteristics using fMRI and graph theoretical measures. NeuroImage. 2012:59(2):1404–1412. [DOI] [PubMed] [Google Scholar]
  9. Brier MR, Thomas JB, Fagan AM, Hassenstab J, Holtzman DM, Benzinger TL, Morris JC, Ances BM. Functional connectivity and graph theory in preclinical Alzheimer's disease. Neurobiol Aging. 2014:35(4):757–768. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Cassady K, Gagnon H, Freiburger E, Lalwani P, Simmonite M, Park DC, Peltier SJ, Taylor SF, Weissman DH, Seidler RD, Polk TA.. Network segregation varies with neural distinctiveness in sensorimotor cortex. NeuroImage. 2020:212:116663. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Cao H, Plichta MM, Schäfer A, Haddad L, Grimm O, Schneider M, Esslinger C, Kirsch P, Meyer-Lindenberg A, Tost H. Test-retest reliability of fMRI-based graph theoretical properties during working memory, emotion processing, and resting state. NeuroImage. 2014:84:888–900. [DOI] [PubMed] [Google Scholar]
  12. Chan MY, Park DC, Savalia NK, Petersen SE, Wig GS. Decreased segregation of brain systems across the healthy adult lifespan. Proc Natl Acad Sci U S A. 2014:111(46):E4997–E5006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Chan MY, Alhazmi FH, Park DC, Savalia NK, Wig GS. Resting-state network topology differentiates task signals across the adult life span. J Neurosci. 2017:37(10):2734–2745. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Chan MY, Na J, Agres PF, Savalia NK, Park DC, Wig GS. Socioeconomic status moderates age-related differences in the brain's functional network organization and anatomy across the adult lifespan. Proc Natl Acad Sci U S A. 2018:115(22):E5144–E5153. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Chan MY, Han L, Carreno CA, Zhang Z, Rodriguez RM, LaRose M, Hassenstab J, Wig GS. Long-term prognosis and educational determinants of brain network decline in older adult individuals. Nat Aging. 2021:1(11):1053–1067. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Chang C, Glover GH. Time-frequency dynamics of resting-state brain connectivity measured with fMRI. NeuroImage. 2010:50(1):81–98. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Chen M, Deem MW. Development of modularity in the neural activity of children's brains. Phys Biol. 2015:12(1):1–12, 016009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Chen B, Xu T, Zhou C, Wang L, Yang N, Wang Z, Dong HM, Yang Z, Zang YF, Zuo XN, et al. Individual variability and test-retest reliability revealed by ten repeated resting-state brain scans over one month. PLoS One. 2015:10(12):1–21. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Choe AS, Nebel MB, Barber AD, Cohen JR, Xu Y, Pekar JJ, Caffo B, Lindquist MA. Comparing test-retest reliability of dynamic functional connectivity methods. NeuroImage. 2017:158:155–175. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Chong JSX, Ng KK, Tandi J, Wang C, Poh JH, Lo JC, Chee MWL, Zhou JH. Longitudinal changes in the cerebral cortex functional organization of healthy elderly. J Neurosci. 2019:39(28):5534–5550. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Ciric R, Wolf DH, Power JD, Roalf DR, Baum GL, Ruparel K, Shinohara RT, Elliott MA, Eickhoff SB, Davatzikos C, et al. Benchmarking of participant-level confound regression strategies for the control of motion artifact in studies of functional connectivity. NeuroImage. 2017:154:174–187. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Dosenbach NU, Nardos B, Cohen AL, Fair DA, Power JD, Church JA, Nelson SM, Wig GS, Vogel AC, Lessov-Schlaggar CN, et al. Prediction of individual brain maturity using fMRI. Science. 2010:329(5997):1358–1361. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Dosenbach NUF, Koller JM, Earl EA, Miranda-Dominguez O, Klein RL, Van AN, Snyder AZ, Nagel BJ, Nigg JT, Nguyen AL, et al. Real-time motion analytics during brain MRI improve data quality and reduce costs. NeuroImage. 2017:161:80–93. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Duncan ES, Small SL. Increased modularity of resting state networks supports improved narrative production in aphasia recovery. Brain Connect. 2016:6(7):524–529. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Eickhoff SB, Yeo BTT, Genon S. Imaging-based parcellations of the human brain. Nat Rev Neurosci. 2018:19(11):672–686. [DOI] [PubMed] [Google Scholar]
  26. Esposito F, Bertolino A, Scarabino T, Latorre V, Blasi G, Popolizio T, Tedeschi G, Cirillo S, Goebel R, Di Salle F. Independent component model of the default-mode brain function: assessing the impact of active thinking. Brain Res Bull. 2006:70(4–6):263–269. [DOI] [PubMed] [Google Scholar]
  27. Ewers M, Luan Y, Frontzkowski L, Neitzel J, Rubinski A, Dichgans M, Hassenstab J, Gordon BA, Chhatwal JP, Levin J, et al. Segregation of functional networks is associated with cognitive resilience in Alzheimer's disease. Brain. 2021:144(7):2176–2185. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Fair DA, Miranda-Dominguez O, Snyder AZ, Perrone A, Earl EA, Van AN, Koller JM, Feczko E, Tisdall MD, Kouwe A, et al. Correction of respiratory artifacts in MRI head motion estimates. NeuroImage. 2020:208:1–17, 116400. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Finn ES, Shen X, Scheinost D, Rosenberg MD, Huang J, Chun MM, Papademetris X, Constable RT. Functional connectome fingerprinting: identifying individuals using patterns of brain connectivity. Nat Neurosci. 2015:18(11):1664–1671. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Fornito A, Harrison BJ, Zalesky A, Simons JS. Competitive and cooperative dynamics of large-scale brain functional networks supporting recollection. Proc Natl Acad Sci U S A. 2012:109(31):12788–12793. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Fransson P. How default is the default mode of brain function?: further evidence from intrinsic BOLD signal fluctuations. Neuropsychologia. 2006:44(14):2836–2845. [DOI] [PubMed] [Google Scholar]
  32. Geerligs L, Renken RJ, Saliasi E, Maurits NM, Lorist MM. A brain-wide study of age-related changes in functional connectivity. Cereb Cortex. 2015:25(7):1987–1999. [DOI] [PubMed] [Google Scholar]
  33. Glasser MF, Sotiropoulos SN, Wilson JA, Coalson TS, Fischl B, Andersson JL, Xu JQ, Jbabdi S, Webster M, Polimeni JR, et al. The minimal preprocessing pipelines for the Human Connectome Project. NeuroImage. 2013:80:105–124. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Gordon EM, Laumann TO, Adeyemo B, Huckins JF, Kelley WM, Petersen SE. Generation and evaluation of a cortical area parcellation from resting-state correlations. Cereb Cortex. 2016:26(1):288–303. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Gordon EM, Laumann TO, Gilmore AW, Newbold DJ, Greene DJ, Berg JJ, Ortega M, Hoyt-Drazen C, Gratton C, Sun H, et al. Precision functional mapping of individual human brains. Neuron. 2017, 2017:95:791–807. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Gratton C, Laumann TO, Nielsen AN, Greene DJ, Gordon EM, Gilmore AW, Nelson SM, Coalson RS, Snyder AZ, Schlaggar BL, et al. Functional brain networks are dominated by stable group and individual factors, not cognitive or daily variation. Neuron. 2018:98(2):439–452.e435. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Gratton C, Dworetsky A, Coalson RS, Adeyemo B, Laumann TO, Wig GS, Kong TS, Gratton G, Fabiani M, Barch DM, et al. Removal of high frequency contamination from motion estimates in single-band fMRI saves data without biasing functional connectivity. NeuroImage. 2020:217:1–14, 116866. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Gu S, Satterthwaite TD, Medaglia JD, Yang M, Gur RE, Gur RC, Bassett DS. Emergence of system roles in normative neurodevelopment. Proc Natl Acad Sci U S A. 2015:112(44):13681–13686. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Guimera R, Mossa S, Turtschi A, Amaral LA. The worldwide air transportation network: anomalous centrality, community structure, and cities' global roles. Proc Natl Acad Sci U S A. 2005:102(22):7794–7799. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Haller S, Monsch AU, Richiardi J, Barkhof F, Kressig RW, Radue EW. Head motion parameters in fMRI differ between patients with mild cognitive impairment and Alzheimer disease versus elderly control subjects. Brain Topogr. 2014:27(6):801–807. [DOI] [PubMed] [Google Scholar]
  41. Han L, Savalia NK, Chan MY, Agres PF, Nair AS, Wig GS. Functional parcellation of the cerebral cortex across the human adult lifespan. Cereb Cortex. 2018:28(12):4403–4423. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Handwerker DA, Roopchansingh V, Gonzalez-Castillo J, Bandettini PA. Periodic changes in fMRI connectivity. NeuroImage. 2012:63(3):1712–1719. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Honey CJ, Sporns O, Cammoun L, Gigandet X, Thiran JP, Meuli R, Hagmann P. Predicting human resting-state functional connectivity from structural connectivity. Proc Natl Acad Sci U S A. 2009:106(6):2035–2040. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Horovitz SG, Fukunaga M, Zwart JA, Gelderen P, Fulton SC, Balkin TJ, Duyn JH. Low frequency BOLD fluctuations during resting wakefulness and light sleep: a simultaneous EEG-fMRI study. Hum Brain Mapp. 2008:29(6):671–682. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Horovitz SG, Braun AR, Carr WS, Picchioni D, Balkin TJ, Fukunaga M, Duyn JH. Decoupling of the brain's default mode network during deep sleep. Proc Natl Acad Sci U S A. 2009:106(27):11376–11381. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Huntenburg JM, Bazin PL, Margulies DS. Large-scale gradients in human cortical organization. Trends Cogn Sci. 2018:22(1):21–31. [DOI] [PubMed] [Google Scholar]
  47. Jones DT, Vemuri P, Murphy MC, Gunter JL, Senjem ML, Machulda MM, Przybelski SA, Gregg BE, Kantarci K, Knopman DS. Non-stationarity in the “resting brain’s” modular architecture. PLoS One. 2012:7(6):e39731. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Kaplan S, Meyer D, Miranda-Dominguez O, Perrone A, Earl E, Alexopoulos D, Barch DM, Day TKM, Dust J, Eggebrecht AT, et al. Filtering respiratory motion artifact from resting state fMRI data in infant and toddler populations. NeuroImage. 2022:247:1–11, 118838. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Khosla M, Jamison K, Ngo GH, Kuceyeski A, Sabuncu MR. Machine learning in resting-state fMRI analysis. Magn Reson Imaging. 2019:64:101–121. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. King BR, Ruitenbeek P, Leunissen I, Cuypers K, Heise KF, Santos Monteiro T, Hermans L, Levin O, Albouy G, Mantini D, et al. Age-related declines in motor performance are associated with decreased segregation of large-scale resting state brain networks. Cereb Cortex. 2018:28(12):4390–4402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Kiviniemi V, Vire T, Remes J, Elseoud AA, Starck T, Tervonen O, Nikkinen J. A sliding time-window ICA reveals spatial variability of the default mode network in time. Brain Connect. 2011:1(4):339–347. [DOI] [PubMed] [Google Scholar]
  52. Kong TS, Gratton C, Low KA, Tan CH, Chiarelli AM, Fletcher MA, Zimmerman B, Maclin EL, Sutton BP, Gratton G, Fabiani M.. Age-related differences in functional brain network segregation are consistent with a cascade of cerebrovascular, structural, and cognitive effects. Netw Neurosci. 2020:4(1):89–114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Laumann TO, Gordon EM, Adeyemo B, Snyder AZ, Joo SJU, Chen MY, Gilmore AW, McDermott KB, Nelson SM, Dosenbach NUF, et al. Functional system and areal organization of a highly sampled individual human brain. Neuron. 2015:87(3):657–670. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Laumann TO, Snyder AZ, Mitra A, Gordon EM, Gratton C, Adeyemo B, Gilmore AW, Nelson SM, Berg JJ, Greene DJ, et al. On the stability of BOLD fMRI correlations. Cereb Cortex. 2017:27(10):4719–4732. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Lewis CM, Baldassarre A, Committeri G, Romani GL, Corbetta M. Learning sculpts the spontaneous activity of the resting human brain. Proc Natl Acad Sci U S A. 2009:106(41):17558–17563. [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Liu H, Stufflebeam SM, Sepulcre J, Hedden T, Buckner RL. Evidence from intrinsic activity that asymmetry of the human brain is controlled by multiple factors. Proc Natl Acad Sci. 2009:106(48):20499–20503. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Lynch CJ, Power JD, Scult MA, Dubin M, Gunning FM, Liston C. Rapid precision functional mapping of individuals using multi-echo fMRI. Cell Rep. 2020:33(12):e1–e3, 108540. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Madden DJ, Jain S, Monge ZA, Cook AD, Lee A, Huang H, Howard CM, Cohen JR.. Influence of structural and functional brain connectivity on age-related differences in fluid cognition. Neurobiol Aging. 2020:96:205–222. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Manza P, Wiers CE, Shokri-Kojori E, Kroll D, Feldman D, Schwandt M, Wang GJ, Tomasi D, Volkow ND.. Brain network segregation and glucose energy utilization: relevance for age-related differences in cognitive function. Cereb Cortex. 2020:30(11):5930–5942. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Marcus DS, Harms MP, Snyder AZ, Jenkinson M, Wilson JA, Glasser MF, Barch DM, Archie KA, Burgess GC, Ramaratnam M, et al. Human connectome project informatics: quality control, database services, and data visualization. NeuroImage. 2013:80:202–219. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Margulies DS, Ghosh SS, Goulas A, Falkiewicz M, Huntenburg JM, Langs G, Bezgin G, Eickhoff SB, Castellanos FX, Petrides M, et al. Situating the default-mode network along a principal gradient of macroscale cortical organization. Proc Natl Acad Sci U S A. 2016:113(44):12574–12579. [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Mars RB, Passingham RE, Jbabdi S. Connectivity fingerprints: from areal descriptions to abstract spaces. Trends Cogn Sci. 2018:22(11):1026–1037. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Miezin FM, Maccotta L, Ollinger JM, Petersen SE, Buckner RL. Characterizing the hemodynamic response: effects of presentation rate, sampling procedure, and the possibility of ordering brain activity based on relative timing. NeuroImage. 2000:11(6):735–759. [DOI] [PubMed] [Google Scholar]
  64. Mills BD, Miranda-Dominguez O, Mills KL, Earl E, Cordova M, Painter J, Karalunas SL, Nigg JT, Fair DA. ADHD and attentional control: impaired segregation of task positive and task negative brain networks. Netw Neurosci. 2018:2(2):200–217. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Mowinckel AM, Espeseth T, Westlye LT. Network-specific effects of age and in-scanner subject motion: a resting-state fMRI study of 238 healthy adults. NeuroImage. 2012:63(3):1364–1373. [DOI] [PubMed] [Google Scholar]
  66. Newman ME. Fast algorithm for detecting community structure in networks. Phys Rev E Stat Nonlinear Soft Matter Phys. 2004:69(6):066133. [DOI] [PubMed] [Google Scholar]
  67. Nickerson LD, Smith SM, Öngür D, Beckmann CF. Using dual regression to investigate network shape and amplitude in functional connectivity analyses. Front Neurosci. 2017:11:115. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Noble S, Scheinost D, Finn ES, Shen X, Papademetris X, McEwen SC, Bearden CE, Addington J, Goodyear B, Cadenhead KS, et al. Multisite reliability of MR-based functional connectivity. NeuroImage. 2017:146:959–970. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Noble S, Scheinost D, Constable RT. A decade of test-retest reliability of functional connectivity: a systematic review and meta-analysis. NeuroImage. 2019:203:116157–116157. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Noble S, Scheinost D, Constable RT. A guide to the measurement and interpretation of fMRI test-retest reliability. Curr Opin Behav Sci. 2021:40:27–32. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. O'Connor D, Potler NV, Kovacs M, Xu T, Ai L, Pellman J, Vanderwal T, Parra LC, Cohen S, Ghosh S, et al. The healthy brain network serial scanning initiative: a resource for evaluating inter-individual differences and their reliabilities across scan conditions and sessions. GigaScience. 2017:6(2):1–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Pardoe HR, Kucharsky Hiess R, Kuzniecky R. Motion and morphometry in clinical and nonclinical populations. NeuroImage. 2016:135:177–185. [DOI] [PubMed] [Google Scholar]
  73. Patriat R, Molloy EK, Meier TB, Kirk GR, Nair VA, Meyerand ME, Prabhakaran V, Birn RM. The effect of resting condition on resting-state fMRI reliability and consistency: a comparison between resting with eyes open, closed, and fixated. NeuroImage. 2013:79:174–188. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Pedersen R, Geerligs L, Andersson M, Gorbach T, Avelar-Pereira B, Wåhlin A, Rieckmann A, Nyberg L, Salami, A.. When functional blurring becomes deleterious: Reduced system segregation is associated with less white matter integrity and cognitive decline in aging. Neuroimage. 2021:242:118449. [DOI] [PubMed] [Google Scholar]
  75. Pedersen M, Omidvarnia AH, Walz JM, Jackson GD. Increased segregation of brain networks in focal epilepsy: an fMRI graph theory finding. Neuroimage Clin. 2015:8:536–542. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Power JD, Cohen AL, Nelson SM, Wig GS, Barnes KA, Church JA, Vogel AC, Laumann TO, Miezin FM, Schlaggar BL, et al. Functional network organization of the human brain. Neuron. 2011:72(4):665–678. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Power JD, Mitra A, Laumann TO, Snyder AZ, Schlaggar BL, Petersen SE. Methods to detect, characterize, and remove motion artifact in resting state fMRI. NeuroImage. 2014:84:320–341. [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Power JD, Plitt M, Laumann TO, Martin A. Sources and implications of whole-brain fMRI signals in humans. NeuroImage. 2017:146:609–625. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Power JD, Plitt M, Gotts SJ, Kundu P, Voon V, Bandettini PA, Martin A. Ridding fMRI data of motion-related influences: removal of signals with distinct spatial and physical bases in multiecho data. Proc Natl Acad Sci U S A. 2018:115(9):E2105–E2114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Power JD, Lynch CJ, Adeyemo B, Petersen SE. A critical, event-related appraisal of denoising in resting-state fMRI studies. Cereb Cortex. 2020:30(10):5544–5559. [DOI] [PubMed] [Google Scholar]
  81. Rosvall M, Bergstrom CT. Maps of random walks on complex networks reveal community structure. Proc Natl Acad Sci U S A. 2008:105(4):1118–1123. [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Rubinov M, Sporns O. Complex network measures of brain connectivity: uses and interpret. NeuroImage. 2010:52(3):1059–1069. [DOI] [PubMed] [Google Scholar]
  83. Sakoglu U, Pearlson GD, Kiehl KA, Wang YM, Michael AM, Calhoun VD. A method for evaluating dynamic functional network connectivity and task-modulation: application to schizophrenia. MAGMA. 2010:23(5–6):351–366. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Satterthwaite TD, Elliott MA, Gerraty RT, Ruparel K, Loughead J, Calkins ME, Eickhoff SB, Hakonarson H, Gur RC, Gur RE, et al. An improved framework for confound regression and filtering for control of motion artifact in the preprocessing of resting-state functional connectivity data. NeuroImage. 2013:64:240–256. [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Savalia NK, Agres PF, Chan MY, Feczko EJ, Kennedy KM, Wig GS. Motion-related artifacts in structural brain images revealed with independent estimates of in-scanner head motion. Hum Brain Mapp. 2017:38(1):472–492. [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Setton R, Mwilambwe-Tshilobo L, Girn M, Lockrow AW, Baracchini G, Hughes C, Lowe AJ, Cassidy BN, Li J, Luh WM, Bzdok D.. Age differences in the functional architecture of the human brain. Cereb Cortex. 2023:33(1):114–134. [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Shah LM, Cramer JA, Ferguson MA, Birn RM, Anderson JS. Reliability and reproducibility of individual differences in functional connectivity acquired during task and resting state. Brain Behav. 2016:6(5):1–15. [DOI] [PMC free article] [PubMed] [Google Scholar]
  88. Shehzad Z, Kelly AMC, Reiss PT, Gee DG, Gotimer K, Uddin LQ, Lee SH, Margulies DS, Roy AK, Biswal BB, et al. The resting brain: unconstrained yet reliable. Cereb Cortex. 2009:19(10):2209–2229. [DOI] [PMC free article] [PubMed] [Google Scholar]
  89. Shrout PE, Fleiss JL. Intraclass correlations: uses in assessing rater reliability. Psychol Bull. 1979:86(2):420–428. [DOI] [PubMed] [Google Scholar]
  90. Siegel JS, Seitzman BA, Ramsey LE, Ortega M, Gordon EM, Dosenbach NUF, Petersen SE, Shulman GL, Corbetta M. Re-emergence of modular brain networks in stroke recovery. Cortex. 2018:101:44–59. [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Sporns O, Betzel RF. Modular brain networks. Annu Rev Psychol. 2016:67(1):613–640. [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Stumme J , Jockwitz C, Hoffstaedter F, Amunts K, Caspers, S. Functional network reorganization in older adults: Graph-theoretical analyses of age, cognition and sex. NeuroImage. 2020:214:116756. [DOI] [PubMed] [Google Scholar]
  93. Stevens WD, Buckner RL, Schacter DL. Correlated low-frequency BOLD fluctuations in the resting human brain are modulated by recent experience in category-preferential visual regions. Cereb Cortex. 2010:20(8):1997–2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Sun FT, Miller LM, Rao AA, D'Esposito M. Functional connectivity of cortical networks involved in bimanual motor sequence learning. Cereb Cortex. 2007:17(5):1227–1234. [DOI] [PubMed] [Google Scholar]
  95. Tagliazucchi E, Wegner F, Morzelewski A, Brodbeck V, Borisov S, Jahnke K, Laufs H. Large-scale brain functional modularity is reflected in slow electroencephalographic rhythms across the human non-rapid eye movement sleep cycle. NeuroImage. 2013:70:327–339. [DOI] [PubMed] [Google Scholar]
  96. Talairach J, Tournoux P. Co-planar stereotaxic atlas of the human brain. New York: Thieme Medical Publishers, Inc.; 1988. [Google Scholar]
  97. Tambini A, Ketz N, Davachi L. Enhanced brain correlations during rest are related to memory for recent experiences. Neuron. 2010:65(2):280–290. [DOI] [PMC free article] [PubMed] [Google Scholar]
  98. Van Dijk K, Hedden T, Venkataraman A, Evans KC, Lazar SW, Buckner RL. Intrinsic functional connectivity as a tool for Human Connectomics: theory, properties, and optimization. J Neurophysiol. 2010:103(1):297–321. [DOI] [PMC free article] [PubMed] [Google Scholar]
  99. Van Dijk KR, Sabuncu MR, Buckner RL. The influence of head motion on intrinsic functional connectivity MRI. NeuroImage. 2012:59(1):431–438. [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. Van Essen DC, Smith SM, Barch DM, Behrens TE, Yacoub E, Ugurbil K, WU-Minn Consortium . The WU-Minn Human Connectome Project: an overview. NeuroImage. 2013:80:62–79. [DOI] [PMC free article] [PubMed] [Google Scholar]
  101. Van Essen DC, Ugurbil K, Auerbach E, Barch D, Behrens TE, Bucholz R, Chang A, Chen L, Corbetta M, Curtiss SW, et al. The Human Connectome Project: a data acquisition perspective. NeuroImage. 2012a:62(4):2222–2231. [DOI] [PMC free article] [PubMed] [Google Scholar]
  102. Van Essen DC, Lewis JW, Drury HA, Hadjikhani N, Tootell RB, Bakircioglu M, Miller MI. Mapping visual cortex in monkeys and humans using surface-based atlases. Vis Res. 2001:41(10–11):1359–1378. [DOI] [PubMed] [Google Scholar]
  103. Van Essen DC, Glasser MF, Dierker DL, Harwell J, Coalson T. Parcellations and hemispheric asymmetries of human cerebral cortex analyzed on surface-based atlases. Cereb Cortex. 2012b:22(10):2241–2262. [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Varangis E, Habeck CG, Razlighi QR, Stern Y. The effect of aging on resting state connectivity of predefined networks in the brain. Front Aging Neurosci. 2019:11:234. [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Watts DJ, Strogatz SH. Collective dynamics of ‘small-world’ networks. Nature. 1998:393(6684):440–442. [DOI] [PubMed] [Google Scholar]
  106. Wig GS. Segregated systems of human brain networks. Trends Cogn Sci. 2017:21(12):981–996. [DOI] [PubMed] [Google Scholar]
  107. Wig GS, Laumann TO, Petersen SE. An approach for parcellating human cortical areas using resting-state correlations. NeuroImage. 2014:93:276–291. [DOI] [PMC free article] [PubMed] [Google Scholar]
  108. Xu T, Opitz A, Craddock RC, Wright MJ, Zuo XN, Milham MP. Assessing variations in areal organization for the intrinsic brain: from fingerprints to reliability. Cereb Cortex. 2016:26(11):4192–4211. [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Yan CG, Cheung B, Kelly C, Colcombe S, Craddock RC, Di Martino A, Li Q, Zuo XN, Castellanos FX, Milham MP. A comprehensive assessment of regional variation in the impact of head micromovements on functional connectomics. NeuroImage. 2013:76:183–201. [DOI] [PMC free article] [PubMed] [Google Scholar]
  110. Yang GJ, Murray JD, Wang XJ, Glahn DC, Pearlson GD, Repovs G, Krystal JH, Anticevic A. Functional hierarchy underlies preferential connectivity disturbances in schizophrenia. Proc Natl Acad Sci U S A. 2016:113(2):E219–E228. [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Yeo BT, Krienen FM, Sepulcre J, Sabuncu MR, Lashkari D, Hollinshead M, Roffman JL, Smoller JW, Zollei L, Polimeni JR, et al. The organization of the human cerebral cortex estimated by intrinsic functional connectivity. J Neurophysiol. 2011:106(3):1125–1165. [DOI] [PMC free article] [PubMed] [Google Scholar]
  112. Zhang Z, Chan MY, Han L, Carreno CA, Winter-Nelson E, Wig GS, The Alzheimer's Disease Neuroimaging Initiative . Dissociable effects of Alzheimer's disease-related cognitive dysfunction and aging on functional brain network segregation. J Neurosci. 2023:43(46):7879–7892. [DOI] [PMC free article] [PubMed] [Google Scholar]
  113. Zuo XN, Xing XX. Test-retest reliabilities of resting-state FMRI measurements in human brain functional connectomics: a systems neuroscience perspective. Neurosci Biobehav Rev. 2014:45:100–118. [DOI] [PubMed] [Google Scholar]
  114. Zuo XN, Kelly C, Adelstein JS, Klein DF, Castellanos FX, Milham MP. Reliable intrinsic connectivity networks: test-retest evaluation using ICA and dual regression approach. NeuroImage. 2010:49(3):2163–2177. [DOI] [PMC free article] [PubMed] [Google Scholar]
  115. Zuo X-N, Xu T, Milham MP. Harnessing reliability for neuroscience research. Nat Hum Behav. 2019:3(8):768–771. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Han_SegregationReliability_SI_20240115_bhad506

Data Availability Statement

The MSC dataset is available via OpenNeuro data repository (https://openneuro.org/datasets/ds000224). The HCP-YA dataset is available via the Human Connectome Project with ConnectomeDB (https://www.humanconnectome.org). The HCP-A dataset is available on the NIMH Data Archive (https://nda.nih.gov/).


Articles from Cerebral Cortex (New York, NY) are provided here courtesy of Oxford University Press

RESOURCES