Skip to main content
Network Neuroscience logoLink to Network Neuroscience
. 2018 Oct 1;3(1):1–26. doi: 10.1162/netn_a_00054

Graph theory approaches to functional network organization in brain disorders: A critique for a brave new small-world

Michael N Hallquist 1,2, Frank G Hillary 3,4,5,*
PMCID: PMC6326733  PMID: 30793071

Abstract

Over the past two decades, resting-state functional connectivity (RSFC) methods have provided new insights into the network organization of the human brain. Studies of brain disorders such as Alzheimer’s disease or depression have adapted tools from graph theory to characterize differences between healthy and patient populations. Here, we conducted a review of clinical network neuroscience, summarizing methodological details from 106 RSFC studies. Although this approach is prevalent and promising, our review identified four challenges. First, the composition of networks varied remarkably in terms of region parcellation and edge definition, which are fundamental to graph analyses. Second, many studies equated the number of connections across graphs, but this is conceptually problematic in clinical populations and may induce spurious group differences. Third, few graph metrics were reported in common, precluding meta-analyses. Fourth, some studies tested hypotheses at one level of the graph without a clear neurobiological rationale or considering how findings at one level (e.g., global topology) are contextualized by another (e.g., modular structure). Based on these themes, we conducted network simulations to demonstrate the impact of specific methodological decisions on case-control comparisons. Finally, we offer suggestions for promoting convergence across clinical studies in order to facilitate progress in this important field.

Keywords: Graph theory, Brain disorders, Network neuroscience, Proportional thresholding

INTRODUCTION

Efforts to characterize a “human connectome” have brought sweeping changes to functional neuroimaging research, with many investigators transitioning from indices of local brain activity to measures of interregional communication (Friston, 2011). The broad goal of this conceptual revolution is to understand the brain as a functional network whose coordination is responsible for complex behaviors (Biswal et al., 2010). The prevailing approach to studying functional connectomes involves quantifying coupling of the intrinsic brain activity among regions. In particular, resting-state functional connectivity (RSFC) methods (Biswal, Yetkin, Haughton, & Hyde, 1995) focus on interregional correspondence in low-frequency oscillations of the BOLD signal (approximately 0.01–0.12 Hz).

Work over the past two decades has demonstrated the value of RSFC approaches for mapping functional network organization, including the identification of separable brain subnetworks (Biswal et al., 2010; Laird et al., 2009; Power et al., 2011; Smith et al., 2009; van den Heuvel & Hulshoff Pol, 2010). Because RSFC methods do not require the study-specific designs and cognitive burden associated with task-based fMRI studies, RSFC data are simple to acquire and have been used in hundreds of studies of human brain function. Nevertheless, there are numerous methodological challenges, including concerns about the quality of RSFC data (Power et al., 2014) and the effect of data processing on substantive conclusions (Ciric et al., 2017; Hallquist, Hwang, & Luna, 2013; Shirer, Jiang, Price, Ng, & Greicius, 2015).

RSFC studies of brain injury or disease typically examine differences in the functional connectomes of a clinical group (e.g., Parkinson’s disease) compared with a matched control group. There are specific methodological and substantive considerations that apply to RSFC studies of brain disorders. For example, differences in the overall level of functional connectivity between patient and control groups could lead to differences in the number of spurious connections in network analyses, potentially obscuring meaningful group comparisons (van den Heuvel et al., 2017). Likewise, there is increasing concern in the clinical neurosciences that an unacceptably small percentage of findings are replicable (Müller et al., 2017). Such concerns echo the growing emphasis on open, reproducible practices in neuroimaging more generally (Poldrack et al., 2017).

In this paper, we review the current state of graph theory approaches to RSFC in the clinical neurosciences. Based on key themes in this literature, we conducted two network simulations to demonstrate the pitfalls of specific analysis decisions that have particular relevance to case-control studies. Finally, we provide recommendations for best practices to promote comparability across studies.

Our review does not directly address many important methodological issues that are active areas of investigation. For example, detecting and correcting motion-related artifacts remains a central challenge in RSFC studies (Ciric et al., 2017; Dosenbach et al., 2017) that is especially problematic in clinical and developmental samples (Greene, Black, & Schlaggar, 2016; Van Dijk, Sabuncu, & Buckner, 2012). The length of resting-state scans (in terms of time, number of measurements, and number of sessions) also influences the reliability of functional connectivity estimates (Birn et al., 2013; Gonzalez-Castillo, Chen, Nichols, & Bandettini, 2017) and likely has downstream consequences on network analyses (Abrol et al., 2017; Yang et al., 2012). Finally, brain parcellation—defining the number and form of brain regions—is one of the most important influences on the composition of RSFC networks. There are numerous parcellation approaches, including anatomical atlases, functional boundary mapping, and data-driven algorithms based on voxelwise BOLD time series (Goñi et al., 2014; Honnorat et al., 2015). To focus on RSFC graph theory research in the clinical literature, where relevant, we refer readers to more focused treatments of important issues that are beyond the scope of this paper.

A Literature Review of Clinical Network Neuroscience Studies

Graph theory is a branch of discrete mathematics that has been applied in numerous studies of brain networks, both structural and functional. A graph is a collection of objects, called vertices or nodes; the pairwise relationships among nodes are called edges or links (Newman, 2010). Graphs composed of RSFC estimates among regions provide a window into the intrinsic connectivity patterns in the human brain. Figure 1 provides a simple schematic of the most common graph theory constructs and metrics reported in RSFC studies. For more comprehensive reviews of graph theory applications in network neuroscience, see Bullmore and Sporns (2009; 2012), or Fornito, Zalesky, and Bullmore (2016).

Figure 1. .

Figure 1. 

A toy graph and related graph theory terminology. Adapted with permission from Hillary and Grafman (2017).

The goal of our literature search was to obtain a representative cross-section of graph-theoretic RSFC studies spanning neurological and mental disorders. We focused on functional connectivity (FC) as opposed to structural connectivity, where a distinct set of methodological critiques are likely relevant. Also, we reviewed fMRI studies only, excluding electroencephalography (EEG) and magnetoencephalography (MEG). Although there are important advantages of EEG/MEG in some respects (Papanicolaou et al., 2017), we focused on fMRI in part because the vast majority of clinical RSFC studies have used this modality. In addition, there are fMRI-specific considerations for network definition and spatial parcellation in RSFC studies. We conducted two related searches of the PubMed database (http://www.ncbi.nlm.nih.gov/pubmed) to identify articles focusing on graph-theoretic approaches to RSFC in mental and neurological disorders (for details on the queries, see Methods).

These searches were run in April 2016 and resulted in 626 potential papers for review (281 from neurological query, 345 from mental disorder query). Studies were excluded if they were reviews, case studies, animal studies, methodological papers, used electrophysiological methods (e.g., EEG or MEG), reported only structural imaging, or did not focus on brain disorders (e.g., healthy brain functioning, normal aging). After exclusions, these two searches yielded distinct 106 studies included in the review (see Table 1). A full listing of all studies reviewed is provided in the Supporting Information (Hallquist & Hillary, 2018).

Table 1. .

Clinical disorders represented in the review of 106 clinical network neuroscience studies

Clinical phenotype n (frequency)
Alzheimer’s disease/MCI 19 (17.9%)
Epilepsy/seizure disorder 13 (12.3%)
Depression/affective 12 (11.3%)
Schizophrenia 11 (10.4%)
Alcohol/substance abuse 7 (6.6%)
Parkinson’s disease/subcortical 6 (5.7%)
Traumatic brain injury 6 (5.7%)
Anxiety disorders 5 (4.7%)
ADHD 5 (4.7%)
Stroke 4 (2.8%)
Cancer 3 (2.8%)
Multiple sclerosis 2 (1.9%)
Autism spectrum disorder 2 (1.9%)
Disorders of consciousness 2 (1.9%)
Somatization disorder 2 (1.9%)
Dual Diagnosis 2 (1.9%)
Other neurological disorder 3 (2.8%)
Other psychiatric disorder 2 (1.9%)

Note. ADHD = attention deficit hyperactivity disorder; MCI = mild cognitive impairment.

Below, we outline important general themes from the literature review (our major concerns are summarized in Box 1), including the heterogeneity of data analytic approaches across graph theoretical studies. We then turn our focus to two critical issues with important implications for interpreting network analyses in case-control studies: (a) network thresholding (i.e., determining how to define a “connection” from a continuous measure of FC) and (b) matching the hypothesis to the level of inquiry in the graph. For each of these issues, we offer network simulations to illustrate the importance of these issues for case-control comparisons.

graphic file with name netn-03-1-i001.jpg

Creating Comparable Networks in Clinical Samples

Defining nodes in functional brain networks.

Graph theory analyses of RSFC data fundamentally depend on the definition of nodes (i.e., brain regions) and edges (i.e., the quantification of functional connectivity). For network analyses to reveal new insights into clinical phenomena, investigators must choose a parcellation scheme that robustly samples the regions and networks of interest. Our literature review revealed that 76% of studies defined graphs based on comprehensive parcellations (i.e., sampling most or all of the brain), whereas 24% analyzed connectivity in targeted subnetworks (e.g., motor regions only; Table 2a). Furthermore, we found substantial heterogeneity in parcellations, ranging from 10 to 67,632 nodes (Mode = 90; M = 1,129.2; SD = 7,035.9). In fact, whereas 25% of studies had 90 nodes (most of these used the AAL atlas; Tzourio-Mazoyer et al., 2002), the frequency of all other parcellations fell below 5%, resulting in at least 50 distinct parcellations in 106 studies.

Table 2. .

Network construction and edge definition

a. Network construction n (frequency)
Comprehensive region sampling 80 (75.5%)
Targeted region sampling 26 (24.5%)
 
b. Edge definition n (frequency)
Weighted network 48 (45.3%)
Binary network 42 (39.6%)
Both 13 (12.3%)
Unknown/unclear 3 (2.8%)
 
c. Edge FC statistic n (frequency)
Correlation (typically, Pearson’s r) 82 (77.3%)
Partial correlation 11 (10.4%)
Wavelet correlation 6 (5.7%)
Causal modeling (e.g., SEM) 3 (2.8%)
Other/unclear 4 (3.7%)
 
d. Treatment of negative edges n (frequency)
Not reported/unclear 61 (57.5%)
Discard negative values 23 (21.7%)
Absolute value 10 (9.4%)
Maintained/analyzed 9 (8.5%)
Other transformation 3 (2.8%)

Note. FC = functional connectivity; SEM = structural equation modeling.

Although it may seem self-evident, it bears mentioning that several popular parcellation schemes provide a broad, but not complete, sampling of functional brain regions. For example, recent parcellations based on the cortical surface of the brain (e.g., Glasser et al., 2016; Gordon et al., 2016) have provided a new level of detail on functional boundaries in the cortex. Yet if a researcher is interested in cortical-subcortical connectivity, it is crucial that the parcellation be extended to include all relevant regions.

There are advantages and challenges to every parcellation approach (e.g., Honnorat et al., 2015; Power et al., 2011); here, we focus on two specific concerns. First, a goal of most clinical network neuroscience studies is to describe group differences in whole-brain connectivity patterns that are reasonably robust to the graph definition. Thus, investigators may wish to use at least two parcellations in the same dataset to determine if the findings are parcellation dependent. Because of the fundamental difficulty in comparing unequal networks (van Wijk, Stam, & Daffertshofer, 2010), one would not expect identical findings. In particular, global topological features such as efficiency or characteristic path length may vary by parcellation, but other features such as modularity and hub architecture are likely to be more robust. Applying multiple parcellations to the same dataset increases the number of analyses multiplicatively, as well as the need to reconcile inconsistent findings. Cassidy et al. (2018) recently demonstrated that persistent homology, a technique from topological analysis, has the potential to quantify the similarity of individual functional connectomes across different parcellations. We believe that such directions hold promise for supporting reproducible findings in clinical RSFC studies (Craddock, James, Holtzheimer, Hu, & Mayberg, 2012; Roy et al., 2017; Schaefer et al., in press).

Second, differences in parcellation fundamentally limit the ability to compare studies, both descriptively and quantitatively. As noted above, our review revealed substantial heterogeneity in parcellation schemes across studies of brain disorders. Extending our concern about parcellation dependence, such heterogeneity makes it impossible to ascertain whether differences between two studies of the same clinical population are an artifact of the graph definition or a meaningful finding. Moreover, whereas meta-analyses of structural MRI and task-based fMRI studies have become increasingly popular (e.g., Goodkind et al., 2015; Müller et al., 2017), such analyses are not currently possible in graph-theoretic studies in part because of differences in parcellation. To resolve this issue, we encourage scientists to report results for a field-standard parcellation, a point we elaborate in our recommendations below. We also believe there is value in investigators having freedom to test and compare additional parcellations that may highlight specific findings.

Defining edges in functional brain networks.

Parcellation defines the nodes comprising a graph, but an equally important decision is how to define functional connections among nodes (i.e., the edges). The vast majority (77%) of the studies reviewed used bivariate correlation, especially Pearson or Spearman, as the measure of FC. In Critical Issue 1 below, we consider how FC estimates are thresholded in order to defines edges as present or absent in binary graphs.

Quantifying functional connectivity.

The statistical measure of FC has important implications for network density and the interpretation of relationships among brain regions. The prevailing application of marginal association (typically, bivariate correlation) does not separate the (statistically) direct connectivity between two regions from indirect effects attributable to additional regions. By contrast, most conditional association methods (e.g., partial correlation) rely on inverting the covariance matrix among all regions, thereby removing common variance and defining edges based on unique connectivity between two regions (Smith et al., 2011).

At the present time, there is no clear resolution on whether FC should be based on marginal or conditional association (for useful reviews, see Fallani, Richiardi, Chavez, & Achard, 2014; Varoquaux & Craddock, 2013). Nevertheless, we wish to highlight that as the average full correlation increases within a network, partial correlation values, on average, must decrease. Furthermore, partial and full correlations reveal fundamentally different graph topologies in resting-state data (Cassidy et al., 2018) and may yield different conclusions about the same scientific question (e.g., identifying functional hubs; see Liang et al., 2012). In cases of neuropathology where one might expect distinct gains or losses of functional connections, the use of partial correlation should be interpreted based on the relative edge density and mean FC. Of the studies reviewed here, 10% used partial correlation, but virtually no study accounted for possible differences in edge density (see Table 2c).

Negative edges.

Another important aspect of defining edges is how to handle negative FC estimates. Full correlations of RSFC data typically yield an FC distribution in which most edges are positive, but an appreciable fraction are negative. By contrast, partial correlation methods often yield a relative balance between positive and negative FC estimates (e.g., Y. Wang, Kang, Kemmer, & Guo, 2016). There remains little consensus for handling or interpreting negative edge weights in RSFC graph analyses (cf. Murphy & Fox, 2017). In our review, 57% of the studies reported insufficient or no information about how negative edges were handled in graph analyses (Table 2d). Twenty-one percent of studies deleted negative edges prior to analysis, and 9% included the negative weights as positive weights (i.e., using the absolute value of FC). As detailed elsewhere, some graph metrics are either not defined or need to be adapted when negative edges are present (Rubinov & Sporns, 2011).

Importantly, the mean of the marginal association RSFC distribution depends on whether global signal regression (GSR) is included in the preprocessing pipeline (Murphy, Birn, Handwerker, Jones, & Bandettini, 2009). When GSR is included, there is often a balance between positive and negative correlations. If GSR is included as a nuisance regressor, a large fraction of FC estimates may simply be discarded as irrelevant to case-control comparisons, which is a major, untested assumption. The meaning of negative FC, however, remains unclear, with several investigators attributing negative correlations to statistical artifacts and GSR (Murphy et al., 2009; Murphy & Fox, 2017; Saad et al., 2012).

Given that negative correlations are observable in the absence of GSR, however, others have examined whether negative weights contribute differentially to information processing within the network (Parente et al., 2017). Negative correlations may also reflect NMDA action in cortical inhibition (Anticevic et al., 2012). These connections bear consideration given that brain networks composed of only negative connections do not retain a small-world topology, but do have properties distinct from random networks (Parente et al., 2017; Schwarz & McGonigle, 2011). Altogether, the omission of methodological details about negative FC in empirical reports severely hampers the resolution of this important choice point in defining graphs.

Degree distribution as a fundamental graph metric.

After resolving the questions of node and edge definition, we also believe it is crucial for studies to report information about global network metrics such as characteristic path length, clustering coefficient (aka transitivity), and degree distribution. As noted in Table 3, local and global efficiency were commonly reported (71% and 74%, respectively) and typically across multiple FC thresholds. However, our review revealed that only 27% of studies provided clear descriptive statistics for mean degree, and 16% plotted the degree distribution. In binary graphs, the degree distribution describes the relative frequency of edges for each node in the network. A similar property can be quantified by examining the strength (also known as cost) distribution in weighted graphs. We argue that presentation of the degree (strength) distribution in published reports is vital to understanding any RSFC network for a few reasons. First, it provides a sanity check on the data. One current perspective is that human brains are organized to maximize communication while minimizing wiring and metabolic cost (Bassett et al., 2009; Betzel et al., 2017; Chen, Wang, Hilgetag, & Zhou, 2013; Tomasi, Wang, & Volkow, 2013), Thus, when examining whole-brain RSFC data, the most highly connected regions are rare and should be evident in the tail of the degree/strength distribution (Achard, Salvador, Whitcher, Suckling, & Bullmore, 2006). Second, reporting details of the degree distribution facilitates comparisons of graph topology across studies, as well as the impact of preprocessing and analytic decisions such as FC thresholding. Finally, examining the degree distribution as a first step in data analysis may offer otherwise unavailable information about the network topology in healthy and clinical samples. For example, in prior work, we isolated edges from the tail of the degree distribution to understand the impact of the most highly connected, and rare, nodes on the network (Hillary et al., 2014).

Table 3. .

Graph metrics commonly reported in clinical RSFC studies

Graph metrics n (frequency)
Degree distribution (plotted) 17 (16.0%)
Mean degree (weighted or binary) 29 (27.4%)
Clustering/local efficiency 76 (71.7%)
Path length/global efficiency 79 (74.5%)
Small-worldness 33 (31.1%)
Modularity (e.g., Q-value) 21 (19.8%)

Note. Total frequency is greater than 100% because some studies reported more than one of these metrics.

Critical Issue 1: Edge Thresholding and Comparing Unequal Networks

We now focus on edge thresholding—that is, how to transform a continuous measure of FC into an edge in the graph. In our review, 39% of studies binarized FC values such that edges were either present or absent in the graph, whereas 45% of studies retained FC as edge weights (see Table 2b). Regardless of whether investigators analyze binary or weighted networks, there are fundamental challenges to comparing networks that differ either in terms of average degree (k) or the number of nodes (N) (Fornito et al., 2016). In particular, comparing groups on graph metrics such as path length and clustering coefficient can be ambiguous because these metrics have mathematical dependencies on both k and N (van Wijk et al., 2010).

Most brain parcellation approaches define graphs with an equivalent number of nodes (N) in each group. On the other hand, connection density is often a variable of interest in clinical studies where the pathology may alter not only connection strength, but also the number of connections. If N is constant, variation in k between groups constrains the boundaries of local and global efficiency. If two groups differ systematically in edge density, this almost guarantees between-group differences in metrics such as clustering coefficient and path length. Determining where to intervene in this circular problem has great importance in clinical network neuroscience, where hypotheses often focus on the number and strength of network connections.

To address this issue, several investigators have recommended proportional thresholding (PT) in which the edge density is equated across networks (Achard & Bullmore, 2007; Bassett et al., 2009; Power, Fair, Schlaggar, & Petersen, 2010). Furthermore, to reduce the possibility that findings are not specific to the chosen density threshold, 29% of studies have tested for group differences across a range (e.g., 5–25%; Table 4). However, we argue that defining edges based on PT may not be ideal for clinical studies, where there are often regional differences in functional coupling or pathology-induced alterations in the number of functional connections (Hillary & Grafman, 2017). For example, in depression, the dorsomedial prefrontal cortex exhibits enhanced connectivity with default mode, cognitive control, and affective networks (Sheline, Price, Yan, & Mintun, 2010). As demonstrated below, when FC differs in selected regions between groups, PT is vulnerable to identifying spurious differences in nodal statistics (e.g., degree). The concerns expressed here extend from van den Heuvel et al. (2017), who demonstrated that PT increases the likelihood of including spurious connections in the network when groups differ in mean FC.

Table 4. .

Thresholding method for defining edges in graphs

Nature of thresholding n (frequency)
All connections retained 6 (5.7%)
Single by value 25 (23.6%)
Multiple by value 28 (26.4%)
Single by density 5 (4.7%)
Multiple by density 31 (29.2%)
Both by value and density 7 (6.6%)
Other (connections lost) 3 (2.8%)
Unknown/unclear 1 (0.9%)

Note. Value: thresholding determined by FC strength, including statistically corrected and uncorrected values; Single: analyses reported at a single threshold value; Multiple: network examined across multiple thresholds.

Simulation to demonstrate a problem with PT in group comparisons: Whack-a-node.

In the first simulation—“whack-a-node”—we examined the consequences of PT on regional inferences when groups differ in FC for selected regions. Unlike empirical resting-state fMRI data, where the underlying causal processes remain relatively unknown, simulations allow one to test the effect of biologically plausible alterations (e.g., hyperconnectivity of certain nodes) on network analyses. Simulations can also clarify the effects of alternative analytic decisions on substantive conclusions in empirical studies. For the details of our simulation approach, see Methods. Briefly, we used a network bootstrapping approach to simulate resting-state network data for a case-control study with 50 patients and 50 controls. We repeated this simulation 100 times, increasing connectivity for patients in three randomly targeted nodes (hereafter called “Positive”) and decreasing connectivity slightly, but nonsignficantly, in three random nodes (“Negative”). Our simulations captured both between- and within-person variation in FC changes for targeted nodes (see Methods, Equations 57). We compared changes in Positive and Negative nodes to three Comparator nodes that did not differ between groups.

Results of Whack-a-Node Simulation

Proportional thresholding.

In a multilevel regression of group difference t statistics (patient–control) on density threshold, we found that PT was sensitive to hyperconnectivity of Positive nodes, reliably detecting group differences, average t = 12.4 (SD = 1.15), average p < .0001 (Figure 2A). Importantly, however, degree was significantly lower in patients than controls for Negative nodes, average t = −6.52 (SD = 1.15), average p < .001. We did not find any systematic difference between groups in Comparator nodes, average t = −.22 (SD = .65), average p = .47.

Figure 2. .

Figure 2. 

The effect of thresholding method on group differences in degree centrality when there is strong hyperconnectivity in three randomly selected nodes (Positive) and weak hypoconnectivity in three nodes (Negative). Three unaffected nodes are depicted for comparison (Comparator). For parameters of this simulation, see Supporting Information Table S1 (Hallquist & Hillary, 2018), Whack-a-node hyperconnectivity. The central bar of each rectangle denotes the median t statistic (patient–control) across 100 replication datasets (patient n = 50, control n = 50), whereas the upper and lower boundaries denote the 90th and 10th percentiles, respectively. The dark line at t = 0 reflects no mean difference between groups, whereas the light gray lines at t = −1.99 and 1.99 reflect group differences that would be significant at p = .05. (A) Graphs binarized at 5%, 15%, and 25% density. (B) Graphs binarized at r = {.2, .3, .4}. (C) Graphs binarized at r = {.2, .3, .4}, with density included as a between-subjects covariate. (D) Strength centrality (weighted graphs).

These group differences were qualified by a significant density x node type interaction (p < .0001) such that group differences for Positive nodes were larger at higher densities (Figure 3, top panel), B = 11.32 (95% CI = 10.59–12.04), t = 30.72, p < .0001. Conversely, we found equal, but opposite, shifts in Negative nodes (Figure 3, middle panel): group differences became increasingly negative at higher densities, B = −11.93, (95% CI = −12.65–−11.2), t = −32.74, p < .0001. However, we did not observe an association between density and group differences in Comparator nodes, B = .31, p = .40 (Figure 3, bottom panel).

Figure 3. .

Figure 3. 

The effect of density threshold on group differences in degree centrality. Dots denote the mean t statistic at a given density, whereas vertical lines denote the 95% confidence interval. All statistics reflect group differences in degree centrality computed on graphs binarized at different densities.

FC thresholding.

In graphs thresholded at differing levels of FC (rs ranging between .2 and .5), we found reliable increases in Positive nodes in patients, average t = 6.37, average p < .0001 (Figure 2B). Negative nodes, however, were not significantly different between groups, average t = −1.75, average p = .23. Neither did Comparator nodes differ by group, average t = .05, p =.50. Unlike PT, for FC-thresholded graphs, we did not find a significant threshold x node type interaction, x2(16) = 13.38, p = .65.

Functional connectivity thresholding with density as covariate.

By definition, FC thresholding cannot handle the problem of group differences in mean FC. Thus, the risk of FC thresholding alone is that nodal statistics may reflect group differences in mean FC that affect interpretation of topological metrics (e.g., small-worldness). To mitigate this concern, one could threshold at a target FC value, then include graph density for each subject as a covariate (as suggested by van den Heuvel et al., 2017). As depicted in Figure 2C, however, although this statistically controls for density, it also reintroduces the constraint that the groups must be equal in average degree. As a result, the pattern of effects mirrors the PT graphs (Figures 2A and 3). More specifically, group differences in Positive nodes were reliably detected, average t = 12.39, average p < .0001. But patients appeared to be significantly more hypoconnected in Negative nodes compared with controls, average t = −6.59, average p < .001.

Weighted analysis.

In analyses of strength centrality, we observed significantly greater degree in Positive nodes, average t = 5.5, average p = .0003 (Figure 2D). As expected, given our simulation design, Negative and Comparator nodes were not significantly different between groups, average ps = .19 and .49, respectively.

Discussion of whack-a-node simulation.

In our whack-a-node simulation, three nodes were robustly hyperconnected in “patients,” while three nodes were weakly hypoconnected. The principal finding of this simulation was that enforcing equal average degree using PT can spuriously magnify changes in group comparisons of nodal statistics. When the groups were otherwise equal, hyperconnectivity in selected nodes was accurately detected using PT across different graph densities. However, nodes that were weakly hypoconnected tended to be identified as statistically significant. The inclusion of low-magnitude, spurious connections into the network inheres from the way in which PT handles the tails of the FC distribution.

By retaining only edges at the high end of the FC distribution, edges that are on the cusp of that criterion are most vulnerable to being removed. For example, at a density of 25%, small variation in FC strength near the 75th percentile could lead to inclusion or omission of an edge. As a result, if FC for edges incident to a given node tends to be weaker in one group than the other, then binary graphs generated using PT will magnify the statistical significance of differences in degree centrality. To the extent that nodal differences in FC strength represent a shift in the central tendency of the distribution, this problem is not solved by applying multiple density thresholds (Figure 3). We observed the same problem if the direction of FC changes was flipped in the simulation: under PT, group comparisons were significant for weakly hyperconnected nodes if some nodes were robustly hypoconnected (see Supporting Information Figure S1, Hallquist & Hillary, 2018). Moreover, our findings were qualitatively unchanged when simulated changes in FC were applied proportionate to the original connection strength, rather than shifting connectivity in correlational units (see Supporting Information Figure S2, Hallquist & Hillary, 2018). This additional simulation respected biologically plausible variation in FC (e.g., greater average connectivity in functional hubs).

We examined FC-based thresholding (here, using Pearson r as the metric) and weighted analyses as comparisons to PT. These methods do not suffer from the spurious detection of nodal differences evident under PT. Rather, FC thresholding accurately detected hyperconnected nodes across different thresholds while not magnifying significance of the weakly hypoconnected nodes. However, as noted elsewhere (van den Heuvel et al., 2017; van Wijk et al., 2010), if two groups differ in mean FC, thresholding at a given level (e.g., r = .3) in both groups will lead to differences in graph density. This could manifest as widespread differences in nodal statistics due to global differences in the number of edges. We considered whether including per subject graph density as a covariate in group difference analyses could retain the desirable aspects of FC thresholding while accounting for the possibility of global differences in FC. We found, however, that statistically covarying for density was qualitatively similar in its effects to PT because it constrains the sum of degree changes across the network to be zero between groups (i.e., equal average degree).

The goal of our simulation was to provide a proof of concept that PT may negatively affect nodal statistics in case-control graph studies by enforcing equal average degree. We did not, however, test a range of parameters to identify the conditions under which this concern holds true. The simulation focused specifically on degree centrality in the binary case and strength in the weighted case. While untested, we anticipate that these effects likely generalize to other nodal measures such as eigenvector centrality. Importantly, the problems with PT highlighted above occur regardless of edge density (Figure 3), so the use of multiple edge densities does not adequately address the whack-a-node issue.

Critical Issue 2: Matching Theory to Scale: Telescoping Levels of Analysis in Graphs

The second major methodological theme from our literature review concerns the alignment between neurobiological hypotheses and graph analyses. We refer to this as theory-to-scale matching. RSFC graphs offer telescoping levels of information about intrinsic connectivity patterns, from global information such as average path length to details such as connectivity differences in a specific edge. For example, in major depression, resting-state studies have focused analyses on specific subnetworks composed of the dorsal medial prefrontal cortex, anterior cingulate cortex (ACC), amygdala, and medial thalamus (for a review, see L. Wang, Hermens, Hickie, & Lagopoulos, 2012).

We propose that graph analyses should be conceptualized and reported in terms of telescoping levels of analysis from global to specific: global topology, modular structure, nodal effects, edge effects. Our review of the clinical network neuroscience literature revealed that the majority of studies (69%; see Table 5) tested whether groups differed on global metrics such as small-worldness. Importantly, however, many studies provided limited theoretical justification for why the pathophysiology of a given brain disorder should alter the global topology of the network. In the following, we provide more specific comments on path length and small-worldness, and how investigators might ideally match hypotheses to levels of analysis.

Table 5. .

Level of group difference hypotheses in graph analyses (i.e., telescoping)

Hypothesis level n (frequency)
Global 73 (68.9%)
Module (subnetwork) 41 (38.7%)
Nodal (region) 52 (49.1%)
Edge 1 (0.9%)

Note. Global: examining whole-graph network features (e.g., small-worldness); Module: examining subnetworks/modules (e.g., default mode network). Total frequency is greater than 100% because some studies tested hypotheses at multiple levels.

The clinical meaningfulness of small-worldness.

In a defining study for network neuroscience, Watts and Strogatz (1998) demonstrated that the organization of the central nervous system in C. elegans reflected a “small-world” topology (cf. Muldoon, Bridgeford, & Bassett, 2016). The impact of this finding continues to resonate in the network neuroscience literature 20 years later (Hilgetag & Kaiser, 2004; Sporns & Zwi, 2004), with many studies focusing on “disconnection” and the loss of network efficiency as quantified by small-worldness (31% of the studies reviewed here). Although small-world topologies have been observed in most studies of brain function (Bassett & Bullmore, 2016), the relevance of this organization for facilitating human information processing remains unclear. Other features of human neural networks, such as modularity, may have more important implications for network functioning (Hilgetag & Goulas, 2015). Higher network modularity reflects a graph in which the connections among nodes tend to form more densely connected communities, which tend to be robust to random network disruption (Shekhtman, Shai, & Havlin, 2015).

It remains uncertain that, as a general rule, brain pathology should be reflected in a measure of small-worldness. For example, a small-world topology is preserved even in experiments that dramatically reduce sensory processing via anesthesia in primates (Vincent et al., 2007) and in disorders of consciousness in humans (Crone et al., 2014). Recognizing that conventional measures of small-worldness (e.g., Humphries & Gurney, 2008) depend on density and do not handle variation in connection strength, recent research has recast this concept and its quantification in terms of the “small world propensity” of a network (Muldoon et al., 2016).

Although we do not dispute the value of global graph metrics such as small-worldness as important descriptors of a network organization, by definition, they provide information at only the most macroscopic level. Group differences in global metrics may largely reflect more specific effects at finer levels of the graph. For example, removing connections in functional hub regions selectively tends to reduce global efficiency and clustering (Hwang, Hallquist, & Luna, 2013). Likewise, failing to identify group differences in global structure does not imply equivalence at other levels of the graph (e.g., nodes or modules). To demonstrate the point that substantial group differences in finer levels of the graph may not be evident in global metrics, we conducted a simulation in which the groups differed substantially in modular connectivity.

Simulation to demonstrate the importance of understanding graphs at multiple levels: Global insensitivity to modular effects.

Extending the basic approach of our whack-a-node simulation, we used a 13-module parcellation of the 264-node groundtruth graph in a simulation of 50 “controls” and 50 “patients” (modular structure from Power et al., 2011). More specifically, we increased FC in the fronto-parietal network (FPN) and dorsal attention network (DAN) in controls, and increased FC in the default mode network (DMN) in patients. The simulation primarily examined group differences in small-worldness (global metric) and within- and between-module degree (nodal metrics). Additional details are provided in the Methods.

Results of global insensitivity simulation.

Consistent with common methods in the field, we applied PT between 7.5% and 25% to each graph. We computed small-worldness, σ, at each threshold (see Methods), then tested for group differences in this coefficient. As depicted in Figure 4, we did not observe any significant group differences in small-worldness at any edge density, average t = .36, ps > .3.

Figure 4. .

Figure 4. 

Group differences in small-worldness (σ) as a function of edge density. The central bar of each rectangle denotes the median σ statistic (patient–control), whereas the upper and lower boundaries denote the 90th and 10th percentiles, respectively.

However, consistent with the structure of our simulations, we found large group differences in within- and between-module degree (Figure 5). In a multilevel regression of within-network degree on group, density, and module, we found a significant DMN increase in patients irrespective of density, B = 1.37, t = 33.25, p < .0001. Likewise, controls had significantly higher within-network degree in the FPN and DAN, ts = 21.67 and 13.52, respectively, ps < .0001. These findings were mirrored in group analyses of between-network degree in the DMN, FPN, and DAN, ps < .0001 (see Figure 5).

Figure 5. .

Figure 5. 

Group differences in z-scored degree statistics at 10% and 20% density. Degree differences for connections between modules are depicted in the top row, whereas within-module differences are in the bottom row. Dots denote the mean z statistic across nodes, whereas the lines represent the 95% confidence interval around the mean. Nonsignificant group differences in the Visual network are depicted for comparison, whereas connectivity in the DAN, FPN, and DMN was focally manipulated in the simulation.

Discussion of global insensitivity simulation.

In the global insensitivity simulation, we induced large group changes in FC at the level of functional modules that represent canonical resting-state networks (e.g., the DMN). The simulation differentially modulated FC within and between regions of the DAN, FPN, and DMN. In group analyses of within- and between-module degree centrality, we detected these large shifts in FC. However, despite robust differences in network structure, the two groups were very similar in the small-world properties of their graphs.

As with the whack-a-node simulation, we did not seek to test the range of conditions under which these findings would hold. Rather, the global insensitivity simulation provides a proof of concept that researchers should be aware that the absence of group differences at a higher level of the graph (here, global topology) does not suggest that the networks are otherwise similar at lower levels (here, modular connectivity). As we have noted above, in graph analyses of case-control resting-state networks, we encourage researchers to state their study goals in terms that clearly match the hypotheses to the scale of the graph.

In the global insensitivity simulation, failing to detect differences in small-worldness should not be seen as an omnibus test of modular or nodal structure. Likewise, if group differences are detected at the global level, there may be substantial value in interrogating finer differences in the networks, even if these were not hypothesized a priori.

General Summary

The overarching goal of this review was to promote shared standards for reporting findings in clinical network neuroscience. Our survey of the resting-state functional connectivity literature revealed the popularity and promise of graph theory approaches to network organization in brain disorders. This potential is evident in large-scale initiatives for acquiring and sharing resting-state data in different populations (e.g., the Human Connectome Project; Barch et al., 2013). Publicly available resting-state data can also support reproducibility efforts by serving as replication datasets to corroborate specific findings in an independent sample (e.g., Jalbrzikowski et al., 2017). We share the field’s enthusiasm for such work and anticipate that with further methodological refinement and standardization, the coupling of network science and brain imaging can provide novel insights into the neurobiological basis of brain disorders.

Our review, however, suggested that the heterogeneity of methods is preventing the field from realizing its potential. Graph analyses across clinical studies varied substantially in terms of brain parcellation, FC quantification, and the use of thresholding methods to define edges in graphs. These decisions are fundamental to graph theory and precede analyses at specific levels such as global topology. In addition, although it was not a focus of our review, there was substantial variation in what network metrics were reported across studies. The lack of standardization in methods at multiple decision points has multiplicative consequences: The likelihood that any two studies used the same parcellation scheme, FC definition, thresholding strategy, and network metric was remarkably low. This makes formal meta-analyses of the clinical network neuroscience literature virtually impossible at the present time, detracting from efforts to distinguish distinct pathophysiological mechanisms or to identify transdiagnostic commonalities. In addition, methodological heterogeneity in graph analyses undercuts the value of data sharing efforts that have made massive datasets available to the network neuroscience community. For the immense potential of data sharing to be realized, standardization must occur not only in data acquisition, but also in data analysis, with a shared framework to guide hypothesis-to-scale matching in graphs.

As summarized in Box 2, we believe that the field should work toward a principled, common approach to graph analyses of RSFC data. This is a challenging proposition because of the rapid and exciting developments in functional brain parcellation (e.g., Schaefer et al., in press), FC definition (e.g., Cassidy, Rae, & Solo, 2015), edge thresholding (van den Heuvel et al., 2017), and network metrics (e.g., Vargas & Wahl, 2014). Such developments highlight both the enthusiasm for, and relative infancy of, network neuroscience as a field. Although we are sensitive to the importance of continuing to refine functional parcellations of the brain, we also see great value in developing field-standard parcellations to promote comparability. Indeed, many aspects of network structure (e.g., homogeneity of functional connectivity patterns within a region) are largely convergent above a certain level of detail (likely 200–400 nodes) in the functional parcellation (Craddock et al., 2012; Schaefer et al., in press). Likewise, the optimal approach for quantifying functional connectivity is an open question (Smith et al., 2011), yet in the absence of methodological convergence, graphs were often not comparable across studies.

graphic file with name netn-03-1-i002.jpg

A related dilemma was that in 57% of studies, little or no detail was provided about how negative FC values were incorporated into graph analyses. This was especially troubling insofar as global signal regression tends to yield an FC distribution in which approximately half of edges are negative (Murphy et al., 2009). Furthermore, FC estimates at the low end of the distribution may have different topological properties such as reduced modularity (Schwarz & McGonigle, 2011). Even among the 21% of studies in which negative edges were explicitly dropped, it remains unclear what consequences this decision has on substantive conclusions about graph structure. In recent years, there have been advances in quantifying common graph metrics such as modularity in weighted networks that include negative edges (Rubinov & Sporns, 2011), as well as increasing calls for weighted, not binary, graph analyses (Bassett & Bullmore, 2016). Regardless, we believe that greater clarity in reporting of negative FC will promote comparisons among clinical studies.

Methodological heterogeneity also resulted in very few graph statistics that were reported in common across studies, an essential ingredient for examining the reproducibility of findings. Networks that are resilient to the deletion of specific edges often have a highly skewed degree distribution (Callaway, Newman, Strogatz, & Watts, 2000) that may relate to small-world network properties (Achard et al., 2006). We propose that studies should routinely depict this distribution. Likewise, broad metrics such as edge density, mean FC, transitivity, and characteristic path length provide important information about the basic properties of graphs that contextualize more detailed inferential analyses. The challenge of developing reporting standards in clinical network neuroscience echoes the broader conversation in neuroimaging about reproducibility, especially the importance of detail and transparency in the analytic approach (Nichols et al., 2017).

In addition to the general issues of standardizing graph analyses and reporting procedures, our review examined two critical issues in greater detail. First, we considered the potential benefits and risks of using PT, a common procedure for equating the number of edges between graphs. Second, we articulated the value of considering the telescoping levels of graph structures in order to match a hypothesis to the corresponding scale of the graph.

Roughly one-third of the studies included in our review applied PT, thresholding graphs at single or multiple edge densities. Although this approach is aligned with prior work highlighting concerns about comparing unequal networks (van Wijk et al., 2010), its application in clinical studies is often conceptually problematic. Many brain pathologies appear to affect targeted regions or networks, while leaving the connectivity of other regions largely undisturbed. For example, although frontolimbic circuitry is heavily implicated in mood disorders (Price & Drevets, 2010), visual networks largely appear unaffected. There is growing evidence that brain disorders alter the strength of functional coupling and potentially the number of functional connections (Hillary & Grafman, 2017). Consequently, if some regions are affected by the pathology, but others are similar to a matched control population, PT may erroneously remove or add connections to graphs in one group in order to maintain equal average degree between groups. Furthermore, if a brain pathology alters the density of functional connections—for example, neurological disruption is associated with hyperconnectivity (Hillary et al., 2015)—PT will preclude the investigator detecting density differences between groups. If, in truth, the groups differ in edge density, artificially equating density also detracts from the interpretability of graph analyses (cf. van den Heuvel et al., 2017).

In addition to these conceptual problems, our whack-a-node simulation demonstrated that PT may result in the detection of spurious group differences (Figure 2A). Altogether, applying PT in clinical studies may be a methodological double jeopardy, characterized by reduced sensitivity to pathology-related differences in connection density and the risk of identifying nodal differences between groups that are a statistical artifact. These risks make PT especially unappealing when one considers that FC-based thresholding and weighted analyses accurately detected group differences (Figure 2B and 2D) while also allowing edge density to vary. We acknowledge, however, that a previous empirical study found that group differences were more consistent across multiple thresholds using proportional, as opposed to FC-based, thresholding (Garrison, Scheinost, Finn, Shen, & Constable, 2015).

We therefore have two recommendations for edge thresholding in case-control comparisons. First, weighted analyses or FC thresholding should typically be preferred to PT if one is interested in nodal statistics. Second, to rule out the possibility that nodal findings reflect global differences in mean FC, one could include mean FC as a covariate in weighted analyses or per subject density in analyses of FC-thresholded binary graphs. Crucially, we propose that these be treated as sensitivity analyses conducted only after establishing a nodal group difference. That is, if one identifies group differences in FC-thresholded graphs (e.g., greater degree in anterior cingulate cortex among patients), does including edge density as a covariate abolish this finding? If so, it suggests that differences in global topology may account for the nodal finding. However, one should not include density as a covariate in FC-thresholded graphs as a first step to identify which nodes differ between groups, as this could fall prey to the whack-a-node problem (i.e., spurious nodal effects).

The second critical issue was that many studies provided limited theoretical justification for the alignment between a given hypothesis and the corresponding graph analysis. A majority of studies (69%) tested whether groups differed in global metrics such as small-worldness, but most pathologies (e.g., brain injury, Alzheimer’s disease) primarily affect regional hubs within networks (Crossley et al., 2014). Our global insensitivity simulation focused on the importance of matching hypotheses to graph analyses, or telescoping. We demonstrated that global graph metrics, specifically small-worldness, may not be sensitive to group differences in module or node centrality. Metrics such as modularity and nodal centrality offer vital information about regional brain organization that can be interpreted in the context of alterations in average degree. The general point is that one cannot generalize findings from one level of a graph to another, nor should null effects at one level be viewed as suggesting that the groups are similar at other levels. By conceptualizing a data analysis plan in terms of the telescoping structure of graphs, researchers can clearly delineate confirmatory from exploratory analyses, which is consistent with the spirit of reproducible science in neuroimaging (Poldrack et al., 2017).

In summary, there is a need for a common framework to inform graph theory analyses of RSFC data in the clinical neurosciences. Any recommendations should emerge organically from a scientific community whose investigators voluntarily adopt procedures that maximize sensitivity to hypothesized effects and simultaneously permit graph comparisons among studies (cf. the COBIDAS effort in neuroimaging more broadly; Nichols et al., 2017). The implementation of best practices in task-based fMRI (e.g., handling autocorrelation; Woolrich, Ripley, Brady, & Smith, 2001) has been supported by powerful, usable, and free software. Although similar software is increasingly available for resting-state studies (e.g., Chao-Gan & Zang Yu-Feng, 2010; Whitfield-Gabrieli & Nieto-Castanon, 2012), best practices are still emerging, and we believe that software developers will be crucial to promoting a standardized graph theory analysis pipeline. We anticipate that a common methodological framework will promote hypothesis-driven research, alignment between theory and graph analysis, reproducibility, data sharing, meta-analyses, and ultimately more rapid progress of clinical network neuroscience.

METHODS

PubMed Search Syntax

Neurological disorders query.

(graph OR graphical OR graph-theor* OR topology) AND (brain OR fMRI OR connectivity OR intrinsic) AND (resting-state OR resting OR rest) AND (neurological OR brain injury OR multiple sclerosis OR epilepsy OR stroke OR CVA OR aneurysm OR Parkinson’s OR MCI OR Alzheimer’s OR dementia OR HIV OR SCI OR spinal cord OR autism OR ADHD OR intellectual disability OR Down syndrome OR Tourette) AND “humans”[MeSH Terms]

Mental disorders query.

(graph OR graphical OR graph-theor* OR topology) AND (brain OR fmri OR connectivity OR intrinsic) AND (resting-state OR resting OR rest) AND (clinical OR psychopathology OR mental disorder OR psychiatric OR neuropsychiatric OR depression OR mood OR anxiety OR addiction OR psychosis OR bipolar OR borderline OR autism) AND “humans”[MeSH Terms]

General Approach to Network Simulations

To approximate the structure of functional brain networks, we identified a young adult female subject who completed a 5-min resting-state scan in a Siemens 3T Trio Scanner (TR = 1.5 s, TE = 29 ms, 3.1 × 3.1 × 4.0 mm voxels) with essentially no head movement (mean framewise displacement [FD] = .08 mm; max FD = .17 mm). We preprocessed the data using a conventional pipeline, including (a) motion correction (FSL mcflirt); (b) slice timing correction (FSL slicetimer); (c) nonlinear deformation to the MNI template using the concatenation of functional → structural (FSL flirt) and structural → MNI152 (FSL flirt + fnirt) transformations; (d) spatial smoothing with a 6-mm full width at half maximum filter (FSL susan); and (e) and voxelwise intensity normalization to a mean of 100. After these steps, we also simultaneously applied nuisance regression and band-pass filtering, where the regressors were six motion parameters, average cerebrospinal fluid, average white matter, and the derivatives of these (16 total regressors). The spectral filter retained fluctuations between .009 and .08 Hz (AFNI 3dBandpass). We then estimated FC of 264 functional regions of interest (ROIs) from the Power et al. (2011) parcellation, where each region was defined by a 5-mm radius sphere centered on a specific coordinate. The activity of a region over time was estimated by the first principal component of voxels in each ROI, and the functional connectivity matrix (264 × 264) was estimated by the pairwise Pearson correlations of time series.

This 264 × 264 adjacency matrix, W, served as the groundtruth for all simulations, with specific within-person, between-person, and between-group alterations applied according to a multilevel simulation of variation across individuals (details of simulation parameters provided in Supporting Information Table S1, Hallquist & Hillary, 2018). More specifically, to approximate population-level variability in a case-control design, we simulated resting-state adjacency matrices for 50 “patients” and 50 “controls” by introducing systematic and unsystematic sources of variability for each simulated participant. Systematic sources were intended to test substantive hypotheses about proportional thresholding (whack-a-node simulation) and insensitivity to global versus modular differences (global insensitivity simulation), whereas unsystematic sources reflected within- and between-person variation in edge strength.

In this model, the simulated edge strength between two nodes, i and j, for a given subject, s, is:

rijs=wij+gijs+aijs+eijs (1)

where wij is the edge strength from the groundtruth adjacency matrix W. Global variation in mean FC is represented by gijs, which reflects contributions of both between- and within-person variation:

gijs=bs+uijs (2)

where bs represents normally distributed between-person variation in mean FC:

bN(0,σb) (3)

while σb controls the level of between-person variation in the sample. The term uijs represents within-person variation of this edge relative to the person mean FC, bs. Within-person variation in FC across all edges is assumed to be normally distributed:

u..sN(0,σw) (4)

with σw scaling the degree of within-person FC variation across all edges.

Node-specific shifts in FC are represented by αijs, which includes both between-person and within-node components. More specifically, the modulation of FC between nodes i and j is given by:

αijs=ais+vijs (5)

where between-person variation in FC for node i is:

ai.N(μai,σai) (6)

with μai and σai capturing the mean and standard deviation in FC shifts for node i across subjects, respectively. Edgewise FC variation of a node i across its neighbors, j, is given by:

vi.sN(0,σvi) (7)

where σvi represents the standard deviation of FC shifts across neighbors of i. When nodes i and j were both manipulated, the shifts were applied sequentially such that FC for the edge between i and j was not allowed to have compounding changes. That is, we set αijs = 0 for i > j.

Finally, eijs represents the random variation in FC for the edge between i and j for subject s. This variation was assumed to be normally distributed across all edges for a subject:

e..sN(0,σe) (8)

where σe controls the standard deviation of edge noise across subjects.

Whack-a-Note Simulation Methods

As mentioned above, PT is often applied in case-control studies to rule out the possibility that network differences between groups reflect differences in the total number of edges. Importantly, in graphs of equal order N (i.e., the same number of nodes), proportional thresholding equates both the density, D, and average degree, 〈k〉 between subjects:

D=2EN(N1),k=2ENk=D(N1) (9)

where E denotes the number of unique edges in an undirected graph with no self-loops. As noted by van den Heuvel et al. (2017), when one applies PT, differences in average connectivity strength can lead to the inclusion of weaker edges in more sparsely connected groups. Furthermore, weak edges estimated by correlation are more likely to reflect an unreliable relationship between nodes. Thus, if one group has lower mean functional connectivity, PT could introduce spurious connections, potentially undermining group comparisons of network topology.

That is, when groups are otherwise equivalent, the sum of increases in degree in hyperconnected nodes for a group must be offset by equal, but opposite, decreases in degree for other nodes in that group. This phenomenon holds because of the mathematical relationship between average degree and graph density (Equation 9). For simplicity, our first simulation represents the scenario where there are meaningful FC increases in patients for selected nodes and unreliable FC decreases in other nodes. This unreliability is intended to represent sampling variability that could lead to erroneous false positives.

Structure of whack-a-node simulation.

As described above, we simulated 50 patients and 50 controls based on a groundtruth FC matrix. We estimated 100 replication samples with equal levels of noise in both groups (see Supporting Information Table S1, Hallquist & Hillary, 2018). In each replication sample, we increased the mean FC in three nodes, selected at random, by r = 0.14. We also applied smaller decreases of r = −.04 to three other randomly targeted nodes1. This level of decrease was chosen such that group differences in analyses of nodal strength (i.e., computed on weighted graphs) were nonsignificant on average (Mp = .19, SDp = .02). We applied PT to binarize graphs in each group, varying density between 5% and 25% in 1% increments. Likewise, for FC thresholding, we binarized graphs at r threshold between r = .2 and .5 in .02 increments. Finally, we retained weighted graphs for all simulated samples to estimate group differences in nodal strength. To ensure that effects were not attributable to particular nodes, we averaged group statistics across the 100 replication samples, where the targeted nodes varied randomly across samples.

In each replication sample, we estimated degree centrality for the Positive (hyperconnected), Negative (weakly hypoconnected), and Comparator nodes. Comparators were three randomly selected nodes in each sample that were not specifically modulated by the simulation. These served as a benchmark to ensure that simulations did not induce group differences in centrality for nodes not specifically targeted. To quantify the effect of PT versus FC thresholding in binary graphs, we estimated group differences (patient–control) on degree centrality using two-sample t tests for each Positive, Negative, and Comparator node. For weighted analyses, we estimated group differences in strength centrality.

Global Insensitivity Simulation Methods: Hypothesis-to-Scale Matching

We simulated a dataset of 50 “controls” and 50 “patients” in which the FPN and DAN were modulated in controls compared with the groundtruth matrix, W. In this simulation, the patient group had increased FC in the DMN, a common finding in neurological disorders (Hillary et al., 2015). In controls, we increased FC strength for edges between FPN/DAN regions and other networks, Mr = 0.2, SD = 0.1. Per module variation in between-network FC changes was assumed to be normally distributed within each subject, SD = 0.1. We also increased controls’ FC on edges within the FPN and DAN, Mr = 0.1, between-subjects SD = 0.05, within-subjects SD = .05. In patients, between-network FC for DMN nodes was increased, Mr = 0.2, between-subjects SD = 0.1, within-subjects SD = 0.1. Likewise, within-network FC in the DMN was increased, Mr = 0.1, between-subjects SD = .05, within-subjects SD = .05. That is, we applied similar levels of FC modulation to the FPN/DAN in controls and the DMN in patients, although these changes largely affected different edges in the networks between groups.

We computed the small-worldness coefficient, s, according to the approach of Humphries and Gurney (2008):

σ=Cg/CrandLg/Lrand

Here, Cg represents the transitivity of the graph, whereas Crand denotes the transitivity of a random graph with an equivalent degree distribution. Likewise, Lg and Lrand represent the characteristic path length of the target graph and randomly rewired graph, respectively. Values of σ much larger than 1.0 correspond to a network with small-world properties. To generate statistics for equivalent random graphs, we applied a rewiring algorithm that retained the degree distribution of the graph while permuting 347,160 edges (10 permutations per edge, on average). This algorithm was applied to the target graph 100 times to generate a set of equivalent random networks. Transitivity and characteristic path length were calculated for each of these, and their averages were used in computing the small-worldness coefficient, σ. We also analyzed within- and between-network degree centrality for each node, z-scoring values within each module and density to allow for comparisons.

ACKNOWLEDGMENTS

We thank Zach Ceneviva, Allen Csuk, Richard Garcia, Melanie Glatz, and Riddhi Patel for their work collecting, organizing, and coding references for the literature review and manuscript.

AUTHOR CONTRIBUTIONS

Michael Hallquist: Frank G Hillary: Conceptualization; Data curation; Formal analysis; Project administration; Supervision; Writing – original draft; Writing – review & editing.

FUNDING INFORMATION

Michael Hallquist, National Institute of Mental Health (http://dx.doi.org/10.13039/100000025), Award ID: K01 MH097091. Frank G Hillary, National Center for Advancing Translational Sciences (http://dx.doi.org/10.13039/100006108), Award ID: UL Tr000127.

Note

1

Results are qualitatively similar using other values for group shifts in FC.

Supplementary Material

TECHNICAL TERMS

Graph theory:

A subdomain of discrete mathematics focused on the study of graphs, mathematical structures used to model pairwise interactions (edges/links) among objects (vertices/nodes). Graphs are ubiquitous models of both natural and artificial systems, and have been used extensively in the physical, biological, and social sciences.

Parcellation (brain):

Investigator decision regarding segregation of the acquired brain space into distinct functional units (nodes), representing a pivotal firststep in creating a brain network.

Proportional thresholding:

A constraint imposed on the network so that the number of edges for all subjects is maintained constant irrespective of the relative network strength.

Functional connectivity thresholding:

The process of determining the meaningfulness of any edge based on a continuous measure of FC (e.g., Pearson r value).

Whack-a-node:

A simulation used within this report to test the consequences of proportional thresholding given local changes in connectivity (e.g., enhanced or diminished connectivity).

Hyperconnectivity:

Common network response observed in clinical samples when the brain structure has been altered, resulting in enhanced functional response locally.

Telescoping:

Graphs provide the opportunity to examine distinct dimensions (e.g., local vs. global patterns of connectivity), and telescoping refers to movement between these dimensions to understand distinct patterns within the network.

Contributor Information

Michael N. Hallquist, Department of Psychology, Pennsylvania State University, University Park, PA, USA Social Life and Engineering Sciences Imaging Center, University Park, PA, USA.

Frank G. Hillary, Department of Psychology, Pennsylvania State University, University Park, PA, USA; Social Life and Engineering Sciences Imaging Center, University Park, PA, USA; Department of Neurology, Hershey Medical Center, Hershey, PA, USA.

REFERENCES

  1. Abrol A., Damaraju E., Miller R. L., Stephen J. M., Claus E. D., Mayer A. R., & Calhoun V. D. (2017). Replicability of time-varying connectivity patterns in large resting state fMRI samples. NeuroImage, 163, 160–176. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Achard S., & Bullmore E. (2007). Efficiency and cost of economical brain functional networks. PLoS Computational Biology, 3(2), 0174–0183. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Achard S., Salvador R., Whitcher B., Suckling J., & Bullmore E. (2006). A resilient, low-frequency, small-world human brain functional network with highly connected association cortical hubs. The Journal of Neuroscience: The Official Journal of the Society for Neuroscience, 26(1), 63–72. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Anticevic A., Cole M. W., Murray J. D., Corlett P. R., Wang X.-J., & Krystal J. H. (2012). The role of default network deactivation in cognition and disease. Trends in Cognitive Sciences, 16(12), 584–592. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Barch D. M., Burgess G. C., Harms M. P., Petersen S. E., Schlaggar B. L., Corbetta M., … WU-Minn HCP Consortium. (2013). Function in the human connectome: Task-fMRI and individual differences in behavior. NeuroImage, 80, 169–189. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Bassett D. S., & Bullmore E. T. (2016). Small-world brain networks revisited. The Neuroscientist. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bassett D. S., Bullmore E. T., Meyer-Lindenberg A., Apud J. A., Weinberger D. R., & Coppola R. (2009). Cognitive fitness of cost-efficient brain functional networks. Proceedings of the National Academy of Sciences of the United States of America, 106(28), 11747–11752. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Betzel R. F., Medaglia J. D., Papadopoulos L., Baum G. L., Gur R., Gur R., … Bassett D. S. (2017). The modular organization of human anatomical brain networks: Accounting for the cost of wiring. Network Neuroscience, 1(1), 42–68. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Birn R. M., Molloy E. K., Patriat R., Parker T., Meier T. B., Kirk G. R., … Prabhakaran V. (2013). The effect of scan length on the reliability of resting-state fMRI connectivity estimates. NeuroImage, 83, 550–558. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Biswal B. B., Mennes M., Zuo X.-N., Gohel S., Kelly C., Smith S. M., … Colcombe S. (2010). Toward discovery science of human brain function. Proceedings of the National Academy of Sciences of the United States of America, 107(10), 4734–4739. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Biswal B. B., Yetkin F. Z., Haughton V. M., & Hyde J. S. (1995). Functional connectivity in the motor cortex of resting human brain using echo-planar MRI. Magnetic Resonance in Medicine, 34(4), 537–541. [DOI] [PubMed] [Google Scholar]
  12. Bullmore E., & Sporns O. (2009). Complex brain networks: Graph theoretical analysis of structural and functional systems. Nature Reviews Neuroscience, 10(3), 186–198. [DOI] [PubMed] [Google Scholar]
  13. Bullmore E., & Sporns O. (2012). The economy of brain network organization. Nature Reviews Neuroscience, 13(5), 336–349. [DOI] [PubMed] [Google Scholar]
  14. Callaway D. S., Newman M. E., Strogatz S. H., & Watts D. J. (2000). Network robustness and fragility: Percolation on random graphs. Physical Review Letters, 85(25), 5468–5471. [DOI] [PubMed] [Google Scholar]
  15. Cassidy B., Bowman D. B., Rae C., & Solo V. (2018). On the reliability of individual brain activity networks. IEEE Transactions on Medical Imaging, 37(2), 649–662. [DOI] [PubMed] [Google Scholar]
  16. Cassidy B., Rae C., & Solo V. (2015). Brain activity: Connectivity, sparsity, and mutual information. IEEE Transactions on Medical Imaging, 34(4), 846–860. [DOI] [PubMed] [Google Scholar]
  17. Chao-Gan Y., & Zang Yu-Feng (2010). DPARSF: A MATLAB toolbox for “pipeline” data analysis of resting-state fMRI. Frontiers in Systems Neuroscience, 4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Chen Y., Wang S., Hilgetag C. C., & Zhou C. (2013). Trade-off between multiple constraints enables simultaneous formation of modules and hubs in neural systems. PLoS Computational Biology, 9(3), e1002937 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Ciric R., Wolf D. H., Power J. D., Roalf D. R., Baum G. L., Ruparel K., … Satterthwaite T. D. (2017). Benchmarking of participant-level confound regression strategies for the control of motion artifact in studies of functional connectivity. NeuroImage, 154(Supplement C), 174–187. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Craddock R. C., James G. A., Holtzheimer P. E., Hu X. P., & Mayberg H. S. (2012). A whole brain fMRI atlas generated via spatially constrained spectral clustering. Human Brain Mapping, 33(8), 1914–1928. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Crone J. S., Soddu A., Höller Y., Vanhaudenhuyse A., Schurz M., Bergmann J., … Kronbichler M. (2014). Altered network properties of the fronto-parietal network and the thalamus in impaired consciousness. NeuroImage. Clinical, 4, 240–248. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Crossley N. A., Mechelli A., Scott J., Carletti F., Fox P. T., McGuire P., & Bullmore E. T. (2014). The hubs of the human connectome are generally implicated in the anatomy of brain disorders. Brain, 137(8), 2382–2395. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Demetriou L., Kowalczyk O. S., Tyson G., Bello T., Newbould R. D., & Wall M. B. (2016). A comprehensive evaluation of multiband-accelerated sequences and their effects on statistical outcome measures in fMRI. BioRxiv, 076307 [DOI] [PubMed] [Google Scholar]
  24. Dosenbach N. U. F., Koller J. M., Earl E. A., Miranda-Dominguez O., Klein R. L., Van A. N., … Nguyen A. L. (2017). Real-time motion analytics during brain MRI improve data quality and reduce costs. NeuroImage, 161, 80–93. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Fallani F. D. V., Richiardi J., Chavez M., & Achard S. (2014). Graph analysis of functional brain networks: Practical issues in translational neuroscience. Philosophical Transactions of the Royal Society B: Biological Sciences, 369(1653), 20130521–20130521. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Fornito A., Zalesky A., & Bullmore E. (2016). Fundamentals of Brain Network Analysis. Amsterdam: Academic Press. [Google Scholar]
  27. Friston K. J. (2011). Functional and effective connectivity: A review. Brain Connectivity, 1(1), 13–36. [DOI] [PubMed] [Google Scholar]
  28. Garrison K. A., Scheinost D., Finn E. S., Shen X., & Constable R. T. (2015). The (in)stability of functional brain network measures across thresholds. NeuroImage, 118, 651–661. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Glasser M. F., Coalson T. S., Robinson E. C., Hacker C. D., Harwell J., Yacoub E., … Van Essen D. C. (2016). A multi-modal parcellation of human cerebral cortex. Nature, 536(7615), 171–178. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Goñi J., van den Heuvel M. P., Avena-Koenigsberger A., de Mendizabal N. V., Betzel R. F., Griffa A., … Sporns O. (2014). Resting-brain functional connectivity predicted by analytic measures of network communication. Proceedings of the National Academy of Sciences of the United States of America, 111(2), 833–838. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Gonzalez-Castillo J., Chen G., Nichols T. E., & Bandettini P. A. (2017). Variance decomposition for single-subject task-based fMRI activity estimates across many sessions. NeuroImage, 154, 206–218. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Goodkind M., Eickhoff S. B., Oathes D. J., Jiang Y., Chang A., Jones-Hagata L. B., … Etkin A. (2015). Identification of a common neurobiological substrate for mental illness. JAMA Psychiatry, 72(4), 305–315. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Gordon E. M., Laumann T. O., Adeyemo B., Huckins J. F., Kelley W. M., & Petersen S. E. (2016). Generation and evaluation of a cortical area parcellation from resting-state correlations. Cerebral Cortex, 26(1), 288–303. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Greene D. J., Black K. J., & Schlaggar B. L. (2016). Considerations for MRI study design and implementation in pediatric and clinical populations. Developmental Cognitive Neuroscience, 18, 101–112. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Hallquist M. N., & Hillary F. G. (2018). Supporting information for “Graph theory approaches to functional network organization in brain disorders: A critique for a brave new small-world.” Network Neuroscience, 3(1), 1–26. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Hallquist M. N., Hwang K., & Luna B. (2013). The nuisance of nuisance regression: Spectral misspecification in a common approach to resting-state fMRI preprocessing reintroduces noise and obscures functional connectivity. NeuroImage, 82, 208–225. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Hilgetag C. C., & Goulas A. (2015). Is the brain really a small-world network? Brain Structure and Function, 221(4), 1–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Hilgetag C. C., & Kaiser M. (2004). Clustered organization of cortical connectivity. Neuroinformatics, 2(3), 353–360. [DOI] [PubMed] [Google Scholar]
  39. Hillary F. G., & Grafman J. H. (2017). Injured brains and adaptive networks: The benefits and costs of hyperconnectivity. Trends in Cognitive Sciences. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Hillary F. G., Rajtmajer S. M., Roman C. A., Medaglia J. D., Slocomb-Dluzen J. E., Calhoun V. D., … Wylie G. R. (2014). The rich get richer: Brain injury elicits hyperconnectivity in core subnetworks. PLoS ONE, 9(8), e104021–e104021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Hillary F. G., Roman C. A., Venkatesan U., Rajtmajer S. M., Bajo R., & Castellanos N. D. (2015). Hyperconnectivity is a fundamental response to neurological disruption. Neuropsychology, 29(1), 59–75. [DOI] [PubMed] [Google Scholar]
  42. Honnorat N., Eavani H., Satterthwaite T. D., Gur R. E., Gur R. C., & Davatzikos C. (2015). GraSP: Geodesic graph-based segmentation with shape priors for the functional parcellation of the cortex. NeuroImage, 106, 207–221. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Humphries M. D., & Gurney K. (2008). Network ‘small-world-ness’: A quantitative method for determining canonical network equivalence. PLoS ONE, 3(4), e0002051–e0002051. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Hwang K., Hallquist M. N., & Luna B. (2013). The development of hub architecture in the human functional brain network. Cerebral Cortex, 23(10), 2380–2393. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Jalbrzikowski M., Larsen B., Hallquist M. N., Foran W., Calabro F., & Luna B. (2017). Development of white matter microstructure and intrinsic functional connectivity between the amygdala and ventromedial prefrontal cortex: Associations with anxiety and depression. Biological Psychiatry, 82(7), 511–521. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Laird A. R., Eickhoff S. B., Li K., Robin D. A., Glahn D. C., & Fox P. T. (2009). Investigating the functional heterogeneity of the default mode network using coordinate-based meta-analytic modeling. Journal of Neuroscience, 29(46), 14496–14505. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Liang X., Wang J., Yan C., Shu N., Xu K., Gong G., & He Y. (2012). Effects of different correlation metrics and preprocessing factors on small-world brain functional networks: A resting-state functional MRI study. PLoS ONE, 7(3), e32766 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Muldoon S. F., Bridgeford E. W., & Bassett D. S. (2016). Small-world propensity and weighted brain networks. Scientific Reports, 6, 22057–22057. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Müller V. I., Cieslik E. C., Serbanescu I., Laird A. R., Fox P. T., & Eickhoff S. B. (2017). Altered brain activity in unipolar depression revisited: Meta-analyses of neuroimaging studies. JAMA Psychiatry, 74(1), 47–55. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Murphy K., Birn R. M., Handwerker D. A., Jones T. B., & Bandettini P. A. (2009). The impact of global signal regression on resting state correlations: Are anti-correlated networks introduced? NeuroImage, 44(3), 893–905. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Murphy K., & Fox M. D. (2017). Towards a consensus regarding global signal regression for resting state functional connectivity mri. NeuroImage, 154(Supplement C), 169–173. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Newman M. (2010). Networks: An Introduction (1st ed.). New York: Oxford University Press. [Google Scholar]
  53. Nichols T. E., Das S., Eickhoff S. B., Evans A. C., Glatard T., Hanke M., … Yeo B. T. T. (2017). Best practices in data analysis and sharing in neuroimaging using MRI. Nature Neuroscience, 20(3), 299–303. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Papanicolaou A. C., Rezaie R., Narayana S., Choudhri A. F., Wheless J. W., Castillo E. M., … Boop F. A. (2017). Clinical applications of functional neuroimaging: Presurgical functional mapping. In The Oxford Handbook of Functional Brain Imaging in Neuropsychology and Cognitive Neurosciences (p. 371). New York: Oxford University Press. [Google Scholar]
  55. Parente F., Frascarelli M., Mirigliani A., Di Fabio F., Biondi M., & Colosimo A. (2017). Negative functional brain networks. Brain Imaging and Behavior. [DOI] [PubMed] [Google Scholar]
  56. Poldrack R. A., Baker C. I., Durnez J., Gorgolewski K. J., Matthews P. M., Munafò M. R., … Yarkoni T. (2017). Scanning the horizon: Towards transparent and reproducible neuroimaging research. Nature Reviews Neuroscience, 18(2), 115–126. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Power J. D., Cohen A. L., Nelson S. M., Wig G. S., Barnes K. A., Church J. A., … Petersen S. E. (2011). Functional network organization of the human brain. Neuron, 72(4), 665–678. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Power J. D., Fair D. A., Schlaggar B. L., & Petersen S. E. (2010). The development of human functional brain networks. Neuron, 67(5), 735–748. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Power J. D., Mitra A., Laumann T. O., Snyder A. Z., Schlaggar B. L., & Petersen S. E. (2014). Methods to detect, characterize, and remove motion artifact in resting state fMRI. NeuroImage, 84, 320–341. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Price J. L., & Drevets W. C. (2010). Neurocircuitry of mood disorders. Neuropsychopharmacology, 35(1), 192–216. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Roy A., Bernier R. A., Wang J., Benson M., French J. J. Jr, Good D. C., & Hillary F. G. (2017). The evolution of cost-efficiency in neural networks during recovery from traumatic brain injury. PloS One, 12(4), e0170541–e0170541. [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Rubinov M., & Sporns O. (2011). Weight-conserving characterization of complex functional brain networks. NeuroImage, 56(4), 2068–2079. [DOI] [PubMed] [Google Scholar]
  63. Saad Z. S., Gotts S. J., Murphy K., Chen G., Jo H. J., Martin A., & Cox R. W. (2012). Trouble at rest: How correlation patterns and group differences become distorted after global signal regression. Brain Connectivity, 2(1), 25–32. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Schaefer A., Kong R., Gordon E. M., Laumann T. O., Zuo X.-N., Holmes A. J., … Yeo B. T. T. (in press). Local-global parcellation of the human cerebral cortex from intrinsic functional connectivity MRI. Cerebral Cortex. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Schönbrodt F. D., & Perugini M. (2013). At what sample size do correlations stabilize? Journal of Research in Personality, 47(5), 609–612. [Google Scholar]
  66. Schwarz A. J., & McGonigle J. (2011). Negative edges and soft thresholding in complex network analysis of resting state functional connectivity data. NeuroImage, 55(3), 1132–1146. [DOI] [PubMed] [Google Scholar]
  67. Shekhtman L. M., Shai S., & Havlin S. (2015). Resilience of networks formed of interdependent modular networks. New Journal of Physics, 17(12), 123007 [Google Scholar]
  68. Sheline Y. I., Price J. L., Yan Z., & Mintun M. A. (2010). Resting-state functional MRI in depression unmasks increased connectivity between networks via the dorsal nexus. Proceedings of the National Academy of Sciences of the United States of America, 107(24), 11020–11025. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Shirer W. R., Jiang H., Price C. M., Ng B., & Greicius M. D. (2015). Optimization of rs-fMRI pre-processing for enhanced signal-noise separation, test-retest reliability, and group discrimination. NeuroImage, 117, 67–79. [DOI] [PubMed] [Google Scholar]
  70. Smith S. M., Fox P. T., Miller K. L., Glahn D. C., Fox P. M., Mackay C. E., … Beckmann C. F. (2009). Correspondence of the brain’s functional architecture during activation and rest. Proceedings of the National Academy of Sciences of the United States of America, 106(31), 13040–13045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Smith S. M., Miller K. L., Salimi-Khorshidi G., Webster M., Beckmann C. F., Nichols T. E., … Woolrich M. W. (2011). Network modelling methods for FMRI. NeuroImage, 54(2), 875–891. [DOI] [PubMed] [Google Scholar]
  72. Sporns O., & Zwi J. D. (2004). The small world of the cerebral cortex. Neuroinformatics, 2(2), 145–162. [DOI] [PubMed] [Google Scholar]
  73. Tomasi D., Wang G. J., & Volkow N. D. (2013). Energetic cost of brain functional connectivity. Proceedings of the National Academy of Sciences of the United States of America, 110(33), 13642–13647. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Tzourio-Mazoyer N., Landeau B., Papathanassiou D., Crivello F., Etard O., Delcroix N., … Joliot M. (2002). Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. NeuroImage, 15(1), 273–289. [DOI] [PubMed] [Google Scholar]
  75. van den Heuvel M. P., de Lange S., Zalesky A., Seguin C., Yeo T., & Schmidt R. (2017). Proportional thresholding in resting-state fMRI functional connectivity networks and consequences for patient-control connectome studies: Issues and recommendations. NeuroImage, 152, 437–449. [DOI] [PubMed] [Google Scholar]
  76. van den Heuvel M. P., & Hulshoff Pol H. E. (2010). Exploring the brain network: A review on resting-state fMRI functional connectivity. European Neuropsychopharmacology: The Journal of the European College of Neuropsychopharmacology, 20(8), 519–534. [DOI] [PubMed] [Google Scholar]
  77. van den Heuvel M. P., Stam C. J., Boersma M., & Hulshoff Pol H. E. (2008). Small-world and scale-free organization of voxel-based resting-state functional connectivity in the human brain. NeuroImage, 43(3), 528–539. [DOI] [PubMed] [Google Scholar]
  78. Van Dijk K. R. A., Sabuncu M. R., & Buckner R. L. (2012). The influence of head motion on intrinsic functional connectivity MRI. NeuroImage, 59(1), 431–438. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. van Wijk B. C. M., Stam C. J., & Daffertshofer A. (2010). Comparing brain networks of different size and connectivity density using graph theory. PLoS ONE, 5(10), e13701–e13701. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Vargas E. R., & Wahl L. M. (2014). The gateway coefficient: A novel metric for identifying critical connections in modular networks. The European Physical Journal B, 87(7), 161 [Google Scholar]
  81. Varoquaux G., & Craddock R. C. (2013). Learning and comparing functional connectomes across subjects. NeuroImage, 80, 405–415. [DOI] [PubMed] [Google Scholar]
  82. Vincent J. L., Patel G. H., Fox M. D., Snyder A. Z., Baker J. T., Essen D. C. V., … Raichle M. E. (2007). Intrinsic functional architecture in the anaesthetized monkey brain. Nature, 447(7140), 83 [DOI] [PubMed] [Google Scholar]
  83. Wang L., Hermens D. F., Hickie I. B., & Lagopoulos J. (2012). A systematic review of resting-state functional-MRI studies in major depression. Journal of Affective Disorders, 142(1), 6–12. [DOI] [PubMed] [Google Scholar]
  84. Wang Y., Kang J., Kemmer P. B., & Guo Y. (2016). An efficient and reliable statistical method for estimating functional connectivity in large scale brain networks using partial correlation. Frontiers in Neuroscience, 10 [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Watts D. J., & Strogatz S. H. (1998). Collective dynamics of “small-world” networks. Nature, 393(6684), 440–442. [DOI] [PubMed] [Google Scholar]
  86. Whitfield-Gabrieli S., & Nieto-Castanon A. (2012). Conn: A functional connectivity toolbox for correlated and anticorrelated brain networks. Brain Connectivity, 2(3), 125–141. [DOI] [PubMed] [Google Scholar]
  87. Woolrich M. W., Ripley B. D., Brady M., & Smith S. M. (2001). Temporal autocorrelation in univariate linear modeling of FMRI data. NeuroImage, 14(6), 1370–1386. [DOI] [PubMed] [Google Scholar]
  88. Yang Z., Zuo X.-N., Wang P., Li Z., LaConte S. M., Bandettini P. A., & Hu X. P. (2012). Generalized RAICAR: Discover homogeneous subject (sub)groups by reproducibility of their intrinsic connectivity networks. NeuroImage, 63(1), 403–414. [DOI] [PubMed] [Google Scholar]
  89. Yeo B. T. T., Krienen F. M., Sepulcre J., Sabuncu M. R., Lashkari D., Hollinshead M., … Buckner R. L. (2011). The organization of the human cerebral cortex estimated by intrinsic functional connectivity. Journal of Neurophysiology, 106(3), 1125–1165. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials


Articles from Network Neuroscience are provided here courtesy of MIT Press

RESOURCES