Abstract
In many biological systems, the network of interactions between the elements can only be inferred from experimental measurements. In neuroscience, non-invasive imaging tools are extensively used to derive either structural or functional brain networks in-vivo. As a result of the inference process, we obtain a matrix of values corresponding to a fully connected and weighted network. To turn this into a useful sparse network, thresholding is typically adopted to cancel a percentage of the weakest connections. The structural properties of the resulting network depend on how much of the inferred connectivity is eventually retained. However, how to objectively fix this threshold is still an open issue. We introduce a criterion, the efficiency cost optimization (ECO), to select a threshold based on the optimization of the trade-off between the efficiency of a network and its wiring cost. We prove analytically and we confirm through numerical simulations that the connection density maximizing this trade-off emphasizes the intrinsic properties of a given network, while preserving its sparsity. Moreover, this density threshold can be determined a-priori, since the number of connections to filter only depends on the network size according to a power-law. We validate this result on several brain networks, from micro- to macro-scales, obtained with different imaging modalities. Finally, we test the potential of ECO in discriminating brain states with respect to alternative filtering methods. ECO advances our ability to analyze and compare biological networks, inferred from experimental data, in a fast and principled way.
Author Summary
Complex brain networks are mainly estimated from empirical measurements. As a result, we obtain networks where everything is connected to everything else through different strengths of interaction. Filtering procedures are typically adopted to prune weakest connections. However, network properties strongly depend on the number of remaining links and how to objectively fix such threshold is still an open issue. Here, we propose a criterion (ECO) to filter connectivity based on the optimization of fundamental properties of complex systems, i.e., efficiency and economy. Using ECO, investigators can analyze and compare connectomes in a fast and principled way, capturing network properties of different brain states to eventually quantify (re)organizational mechanisms underlying cognition and disease. Given its generality, we anticipate that ECO can also facilitate the study of networks in other fields, such as system biology.
This is a PLOS Computational Biology Methods paper.
Introduction
Network science has provided a breakthrough in the analysis and modeling of biological systems with the aim to unlock molecular mechanisms behind human disease [1–3] and quantify brain (re)organization underlying behavior, cognition and mental disorders [4–6].
In part, this has been made possible by the increasing availability of tools that indirectly infer the structure of those networks from empirical measurements, thus bypassing the current lack of accurate and complete interaction maps [3, 7]. In system biology, functional links are estimated from transcriptional or phenotypic profiling, and genetic interactions by using measures such as Pearson correlation [8] or Granger causality [9].
In neuroscience, imaging tools such as magnetic resonance imaging (MRI) and electro/magnetoencephalography (E/MEG), are extensively used to map connections and/or interactions between different brain sites, i.e., the connectome [7, 10]. Brain connectivity methods are typically used to estimate the links between the nodes. While structural connectivity (SC) measures the probability to find axonal pathways between brain areas, typically from diffusion MRI, functional connectivity (FC) rather calculates the temporal dependence between remote neural processes as recorded, for instance, by functional MRI, EEG or MEG [4, 7].
At this stage, the resulting networks correspond to maximally dense graphs whose weighted links code for the strength of the connections between different nodes. Common courses in brain network analysis use thresholding procedures to filter information in these raw networks by retaining and binarizing a certain percentage of the strongest links (S1 Fig). Despite the consequent information loss, these procedures are often adopted to mitigate the incertainty of the weakest links, reduce the false positives, and facilitate the interpretation of the inferred network topology [3, 11].
At present, there’s no objective way to fix the value of such threshold. Because network properties significantly depend on the number of remaining links, scientists are obliged to explore brain network properties across a wide range of different candidate thresholds and eventually select one representative a-posteriori [12]. Concurrently, alternative approaches can be adopted to cancel spurious links emerging from third-party interactions [13–15], or statistically validate the estimated connections [7, 16, 17]. However, these procedures lack of precise rationale, are subject to arbitrariness (e.g., the choice of the statistical significance) and make difficult the comparison of network properties between many individuals or samples [11, 18]. Furthermore, these become extremely time-consuming when considering several large connectomes due to the computational complexity of graph quantities based on paths between nodes or on communities detection [19].
To circumvent these issues, we propose a topological criterion for selecting a threshold which captures the essential structure of a network while preserving its sparsity. Based on the optimal trade-off between two desirable but incompatible features—namely high global and local integration between nodes, and low connection density—this method is inherently motivated by the principle of efficiency and economy observed in many complex systems [20], including the brain [21].
Results
Filtering information as a network optimization problem
Global- and local-efficiency have revealed to be important graph quantities to characterize the structure of complex systems in terms of integration and segregation of information [22, 23].
Both structural and functional brain networks tend to exhibit relatively high values of global- and local-efficiency. At the same time they also tend to minimize, for economical reasons, the number of their links leading to sparse networks [21].
Thus, we propose to determine a density threshold that filters out the weakest links and maximizes the ratio between the overall efficiency of a network and its wiring cost. Notice that the definition of cost can have different connotations, e.g., the spatial distance between connected nodes [21]. Here, the cost in terms of number of links is a more general definition which also applies to non-spatially embedded networks (e.g., molecular interaction networks). We formally introduce a criterion to filter information in a given network by finding the connection density ρ that maximizes the quality function:
(1) |
where Eg and El represent respectively the global- and local-efficiency of a network. By definition, the three quantities Eg, El and ρ are normalized in the range [0, 1], and both Eg and El are non-decreasing functions of ρ. More details about the formulation of J can be found in the Material and Methods.
For both regular lattices and random networks, we proved analytically that the optimal density that maximizes J follows a power-law ρ = c/(n − 1), where c is a constant and n is the network size, i.e., the number of nodes in the network. More specifically, c = 3.414 for lattices and c = e = 2.718 for random networks, so that we have approximately ρ ≃ 3/(n − 1). Hence, to maximize J, these networks have to be sparse with an average node degree k ≃ 3 or, equivalently, with a total number of links m that scales as (S1 Appendix).
We confirmed this result (S2a and S2b Fig) through extensive numerical simulations (Materials and Methods), showing that it held true also in more realistic network models, such as in small-world networks [24] (Fig 1a) and in scale-free networks [25] (Fig 1b). For these simulated networks the fitted values varied progressively from c = 3.265, in lattices, to c = 2.966, in random networks, thus falling within the theoretical range found analytically (S1 Table).
Notably, the optimal density values maximizing J emphasized the intrinsic properties (random or regular) of all the implemented synthetic networks in terms of global- and local-efficiency (Fig 1d and 1e and S2d and S2e Fig).
Density threshold in networks inferred from neuroimaging data
We computed the quality function J in both micro- and macro-scale brain networks and we evaluated how the density maximizing J scaled as a function of the network size. We considered connectomes used in previously published studies that were obtained with different imaging modalities, from calcium imaging to EEG, and constructed with disparate brain connectivity methods (Table 1).
Table 1. Experimental details and network characteristics of imaging connectomes.
Imaging modality | Group(s) | Species | Samples x Group | Condition | Nodes | Connectivity method | Domain | Links | Ref. |
---|---|---|---|---|---|---|---|---|---|
Ca+S | Healthy | Zebrafish | 5 | Spontaneuous | [9,21] | Granger causality | Time | Directed | [58] |
eEEG | Healthy | Rodent | 1 | Evoked potential | 15 | Partial directed coherence | Time/Freq. (8 ms/14–29 Hz) | Directed | [59] |
Ca+M | - | Culture | 2 | Spontaneous | [19,32] | Time delay | Time | Directed | [60] |
fNIRS | Healthy | Human | 2 | Resting state | 46 | Pearson’s correlation | Time | Undirected | [61] |
fMRIS | Healthy | Primate | 3 | Resting state | 56 | Pearson’s correlation | Time | Undirected | [62] |
EEG | Healthy, Stroke | Human | 20 | Motor imagery | 61 | Imaginary coherence | Frequency (14–29 Hz) | Undirected | [63] |
fMRI | Healthy, Coma | Human | 17 | Resting state | 90 | Wavelet correlation | Time | Undirected | [64] |
MEG | Healthy, Epilepsy | Human | 5 | Resting state | 149 | Spectral coherence | Frequency (5–14 Hz) | Undirected | [65] |
DTI | Healthy, Epilepsy | Human | 19 | - | 164 | Fractional anisotropy | - | Undirected | [66] |
fMRIL | Healthy, Coma | Human | 17 | Resting state | 417 | Wavelet correlation | Time | Undirected | [64] |
Ca+L | - | Culture | 6 | Spontaneous | [562,1107] | Time delay | Time | Directed | [60] |
EEGL | Healthy | Human | 5 | Motor execution | 4094 | Imaginary Coherence | Frequency (13–30 Hz) | Undirected | [67] |
For each connectome we applied a standard density-based thresholding. We started with the empty network by removing all the links (ρ = 0). Then, we reinserted and binarized one link at time, from the strongest to the weakest, until we obtained the maximally dense network (ρ = 1). At each step we computed J and we recorded its profile as a function of ρ. The pooled density values, as returned by the maximization of the healthy group-averaged J profile in each modality (see Fig 1f for one representative), followed a power law comparable to the one that we reported for synthetic networks (Fig 1c). In particular, the fit ρ = c/(n − 1) to the data gave c = 3.06 with an adjusted r-square R2 = 0.994. Notably, we obtained a similar scaling (c = 2.87 adjusted R2 = 0.946, S2c Fig) when considering individual J profiles (S2f Fig). These results confirm that also for brain networks we can assume that the optimal density threshold maximizing J only depends on the network size according to the same rule ρ ≃ 3/(n − 1).
In conclusion, we introduced a criterion, named efficiency cost optimization (ECO), to select a threshold leading to sparse, yet informative brain networks. Such a threshold is relatively independent of the connectome’s construction and invariant to the underlying network topology so that it can be selected a-priori once the number of nodes is known.
ECO discriminated network properties of different brain states
To illustrate the methodology, we considered connectomes from four different imaging modalities, namely EEG, MEG, fMRI, and DTI (Table 1). Because we do not know the true structure for these connectomes, we evaluated the ability of ECO to discriminate network properties of different brain states, i.e., healthy versus diseased, at individual level.
We characterized brain networks by calculating graph quantities at different topological scales, i.e., large (global- and local-efficiency, Eg and El), intermediate (community partition, P; and modularity, Q), and small (node degree, ki; and betwenness, bi) (Materials and Methods). To assess network differences between brain states, we measured distances between the values of the graph quantities obtained in the healthy group and those in the diseased group. We adopted the Mirkin index (MI) to measure distances between community partitions, and the divergent coefficient (D) for other graph quantities (Materials and Methods).
We explored a wide range of density thresholds and, as expected, the value of the threshold affected the ability to separate network properties of different brain states (Kruskalwallis tests P < 0.01, S2 Table). Notably, the choice ρ = 3/(n − 1) resulted among the best candidates in producing larger distances regardless of the graph quantity (Tukey-Kramer post hoc tests P < 0.05, Fig 2 and S3 Fig). This outcome was not associated to the possible presence of disconnected components. In all the filtered brain networks the size of the largest component (> 50% of the nodes) did not differ between groups for any threshold value (Wilcoxon rank-sum tests P ≥ 0.01, Fig 3). Furthermore, ECO overall outperformed alternative methods, such as the minimum spanning tree (MST) and the planar maximally filtered graph (PMFG) [26], in giving larger distances (Tukey-Kramer post hoc tests P < 0.05, Fig 4, S4b Fig, S2 and S3 Tables). Notably, we reported good performance with respect to a hybrid method, named MST+ECO, where we added the remaining strongest links to the backbone obtained with MST, in order to reach the same average node degree as ECO, i.e. k = 3 (Tukey-Kramer post hoc tests P < 0.05, S3 Table).
Finally, brain networks filtered with ECO were more efficient (Fig 5a) and exhibited J values that better separated different brain states (Fig 5b) as compared to the other filtering methods (Tukey-Kramer post hoc tests P < 0.05, S3 Table).
Discussion
We introduced ECO to filter information in networks whose links are predictions, and not direct measures, of connectivity between biological components, such as brain regions. Conventional approaches evaluate brain network properties across a large and arbitrary number of thresholds [27]. Eventually, they select a representative threshold a-posteriori that maximizes the separation between different brain states [11]. ECO allows to select an objective threshold a-priori, thus reducing the computational burden associated with typical iterative approaches.
Other methods, similar in purpose to ECO, impose unnatural constraints on the filtered network. The minimum spanning tree (MST), for instance, leads to brain networks with a null clustering coefficient [28]. The planar maximally filtered graph (PMFG) tries to alleviate this bias by allowing closed loops, but still forces planarity [26]. Conversely, ECO does not impose structural constraints, apart from favoring sparsity, and lets the intrinsic structure to emerge as illustrated in synthetic networks with known topological organization (Fig 1d and 1e, S2d and S2e Fig). This appears an important feature as different brain states (e.g., diseased versus healthy) are often characterized by networks with different topological orders (more random or more regular) depending on the underlying physiopathological neural mechanism [6].
Overall, results obtained with ECO improved the separation of all the considered network properties between different brain states as compared to other thresholds or filtering methods. In general, this does not necessarily imply a significant group difference for each graph quantity. Instead, it means that if there are underlying network differences, then ECO would be able to point them out.
Maximizing global- and local-efficiency with respect to connection density can be seen as a way to emphasize the integration and segregation properties of a connectome [29] while keeping a biologically plausible wiring cost. This rationale dovetails with current evidence showing that advantageous topological properties, such as economic small-world architectures [21], tend to be maximized in brain networks, and that, in general, sparsity increases robustness of complex systems [30].
Using ECO, networks will have a total number of links m that scales with the number of nodes as m = cn, with c ≃ 3/2. Put differently, the resulting connection density follows a fractal scaling regardless of the network size according to the power-law ρ ≃ 3n−1. Fractal scaling of size and density in self-organized systems has been recently reported and advocated as an important organizational principle to ensure optimal network functioning [31]. Although beyond the scope of this methodological study, we speculate that such characteristic scaling could result, at least for neuronal systems, from a natural optimization of the network efficiency and cost [21].
ECO makes use of density thresholds. Hence, networks having same number of nodes, will have, after pruning, the same number of links. On the one hand, this ensures that differences between network properties are not merely due to differences in the connection density [18]. On the other hand, ECO does not allow a direct evaluation of neural processes altering the number of links; however it does inform on the possible (re)organizational mechanisms.
Finally, it is important to notice that while ECO exhibits several advantageous features, it also has some limitations as described in the following section.
Methodological considerations
ECO is based on a graph theoretic approach and cannot filter out possible false positives (i.e., spurious links) resulting from biased brain connectivity estimates [7, 11]. Our criterion admits that the weighted links of the raw networks had been previously validated, either maintained or canceled. Some inference methods [32, 33] and group-based approaches [34] naturally produce sparse brain networks. In these cases, ECO would still apply as long as there is enough information to filter, i.e., a number of links .
By construction, brain networks filtered with ECO (k ≃ 3) are less sparse than networks filtered with MST (k ≤ 2). However, differently from MST and PMFG, ECO does not guarantee the connectedness of the pruned networks, which can be indeed fragmented (S5 Fig). Whether this condition leads to a more realistic representation of connectomes, especially for large n, we cannot say. Current literature tends to focus on thresholded brain networks which are slightly denser than ECO, with 0.05 ≤ ρ ≤ 0.3 [35]. However, little is known on how this range depends on the number of brain nodes and future studies will have to ascertain if and how the choice of a specific threshold can give more accurate results. Here, we showed that the size of the largest components contained in average more than the 50% of the nodes (Fig 3). Therefore, caution should be used in the evaluation of the resulting network properties and, whenever possible, using graph quantities that can handle networks with disconnected nodes (e.g., the harmonic mean of the shortest path lengths [36]) appears more appropriate.
Finally, other combinations could have been considered when conceiving the quality function J. For example, in [37] authors introduced the cost-efficiency Eg − ρ, which, however, did not include the clustering counterpart. This quality function, as well as other ones that we investigated, did not exhibit meaningful analytic solutions and was therefore excluded as a possible alternative (S2 Appendix). A more general formulation would include a scaling factor in the numerator, like for example 2[αEg + (1 − α)El] where α is a control parameter ranging from 0 to 1. We proved analytically that, for both regular lattices and random graphs, the optimal density that maximizes the corresponding quality function remained regardless of the α value (S3 Appendix). We confirmed this result through numerical simulations also in small-world and scale-free networks (S4 Fig) where the optimal density maximizing J corresponded to an average node degree k ≃ 3, except when α → 1 in lattices and α → 0 in random networks.
Taken together, these findings indicate that the density threshold given by ECO is relatively invariant to the specific value we assigned to the parameter α. The advantage of considering our quality function is that i) it did not depend on external parameters, ii) we could derive analytically the optimal ρ values for lattices and random networks, and iii) the density values obtained by maximizing J in real brain networks fitted the power-law that we found analytically and were able to separate different brain states. Despite these advantages, we notice that ECO could not be the definitive solution to the problem of thresholding in imaging connectomics. Other methods, possibly inspired by biology, are likely to be developed in the future and validation benchmarks will be crucial to evaluate their potential.
Future directions
ECO is founded on asymptotic results in unweighted network models. Its natural application implies binarization after thresholding, a procedure widely adopted to mitigate the uncertainty carried by the weights estimated from neuroimaging data [4, 11]. Further work is needed to clarify how ECO can be extended to weighted networks, where the asymptotic expression of topological properties is less straightforward.
Interactions between biological components are not constant and need to dynamically vary to accomplish internal regulation and external function [38–40]. In neuroscience, functional brain connectivity exhibits rich temporal dynamics that are fundamental for human cognition and complex behavior [41–44]. Further studies should aim to elucidate if and how brain network differences highlighted by ECO change over time.
Conclusion
We introduced ECO as a possible method for filtering information in imaging connectomes. Concrete applications range from cognitive to clinical and computational neuroscience. Given its generality, we anticipate that ECO can also serve to facilitate the analysis of interconnected systems where the need of sparsity is plausible and the links are weighted estimates of connectivity. This is, for example, the case of functional networks in system biology, where links are typically derived from transcriptional or phenotypic profiling, and genetic interactions [3].
Materials and Methods
On the quality function J
The expression of J can be seen as a particular case of a general family of functions of the form f(Eg, El, ρ). Here, we defined J as a ratio to measure the incidence of the density on the network efficiency both at global and local scale. Indeed, we were interested in a relative measure that could tell the network efficiency changes per unit of density. In addition, we did not weight the global- and local-efficiency in the numerator. While, in general, a scaling factor might be necessary to normalize changes between different graph quantities [45], here both Eg and El range between 0 and 1 and are formulated in terms of the same concept, namely the efficiency (at global and local scale) between nodes [22]. We remind to S3 Appendix and S4 Fig for more details on the introduction of a scaling parameter.
By looking at Eq (1), we have that when ρ = 0, then both global- Eg and local-efficiency El are null leading to an indefinite form. As density slightly increases (0 < ρ < ϵ, with ϵ sufficiently small) it can be demonstrated that J tends to 1. Indeed, in this range, the probability to find at least three nodes connected together (a triangle) is extremely low. By definition, El = 0 in absence of at least one triangle [22] and therefore J ≃ Eg/ρ. By considering the definitions of Eg and ρ, this quantity can be rewritten as , where m is the number of existing links and di,j is the distance between the nodes i and j. In a generic network with m links there are at least m pairs of nodes directly connected (i.e., di,j = 1). This means that the sum in the latter equation is bounded from below by m in the case of isolated pairs of connected nodes (m = n/2) or in the trivial case of m = 1. It follows that J → 1 when there are relatively few links in a network.
When ρ tends to 1, it is trivial to see from Eq (1) that J → 2, as both Eg and El tend to one. For intermediate density ranges (ϵ < ρ ≪ 1 − ϵ) the analytic estimate of J is not trivial since Eg and El depend on the network topology which is, in general, unknown.
Numerical simulations for small-world and scale-free networks
Small-world networks were generated according to the Watts-Strogatz (WS) model [24] with a rewiring probability pws = 0.1. Scale-free networks were generated according to the Barabasi-Albert (BA) model [25].
In the first simulation, we considered undirected networks. We varied both the network size and the average node degree, i.e., n = 16, 128, 1024, 16384 and k = 1, 2, 3, 4, 5. In the WS models, k is even accounting for the number of both left and right neighbors of the nodes in the initial lattice. To obtain small-world networks with k odd, we first generated lattices with k even and then, for each odd node (e.g., 1, 3, …), we removed the link with its left farthest neighbor. This procedure removes in total n/2 links leading to a new average node degree k′ = k − 1, while keeping a regular structure. As for BA models, we set the number of links in the preferential attachment mba = 3 and the initial seed was a fully connected network of n0 = mba nodes. This setting generated scale-free networks with k = 6 − 12/n, that is k ≥ 5 regardless of the selected network size. We then removed at random the exceeding number of links until we reached the desired k value. This procedure had the advantage to preserve the original scale-free structure.
In the second simulation, we considered directed networks to confirm and extend the results we obtained for undirected WS and BA networks. We selected eight representative network sizes, i.e., n = 8, 16, 32, 64, 128, 256, 512, 1024 covering the typical size of most current imaging connectomes, and we varied the connection density. Specifically, we performed a two-step procedure:
We fixed one-hundred ρ values quadratically spaced within the entire available density interval.
After having identified the optimal ρ* maximizing J, we performed a refined research among one-hundred new values linearly spaced between the density values, in step 1, before and after ρ*.
For WS models, initial lattices had k equal to the nearest even integer equal or higher than ρ(n − 1), with ρ ∈ (0, 1). For BA models, the number of attaching links was mba = log2 n to ensure an initial relatively high density; the seed was a fully connected network of n0 = mba nodes. By construction , where m0 = n0(n0 − 1)/2 is total number of links in the initial seed. For both models, we then removed at random the exceeding links until we reached the desired density value. For both simulation we generated one-hundred sample networks.
Graph analysis of brain networks
Complex networks can be analyzed by a plethora of graph quantities characterizing different topological properties [46]. Here, we considered a subset of representative ones which have been shown to be relevant for brain network analysis [47]. To characterize the entire brain network (i.e., large-scale topology), we used global- and local-efficiency, which respectively read:
(2) |
where dij is the length of the shortest path between nodes i and j, and Eg(i) is the global-efficiency of the ith subgraph of the network [22].
To characterize modules, or clusters, of brain regions with dense connections internally and sparser connections between groups (i.e., mid-scale topology), we evaluated the community structure of the brain network [4]. We extracted the partition P of the network into modules by means of the Newman’s spectral algorithm maximizing the modularity:
(3) |
where G is the (non-square) matrix having elements Gig = 1 if node i belongs to cluster g and zero otherwise, and M is the so-called modularity matrix [48].
To characterize individual brain areas (i.e., small-scale topology), we measured the centrality of the nodes in the brain network by means of the node degree and of the node betwenness, which respectively read:
(4) |
where the element of the adjacency matrix Aij = 1 if there is a link between node i and j, zero otherwise; and where σjh is the total number of shortest paths between nodes j and h, while σjh(i) is the number of those paths that pass through i.
These quantities represent a small subset of all the possible metrics available in the market. Nevertheless, these are among the most adopted in network neuroscience thanks to their interpretability in terms of connectivity at different topological levels (e.g., network, modules, nodes) [4, 11, 27, 49–51].
Distances between samples and statistical analysis
To assess brain network differences between individuals (or samples) in the two groups, we measured the distance between the respective values obtained for each graph quantity. We used the Mirkin index to compute distances between two network partitions Pu and Pv:
(5) |
where n01 is the number of pairs of nodes in the same cluster under Pv but not under Pu; and n10 is the number of pairs in the same cluster under Pu but not under Pv [52]. The Mirkin index is an adjusted form of the well-known Rand index and it assumes null value for identical clusterings and 1 for totally different clusterings [52]. It corresponds to the Hamming distance between the binary vector representation of each partition. Although this measure can be sensitive to the cluster sizes, it has the advantage of being a metric on the space of the clustering partitions [53].
For all other graph quantities, we used the divergent coefficient [54]:
(6) |
where Xu = [xu,1, xu,2, …, xu,M] and Xv = [xv,1, xv,2, …, xv,M], contain the value(s) of the graph quantity for the uth and vth sample. Notably, M = 1 for global-, local-efficiency and modularity (i.e., Eg, El, Q). M = n for the node degree vector K = [k1, k2, …, kn] and the node betweenness vector B = [b1, b2, …, bn]. The divergent coefficient is a L2-norm distance similar to Euclidean distance but with a normalizing factor which is used for multidimensional scaling [55]. It ranges between 0 (equal multidimensional distribution of the features) and 1 (totally heterogeneous multidimensional distribution). This coefficient is a metric in the Euclidean space when all the values of the features are positive, as for our graph quantities [56]. Both Mirkin index and divergent coefficient are therefore metrics normalized between 0 and 1, allowing for a coherent analysis across different imaging modalities and threshold values.
We used Kruskal–Wallis one-way analysis of variance, with a 0.01 statistical threshold, to evaluate the overall effect of different thresholds, or filtering methods (i.e., MST, PMFG) on distances between individuals. A Tukey-Kramer multiple comparison post hoc test was then used to determine specific differences between pairs of thresholds or methods [57]. Here the statistical threshold was fixed to 0.05.
Supporting Information
Acknowledgments
We thank the colleagues who kindly provided the connectivity data used in this work (Table 1). We are grateful to S. Achard, V. Nicosia, J. Richiardi, M. Valencia for their useful suggestions.
Data Availability
In this study, the data we used to validate our method consisted in the individual connectivity matrices that were kindly provided by the authors of previously published studies (see Table 1 in the main text). We do neither have access to the raw brain signals/images. While those data can be obtained by addressing specific requests to the corresponding authors, we share our Matlab code to filter connectivity matrices, as well as the obtained results, in the third party repository https://github.com/devuci/3n.git.
Funding Statement
VL and MC acknowledge support by the European Commission Project LASAGNE Grant 318132; VL acknowledges support from EPSRC project GALE Grant EP/K020633/1; FDVF and MC acknowledge support by French program “Investissements d’avenir” ANR-10-IAIHU-06; FDVF acknowledges support by the “Agence Nationale de la Recherche” through contract number ANR-15-NEUC-0006-02. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1. Albert R. Scale-free networks in cell biology. J Cell Sci. 2005;118(Pt 21):4947–4957. 10.1242/jcs.02714 [DOI] [PubMed] [Google Scholar]
- 2. Ideker T, Sharan R. Protein networks in disease. Genome Res. 2008;18(4):644–652. 10.1101/gr.071852.107 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Vidal M, Cusick M, Barabási AL. Interactome Networks and Human Disease. Cell. 2011;144(6):986–998. 10.1016/j.cell.2011.02.016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Bullmore E, Sporns O. Complex brain networks: graph theoretical analysis of structural and functional systems. Nature Reviews Neuroscience. 2009;10(3):186–198. 10.1038/nrn2575 [DOI] [PubMed] [Google Scholar]
- 5. Park HJ, Friston K. Structural and Functional Brain Networks: From Connections to Cognition. Science. 2013;342(6158):1238411 10.1126/science.1238411 [DOI] [PubMed] [Google Scholar]
- 6. Stam CJ. Modern network science of neurological disorders. Nat Rev Neurosci. 2014;15(10):683–695. 10.1038/nrn3801 [DOI] [PubMed] [Google Scholar]
- 7. Craddock RC, Jbabdi S, Yan CG, Vogelstein JT, Castellanos FX, Martino AD, et al. Imaging human connectomes at the macroscale. Nature Methods. 2013;(6):524 10.1038/nmeth.2482 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Eisen MB, Spellman PT, Brown PO, Botstein D. Cluster analysis and display of genome-wide expression patterns. Proc Natl Acad Sci USA. 1998;95(25):14863–14868. 10.1073/pnas.95.25.14863 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Guo S, Wu J, Ding M, Feng J. Uncovering interactions in the frequency domain. PLoS Comput Biol. 2008;4(5):e1000087 10.1371/journal.pcbi.1000087 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Sporns O. Making sense of brain network data. Nat Meth. 2013;10(6):491–493. 10.1038/nmeth.2485 [DOI] [PubMed] [Google Scholar]
- 11. De Vico Fallani F, Richiardi J, Chavez M, Achard S. Graph analysis of functional brain networks: practical issues in translational neuroscience. Phil Trans R Soc B. 2014;369(1653):20130521 10.1098/rstb.2013.0521 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Garrison KA, Scheinost D, Finn ES, Shen X, Constable RT. The (in)stability of functional brain network measures across thresholds. Neuroimage. 2015;118:651–661. 10.1016/j.neuroimage.2015.05.046 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Kus R, Kaminski M, Blinowska KJ. Determination of EEG activity propagation: pair-wise versus multichannel estimate. IEEE Transactions on Biomedical Engineering. 2004;51(9):1501–1510. 10.1109/TBME.2004.827929 [DOI] [PubMed] [Google Scholar]
- 14. Marrelec G, Krainik A, Duffau H, Pélégrini-Issac M, Lehéricy S, Doyon J, et al. Partial correlation for functional brain interactivity investigation in functional MRI. NeuroImage. 2006;32(1):228–237. 10.1016/j.neuroimage.2005.12.057 [DOI] [PubMed] [Google Scholar]
- 15. Barzel B, Barabási AL. Network link prediction by global silencing of indirect correlations. Nat Biotech. 2013;31(8):720–725. 10.1038/nbt.2601 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Gourévitch B, Bouquin-Jeannès RL, Faucon G. Linear and nonlinear causality between signals: methods, examples and neurophysiological applications. Biol Cybern. 2006;95(4):349–369. 10.1007/s00422-006-0098-0 [DOI] [PubMed] [Google Scholar]
- 17. Toppi J, De Vico Fallani F, Vecchiato G, Maglione AG, Cincotti F, Mattia D, et al. How the Statistical Validation of Functional Connectivity Patterns Can Prevent Erroneous Definition of Small-World Properties of a Brain Connectivity Network. Computational and Mathematical Methods in Medicine. 2012;2012 10.1155/2012/130985 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. van Wijk BCM, Stam CJ, Daffertshofer A. Comparing Brain Networks of Different Size and Connectivity Density Using Graph Theory. PLoS ONE. 2010;5(10):e13701 10.1371/journal.pone.0013701 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Thulasiraman K, Swamy MNS. Graph Algorithms In: Graphs: Theory and Algorithms. John Wiley & Sons, Inc.; 1992. p. 306–389. [Google Scholar]
- 20. Latora V, Marchiori M. Economic small-world behavior in weighted networks. The European Physical Journal B—Condensed Matter. 2003;32(2):249–263. [Google Scholar]
- 21. Bullmore E, Sporns O. The economy of brain network organization. Nat Rev Neurosci. 2012;13(5):336–349. [DOI] [PubMed] [Google Scholar]
- 22. Latora V, Marchiori M. Efficient behavior of small-world networks. Physical Review Letters. 2001;87(19):198701/1–198701/4. 10.1103/PhysRevLett.87.198701 [DOI] [PubMed] [Google Scholar]
- 23. Bassett DS, Bullmore E. Small-world brain networks. Neuroscientist. 2006;12(6):512–523. 10.1177/1073858406293182 [DOI] [PubMed] [Google Scholar]
- 24. Watts DJ, Strogatz SH. Collective dynamics of ‘small-world’ networks. Nature. 1998;393(6684):440–442. 10.1038/30918 [DOI] [PubMed] [Google Scholar]
- 25. Barabási AL, Albert R. Emergence of Scaling in Random Networks. Science. 1999;286(5439):509–512. 10.1126/science.286.5439.509 [DOI] [PubMed] [Google Scholar]
- 26. Tumminello M, Aste T, Matteo TD, Mantegna RN. A tool for filtering information in complex systems. PNAS. 2005;102(30):10421–10426. 10.1073/pnas.0500298102 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Fornito A, Zalesky A, Breakspear M. Graph analysis of the human connectome: Promise, progress, and pitfalls. NeuroImage. 2013;80:426–444. 10.1016/j.neuroimage.2013.04.087 [DOI] [PubMed] [Google Scholar]
- 28. Tewarie P, van Dellen E, Hillebrand A, Stam CJ. The minimum spanning tree: An unbiased method for brain network analysis. Neuroimage. 2015;104:177–188. 10.1016/j.neuroimage.2014.10.015 [DOI] [PubMed] [Google Scholar]
- 29. Tononi G, Sporns O, Edelman GM. A measure for brain complexity: Relating functional segregation and integration in the nervous system. Proceedings of the National Academy of Sciences. 1994;91(11):5033–5037. 10.1073/pnas.91.11.5033 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Leclerc RD. Survival of the sparsest: robust gene networks are parsimonious. Molecular Systems Biology. 2008;4(1):213 10.1038/msb.2008.52 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31. Laurienti PJ, Joyce KE, Telesford QK, Burdette JH, Hayasaka S. Universal fractal scaling of self-organized networks. Physica A. 2011;390(20):3608–3613. 10.1016/j.physa.2011.05.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32. Descoteaux M, Deriche R, Knösche TR, Anwander A. Deterministic and probabilistic tractography based on complex fibre orientation distributions. IEEE Trans Med Imaging. 2009;28(2):269–286. 10.1109/TMI.2008.2004424 [DOI] [PubMed] [Google Scholar]
- 33. Ryali S, Chen T, Supekar K, Menon V. Estimation of functional connectivity in fMRI data using stability selection-based sparse partial correlation with elastic net penalty. NeuroImage. 2012;59(4):3852–3861. 10.1016/j.neuroimage.2011.11.054 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Roberts JA, Perry A, Roberts G, Mitchell PB, Breakspear M. Consistency-based thresholding of the human connectome. Neuroimage. 2017;145(A):118–129. 10.1016/j.neuroimage.2016.09.053 [DOI] [PubMed] [Google Scholar]
- 35. Bassett DS, Bullmore ET. Small-World Brain Networks Revisited. Neuroscientist. 2016; p. 1073858416667720 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36. Boccaletti S, Latora V, Moreno Y, Chavez M, Hwang DU. Complex networks: Structure and dynamics. Physics Reports. 2006;424(4–5):175–308. 10.1016/j.physrep.2005.10.009 [DOI] [Google Scholar]
- 37. Achard S, Salvador R, Whitcher B, Suckling J, Bullmore E. A resilient, low-frequency, small-world human brain functional network with highly connected association cortical hubs. Journal of Neuroscience. 2006;26(1):63–72. 10.1523/JNEUROSCI.3874-05.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38. Han JDJ, Bertin N, Hao T, Goldberg DS, Berriz GF, Zhang LV, et al. Evidence for dynamically organized modularity in the yeast protein–protein interaction network. Nature. 2004;430(6995):88–93. 10.1038/nature02555 [DOI] [PubMed] [Google Scholar]
- 39. Luscombe NM, Madan Babu M, Yu H, Snyder M, Teichmann SA, Gerstein M. Genomic analysis of regulatory network dynamics reveals large topological changes. Nature. 2004;431(7006):308–312. 10.1038/nature02782 [DOI] [PubMed] [Google Scholar]
- 40. Przytycka TM, Singh M, Slonim DK. Toward the dynamic interactome: it’s about time. Brief Bioinform. 2010;11(1):15–29. 10.1093/bib/bbp057 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Bassett DS, Wymbs NF, Porter MA, Mucha PJ, Carlson JM, Grafton ST. Dynamic reconfiguration of human brain networks during learning. PNAS. 2011; 10.1073/pnas.1018985108 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42. Hutchison RM, Womelsdorf T, Allen EA, Bandettini PA, Calhoun VD, Corbetta M, et al. Dynamic functional connectivity: Promise, issues, and interpretations. NeuroImage. 2013;80:360–378. 10.1016/j.neuroimage.2013.05.079 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Zalesky A, Fornito A, Cocchi L, Gollo LL, Breakspear M. Time-resolved resting-state brain networks. PNAS. 2014;111(28):10341–10346. 10.1073/pnas.1400181111 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44. Kucyi A, Hove MJ, Esterman M, Hutchison RM, Valera EM. Dynamic Brain Network Correlates of Spontaneous Fluctuations in Attention Cereb Cortex. 2016; [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45. Souza FSH, Mateus GR, Cunha ASd. Optimization in Designing Complex Communication Networks In: Thai MT, Pardalos PM, editors. Handbook of Optimization in Complex Networks. No. 57 in Springer Optimization and Its Applications. Springer; US; 2012. p. 3–37. [Google Scholar]
- 46. Costa LDF, Oliveira ON, Travieso G, Rodrigues FA, Villas Boas PR, Antiqueira L, et al. Analyzing and modeling real-world phenomena with complex networks: A survey of applications. Advances in Physics. 2011;60(3):329–412. 10.1080/00018732.2011.572452 [DOI] [Google Scholar]
- 47. Rubinov M, Sporns O. Complex network measures of brain connectivity: Uses and interpretations. NeuroImage. 2010;52(3):1059–1069. 10.1016/j.neuroimage.2009.10.003 [DOI] [PubMed] [Google Scholar]
- 48. Newman MEJ. Modularity and community structure in networks. PNAS. 2006;103(23):8577–8582. 10.1073/pnas.0601602103 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49. Stam CJ, Reijneveld JC. Graph theoretical analysis of complex networks in the brain. Nonlinear Biomedical Physics. 2007;1 10.1186/1753-4631-1-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50. He Y, Evans A. Graph theoretical modeling of brain connectivity. Current Opinion in Neurology. 2010;23(4):341–350. [DOI] [PubMed] [Google Scholar]
- 51. Wig GS, Schlaggar BL, Petersen SE. Concepts and principles in the analysis of brain networks: Brain networks. Annals of the New York Academy of Sciences. 2011;1224(1):126–146. 10.1111/j.1749-6632.2010.05947.x [DOI] [PubMed] [Google Scholar]
- 52. Meilă M. Comparing clusterings—an information based distance. Journal of Multivariate Analysis. 2007;98(5):873–895. 10.1016/j.jmva.2006.11.013 [DOI] [Google Scholar]
- 53. Wagner S, Wagner D. Comparing clusterings: an overview. Univ., Fak. für Informatik; 2007. [Google Scholar]
- 54. Cox MAA, Cox TF. Multidimensional Scaling In: Handbook of Data Visualization. Springer Handbooks Comp.Statistics. Springer; Berlin Heidelberg; 2008. p. 315–347. [Google Scholar]
- 55. Clark PJ. An Extension of the Coefficient of Divergence for Use with Multiple Characters. Copeia. 1952;1952(2):61 10.2307/1438532 [DOI] [Google Scholar]
- 56. Gower JC, Legendre P. Metric and Euclidean properties of dissimilarity coefficients. Journal of Classification. 1986;3(1):5–48. 10.1007/BF01896809 [DOI] [Google Scholar]
- 57. McDonald J. Handbook of Biological Statistics. Sparky House Publishing; 2009. [Google Scholar]
- 58. De Vico Fallani F, Corazzol M M, Sternberg J, Wyart C, Chavez M. Hierarchy of Neural Organization in the Embryonic Spinal Cord: Granger-Causality Graph Analysis of Calcium Imaging Data. IEEE Transactions on Neural Systems and Rehabilitation Engineering. 2014;PP(99):1–1. 10.1109/TNSRE.2014.2341632 [DOI] [PubMed] [Google Scholar]
- 59. Plomp G, Quairiaux C, Michel CM, Astolfi L. The physiological plausibility of time-varying Granger-causal modeling: Normalization and weighting by spectral power. NeuroImage. 2014;97:206–216. 10.1016/j.neuroimage.2014.04.016 [DOI] [PubMed] [Google Scholar]
- 60. Teller S, Granell C, De Domenico M, Soriano J, Gómez S, Arenas A. Emergence of Assortative Mixing between Clusters of Cultured Neurons. PLoS Comput Biol. 2014;10(9):e1003796 10.1371/journal.pcbi.1003796 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61. Niu H, Li Z, Liao X, Wang J, Zhao T, Shu N, et al. Test-Retest Reliability of Graph Metrics in Functional Brain Networks: A Resting-State fNIRS Study. PLoS ONE. 2013;8(9):e72425 10.1371/journal.pone.0072425 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62. O’Reilly JX, Croxson PL, Jbabdi S, Sallet J, Noonan MP, Mars RB, et al. Causal effect of disconnection lesions on interhemispheric functional connectivity in rhesus monkeys. PNAS. 2013;110(34):13982–13987. 10.1073/pnas.1305062110 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63. De Vico Fallani F, Pichiorri F, Morone G, Molinari M, Babiloni F, Cincotti F, et al. Multiscale topological properties of functional brain networks during motor imagery after stroke. NeuroImage. 2013;83:438–449. 10.1016/j.neuroimage.2013.06.039 [DOI] [PubMed] [Google Scholar]
- 64. Achard S, Delon-Martin C, Vértes PE, Renard F, Schenck M, Schneider F, et al. Hubs of brain functional networks are radically reorganized in comatose patients. PNAS. 2012;109(50):20608–20613. 10.1073/pnas.1208933109 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65. Chavez M, Valencia M, Navarro V, Latora V, Martinerie J. Functional Modularity of Background Activities in Normal and Epileptic Brain Networks. Phys Rev Lett. 2010;104(11):118701 10.1103/PhysRevLett.104.118701 [DOI] [PubMed] [Google Scholar]
- 66. Besson P, Dinkelacker V, Valabregue R, Thivard L, Leclerc X, Baulac M, et al. Structural connectivity differences in left and right temporal lobe epilepsy. NeuroImage. 2014;100:135–144. 10.1016/j.neuroimage.2014.04.071 [DOI] [PubMed] [Google Scholar]
- 67. De Vico Fallani F, Chessa A, Valencia M, Chavez M, Astolfi L, Cincotti F, et al. Community structure in large-scale cortical networks during motor acts. Chaos, Solitons & Fractals. 2012;45(5):603–610. 10.1016/j.chaos.2012.02.006 [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
In this study, the data we used to validate our method consisted in the individual connectivity matrices that were kindly provided by the authors of previously published studies (see Table 1 in the main text). We do neither have access to the raw brain signals/images. While those data can be obtained by addressing specific requests to the corresponding authors, we share our Matlab code to filter connectivity matrices, as well as the obtained results, in the third party repository https://github.com/devuci/3n.git.