Abstract
The connectome, or the entire connectivity of a neural system represented by a network, ranges across various scales from synaptic connections between individual neurons to fibre tract connections between brain regions. Although the modularity they commonly show has been extensively studied, it is unclear whether the connection specificity of such networks can already be fully explained by the modularity alone. To answer this question, we study two networks, the neuronal network of Caenorhabditis elegans and the fibre tract network of human brains obtained through diffusion spectrum imaging. We compare them to their respective benchmark networks with varying modularities, which are generated by link swapping to have desired modularity values. We find several network properties that are specific to the neural networks and cannot be fully explained by the modularity alone. First, the clustering coefficient and the characteristic path length of both C. elegans and human connectomes are higher than those of the benchmark networks with similar modularity. High clustering coefficient indicates efficient local information distribution, and high characteristic path length suggests reduced global integration. Second, the total wiring length is smaller than for the alternative configurations with similar modularity. This is due to lower dispersion of connections, which means each neuron in the C. elegans connectome or each region of interest in the human connectome reaches fewer ganglia or cortical areas, respectively. Third, both neural networks show lower algorithmic entropy compared with the alternative arrangements. This implies that fewer genes are needed to encode for the organization of neural systems. While the first two findings show that the neural topologies are efficient in information processing, this suggests that they are also efficient from a developmental point of view. Together, these results show that neural systems are organized in such a way as to yield efficient features beyond those given by their modularity alone.
Keywords: connectome, network analysis, modularity, efficiency, brain development, brain evolution
1. Introduction
In network representation, neural networks at different levels of organization, ranging from connections between individual neurons to connections between brain regions, can be described coherently if the individual neurons or brain regions are substituted by the nodes and the connection between them by the links. Also, the modular organization found in different levels of neural networks can be exhibited by network modules, where a module is a subset of the nodes having many connections among them and few to the rest of the network [1].
The first species to show neural networks are coelenterates such as Cnidaria [2,3]. These animals show a diffuse two-dimensional nerve network called a lattice network. In such networks, neighbours are well connected, but there are no long-distance connections. For functionally specialized circuits, however, a regular organization is unsuitable. Starting with the formation of sensory organs and motor units, neurons segregate in modules, e.g. forming ganglia in the roundworm Caenorhabditis elegans [4]. By forming such modules, ganglia can process one modality with little interference from neurons processing different kinds of information. At one point of growing complexity of organisms, having one module for one modality or function is not sufficient. An example is processing of visual information in primates where the visual module consists of two network components: nodes that form the dorsal pathway for processing object position and movement, and nodes of the ventral pathway for processing object features such as colour and form. These networks, where smaller sub-modules are nested within modules, are a type of hierarchical network [5–7].
The modularity Q measures how modular a given network is [8]. The human brain network for the connections between brain regions or region of interests (ROIs) as well as the neuronal network for the connections between neurons shows a high modularity compared with randomly connected networks [6], and this modularity is preserved from at least 4 to 40 years [9]. However, there are numerous ways of constructing modular networks with a given value of modularity. What are specific to the chosen biological organizations over alternative modular arrangements and what are the advantages of them? In this article, we address these questions on two different levels of organization: the connections between individual neurons in C. elegans, the level of the micro-connectome [10], and the connections between different human brain regions, the level of the macro-connectome [11]. To investigate the connection specificity of these networks over alternative arrangements, we employ benchmark networks generated by a link swapping process, which is controlled by the simulated annealing algorithm. Such rewired networks can serve as control groups, where the number of connections for each node and the modularity of networks are kept constant.
First, at both levels we find that the clustering coefficient (indicating how well information can be distributed locally) and the characteristic path length (indicating how difficult global integration is) are high compared with alternative networks of similar modularity. This shows a balance between the need for communication within local circuits (high neighbourhood connectivity within modules) and the reduction of interference between modules (fewer shortcuts linking different modules). Indeed, brain disorders such as schizophrenia and epilepsy [12] can be linked to changes in local and global efficiency. Second, the total wiring length is smaller compared with the alternative networks of similar modularity. The connectivity of the original network and alternative networks are compared through their network of modules, the coarse-grained network obtained when human brain areas are considered as new nodes instead of the ROIs. We find that the formation of fibre bundles, or the fasciculation, is correlated with the reduced total wiring length. A similar behaviour is observed from the network of neurons and the network of ganglia in C. elegans. To quantify this bundling behaviour, we introduce the novel measure of dispersion, indicating how widely individual nodes are connected to different modules of the network. Third, both neural networks show lower algorithmic entropy than their alternative arrangements. As the algorithmic entropy quantifies the amount of information needed to construct an object, this suggests that fewer genes are needed to encode for the organization of neural networks and the neural systems are efficiently organized from a developmental point of view [13].
2. Material and methods
(a). Data
The human brain network used in this paper was from the study of Hagmann et al. [14]. The connectivity was obtained from five individual subjects using diffusion spectrum imaging (DSI). DSI is one of the protocols of diffusion magnetic resonance imaging (dMRI), which detects the diffusion pattern of water molecules in the brain to predict the trajectory of fibre tracts. In DSI, first the brain was partitioned into anatomical areas called Brodmann areas, and then each of them was subdivided into a certain number of ROIs in such a way that each ROI has a similar surface area. The number of brain areas were chosen to be R = 66 and the number of ROIs resulted in N = 998. The ROIs were regarded as nodes and the brain areas as modules. Next, the tractography was constructed from the diffusion pattern and a link was assigned between two ROIs that were connected by the predicted fibre tract. The total number of links was E = 17 865.
For C. elegans, a total of N = 279 neurons and corresponding E = 2990 connections were used. These included 1584 unidirectional and 1406 bidirectional connections. Biologically, they represent 672 gap junctions, 1962 chemical synapses and 376 connections where both gap junctions and chemical synapses exist between the neuron pairs. As some network measures are defined only for undirected networks, all the unidirectional connections were replaced by bidirectional ones, leading to a total of 2287 bidirectional links. Three-dimensional neuron coordinates were used as described in [15]. The information about the R = 10 ganglia membership for modules was taken from the study of Achacoso & Yamamoto [16].
The network of modules was defined as follows. The modules, corresponding to the anatomical areas for brain or the ganglia for C. elegans, were regarded as nodes in place of the ROIs or neurons. Correspondingly, two modules were assigned with a link between them only if there was at least one link between a pair of nodes each of which was contained by each module (see the electronic supplementary material, figure S1).
(b). Network measures
All the calculations, including measurement of modularity and simulated annealing procedure (see §2c), were performed by custom-built codes in C programming language and MATLAB (routines are available at http://www.dynamic-connectome.org/). The characteristic path length (L) was the average number of connections that have to be passed on the shortest paths between all pairs of network nodes. The clustering coefficient (C) was the proportion of connections actually present out of all possible connections among network nodes directly connected to a node. It was calculated as the average over all individual nodes of the network [17]. The small-world index was calculated as σsw = (C/Crand)/(L/Lrand) or equivalently σsw = (C/L)/(Crand/Lrand), where C and L defined as above were measured from the observed network and Crand and Lrand were the average values from 100 Erdős–Rényi (ER) random networks [18]. The rules for generating the ER network are as follows. Initially, N nodes are given without any connection. At each time step, a link is added between a pair of nodes which are selected among the N nodes at random, avoiding multiple times of selection. This step is repeated until the number of links becomes E. The small-worldness σsw is larger than unity for small-world networks, equal to unity if the ratio between C/Crand and L/Lrand is the same as for random networks (note that absolute C and L might still differ from those of random networks), and smaller than unity when the clustering coefficient is smaller and/or the characteristic path length is larger than for random networks. The total wiring length (W) is the sum of the Euclidean distance between all connections of a network when the network nodes are provided with spatial locations.
(c). Modularity and link swapping
For a network of N-nodes, E-links and R-modules, whose node index, i, runs from unity to N and the module to which node i belongs, qi, can take value from unity to R, the modularity was defined as
![]() |
2.1 |
where Aij is the (i,j) element of the adjacency matrix, ki is the number of connections, or the degree, of node i, and δ is Kronecker's delta function [8]. The modularity measures what fraction of the links connects two nodes within one module and its deviation from the case when the links are distributed at random. The modularity can be used for finding the modular structure of a given network when it is unknown. In such a setting, an optimal partitioning of the network nodes is searched, which maximizes the modularity of the given network. Therefore, the assignments of nodes to modules are varied while the connections of the nodes are fixed. In this study, however, the predefined modules of respective networks, i.e. the anatomical areas of human brain and the ganglia of C. elegans, were regarded as fixed, each node already having its intrinsic module membership. Instead, the connections between nodes were varied by link swapping controlled by simulated annealing.
Link swapping is a process in which a pair of links is selected, and then two nodes at an arbitrary end of each link are exchanged. Whereas the degree of each node, as well as its distribution for the entire network, is preserved before and after the manipulation, the modularity of the network can be increased, decreased or remain the same depending on the selected pair of links. It is increased if a pair of links is selected in such a way that at least two nodes at the ends of different links lie in one module, and swapping is carried out to connect those two nodes. Likewise, the number of intra-module links determines the modularity of the rewired network after swapping (see the electronic supplementary material, figure S1).
To alter the modularity of the networks to have desired values, the selection of link pairs for the swapping process was controlled by simulated annealing as follows [19]. At each step, the link swapping is attempted and the amount of change in modularity for the attempt, ΔQ, is calculated. The attempt is accepted with probability unity if ΔQ ≥ 0 or with probability eΔQ/T if ΔQ < 0, where T is the control parameter or temperature. Otherwise, the attempt is rejected and the swapping is reversed to recover the original connectivity. When T → 0, link swappings are accepted only when the modularity increases and the simulated annealing becomes equivalent to the greedy algorithm for finding the maximum modularity. Originally, the simulated annealing was devised to avoid trapping into local extrema as the greedy algorithm often does, and T is incrementally decreased from a finite value to infinitesimal so that the swapping happens a certain number of times at each T value. The consequent maximum value during all the time steps is expected to be the global maximum. In a similar manner, to obtain a network with the desired modularity Qd, one can set the problem to minimize |Q − Qd|.
However, in this study, we employed a simpler method since the minimization procedure is computationally expensive and the networks from the two different methods are theoretically equivalent to each other. The alternative method took advantage of the fact that the modularity, unless there are small fluctuations, converges to a single value for a given temperature. After a sufficient number of link swappings are performed, the connection specificity of the original network is lost and the resulting network has the desired modularity but is otherwise maximally random. Any choice of network snapshot at this state is statistically identical to every other, and the entire set of such networks is the ensemble of networks with the given modularity. In practice, we first performed 800 × E link swaps for a given temperature, and then sampled 100 network snapshots during additional 200 × E steps.
(d). Dispersion
We introduced the novel measure, dispersion D, of a network which shows how widely the connections are distributed across different modules. The dispersion of an individual node i was defined as Di = Ri/R, where Ri is the number of different modules to which the node is connected (brain areas for the human connectome or ganglia for the C. elegans connectome), and R is the total number of modules (66 and 10, respectively). The maximum dispersion of a node is unity in the case where the node is connected to at least one node in all other modules of the network. The dispersion of a network is the average dispersion for all nodes: D = ∑Di/N, where N is the number of nodes (998 ROIs for the human connectome and 279 neurons for the C. elegans connectome). Note that the modules in this study are anatomical units (brain areas or ganglia) and not the modules defined by network analysis module detection algorithms [1]. However, alternative definitions for a module can also be applied, and dispersion could serve as a useful measure for future studies.
(e). Algorithmic entropy
Algorithmic entropy was used as a measure for the amount of information the networks bear. It was originally introduced as a conceptual measure for any kind of physical or abstract objects, and later a practical way to quantify it was devised [20]. Assume an object saved in a computer storage device. If the object contains regularities, it can be described by a shorter message leading to less storage usage. A compression algorithm is a standard way to detect such regularities and reduce storage usage, and the compressed data size can give an estimate of the amount of information. To apply this to the neural networks, we saved the networks in the format of unweighted adjacency matrices into N × N int8 arrays, whose (i,j) element takes value unity if nodes i and j have a connection to each other and otherwise zero. Any configuration of networks with the same number of nodes N has N2 bytes of data size. Then, the minimum compression size for each of the adjacency matrix arrays for the original connectomes together with the rewired network ensembles for different values of Q were found by the simulated annealing method similar to §2c. The compression was performed by the gzip library which uses Lempel–Ziv coding [21]. The compression ratio, the ratio of the compressed data size to the original size of the array in bytes, was measured to indicate the relative amount of information in the networks. As the adjacency matrix of the networks is symmetric and sparse, a more efficient data storing strategy could be devised. Although this can change the quantitative values of the compression ratio, it is unlikely that the qualitative trend of the result from the original and alternative networks would change.
For simulated annealing, the objective measure to minimize was the compression size Z and the variable was node index assignment. Whereas the assignment of node index, i.e. which node becomes node i, is arbitrary, the shape of the adjacency matrix depends on the index assignment and in turn the compression size depends on the shape. As the algorithmic entropy, by definition, aims to measure the upper bound of the amount of information, the node index assignment needs to minimize the size of the compressed array. At each time step of simulated annealing, the node indices were reassigned by exchanging the indices of two nodes i and j, which is equivalent to exchanging the ith and jth column and row of the adjacency matrix. Then ΔZ was measured by comparing the Z values before and after the reassignment, to determine whether such a reassignment should be kept or reverted with the probability of unity when ΔZ ≤ 0 or with probability of e−ΔZ/T when ΔZ > 0. T was incrementally decreased from a finite value to infinitesimal so that the index reassignment happened a certain number of time steps at each T value. The global minimum Z during all the time steps was recorded.
3. Results
To illustrate the connectivity of the neural networks, we calculated the network measures of the human and C. elegans connectome. Two relevant measures, L and C, were compared to those of the ER random networks with the same number of nodes and links (table 1). First, the characteristic path length L, related to the global efficiency of reaching other nodes at the global level, shows the average number of connections that need to be crossed to go from one network node to another. Second, the clustering coefficient C, related to the local efficiency of reaching nearby nodes, indicates how well neighbours of a node are connected, i.e. what proportion of potential links between neighbours actually exists. Third, the small-world index σsw indicates to what extent the fraction of two small-world measures, C/L, of a network deviates from that of random networks. Finally, we observed the total wiring length, which is the sum of the approximated metric lengths of all individual connections. Note that the Euclidean distance in three dimensions gives an estimate or lower bound of the length of a connection, as the curvature in actual wiring between nodes makes the real distance longer. More information on network measures can be found in [1,22].
Table 1.
Network measures depicting the connectomes. Network measures for the human brain network with 998 nodes and the C. elegans neuronal network with 279 nodes: Q modularity, D dispersion, L characteristic path length, C clustering coefficient, σsw small-world index. The values for ER random networks, Lrand and Crand, show the average and the standard deviation over 100 ER networks.
Q | D | L | Lrand | C | Crand | σsw | |
---|---|---|---|---|---|---|---|
human | 0.26 | 0.12 | 3.07 | 2.231 ± 0.001 | 0.47 | 0.036 ± 0.002 | 9.27 |
C. elegans | 0.15 | 0.46 | 2.43 | 2.300 ± 0.002 | 0.34 | 0.059 ± 0.001 | 5.37 |
The human macro-connectome consists of R = 66 brain areas (modules), N = 998 ROIs (nodes) and E = 17 865 connections (links) between ROIs in total. The average degree, , is 35.80. The characteristic path length, L, is 3.07 and the clustering coefficient, C, is 0.47. For comparison, the ER networks with the same number of nodes and links yield L = 2.22 and C = 0.036 (average over 100 generated networks). The high small-world index σsw value of 9.27, as well as the high C value compared with Crand, suggests that the human brain connectome is a small-world network. However, it is interesting to note that L is slightly larger than Lrand which suggests the opposite. It is due to the fact that L can be reduced drastically by only a few extremely long-range connections. While the ER networks can have such long-range connections, the connection range of human connectome is relatively limited. The total wiring length, W, is 493.5 m. The modularity, Q, is 0.26.
For the C. elegans micro-connectome of N = 279 neurons and E = 2287 links, L = 2.43 and C = 0.34, whereas Lrand = 2.30 and Crand = 0.059, which gives a small-world index, σsw, of 5.37. Similar observations can be made as for human connectome: C and σsw indicate that the C. elegans connectome is a strongly small-world network, but its L is slightly larger than Lrand because of the lack of extremely long-range connections. The total wiring length, W, is 588.2 mm. The modularity, Q, is 0.15.
From these basic measures, the connection specificity of the networks can be roughly depicted. Both networks are small-world with few long-range connections and have modular organization. Since the modularity values (0.26 for human and 0.15 for C. elegans) are small compared with those of other networks known to have modular structure, the significance of the modular organization could be questioned. However, these networks, though small, do have modularity indicated by the Q values when compared to those for the completely randomized, zero-modularity networks obtained by link swapping as seen below.
To understand the connectivity in detail, next we compared the network measures of the connectome to their respective benchmark networks which were generated through the link swapping process controlled by simulated annealing as described in §2c. Each node of the benchmark networks has one-to-one correspondence to a node of the original network and has the same degree and membership to a module as the original node. By changing the control parameter, T, of the simulated annealing process, the resulting benchmark networks were obtained with varying modularities. The relationship between T and resulting Q values is given in the electronic supplementary material, figure S2 and table S1. Figure 1 visualizes the original neural networks and corresponding benchmark networks with different modularities.
Figure 1.
Adjacency matrices of the connectomes. The matrices represent the network of ROIs for the human brain (a) and the network of neurons for C. elegans (b), respectively. Each dot represents a fibre tract between ROIs in (a) or an axonal connection between neurons in (b). For both humans and C. elegans, we analysed benchmark networks with similar (=), increased (>) or decreased (<) modularity Q relative to the original neural networks. (Online version in colour.)
The network measures of the original and benchmark networks are shown in figure 2. The quantities L, C, σsw and W show strong positive or negative correlations to Q for the benchmark networks, whereas the values from the original network deviate from the trends of the curves in all cases. In general, as the modularity grows, the number of local loops increases and the number of long-range connections decreases. Therefore, the increase in L and C, as well as the decrease in W, with respect to growing Q is easily understood. For all the network measures, the original neural networks show marked differences to alternative arrangements with the same modularity. In addition, some values for the original networks can only be reached for much higher modularity in alternative networks or cannot be reached at all (L, C and W for the human connectome). Note that the clustering coefficients of the original networks are higher than those of alternative networks of the same modularity, which suggests better local interaction efficiency. The high characteristic path length, in contrast, suggests reduced global communication efficiency.
Figure 2.
(a–d) Small-worldness is different in the connectomes. The small-world measures, characteristic path length L, clustering coefficient C, small-world index σsw and total wiring length W, of human (circles) and C. elegans (squares) connectomes with respect to modularity, Q, which is varied by link swapping. Note that W is normalized with respect to the values of the original neural networks. Unobservable error bars lie within the symbols. The vertical dashed lines denote the values of the original networks. The original networks show more global segregation (higher L suggests lower global efficiency) and more local integration (higher C suggests higher local efficiency) at the same time. (Online version in colour.)
What made the original neural networks deviate from the tendency of alternative benchmark networks, or what is specific to the connectivity of the original networks? The answer is that one module of the neural networks is connected only to a small number of other modules, and as a result, a pair of modules is connected by a redundant number of links. Two modules are considered to be connected to each other if any of their member nodes are connected. To test such connectivity between modules, the network of modules for both connectomes and examples of their benchmark networks are visualized in figure 3, inspection of which immediately shows that the network for the human connectome is sparse (figure 3a) and the benchmark network of it is dense (figure 3b). This effect is also visible, though less apparent, for the C. elegans connectome (figure 3d) and for the corresponding benchmark network (figure 3e).
Figure 3.
Connectivity of modules and the dispersion. The networks of modules for (a) the human (brain areas; horizontal plane) and (d) the C. elegans (ganglia; lateral view) connectomes, and those of the benchmark network snapshots with similar modularity, (b) and (e), respectively. Each node is a module of the networks, whose size is proportional to the square-root of the number of nodes in the module. The locations are given by the centres of mass of its constituent ROIs or neurons. Note that the node locations for C. elegans are scaled differently in x- and y-axes for visualization and do not represent the actual coordinates. The dispersion D of human (c) and C. elegans (f) networks (data point on the dashed vertical line) is much lower than those of the benchmark networks. (Online version in colour.)
As discussed above, the number of links before and after link swapping does not change. Therefore, the observed difference in link density must have occurred during the process of coarse-graining the network of nodes into the network of modules. Note that the multiple links between a pair of modules converge into a single link on the network of modules. Accordingly, the number of links on the network of modules is determined by the number of other modules to which the modules are connected. Sparse connectivity of the network of modules implies that each module is connected to only a small number of other modules on the network of nodes and that two modules are connected to each other by a redundant number of links. This is observed as bundling of fibres towards relatively few target nodes in the brain connectome, and is also found in the C. elegans connectome where neurons are able to follow early established pathways, e.g. in the ventral cord [15]. However, the benchmark networks lose such connection specificity during link swapping. Some of the multiple links from one module to another in the original networks are redirected to many new modules during link swapping, increasing the number of modules to which they connect but decreasing the number of links between a given pair of modules.
As a way to measure this, we introduced a novel network property called dispersion, D, that measures the average proportion of modules to which a network node is connected. Note that this is different from an existing measure, the participation coefficient, which is the proportion of a node's connections that connect to other modules, as the dispersion also indicates to how many other modules a node is connected. For the human connectome, the dispersion is 0.12, indicating that each ROI is, on average, connected to 12% of all anatomical brain areas (figure 3c). For C. elegans connectome, with a dispersion of 0.46, each neuron is, on average, connected to 46% of all ganglia (figure 3f). These values for the connectomes are much lower than those of the benchmark networks with similar modularity. Human benchmark networks with Q = 0.25 have D = 0.31 (larger than the value of human connectome by factor of 2.6) and C. elegans benchmark networks with Q = 0.15 have D = 0.60 (factor of 1.3). In addition, such low dispersion values can only be reached for much higher modularities in alternative networks of C. elegans, or cannot be achieved at all for alternatives of the human connectome. Fewer distributed fibres also reduce the total wiring length, meaning that less energy is needed for connection establishment (myelination) and maintenance (recovery to the resting potential after transmitting an action potential) [23–25].
These considerations on the costs of material and energy can be seen as being related to the physical structure or ‘hardware’ of neural networks. However, costs of the neural ‘hardware’ are not the only potential evolutionary constraint [26]. Complementary to the concept of ‘hardware’, the rules for changing the pattern of connections and connection weights can be considered as the ‘software’ of the brain. Connection weights can adapt through learning, and connections can be rewired after a lesion or traumatic brain injury [27]. However, looking at changes during brain development, the early perinatal large-scale architecture seems to be remarkably stable. Eliminating activity propagation by blocking neurotransmitter release has little effect on the layer and cortico-cortical architecture [28]. Such invariance in the organization of neural systems could be considered as being determined by genetics factors.
Hence, the following question can be raised: how much genetic information is needed to encode the connectivity patterns in human and C. elegans? One estimate, based on earlier studies in metabolic networks [29], is the algorithmic entropy or Kolmogorov complexity [20]. The algorithmic entropy is the length of a ‘sentence’ describing an object in a ‘language’. The upper bound of the amount of information embedded in any type of data, here the connectivity matrix, can be approximated by the size of compressed data compared to the size of original data. It can be simply calculated by saving the data in a standard format and then applying a data compression. The compression ratio is the size of the compressed data divided by the size of the original data in bytes. The compression ratio approaches unity when almost the same amount of information is needed to describe a network structure, whereas the ratio is close to zero when little information is needed to encode the connectivity. In biological terms, we can think of the compressed data as the genetic information, the decompression algorithm as the pattern formation mechanism that is guided through genetic factors, and the uncompressed connectivity matrix as the organization of neural systems that follows neural development.
As shown in figure 4, the amount of information in the benchmark networks decreases as modularity grows larger. The networks with locally constrained connections are easier to describe than those with many long-range connections, and thus have less information. The original networks, however, largely deviate from the curve. The values are comparable to, or even smaller than, the case of maximum modularity. This is also a consequence of the abundant connections between modules. Even when there are a considerable number of connections that are not locally confined, they can be easily described when the connections align towards similar destinations. The connection specificity of the human connectome, which is locally dense and has only a limited number of global connections between brain areas, requires less information in describing the topology. Similar observations and arguments are applied to the C. elegans connectome as well.
Figure 4.
Compression ratio as a function of modularity. The compression ratio is defined as the size of the compressed network divided by the size of the original network in bytes when the networks are represented by the adjacency matrices. It is shown for the original (vertical dashed lines) and rewired networks of human (circles, left axis) and C. elegans (squares, right axis). (Online version in colour.)
4. Discussion
Neural systems show a modular architecture at different hierarchical levels, ranging from the network of individual neurons to the network of brain regions. Observing human and C. elegans neural networks, we showed that the original networks are markedly different from the alternative benchmark networks. From both of the connectomes, we found that local information distribution is more efficient but global integration is less so by studying the clustering coefficient and the characteristic path length, respectively. We also found that metabolic costs for establishing neural connections are low, which is suggested by relatively small total wiring length. To explain these results with the connection specificity of the neural networks, we introduced the novel measure dispersion, the ratio of modules to which an individual node is connected on average. By quantifying the distribution of connections across the modules, we found that smaller dispersion is specific to the original neural networks. Third, both neural networks showed a low algorithmic entropy, which indicates lower requirements for genetic regulation to organize the architecture of neural networks.
(a). Increased separation reduces spreading and interference
Characteristic path lengths of neural networks were high compared with benchmark networks with the same modularity. Relatively high path length makes rapid spreading of activity, such as during epileptic seizures, less likely [30]. Sparse connectivity between modules can become a bottleneck for information flow. In contrast, higher connectivity between modules or merging of modules can enhance the likelihood of activity propagation—we previously described this bottleneck behaviour as topological inhibition [31]. Recent studies of functional connectivity in epilepsy indeed found a reduced path length, measured through increased global efficiency, and more connections between modules [12]. Therefore, a relatively large characteristic path length might be one of the features that supports healthy cognitive functions [32–34]. At the same time, increased neighbourhood connectivity, as measured by high clustering coefficient, renders a strong local interaction possible within a functionally related brain area or ganglion. In a similar line of argument, a study of oscillatory dynamics on neural networks has shown that the modular structure enables strong synchronization within modules and weak synchronization between them [35].
(b). Reduced dispersion decreases total wiring length
Short total wiring length reduces metabolic costs for connection establishment and at the same time limits activity propagation in neural systems [32,33,36]. For both the human and C. elegans connectome, we saw reduced dispersion of connections which is linked to the decreased total wiring length. As primate and nematode systems are close to the optimal arrangement for reducing wiring length [32,37], any re-arrangement of connections to spread more widely throughout the network will lead to the formation of longer connections in the system. A mechanism that can limit dispersion in fibre tract systems is fasciculation of axons. Fasciculation is a mechanism where a small number of pioneer neurons form pathways that guide the axons of the following neurons, resulting in a bundle of axon fibres. This might also be the case for C. elegans where some neurons in the ventral cord are formed early on [15], providing a pathway between anterior and posterior parts of the worm. The reliance on pioneer fibres might prevent more diverse connectivity to other areas located afar.
Given the relationship between dispersion and other network properties that change in schizophrenia [38], autism [39] or epilepsy [40], a reduced coherence of fibre tracts might be an important component in the path towards developmental diseases. Moreover, the dispersion might be related to changes in diffusion imaging, since a more distributed pattern of connectivity would break apart the fascicular pattern of fibre tracts. Therefore, we would expect that higher values of dispersion will be associated with lower values of fractional anisotropy (FA) and to a shift towards more regular networks with higher characteristic path length as well as clustering coefficient. For neural disorders, for example, a shift towards regular networks has been reported for epilepsy [41] and lowered FA was reported for schizophrenia [42,43]. Note, however, that lower FA might not only result from more diffuse fibre tracts within a voxel but also from reduced myelination.
(c). Development of modular neural networks
Both of the connectomes showed higher Kolmogorov complexity as measured through the compression ratio. This algorithmic entropy is different from the information theory inspired entropy, which has been applied to brain networks [44]. Kolmogorov complexity shows how much code is needed to generate an object. The generation of neural networks is the process of neural development. It can be driven by several factors, including genes, epigenetic factors and self-organization. Although we are only beginning to understand the relationship between genes and connectomes [45,46], it has been pointed out that gene expression patterns which mediate growth factors and guidance cues play an important role in determining the connectivity of neural systems [47]. However, gene expression and the inclusion of genes into the genome are costly endeavours that would be expected to be under evolutionary pressure. Indeed, neural systems try to reduce the amount of genetic encoding that is needed for neural networks. At early stages of development in C. elegans, most long-distance connections can be established when the connected neurons are still spatially adjacent [15]. This can reduce the need to control axon growth over long distances. The lower dispersion, which we found in both connectomes, might be another mechanism to reduce the amount of code required. Altogether, this suggests that the neural system might be efficient not only for the metabolic ‘running costs’ [23] but also in terms of their developmental mechanisms.
Which developmental mechanism could influence the modular organization of neural systems? Several potential biological mechanisms for generating hierarchical modular networks have been described. One way is to start with an existing network and generate copies of the network where the copies retain the same internal connectivity as the original network but also establish connections directly to the original network. Variations of this method can be used to generate hierarchical scale-free networks [48] and were also thought to lead to cortical connectivity-like networks [49]. The timing of synaptogenesis and cell birth can also be crucial for development [15,50]. For modular networks, time windows during development can lead to multiple modules where the module number, module size and inter-module connectivity are determined by the number, width and overlap of developmental time windows for synaptogenesis, respectively [51,52].
(d). Link swapping perturbs lattice structure
Neural systems can be seen as lattice networks, using two-dimensional sheets of tissue preferring to connect to nearby nodes [53] with additional long-distance shortcuts to promote rapid processing and integration of information [32]. The connections are established with geometrical constraints [54]. Previous studies have shown that lattice networks show a low compression ratio compared with other topologies [55]. During the link swapping, however, such geometrical constraints become relaxed. A rewired link can establish a new connection with any node in the module (intra-module link) or any node in the entire network (inter-module link). As a result, the lattice structure of the original network is perturbed which leads to higher dispersion. High dispersion prohibits efficient compression and the Kolmogorov complexity of the perturbed networks becomes high. It has been claimed that other measures of the neural networks, such as characteristic path length, clustering coefficient and modularity, can also be interpreted as those of regular networks [55]. This study rediscovers such findings by showing that perturbation in lattice structure makes those measures deviate from the original values.
5. Conclusion
In summary, both C. elegans and the human connectome show reduced global efficiency (higher characteristic path length), increased local efficiency (higher clustering coefficient) and reduced metabolic cost (lower total wiring length) compared with random modular networks. A marked difference in the organization of the connectomes that is relevant to those properties is their low dispersion. The specific modular organization of the connectomes requires fewer rules to construct it (lower algorithmic entropy), or fewer genetic factors to develop such a neural system. Together, these results show that neural systems across different levels, from the network of neurons to the network of brain regions, commonly have efficiencies in the multiple aspects listed above. The hierarchical natures of the modular organization of these connectomes and how they can be understood with respect to the multiple constraints given by various network measures [6,7] remain a topic for future studies.
Supplementary Material
Acknowledgment
The authors are grateful to Olaf Sporns for sharing the human connectome data.
Funding statement
This research was supported by the WCU program through the KOSEF funded by the MEST (R31-10089) and the Human Green Brain Project (http://www.greenbrainproject.org/) funded by EPSRC (EP/K026992/1).
References
- 1.Kaiser M. 2011. A tutorial in connectome analysis: topological and spatial features of brain networks. Neuroimage 57, 892–907. ( 10.1016/j.neuroimage.2011.05.025) [DOI] [PubMed] [Google Scholar]
- 2.Mackie GO. 1988. Nerve nets. In Comparative neuroscience and neurobiology (ed. LN Irwin), pp. 84–86. New York, NY: Springer. [Google Scholar]
- 3.Arendt D, Denes AS, Jekely G, Tessmar-Raible K. 2008. The evolution of nervous system centralization. Phil. Trans. R. Soc. B 363, 1523–1528. ( 10.1098/rstb.2007.2242) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.White JG, Southgate E, Thomson JN, Brenner S. 1986. The structure of the nervous system of the nematode C. elegans. Phil. Trans. R. Soc. Lond. B 314, 1–340. ( 10.1098/rstb.1986.0056) [DOI] [PubMed] [Google Scholar]
- 5.Kaiser M, Hilgetag CC, Kötter R. 2010. Hierarchy and dynamics of neural networks. Front Neuroinform. 4, 112 ( 10.3389/fninf.2010.00112) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Meunier D, Lambiotte R, Bullmore ET. 2010. Modular and hierarchically modular organization of brain networks. Front Neurosci. 4, 200 ( 10.3389/fnins.2010.00200) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Krumnack A, Reid AT, Wanke E, Bezgin G, Kötter R. 2010. Criteria for optimizing cortical hierarchies with continuous ranges. Front Neuroinform. 4, 7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Newman MEJ. 2004. Fast algorithm for detecting community structure in networks. Phys. Rev. E Stat. Nonlinear Soft Matter Phys. 69, 066133 ( 10.1103/PhysRevE.69.066133) [DOI] [PubMed] [Google Scholar]
- 9.Lim S, Han C, Uhlhaas PJ, Kaiser M. 2013. Preferential detachment during human brain development: age- and sex-specific structural connectivity in Diffusion Tensor Imaging (DTI). Cereb Cortex. ( 10.1093/cercor/bht333) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Seung HS. 2009. Reading the book of memory: sparse sampling versus dense mapping of connectomes. Neuron 62, 17–29. ( 10.1016/j.neuron.2009.03.020) [DOI] [PubMed] [Google Scholar]
- 11.DeFelipe J. 2010. From the connectome to the synaptome: an epic love story. Science 330, 1198–1201. ( 10.1126/science.1193378) [DOI] [PubMed] [Google Scholar]
- 12.Chavez M, Valencia M, Navarro V, Latora V, Martinerie J. 2010. Functional modularity of background activities in normal and epileptic brain networks. Phys. Rev. Lett. 104, 118701 ( 10.1103/PhysRevLett.104.118701) [DOI] [PubMed] [Google Scholar]
- 13.Roberts A, Conte D, Hull M, Merrison-Hort R, al Azad AK, Buhl E, Borisyuk R, Soffe SR. 2014. can simple rules control development of a pioneer vertebrate neuronal network generating behavior? J. Neurosci. 34, 608–621. ( 10.1523/JNEUROSCI.3248-13.2014) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Hagmann P, Cammoun L, Gigandet X, Meuli R, Honey CJ, Wedeen VJ, Sporns O, Friston KJ. 2008. Mapping the structural core of human cerebral cortex. PLoS Biol. 6, e159 ( 10.1371/journal.pbio.0060159) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Varier S, Kaiser M. 2011. Neural development features: spatio-temporal development of the C. elegans neuronal network. PLoS Comput. Biol. 7, e1001044 ( 10.1371/journal.pcbi.1001044) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Achacoso TB, Yamamoto WS. 1992. AY‘s Neuroanatomy of C. elegans for computation. Boca Raton, FL: CRC Press. [Google Scholar]
- 17.Watts DJ, Strogatz SH. 1998. Collective dynamics of ‘small-world’ networks. Nature 393, 440–442. ( 10.1038/30918) [DOI] [PubMed] [Google Scholar]
- 18.Humphries M, Gurney K. 2008. Network ‘small-worldness’: a quantitative method for determining canonical network quivalence. PLoS ONE 3, e0002051 ( 10.1371/journal.pone.0002051) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Guimera R, Amaral LA. 2005. Cartography of complex networks: modules and universal roles. J. Stat. Mech. 2005, P02001 ( 10.1088/1742-5468/2005/02/P02001) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Li M, Vitanyi P. 1997. An introduction to Kolmogorov complexity and its applications, 2nd edn New York, NY: Springer. [Google Scholar]
- 21.Ziv J, Lempel A. 1977. Universal algorithm for sequential data compression. IEEE Trans. Inform. Theory 23, 337–343. ( 10.1109/TIT.1977.1055714) [DOI] [Google Scholar]
- 22.Rubinov M, Sporns O. 2010. Complex network measures of brain connectivity: uses and interpretations. Neuroimage 52, 1059–1069. ( 10.1016/j.neuroimage.2009.10.003) [DOI] [PubMed] [Google Scholar]
- 23.Laughlin SB, de Ruyter Van Steveninck RR, Anderson JC. 1998. The metabolic cost of neural information. Nat. Neurosci. 1, 36–41. ( 10.1038/236) [DOI] [PubMed] [Google Scholar]
- 24.Chklovskii DB, Koulakov AA. 2004. Maps in the brain: what can we learn from them? Annu. Rev. Neurosci. 27, 369–392. ( 10.1146/annurev.neuro.27.070203.144226) [DOI] [PubMed] [Google Scholar]
- 25.Karbowski J. 2014. Constancy and trade-offs in the neuroanatomical and metabolic design of the cerebral cortex. Front Neural Circuits 8, 9 ( 10.3389/fncir.2014.00009) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Kaiser M, Varier S. 2011. Evolution and development of brain networks: from Caenorhabditis elegans to Homo sapiens. Network: Comput. Neural Syst. 22, 143–147. [DOI] [PubMed] [Google Scholar]
- 27.Johansen-Berg H. 2007. Structural plasticity: rewiring the brain. Curr. Biol. 17, R141– R144. ( 10.1016/j.cub.2006.12.022) [DOI] [PubMed] [Google Scholar]
- 28.Verhage M, et al. 2000. Synaptic assembly of the brain in the absence of neurotransmitter secretion. Science 287, 864–869. ( 10.1126/science.287.5454.864) [DOI] [PubMed] [Google Scholar]
- 29.Nykter M, Price ND, Larjo A, Aho T, Kauffman SA, Yli-Harja O, Shmulevich I. 2008. Critical networks exhibit maximal information diversity in structure–dynamics relationships. Phys. Rev. Lett. 100, 058702 ( 10.1103/PhysRevLett.100.058702) [DOI] [PubMed] [Google Scholar]
- 30.Kaiser M, Hilgetag CC. 2010. Optimal hierarchical modular topologies for producing limited sustained activation of neural networks. Front Neuroinform. 4, 8 ( 10.3389/fninf.2010.00008) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Kaiser M, Görner M, Hilgetag CC. 2007. Functional criticality in clustered networks without inhibition. New J. Phys. 9, 110 ( 10.1088/1367-2630/9/5/110) [DOI] [Google Scholar]
- 32.Kaiser M, Hilgetag CC. 2006. Nonoptimal component placement, but short processing paths, due to long-distance projections in neural systems. PLoS Comput. Biol. 2, e95 ( 10.1371/journal.pcbi.0020095) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Chen Y, Wang S, Hilgetag CC, Zhou C. 2013. Trade-off between multiple constraints enables simultaneous formation of modules and hubs in neural systems. PLoS Comput. Biol. 9, e1002937 ( 10.1371/journal.pcbi.1002937) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Vértes PE, Alexander-Bloch AF, Gogtay N, Giedd JN, Rapoport JL, Bullmore ET. 2012. Simple models of human brain functional networks. Proc. Natl Acad. Sci. USA 109, 5868–5873. ( 10.1073/pnas.1111738109) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Zhao M, Zhou C, Lü J, Lai CH. 2011. Competition between intra-community and inter-community synchronization and relevance in brain cortical networks. Phys. Rev. E 84, 016109 ( 10.1103/PhysRevE.84.016109) [DOI] [PubMed] [Google Scholar]
- 36.Cherniak C. 1994. Component placement optimization in the brain. J. Neurosci. 14, 2418–2427. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Markov NT, et al. 2014. A weighted and directed interareal connectivity matrix for macaque cerebral cortex. Cereb. Cortex 24, 17–36. ( 10.1093/cercor/bhs270) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Zalesky A, Fornito A, Seal ML, Cocchi L, Westin CF, Bullmore ET, Egan GF, Pantelis C. 2011. Disrupted axonal fiber connectivity in schizophrenia. Biol. Psychiatry 69, 80–89. ( 10.1016/j.biopsych.2010.08.022) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Courchesne E, Pierce K. 2005. Why the frontal cortex in autism might be talking only to itself: local over-connectivity but long-distance disconnection. Curr. Opin. Neurobiol. 15, 225–230. ( 10.1016/j.conb.2005.03.001) [DOI] [PubMed] [Google Scholar]
- 40.Walker MC, Daunizeau J, Lemieux L. 2011. Concepts of connectivity and human epileptic activity. Front. Syst. Neurosci. 5, 12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Ponten S, Bartolomei F, Stam C. 2007. Small-world networks and epilepsy: graph theoretical analysis of intracerebrally recorded mesial temporal lobe seizures. Clin. Neurophysiol. 118, 918–927. ( 10.1016/j.clinph.2006.12.002) [DOI] [PubMed] [Google Scholar]
- 42.Skudlarski P, Jagannathan K, Anderson K, Stevens MC, Calhoun VD, Skudlarska BA, Pearlson G. 2010. Brain connectivity is not only lower but different in schizophrenia: a combined anatomical and functional approach. Biol. Psychiatry 68, 61–69. ( 10.1016/j.biopsych.2010.03.035) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.van den Heuvel MP, Mandl RCW, Stam CJ, Kahn RS, Pol HEH. 2010. Aberrant frontal and temporal complex network structure in schizophrenia: a graph theoretical analysis. J. Neurosci. 30, 15 915–15 926. ( 10.1523/JNEUROSCI.2874-10.2010) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Sporns O. 2006. Small-world connectivity, motif composition, and complexity of fractal neuronal connections. Biosystems 85, 55–64. ( 10.1016/j.biosystems.2006.02.008) [DOI] [PubMed] [Google Scholar]
- 45.Kaufman A, Dror G, Meilijson I, Ruppin E. 2006. Gene expression of C. elegans neurons carries information on their synaptic connectivity. PLoS Comput. Biol. 2, e167 ( 10.1371/journal.pcbi.0020167) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Fornito A, et al. 2011. Genetic influences on cost-efficient organization of human cortical functional networks. J. Neurosci. 31, 3261–3270. ( 10.1523/JNEUROSCI.4858-10.2011) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.van Ooyen A, Willshaw DJ. 2000. Development of nerve connections under the control of neurotrophic factors: parallels with consumer–resource systems in population biology. J. Theor. Biol. 206, 195–210. ( 10.1006/jtbi.2000.2114) [DOI] [PubMed] [Google Scholar]
- 48.Ravasz E, Somera AL, Mongru DA, Oltvai ZN, Barabási AL. 2002. Hierarchical organization of modularity in metabolic networks. Science 297, 1551–1555. ( 10.1126/science.1073374) [DOI] [PubMed] [Google Scholar]
- 49.Krubitzer L, Kahn DM. 2003. Nature versus nurture revisited: an old idea with a new twist. Prog. Neurobiol. 70, 33–52. ( 10.1016/S0301-0082(03)00088-1) [DOI] [PubMed] [Google Scholar]
- 50.Rakic P. 2002. Neurogenesis in adult primate neocortex: an evaluation of the evidence. Nat. Rev. Neurosci. 3, 65–71. ( 10.1038/nrn700) [DOI] [PubMed] [Google Scholar]
- 51.Nisbach F, Kaiser M. 2007. Developmental time windows for spatial growth generate multiple-cluster small-world networks. Euro Phys. J. B 58, 185–191. ( 10.1140/epjb/e2007-00214-4) [DOI] [Google Scholar]
- 52.Kaiser M, Hilgetag CC. 2007. Development of multi-cluster cortical networks by time windows for spatial growth. Neurocomputing 70, 1829–1832. ( 10.1016/j.neucom.2006.10.060) [DOI] [Google Scholar]
- 53.Kaiser M, Hilgetag CC, van Ooyen A. 2009. A simple rule for axon outgrowth and synaptic competition generates realistic connection lengths and filling fractions. Cereb. Cortex 19, 3001–3010. ( 10.1093/cercor/bhp071) [DOI] [PubMed] [Google Scholar]
- 54.Henderson JA, Robinson PA. 2011. Geometric effects on complex network structure in the cortex. Phys. Rev. Lett. 107, 018102 ( 10.1103/PhysRevLett.107.018102) [DOI] [PubMed] [Google Scholar]
- 55.Sun J, Bollt EM, ben Avraham D. 2008. Graph compression—save information by exploiting redundancy. J. Stat. Mech. Theor. Exp. 06, P06001. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.