Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2012 Oct 5;7(10):e40483. doi: 10.1371/journal.pone.0040483

Analyzing Self-Similar and Fractal Properties of the C. elegans Neural Network

Tyler M Reese 1, Antoni Brzoska 1, Dylan T Yott 2, Daniel J Kelleher 1,*
Editor: Xi-Nian Zuo3
PMCID: PMC3465333  PMID: 23071485

Abstract

The brain is one of the most studied and highly complex systems in the biological world. While much research has concentrated on studying the brain directly, our focus is the structure of the brain itself: at its core an interconnected network of nodes (neurons). A better understanding of the structural connectivity of the brain should elucidate some of its functional properties. In this paper we analyze the connectome of the nematode Caenorhabditis elegans. Consisting of only 302 neurons, it is one of the better-understood neural networks. Using a Laplacian Matrix of the 279-neuron “giant component” of the network, we use an eigenvalue counting function to look for fractal-like self similarity. This matrix representation is also used to plot visualizations of the neural network in eigenfunction coordinates. Small-world properties of the system are examined, including average path length and clustering coefficient. We test for localization of eigenfunctions, using graph energy and spacial variance on these functions. To better understand results, all calculations are also performed on random networks, branching trees, and known fractals, as well as fractals which have been “rewired” to have small-world properties. We propose algorithms for generating Laplacian matrices of each of these graphs.

Introduction

Fractal theory has become an increasingly prevalent topic of both debate and research in recent years. Beginning with Mandelbrot's discussion of Britain's immeasurable coastline [1], fractal analysis has found applications in both the mathematics and scientific communities. In the geometric sense, fractals are objects that contain self-symmetry: they exhibit the same pattern on increasingly smaller scales.

More recently, fractal theory has found applications in the biological realm. Kinetics of ion channels have been modeled with fractal structures [2], [3]. Fractal dimension has been used to analyze human EEG signals [4] as well as the complex morphology of living cells [5], [6]. The applications of fractal theory in neuroscience have been a particularly prevalent topic of research [7][9]. Glial cells have been analyzed in-depth using fractal dimensions and modeling [10][12]. Dendritic branching has been shown to exhibit self-similarity [13], [14], and three-dimensional fractal structures have been used to approximate the white matter surface of the human brain, based on MRI images [15]. Nevertheless, some have warned against the possible misuses of fractal theory in neuroscience [16], [17]. In particular, calculations on fractal dimensions of biological systems have been called into question, where some studies have attempted to use this measurement as an overly-generalized tool which lacks definite relation to actual biological mechanisms.

In this paper we use a graph-theoretical approach to probe the structure of the Caenorhabditis elegans neural network for self-similarity. Similar work was done by Sporns in examining the presence of fractal patterns in neuron connectivity: in [18], fractal networks were generated and structural measures were calculated, including small-world properties, complexity, and motif composition. Advances in graph theory have proven useful in analyzing complex neural networks [19][21] as well as underlying motifs in the brain [22], [23]. In this paper we apply mathematical techniques to a physical map of the C. elegans connectome. With a well-connected component of only 279 neurons, it is an excellent candidate for graph theoretical research on a complete-brain model. While [24] presents a geometric structure of this system, our research builds upon that of [25] in which Varshney et al. propose a finalized schematic of the C. elegans neural network.

The C. elegans brain is composed of three types of neuronal cells: sensory neurons, motor neurons, and interneurons. Two types of connection exist between these neurons: chemical synapses and gap junctions. The gap junction network, which sends electrical signals via ion transport, is an undirected system. Conversely, chemical synapses possess clear directionality [25]. In this paper we study the overarching connectivity between neurons. In order to analyze the fundamental network-structure of the C. elegans neurons, we consider only the skeleton of the brain's organization. Although some neurons share multiple points of contact and chemical synapses send directional signals, we only observe that two neurons are connected. As a result, we study an undirected and unweighted network combining the chemical and gap junctions, representing only the framework of connections (See Methods). In the process, this analysis loses many of the biological details which correspond to functionality and neuron hierarchy. While such information is vital in understanding the mechanisms acting within the brain, the goal of this study is to mathematically analyze the system's structural connectivity, whereby this simplified network is sufficient.

In order to index each of these connections we use the graph Laplacian matrix, Inline graphic. For a graph, Inline graphic, we define Inline graphic as the degree of vertex Inline graphic: the number of total connections. If vertex Inline graphic is connected to vertex Inline graphic then Inline graphic and Inline graphic, where Inline graphic corresponds to the entry in the Inline graphic row and Inline graphic column. Furthermore, Inline graphic, and all other entries of matrix Inline graphic are 0. The original goal of this study was to examine the structure of the C. elegans neural network for self-similarity, and results were compared to identical calculations on other graphs. One should note that although fractal theory has repeatedly been applied in neuroscience, in studying the structure of a network the results are not as simple as saying “fractal” or “not-fractal.” Instead we search specifically for self-similar structures in the network's organization.

Results and Discussion

For the C. elegans model, we derived a Laplacian matrix from the adjacency matrices used in [25]. Algorithms were developed to produce similar matrices representing random graphs, branching trees, and rewired fractal geometries. These graphs were generated with similar properties as the C. elegans neural network, including number of vertices and probability of connection. Details on matrix generation can be found in Methods.

The Eigenvalue Counting Function

The eigenvalue counting function is a cumulative distribution function on the spectrum of a matrix, in this case the Laplacian (see Methods). Plotting this function gives an expedient way to analyze the spectrum of the graph Laplacian [26]. It is known that this function exhibits spectral “gaps” when applied to fractal geometries, corresponding to sections of slope-zero in the plot [27]. The asymptotics of this function have also been shown to be linked with heat dissipation in networks [28]. Figure 1 shows plots of the eigenvalue counting function on the Laplacian matrices.

Figure 1. Weyl Ratios.

Figure 1

(a) C. elegans neural network (b) Sierpinski Gasket, Level 5 (c) Random Tree Inline graphic (d) Hexacarpet Level 3 (e) Random Network Inline graphic (f) Sierpinski Gasket Rewiring Inline graphic.

There is a clear presence of step-like portions of those graphs corresponding to known fractals. These sections of slope-zero correspond to spectral gaps, consistent with expected results. The eigenvalue counting function plot of the C. elegans connectome (Fig. 1 (a)) does not show definitive spectral gaps, indicating that the nematode brain is not strictly fractal in structure. This, however, does not eliminate the possibility of some degree of self-similarity. Figure 1 (c) contains the graph corresponding to a random-branching tree. The large vertical jump at Inline graphic, with a change on the y-axis of approximately 200, indicates that the eigenvalue 1 occurs with extremely high multiplicity. This is caused by the nature of the tree's organization. There is a large number of endpoints: vertices at which no further branching occurs, connected only to the “parent” vertex. As the highly interconnected neural network is not tree-like, dissimilarity in observed eigenvalue counting patterns is consistent with expected results. Although the eigenvalue counting function of the C. elegans neural network does resemble those of the random network (Fig. 1 (e)) and the rewired Sierpinski Gasket (Fig. 1 (f)), this cannot conclusively point to similar structural organization, whereas drastic dissimilarity would point to fundamental differences.

Weyl Ratios

The Weyl ratio of a graph is defined as

graphic file with name pone.0040483.e018.jpg

where Inline graphic is the eigenvalue counting function. Inline graphic is determined by the logarithmic asymptotics of the eigenvalue counting function. One way to determine this Inline graphic is via a linear regression on Inline graphic when plotted on logarithmic axes. Log-log periodicity of Weyl ratios is present in fractal geometries, and has been observed in graph-approximations of fractals. Similar periodicity in Weyl ratio patterns of other networks can point to self-similarity in these graphs. For more on Weyl ratio analysis on fractals, see [29].

As expected, the Weyl ratios of known self-similar fractals show a high degree of organization. That of the Sierpinski gasket in particular (Fig. 2 (b)) shows unmistakable periodicity. It should be noted that the Weyl ratio graph of the branching-tree (Fig. 2 (c)) is again different from those of other networks, arising from a lack of what we call “looping.” In highly interconnected networks, many cyclic paths exist, allowing a signal to arrive back at a starting vertex by traveling through a series of other vertices. Trees, on the other hand, lack this feature: only one path exists between any two points, helping to create a unique Weyl ratio pattern.

Figure 2. Weyl Ratios.

Figure 2

(a) C. elegans neural network (b) Sierpinski Gasket, Level 5 (c) Random Tree Inline graphic (d) Hexacarpet Level 3 (e) Random Network Inline graphic (f) Sierpinski Gasket Rewiring Inline graphic.

While several cases of slight periodicity could be argued for, this evidence is not definitive enough to indicate self-similarity in the C. elegans neural network (Fig. 2 (a)). However, there exists similarity between the Weyl ratio patterns generated by the spectrum of the C. elegans neural network, the random network (Fig. 2 (e)), and the rewiring of the Sierpinski Gasket (Fig. 2 (f)). The significance of examining a “rewired” fractal structure will become clear later in this paper. It is important to note that this likeness in Weyl ratio patterns can suggest some structural similarity.

The Eigen-Projection Method

We replicated the network visualization performed in [25] and extended this technique to our other networks. This was done in Euclidean space via the eigen-projection method explained in [30], similar to those processes described in [31], [32]. This spectral approach to visualizing graphs utilizes the eigenfunctions of degree-normalized Laplacian matrices (see Methods). The eigen-projection method, also known as “plotting in eigenfunction coordinates,” plots the vertices of a graph using the eigenfunctions of its Laplacian matrix as a coordinate basis. See Methods for a more rigorous description.

After embedding each vertex in either 2- or 3-dimensional space, neuronal or network connections were represented with line segments between the appropriate points. In the case of the C. elegans diagram, the same color-coding as [25] was utilized: where red represents sensory neurons, green are motor neurons, and blue indicates interneurons. Lastly, points were labeled with the corresponding neuron abbreviations. This was done using a slight variation of the VISUALIZE program used by Chklovski and Varshney, available at [33].

The eigen-projection visualizations (Figure 3) allow us to make further qualitative distinctions between the C. elegans brain and other networks. In support of previous observations, it is again clear that the nematode connectome is not strictly fractal in structure. On the contrary, the eigenfunction graphs of the Sierpinski Gasket once again display characteristics expected of self-similar fractals: a high degree of ordering and self-symmetry. While the eigenvalue counting function and Weyl ratios showed little distinction between the C. elegans brain and a random graph, eigen-projections provide differentiation between the two. The random graph appears, as expected, more or less a scatter of points. The C. elegans brain, however, shows a definite structure with organized connectivity, suggesting that the C. elegans neural network is not a randomly connected system of neurons. On the other hand, the C. elegans neural network maintains its resemblance to a rewired Sierpinski gasket when plotted in eigenfunction coordinates. While there is no effective way to quantify this heuristic similarity in a relevant manner, it sustains its interest experimentally and continues to suggest the presence of some structural parallels.

Figure 3. The Eigen Projection Method.

Figure 3

(a) C. elegans neural network, Inline graphic (b) C. elegans neural network, Inline graphic (c) Sierpinski Gasket, Level 5, Inline graphic (d) Sierpinski Gasket, Level 5, Inline graphic (e) Random Network Inline graphic, Inline graphic (f) Random Network Inline graphic, Inline graphic (g) Sierpinski Gasket Rewiring Inline graphic, Inline graphic (h) Sierpinski Gasket Rewiring Inline graphic, Inline graphic.

The eigen-projections display some of the functional organization of the C. elegans neural network. It is clear that the neurons are arranged roughly by neuron type. There is a distinctive cluster of motor neurons (green), a larger sub-component of sensory neurons (red), and interneurons interspersed throughout the network (blue). This indicates that although the brain may not posses the strict self-similarity of a fractal structure, it is indeed highly organized as one would anticipate, developed for entirely functional purposes.

Small-World Network Properties

We consider two functions defined on graphs: average clustering coefficient and average path length. The clustering coefficient of a vertex Inline graphic, Inline graphic, is the probability that any two vertices neighboring Inline graphic are also connected to each other. The path length between two vertices Inline graphic and Inline graphic is the shortest path along the graph's edges connecting Inline graphic and Inline graphic (Note that this path usually travels through a number of other vertices). Using Djisktra's algorithm, it is possible to rigorously determine the shortest path between a given vertex and each other vertex on the graph. By repeating the algorithm for each node on the graph, it is possible to determine the shortest path between each pair of vertices. The average path length, Inline graphic, is calculated by finding the arithmetic mean of the shortest paths between each pair of vertices on the graph. Small-world networks are (generally) defined as networks which have a much higher Inline graphic value than random networks, but maintain a value of Inline graphic only slightly larger than that of a random network [34].

Small-world networks arise quite often in the natural sciences, as they allow for the efficient transfer of information while maintaining a certain level of complexity. There is a great deal of research which suggests that neural networks possess small-world properties [35], [36]. In fact, [25] and [34] have previously demonstrated that the C. elegans neural network is small-world in nature. Our calculations of clustering coefficient and average path length confirm these findings. As Table 1 shows, the C. elegans neural network has an average path length only slightly larger than that of its associated random network (see Methods for how these ‘associated random networks’ were constructed). At the same time, the clustering coefficient for C. elegans is six times larger than that of its associated random network, meaning the neural network of C. elegans satisfies the small-world properties as defined by [34].

Table 1. Clustering coefficient and path length.

Graph Clustering Coefficient Average Path Length
Sierpinski Gasket, Level 5 0.4495 17.3721
Random(Sierpinski Gasket) 0.0104 5.748
Sierpinski Gasket Rewire Inline graphic 0.2843 7.3833
Random(SG Rewire) 0.0104 5.748
C. elegans Neural Network 0.3371 2.5377
Random(C. elegans Neural Network) 0.0581 2.3458

This motivated our work with network-rewiring, related to that done by Watts and Strogatz. In [34] they showed that moving connections in an ordered network, with a certain probability Inline graphic, led to some interesting changes in graph structure. Namely, when Inline graphic is small, a slight increase in Inline graphic causes a large drop in Inline graphic but does not change Inline graphic appreciably: the network takes on small-world characteristics. Intuitively this can be explained by the fact that these sparse random connections don't change a graph's strong localized structure, but it becomes easier to travel long distances via these new connections which can span large gaps. It is clear from Table 1 that the Sierpinski Gasket does not possess small-world characteristics. This supports the propositions of [37], which describes the existence of a dichotomy between fractal structures and small world networks. As a result, the Laplacian matrix of a small-world rewiring of the Sierpinski Gasket was included throughout this study. Sporns used a similar method in [18] by generating fractal connections and “rewiring” such networks for small-world properties in studying complexity and self similarity in neuron connectivity.

Energies and Spacial Variances

Using an eigenfunction of a graph's Laplacian, Inline graphic, one can calculate a graph energy specific to Inline graphic. Knowing the resistance between any two vertices and a constant Inline graphic, these can be used to calculate the spacial variance of Inline graphic (See Methods). Variance is a measure of how localized a function is, and functions which possess a low spacial variance are said to be localized. In particular, localization occurs when an eigenfunction is approximately zero except for in a a small (localized) region: that is, the eigenfunction takes on most of its values inside a small number of connected regions on the graph. Localized eigenfunctions are known to be present in fractals but not in Euclidean or other smooth spaces [38]. Figure 4 shows distributions of the spacial variance of all eigenfunctions on the graphs used previously.

Figure 4. Spacial Variance.

Figure 4

(a) C. elegans neural network (b) Random Graph (Inline graphic) (c) Sierpinski Gasket, Level 5 (d) Sierpinski Gasket Rewiring (Inline graphic).

Of the four graphs considered here, the eigenfunctions of the random network possess the largest spacial variances. As this system was designed to lack general organization, non-localized eigenfunctions were both expected and observed. The eigenfunctions of the Sierpinski Gasket possess both highly concentrated and low-valued spacial variances. Such trends correspond to a high degree of localization, as anticipated in approximations of fractal geometries. Eigenfunctions of the rewired Sierpinski Gasket demonstrate a similar concentration pattern with slightly higher spacial variance, indicating a slightly lower degree of eigenfunction localization.

Of particular interest are the spacial variances of eigenfunctions of the C. elegans neural network. Although these spacial variances do not show the same degree of concentration as those of the Sierpinski Gasket, the values of these variances are a power of 10 less than those of the fractal network. Whereas spacial variances on eigenfunctions of the Sierpinski Gasket are concentrated around Inline graphic the majority of eigenfunctions of the neural network lie below Inline graphic. These comparatively lower spacial variances indicate a high level of localization in the eigenfunctions. Such highly localized eigenfunctions can indicate the presence of self-similarity in the network, although further tests would be required to determine the absolute origin of this localization.

In [25] sparsity of eigenfunctions suggests the presence of subcircuits with a specific function. As localization suggests that the value of of an eigenfunction is concentrated on a few vertices, it is reasonable that localized eigenfunction may be used in place of sparse ones. In light of this, the existence of localized eigenfunctions is not entirely unexpected. However [25] looks for sparsity in eigenfunctions of the gap junction network only.

Conclusions

In this paper we used a variety of graph theoretic and mathematical techniques to probe the structural framework of the C. elegans connectome. To better understand results, calculations were also performed on random networks, finite approximations of fractal geometries, and small-world “rewired” graphs. This study confirmed previous results, demonstrating that the neural network exhibits small-world characteristics. Furthermore, the network has highly localized eigenfunctions, which could suggest the presence of self-similar structural motifs. Further research would be required to determine the nature of this localization. Although the C. elegans neural network is not random, tree-like, nor fractal in structure, it is certainly highly ordered, aiding in functional efficiency of the system. Although C. elegans has proven to be a useful model organism, with a well-defined map of its neural network, this network consists of only 279 nodes. While this makes the system fairly efficient to study computationally, this small number of nodes makes network-analysis rather limited. Although ideally a much larger map would be used, due to the difficulty in determining the exact layout of each neuron in a network, few consistent complete-brain maps exist at this time. In the end, this paper presents a type of analytic “toolbox,” offering many mathematical techniques which can be used to search for structural self-similarity within networks. In particular, the tools presented here can be used to study higher-order brains as more complex neural networks become well-understood.

Methods

In order to analyze only the framework of the C. elegans neural network, we constructed a Laplacian matrix derived form the adjacency matrices in [33]. The network of chemical synapses sends signals in one direction only, resulting in a non-symmetric adjacency matrix, Inline graphic. To disregard this directionality, we added this matrix to its own transpose, Inline graphic, creating a symmetric matrix indexing all chemical connections. We added this matrix to the adjacency matrix of the gap junction system, Inline graphic (already symmetric as these connections are bidirectional).

graphic file with name pone.0040483.e065.jpg

All non-zero entries of this combined matrix, Inline graphic, were normalized to be 1, avoiding multiplicity of connection, resulting in matrix Inline graphic.

graphic file with name pone.0040483.e068.jpg

It is then simple to produce a Laplacian matrix, Inline graphic, as shown below:

graphic file with name pone.0040483.e070.jpg

Note that Inline graphic is the degree of each vertex Inline graphic. The degree matrix, Inline graphic, is now defined as:

graphic file with name pone.0040483.e074.jpg

Then the Laplaican matrix, Inline graphic, is given by:

graphic file with name pone.0040483.e076.jpg

Generating Random Graphs and Trees

In order to generate a Laplacian matrix representation of the random graphs, we used the following algorithm:

First, fix the number of vertices, Inline graphic, and the probability of connection, Inline graphic, and construct an empty Inline graphic matrix, Inline graphic.

For each Inline graphic such that Inline graphic assign a random value Inline graphic such that Inline graphic for all Inline graphic. If Inline graphic then Inline graphic, otherwise Inline graphic.

To produce an adjacency matrix of this graph, Inline graphic, add Inline graphic to its own transpose:

graphic file with name pone.0040483.e091.jpg

Using this adjacency matrix, construct a Laplacian matrix using the method described previously.

The algorithm used for producing the Laplacian matrix of a random-branching tree is more involved. Again, fix the number of vertices, Inline graphic, and also specify the maximum number of “children” from any given branch-point, Inline graphic. Create an empty Inline graphic matrix, Inline graphic

Begin by generating a random integer Inline graphic such that Inline graphic, and take Inline graphic. This corresponds to the first vertex having Inline graphic branches. To represent these branches in the matrix, take Inline graphic for Inline graphic and Inline graphic for Inline graphic.

Next move to all subsequent vertices. Because no “looping” exists in the structure of the tree, each node can only be connected to its parent vertex and its “children” vertices. We take Inline graphic Then Inline graphic, where Inline graphicmin(Inline graphic) is the smallest-labeled node which does not have a parent vertex, i.e. the first column with all 0 entries corresponds to the first point not yet connected. (Note in the case of vertex 2, Inline graphic). This vertex Inline graphic is the first “offspring” from the next branch-point.

Now, as above, for each remaining vertex Inline graphic we choose another random integer, Inline graphic, such that Inline graphic and take Inline graphic. (Note that vertex Inline graphic has Inline graphic children, however Inline graphic is the degree of node Inline graphic, taking into account its parent-connection). To represent the “offspring” branches of this vertex Inline graphic, use the following:

graphic file with name pone.0040483.e119.jpg

and

graphic file with name pone.0040483.e120.jpg

Use minInline graphic when choosing Inline graphic to avoid adding more vertices than the Inline graphic which was originally fixed.

The Eigenvalue Counting Function and Weyl Ratios

For a given graph Laplacian matrix, Inline graphic, the eigenvalue counting function, Inline graphic is a cumulative frequency function on the spectrum of the matrix where:

graphic file with name pone.0040483.e126.jpg

The growth of Inline graphic is approximately Inline graphic, thus the relevant portion of each graph, when using a logarithmic scale, appears linear. A linear regression was found for each relevant interval, and the slope, Inline graphic, calculated. Using this Inline graphic, we plotted the Weyl ratio, Inline graphic, such that:

graphic file with name pone.0040483.e132.jpg

These Weyl ratios allow us to examine the spectrum of each matrix, looking for elements such as symmetry and periodicity [29].

Normalizing a Laplacian Matrix

We used two different forms of the graph-Laplacian matrix: the standard Laplacian and the degree-normalized Laplacian. In the case of eigen-projections, we utilize the degree-normalized matrix. We define the degree matrix, Inline graphic, as before: a diagonal matrix whose non-diagonal elements are 0, and each entry Inline graphic is the degree of the Inline graphic vertex. Using both the standard Laplacian, Inline graphic and its corresponding degree matrix, Inline graphic, produce the degree-normalized Laplacian, Inline graphic:

graphic file with name pone.0040483.e139.jpg

Using the degree-normalized Laplacian has many aesthetic advantages, as shown in Figure 5. The normalized matrix also has all eigenvalues Inline graphic such that Inline graphic.

Figure 5. Normalizing the Laplacian.

Figure 5

C. elegans neural network: (a) Un-normalized Laplacian, Inline graphic (b) Normalized Laplacian Inline graphic.

Graphing in Eigenfunction Coordinates

We found all eigenvalues, Inline graphic, and their corresponding eigenfunctions, Inline graphic, for each matrix. Given two eigenfunctions Inline graphic and Inline graphic, (such that Inline graphic) we then plotted the ordered pair Inline graphic for each Inline graphic from 1 to 279, as described in [39]. The first eigenvalue of any Laplacian matrix is always 0, corresponding to a constant eigenfunction. Thus we only consider Inline graphic and Inline graphic with Inline graphic. Edges were then added between points to represent relevant connections, and the same color-coding as [25] was used. The same process was then repeated in three dimensions, plotting Inline graphic for some Inline graphic, such that Inline graphic.

Clustering Coefficient

The clustering coefficient measures the probability that two neighbors of a given vertex are also connected to one another. For a graph Inline graphic and a given vertex Inline graphic, let Inline graphic denote the number of connections that exist between the neighbors of Inline graphic. Take Inline graphic as the number of neighbors of Inline graphic (the degree of vertex Inline graphic). Then the clustering coefficient of vertex Inline graphic, Inline graphic, is given by:

graphic file with name pone.0040483.e166.jpg

Note that total number of possible connections among neighbors of Inline graphic is Inline graphic.

For a graph Inline graphic with Inline graphic vertices, the average clustering coefficient, Inline graphic, is defined as:

graphic file with name pone.0040483.e172.jpg

Generating a Related Random Graph for Small-World Analysis

In order to analyze our networks for small-world properties, it was useful to compare these graphs to those of similar networks with randomly assigned edges. Small-world networks are nearly as well-connected as random graphs, but possess a well-localized structure. We developed the following algorithm for this process:

For a graph Inline graphic with Inline graphic vertices, let Inline graphic be the number of edges on Inline graphic. The average number of edges per vertex is Inline graphic. Furthermore, the probability that any two random vertices are connected, Inline graphic, is given by the number of existing connections divided by the total possible connections:

graphic file with name pone.0040483.e179.jpg

Next we generate a random graph, Inline graphic, with Inline graphic vertices and a Inline graphic probability of connection between two vertices (See Methods). We then compute Inline graphic and Inline graphic for Inline graphic and Inline graphic.

Graph Rewiring

First number each vertex in Inline graphic from 1 to Inline graphic, the total number of vertices. If there is a connection between vertices Inline graphic and Inline graphic in Inline graphic, we generate a random number between 0 and 1. If this random number is less than a given probability Inline graphic, then the connection will be rewired. Without loss of generality assume Inline graphic. We then fix the connection to vertex Inline graphic, and move the connection to another vertex, Inline graphic, such that Inline graphic and Inline graphic are now connected whereas they were not previously.

Graph Energy

For a graph Inline graphic) where Inline graphic is the set of vertices and Inline graphic is the set of edges, one can define an arbitrary scalar function Inline graphic. The energy of Inline graphic associated with Inline graphic, Inline graphic is defined as:

graphic file with name pone.0040483.e205.jpg

We analyzed the energies of the Laplacian matrix eigenfunctions, thus Inline graphic.

Spacial Variance

In order to discuss spacial variance, we must first define the resistance between two vertices on a graph. Let Inline graphic be a graph and Inline graphic. Then the resistance between Inline graphic and Inline graphic, Inline graphic, is given by:

graphic file with name pone.0040483.e212.jpg

Where Inline graphic is a harmonic function defined as follows:

Let Inline graphic be a graph and Inline graphic. Then the harmonic function corresponding to Inline graphic is a scalar function Inline graphic such that:

1.

graphic file with name pone.0040483.e218.jpg

2.

graphic file with name pone.0040483.e219.jpg
  1. Inline graphic.

Finding the harmonic function is equivalent to finding a vector Inline graphic such that Inline graphic, where Inline graphic is a vector whose entries are all 0 except for those entries corresponding to Inline graphic and Inline graphic. This is analogous to what “harmonic” means in Euclidean space. This changes the maximization problem in condition 3 to solving a system of linear equations.

Using these we can now define the spacial variance of a graph. Again, let Inline graphic be a graph with Inline graphic vertices and Inline graphic be a scalar function on Inline graphic. Let Inline graphic be a constant. Then the Inline graphic spacial variance of Inline graphic over Inline graphic is given by:

graphic file with name pone.0040483.e234.jpg

In this paper, the spacial variances of eigenfunctions of Laplacian matrices were evaluated at Inline graphic.

Acknowledgments

We thank Dr. Alexander Teplyaev, Department of Mathematics, University of Connecticut, for his guidance in organizing and overseeing our project. We thank Dr. Dmitri Chklovskii, Howard Hughes Medical Institute, Janiela Farm Research Campus, for allowing us to extend his work and use his adjacency matrices. We thank Matthew Begué, Department of Mathematics, University of Maryland, for his help with our MATLAB code.

Funding Statement

Research was supported in part by NSF (National Science Foundation) grant DMS-0505622. No additional external funding was received for this study. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Mandelbrot BB (1967) How long is the coast of Britain? Statistical self-similarity and fractal dimension. Science 156: 636–638. [DOI] [PubMed] [Google Scholar]
  • 2. Liebovitch LS, Fischbarg J, Koniarek JP, Todorova I, Wang M (1987) Fractal model of ion-channel kinetics. Biochim Biophys Acta 896: 173–180. [DOI] [PubMed] [Google Scholar]
  • 3. Lowen SB, Liebovitch LS, White JA (1999) Fractal ion-channel behavior generates fractal firing patterns in neuronal models. Phys Rev E Stat Phys Plasmas Fluids Relat Interdiscip Topics 59: 5970–5980. [DOI] [PubMed] [Google Scholar]
  • 4. Paramanathan P, Uthayakumar R (2008) Application of fractal theory in analysis of human elec-troencephalographic signals. Comput Biol Med 38: 372–378. [DOI] [PubMed] [Google Scholar]
  • 5. Bernard F, Bossu JL, Gaillard S (2001) Identification of living oligodendrocyte developmental stages by fractal analysis of cell morphology. J Neurosci Res 65: 439–445. [DOI] [PubMed] [Google Scholar]
  • 6. Smith TG, Lange GD, Marks WB (1996) Fractal methods and results in cellular morphology–dimensions, lacunarity and multifractals. J Neurosci Methods 69: 123–136. [DOI] [PubMed] [Google Scholar]
  • 7. Fernandez E, Jelinek HF (2001) Use of fractal theory in neuroscience: methods, advantages, and potential problems. Instituto de Bioingenier ia 24: 309–321. [DOI] [PubMed] [Google Scholar]
  • 8. Kiselev VG, Hahn KR, Auer DP (2003) Is the brain cortex a fractal? Neuroimage 20: 1765–1774. [DOI] [PubMed] [Google Scholar]
  • 9. Werner G (2010) Fractals in the nervous system: conceptual implications for theoretical neuro-science. Front Physiol 1: 15. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Smith TG, Behar TN (1994) Comparative fractal analysis of cultured glia derived from optic nerve and brain demonstrate different rates of morphological differentiation. Brain Res 634: 181–190. [DOI] [PubMed] [Google Scholar]
  • 11. Smith TG, Behar TN, Lange GD, Marks WB, Sheriff WH (1991) A fractal analysis of cultured rat optic nerve glial growth and differentiation. Neuroscience 41: 159–166. [DOI] [PubMed] [Google Scholar]
  • 12. Reichenbach A, Siegel A, Senitz D, Smith TG (1992) A comparative fractal analysis of various mammalian astroglial cell types. Neuroimage 1: 69–77. [DOI] [PubMed] [Google Scholar]
  • 13. Caserta F, Stanley HE, Eldred WD, Daccord G, Hausman RE, et al. (1990) Physical mechanisms underlying neurite outgrowth: A quantitative analysis of neuronal shape. Phys Rev Lett 64: 95–98. [DOI] [PubMed] [Google Scholar]
  • 14. Bieberich E (2002) Recurrent fractal neural networks: a strategy for the exchange of local and global information processing in the brain. BioSystems 66: 145–164. [DOI] [PubMed] [Google Scholar]
  • 15. Free SL, Sisodiya SM, Cook MJ, Fish DR, Shorvon SD (1996) Three-dimensional fractal analysis of the white matter surface from magnetic resonance images of the human brain. Cereb Cortex 6: 830–836. [DOI] [PubMed] [Google Scholar]
  • 16. Jelinek HF, Fernández E (1998) Neurons and fractals: how reliable and useful are calculations of fractal dimensions? J Neurosci Methods 81: 9–18. [DOI] [PubMed] [Google Scholar]
  • 17. Murray J (1995) Use and Abuse of Fractal Theory in Neuroscience. The Journal of Comparative Neurology 361: 369–371. [DOI] [PubMed] [Google Scholar]
  • 18. Sporns O (2006) Small-world connectivity, motif composition, and complexity of fractal neuronal connections. BioSystems 85: 55–64. [DOI] [PubMed] [Google Scholar]
  • 19. Bullmore E, Sporns O (2009) Complex brain networks: graph theoretical analysis of structural and functional systems. Nat Rev Neurosci 10: 186–198. [DOI] [PubMed] [Google Scholar]
  • 20. Stam CJ, Reijneveld JC (2007) Graph theoretical analysis of complex networks in the brain. Non-linear Biomed Phys 1: 3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Fallani FV, Costa LF, Rodriguez FA, Astolfi L, Vecchiato G, et al. (2010) A graph-theoretical approach in brain functional networks. Possible implications in EEG studies. Nonlinear Biomed Phys 4 Suppl 1: S8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Sporns O, Kotter R (2004) Motifs in brain networks. PLoS Biol 2: 1910–1918. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Itzkovitz S, Alon U (2005) Subgraphs and network motifs in geometric networks. Phys Rev E Stat Nonlin Soft Matter Phys 71: 026117. [DOI] [PubMed] [Google Scholar]
  • 24. Morita S, Oshio Ki, Osana Y, Funabashi Y, Oka K, et al. (2001) Geometrical structure of the neuronal network of Caenorhabditis elegans. Physica A 298: 553–561. [Google Scholar]
  • 25. Varshney LR, Chen BL, Paniagua E, Hall DH, Chklovskii DB (2011) Structural Properties of the Caenorhabditis elegans Neuronal Network. PLoS Comput Bio 7 7: e1001066. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Das KC (2004) The Laplacian spectrum of a graph. Comput Math Appl 48: 715–724. [Google Scholar]
  • 27.Zhou D (2008) Spectral analysis of Laplacians on certain fractals. ProQuest LLC, Ann Arbor, MI, 109 pp. Thesis (Ph.D.)–University of Waterloo (Canada).
  • 28. Begue M, DeValve L, Miller D, Steinhurst B (2009) Spectrum and Heat Kernel Asymptotics on General Laakso Spaces. Fractals (to appear) [Google Scholar]
  • 29. Berry T, Heilman SM, Strichartz RS (2009) Outer approximation of the spectrum of a fractal Laplacian. Experiment Math 18: 449–480. [Google Scholar]
  • 30. Koren Y (2005) Drawing graphs by eigenvectors: theory and practice. Comput Math Appl 49: 1867–1888. [Google Scholar]
  • 31.Mohar B (1991) The Laplacian spectrum of graphs. In: Graph theory, combinatorics, and applications. Vol. 2 (Kalamazoo, MI, 1988), New York: Wiley, Wiley-Intersci. Publ. pp. 871–898.
  • 32. Pisanki T, Shawe-Taylor J (2000) Characterizing Graph Drawing with Eigenvectors. J Chem Inf Comput Sci 40: 567–571. [DOI] [PubMed] [Google Scholar]
  • 33.Varshney L, Chen B, Paniagua E, Hall D, Chklovskii D (2011) Structural properties of the Caenorhabditis elegans neuronal network. Available: http://mitedu/lrv/www/elegans/. [DOI] [PMC free article] [PubMed]
  • 34. Watts DJ, Stogatz SH (1998) Collective dynamics of ‘small-world’ networks. Nature 393: 440–442. [DOI] [PubMed] [Google Scholar]
  • 35. Bassett DS, Meyer-Lindenberg A, Achard S, Duke T, Bullmore E (2006) Adaptive reconfiguration of fractal small-world human brain functional networks. Proc Natl Acad Sci USA 103: 19518–19523. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Sporns O, Honey CJ (2006) Small worlds inside big brains. Proc Natl Acad Sci 103: 19219–19220. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Csányi G, Szendröi B (2004) Fractal-small-world dichotomy in real-world networks. Phys Rev E Stat Nonlin Soft Matter Phys 70: 016122. [DOI] [PubMed] [Google Scholar]
  • 38. Okoudjou KA, Saloff-Coste L, Teplyaev A (2008) Weak uncertainty principle for fractals, graphs and metric measure spaces. Trans Amer Math Soc 360: 3857–3873. [Google Scholar]
  • 39. Begue M, Kelleher DJ, Nelson A, Panzo H, Pellico R, et al. (2011) Random walks on barycentric subdivisions and the Strichartz hexacarpet. ArXiv e-prints [Google Scholar]

Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES