Skip to main content
National Science Review logoLink to National Science Review
. 2021 Jun 10;9(3):nwab102. doi: 10.1093/nsr/nwab102

Less is more: wiring-economical modular networks support self-sustained firing-economical neural avalanches for efficient processing

Junhao Liang 1,b, Sheng-Jun Wang 2,b, Changsong Zhou 3,4,5,
PMCID: PMC8962757  PMID: 35355506

Abstract

The brain network is notably cost-efficient, while the fundamental physical and dynamic mechanisms underlying its economical optimization in network structure and activity have not been determined. In this study, we investigate the intricate cost-efficient interplay between structure and dynamics in biologically plausible spatial modular neuronal network models. We observe that critical avalanche states from excitation-inhibition balance under modular network topology with less wiring cost can also achieve lower costs in firing but with strongly enhanced response sensitivity to stimuli. We derive mean-field equations that govern the macroscopic network dynamics through a novel approximate theory. The mechanism of low firing cost and stronger response in the form of critical avalanches is explained as a proximity to a Hopf bifurcation of the modules when increasing their connection density. Our work reveals the generic mechanism underlying the cost-efficient modular organization and critical dynamics widely observed in neural systems, providing insights into brain-inspired efficient computational designs.

Keywords: neural network, cost efficiency, critical avalanche, modular network, mean-field theory


This modelling, computational and theoretical study offers a fundamental understanding of the remarkable cost efficiency in neural systems that surprisingly achieves much more functional value with much less wiring and firing costs.

INTRODUCTION

The interplay between the structure and dynamics of complex networked systems is a long-standing area of investigation, covering applications in complex systems from diverse scientific fields. Currently, research on this topic and applications in the brain and neuroscience are experiencing rapid growth.

Neurons in the human brain form a very huge and complex dynamic network for efficient functional processing with remarkable cost efficiency. The principles underlying its efficiency have been actively studied in recent years, either from a structural or dynamic aspect.

The brain network is very sparse globally: ∼100 billion neurons with ∼1014 synaptic connections each so that the overall density is ∼10–8 in the human brain [1]. However, the overall low-density connectivity is organized in a hierarchical manner from local circuits and cortical sheets to the whole-brain connectome [2,3]. Thus, a prominent feature of brain organization is that it is globally sparse with hierarchical, relatively dense, modular architectures [4–6], which is economical in network wiring since most of the connections are in the short range. There is ample evidence that brain networks can achieve local wiring cost minimization from the brain structure [7,8] and a trade-off between global wiring cost and processing efficiency [9,10].

Brain activities consume a low energy power of only 20 W, which is remarkably energy efficient when compared to digital computers [11]. Dynamically, the irregular and sparse firing of neurons [12] can be collectively organized as oscillations and critical avalanches across different scales [13–16]. Such ‘scale-free’ dynamic activities were originally explained by critical branching theory [13], in which critical avalanches emerge near the transition point between a silent and an overactive phase. Later experimental evidence [14,17] supports that the transition point between an asynchronous and a synchronous phase better explains the observed critical avalanches, especially in terms of the existence of different critical exponents [14,18,19] satisfying scaling relations [20]. Functionally meaningful avalanche dynamics in critical synchronous transition states enable neurons to fire at a low rate [21]. Since cortical metabolic energy usage is dominated by action potentials and synaptic transmission [22–26], the avalanche dynamic is also energy economical to maintain the sustained spontaneous (resting) state, which consumes the majority of brain metabolic cost [27]. Finally, critical states are functionally beneficial by providing a broad dynamic range in response to stimulations [28,29] and thus a sensitive standby state for the brain to respond to constantly changing environments [30].

Although it is recognized that metabolic cost is a unifying principle governing neuronal biophysics [31], the fundamental mechanism underlying the economical interaction between structures and dynamic modes at the neural circuit level is not well understood. Specifically, how do the modular network (MN) structure and critical dynamics jointly achieve structural and dynamical optimization for energy-efficient processing? Deciphering these mechanisms is also important for developing brain-inspired efficient computing. Here, we address these questions with a biologically realistic neural dynamic model of excitation-inhibition (E–I)-balanced [32,33] spiking neuronal networks clustered in two-dimensional (2D) space to represent the resting-state dynamics on a cortical sheet composed of microcolumns. Interestingly, when rewiring the initial globally sparse random network (RN) into locally dense MNs, the firing rates decrease, the self-sustained dynamics change from asynchronous states to critical avalanche states, and the response sensitivity to weak transient external stimuli is greatly enhanced. Theoretically, we reveal the enhanced response of neurons by clustered firing and elucidate the dynamic transition via a Hopf bifurcation induced by denser connections within modules during rewiring through a novel mean-field theory. Overall, our integrative study of cost-structure-dynamics-function relationships in neural networks finds that locally dense connectivity under E–I-balanced dynamics appears to be the key ‘less-is-more’ solution to achieving cost-efficient organization.

RESULTS

Dynamic transition from RN to MN

We study a model of Inline graphicneurons spread on a 2D plane. Considering that other tissues, such as vessels, may separate microcolumns of neurons, neurons are randomly placed on Inline graphic square regions (modules). Modules are separated by blank space (Fig. 1 top), and each of the modules contains 500 neurons (80% excitatory and 20% inhibitory). Initially, we construct an RN by randomly connecting each neuron pair with a probability Inline graphic ( Inline graphic in the main text). To build an MN, inter-modular links are rewired, with a probability Inline graphic, into the same module to become intra-modular links. The rewiring method is equivalent to constructing small-world networks, an essential feature of brain networks [34]. Here, the essential structural property captured in our model is that the network consists of coupled modules embedded in space [3,5,9]. The voltage (membrane potential) Inline graphic of a neuron in the E–I network is governed by conductance-based (COB) leaky integrate-and-fire (IF) dynamics [35],

graphic file with name TM0007.gif (1)

where Inline graphic are the membrane time constant, resting (leaky) potential, and excitatory and inhibitory reversal potential, respectively. When a neuron receives a spike from an E, I neuron, its E, I conductance Inline graphic,Inline graphic is changed as Inline graphic, Inline graphic, respectively, followed by exponential decay, Inline graphic and Inline graphic. The network does not receive other external inputs. To launch the network activity, Gaussian white noise (GWN) is added to Equation (1) in the initial 200 ms and then removed. Then, we study the properties of its self-sustained dynamics without any external inputs. Details of the model parameters and simulation methods are provided in Supplementary Notes II. Neural Dynamics.

Figure 1.

Figure 1.

Diagram of the model neural network architecture. Neurons are placed in separated square modules in 2D space (top), mimicking the cortical sheet. Two network wiring patterns are distinguished: globally random topology (left) with equally dense intra- and inter-module coupling and modular topology (right) with dense intra-module coupling but sparse inter-module coupling. In this illustration of the model, the strength of coupling denotes the number of connections.

As the initial RN is rewired into the MN, it is interesting to find that the spike-generating and spike-transmission costs of the network are significantly decreased by orders of magnitude (Fig. 2(a)). Here, the spike-generating cost is defined as the average firing rate, and the spike-transmission cost of a neuron is defined as the product of its firing rate and the total length of its outgoing synapses (the average spike-transmission cost shown in Fig. 2(a) is normalized by the value of RN). These two measures for running costs mimic the energy cost for generating spikes and transmitting spikes [36,37]. In addition, since many links become local short-range connections, the normalized wiring cost (defined as the total Euclidian length of all links, normalized by the value of RN) is decreased by orders of magnitude, and the connection density within modules increases, approaching a value of Inline graphic (Inline graphic times the whole network density Inline graphic, refer to Equation (4) below). Such smaller wiring length is desirable, as the reduced membrane areas of the fibres can reduce metabolic cost and reduce the transmission delay (although synaptic delay is not considered in our model for simplicity). The wiring cost reduction is more pronounced for larger networks with more modules (see Fig. S1 and detailed analysis in Supplementary Notes I. Network Setting). Thus, MN structures reduce both the wiring and running costs.

Figure 2.

Figure 2.

Wiring-economical modular networks support firing-economical avalanches and greatly enhance response sensitivity. (a) Spontaneous dynamic properties of the network during rewiring. From top to bottom: the average firing rate representing spike-generating cost; spike-transmission cost; self-sustained probability Inline graphic; current ratio Inline graphic; CV of activity; and the MSD of avalanche distribution from power-law function. (b) Topological properties of the network during rewiring: normalized wiring cost of the whole network and connection density within a module. (c) Stimulus-response properties. The upper two panels show the response size of the membrane potential and firing rate during rewiring, where the stimulus strength is 1%. The lower two panels compare the responses of RN (Inline graphic) and MN (Inline graphic) under different stimulus strengths.

The initial large and sparse RN can self-sustain asynchronous activity without external input, giving a sustained probability Inline graphic, which is maintained as the network is rewired into the MN (Fig. 1). This self-sustained activity [38] resembles the resting states of the brain and thus may play a functional role. The results are similar when the overall connection density Inline graphic changes (Fig. S4). However, a denser MN (larger Inline graphic) with too weak inter-modular connections (Inline graphic) may not maintain self-sustained activities (Fig. S4). This breakdown of self-sustainability can be understood later by dynamic analysis of a separate module.

Interestingly, the dynamic modes of networks also covary during rewiring. RNs exhibit a classical E–I balanced asynchronous state with Poisson-like neuronal spiking [32]. We measure the balance by the net synaptic input current rescaled by the excitatory synaptic current (Inline graphic) averaged over time and neurons. It maintains 0 when an RN is changed into an MN (Fig. 2(a)), suggesting the maintenance of overall balance. The asynchronous state in the RN has a noisy fluctuation of the mean voltages around an equilibrium value (Fig. 3(a), upper panel). Spikes in the MN are clustered yet preserve irregular features and are interrupted by temporally silent periods (refer to Fig. S3(d) for raster plots of the spiking time in a module in an MN), exhibiting temporal dynamic variability in mean voltages (Fig. 3(a), lower panel), which can be measured from the CV (coefficient of variability, defined as standard deviation over absolute value of the mean) of the mean voltages of modules in each millisecond (Fig. 2(a)).

Figure 3.

Figure 3.

Dynamic comparison between RN (Inline graphic) and MN (Inline graphic). (a) Spontaneous activity (the average membrane potential) in a module. (b) The distributions of avalanche size Inline graphic, avalanche duration Inline graphic and average size Inline graphic given duration Inline graphic. The upper/lower rows are the results for RN and MN, respectively. The green lines on top of the avalanche size and duration distribution under critical dynamics indicate the ranges of estimated power-law distributions. (c) Trial-averaged mean membrane potential (left) and mean firing rate (right) of a module, with transient stimuli (strength 1%) applied at the time marked by the arrows.

Importantly, MNs support critical neuronal avalanches in modules. Here, the time bin for measuring avalanches is the average inter-spike interval (ISI) of the merged spiking train [13] in a module. When rewiring an RN into a strong MN (e.g. Inline graphic = 0.995), the avalanche size and duration distribution of a module changes from exponential decay to a power-law (Fig. 3(b)). Statistical tests and estimations of critical exponents are made by an accustomed truncation algorithm [39]. Power-law avalanche size and duration distributions Inline graphicInline graphic and Inline graphic (Inline graphic, Inline graphic, Inline graphic with Inline graphic value Inline graphic) are found in the truncated ranges, where scaling relation Inline graphic[20] approximately holds (error ∼0.15). The size distribution is fitted into a power-law function [39], and its mean square deviation (MSD) from the fitted curve in Fig. 2(a) bottom shows that modules in an MN with Inline graphic have avalanches with power-law distributions, exhibiting features of criticality. Other measurements of avalanches in MNs associated with the threshold of the average membrane potential are presented in Fig. S5, which also exhibits power-law distributions. This transition from asynchronous spiking to critical avalanche dynamics is the approach to a continuous synchronous transition point, as seen from the increase in the CV of activity (Fig. 2(a)). The self-sustained activity of coupled modules in the critical states provides the ideal scheme in which networks can work with a low firing rate. The reduced firing rate at criticality is a feature of the critical synchronous transition model [21], whereas the traditional branching process model does not exhibit this property—the firing rate of branching processes at critical states should be larger than that at subcritical states.

Critical states also induce greatly enhanced response sensitivity to transient stimuli. Here, the stimulus is modelled by raising the voltage Inline graphic to Inline graphic mV of a proportion Inline graphic of the neurons in all modules. We callInline graphic the stimulus strength. These neurons driven above the firing threshold emit spikes immediately, similar to optogenetic stimulation in experiments. The response sensitivity of a system can be reflected by the returning process of a signal to its baseline value after a transient perturbation. We measured the response in membrane potential and firing rate of the network modules. The size of the response is defined as Inline graphic, which is the area between the signal Inline graphic (the trial-average voltage or firing rate of the network) and its resting valueInline graphic, within a window of Inline graphic, beginning from stimulus onset at Inline graphic (see also details in Supplementary Notes II. Neural Dynamics). As shown by the stimulus-induced trial-average voltage and firing rate of a module (Fig. 3(c)), the response of MNs is much larger and more pronounced than that of RNs. Interestingly, MNs show a stronger damped oscillation-like response pattern, which is a characteristic of event-related potentials in electroencephalogram signals of the brain's response to stimuli [40]. Importantly, both in membrane potential and in firing rate, MNs with critical avalanches exhibit response sensitivity that is much higher than RNs with asynchronous spiking activity for small stimulus strength (Fig. 2(c)). Furthermore, we check the dynamic range, defined as Inline graphic, where Inline graphic are the stimulus strengths that induce 90% and 10% responses between the minimum and maximum values on a logarithmic scale, as in [28,29]. The dynamic ranges of MNs (Inline graphic,Inline graphic) are greater than those of RNs (Inline graphic,Inline graphic).

The above structure-dynamics relationships are robust with respect to the overall connection density Inline graphic (see Fig. S4) and hold in an extended modelling procedure where the number of inter-modular links decays with distance (see Fig. S6). In a word, MNs can support cost-efficient critical dynamical modes with greatly enhanced response sensitivity to encode variable input strength, whereas globally sparse RNs are both costly in architecture and in running and cannot properly respond to weak input signals.

Structural correlation and dynamic transition of a single separate module

The key features in the structure-dynamics relationship can be understood from an isolated module separate from the whole network but subjected to a background excitatory Poisson input train with rate Inline graphic. Here, we use Inline graphicHz to approximate the weak input received by a neuron from other modules in the highly rewired region in the original network. As the rewiring probability Inline graphic increases, the local connection within a module becomes denser (Fig. 1 and Fig. 2(b)). For a separate module, as its connection density Inline graphic increases, neurons tend to have more common neighbours in the module, the common signal received by a pair of neurons becomes stronger, and their output spikes can be more correlated, as shown in Fig. 4(a).

Figure 4.

Figure 4.

Spontaneous dynamic of a separate module. (a) The change in properties versus the density Inline graphic. From top to bottom: topological correlation; spike correlation; sustained probability; averaged maximum silent period; the MSD of avalanche distribution from power-low fitting function and the numerically estimated effective parameters Inline graphic, Inline graphic through Equation (3). (b) Raster plot of spiking time (blue, red for E, I cells) and synaptic currents (blue, red, black for E, I, net current) received by a neuron. The upper and lower panels are for Inline graphic and Inline graphic, respectively. (c) The distribution of net synaptic current received by a neuron for Inline graphic= 0.05, 0.17 and 0.20. (d) Avalanche distributions (as Fig. 3(b)) for Inline graphic (red), Inline graphic (blue) and Inline graphic (purple).

This correlation in spiking changes the internal interactions in the network. Figure 4(b) shows the synaptic current of a randomly selected neuron. For low density, the net input current fluctuates slightly around zero due to strong E–I balance (Fig. 4(b), upper panels), and the distribution is close to a normal distribution (Fig. 4(c)). With higher density where spike correlation becomes prominent, correlated excitatory spikes induce quick activation of the network, followed by the activation of inhibitory neurons after an effective delay (due to slower inhibitory synaptic time), and then the activity is depressed. Thus, the net current exhibits oscillations around zero (thus, the network maintains the E–I balance on average), as shown in the lower panels of Fig. 4(b), and its distribution has a large tail on the positive side (Fig. 4(c)). The dynamic pattern is an alternation between synchronized firing and quiescent states with no spikes (Fig. 4(b), lower panels).

Furthermore, as the module becomes denser, the self-sustainability of the module decreases (Fig. 4(a)). Here, self-sustainability is tested by turning off the external inputs, i.e. letting Inline graphic after the initial 200 ms. The activity almost cannot be sustained when Inline graphic, which is the module density in the original MN when Inline graphic (Fig. 2(b); see also Equation (4) below). The weaker sustainability of a denser network is a result of the clustered firing dynamic mode (Fig. 4(b), lower panel). The silent period during which no neuron fires increases with Inline graphic(Fig. 4(a)). If this period is too long, all recurrent inputs drop out, and the network activity dies out since there is no external driving.

Under fixed weak external background inputs Inline graphic Hz, as the density Inline graphic increases, the network dynamics undergo a transition from asynchronous firing patterns (Fig. 4(b), upper panel) to critical avalanches (Fig. 4(b), lower panel) with reduced firing rates (Fig. 6(b)). The MSD of the avalanche size distribution from its best-fitted power-law function in Fig. 4(a) shows a minimum at Inline graphic, close to the transition point of self-sustainability. Typical avalanche distributions for subcritical, critical and supercritical dynamic modes for Inline graphic and Inline graphic are shown in Fig. 4(d). Power-law avalanche size and duration distributions Inline graphicInline graphic and Inline graphic (Inline graphic, Inline graphic, Inline graphic with Inline graphic value Inline graphic) are found for Inline graphic, where the scaling relation Inline graphic[20] approximately holds (error ∼0.05).

Figure 6.

Figure 6.

Mean-field theory to understand the dynamic transition. (a) Examples of the mean membrane potential and mean firing rate evolution of the single-module field equations when Inline graphic and Inline graphic, with transient stimuli applied at the time marked by the arrows. (b) Properties of the single-module field equations with different densities Inline graphic. From top to bottom: the firing rate (circles are results from spiking network simulation, curves are results by fixing Inline graphic, and by ‘optimal’ Inline graphic given in Fig. 4(a)); real part of the eigenvalue of the fixed point; CV of activity; response size of membrane potential. (c) Properties of the coupled field equations with different rewiring probabilities Inline graphic. From top to bottom: the corresponding density within a module; the firing rate; CV of activity; response size of membrane potential.

Extended simulation of a separate module (Fig. S7) shows that this dynamic change is independent of input strength Inline graphic, while the critical density Inline graphic, where critical avalanches emerge, depends on Inline graphic. Thus, the emergence of neuronal avalanches can be understood from the large transient fluctuations in the postsynaptic currents induced by correlation, leading to intermittent activities with lower rates.

Effect of correlation: insight from a simplified model

To quantitatively illustrate the impact of input correlation on response sensitivity under E–I-balanced dynamics, we can consider a simplified model as follows. A single neuron receives spike inputs from other K = 200 Poisson excitatory spike trains. Each of the received spikes generates a unit of postsynapse current lasting for Inline graphic. The input signal of the neuron is the summation of these arriving spikes minus a constant equal to the mean current generated by these spikes to mimic the E–I balance. The input correlation is introduced by copying a common Poisson spike train into all input trains [41]; see Fig. 5(a) for a paradigm illustration. To construct spike trains with rate Inline graphic and correlation Inline graphic, the common spike train has a rate Inline graphic, and independent spike trains have rates Inline graphic. Assuming a threshold (Inline graphic) of the input signal above which the neuron fires a spike, we can numerically obtain the input-output rate response curves (Fig. 5(c)). Compared with independent input trains (Inline graphic), the correlation in inputs induces a positive tail in the distribution of the input signal (Fig. 5(b)), qualitatively capturing the feature of the IF module (Fig. 4(c)). In the simulation example of Fig. 5(b), the common spike train is copied into a portion of randomly selected input synapses at different times (such that the probability that more synapses receive spikes simultaneously is lower), resulting in a decaying positive tail in the distribution of the input signal when Inline graphic, which quantitatively resembles the observation from the spiking neural network simulations (Fig. 4(c)). We can see that the correlation increases the output rate when the input rate is the same (Fig. 5(c)). Moreover, this simplified model allows an analytic treatment to explain the effect of the correlation on the response rate (see Supplementary Notes III. Analysis of the Simplified Model with Correlated Inputs for details). The theoretical results (red dashed line for Inline graphic and blue solid line for Inline graphic in Fig. 5(c)) fit well to the simulation results of this simplified model.

Figure 5.

Figure 5.

The simplified illustrative model. (a) The paradigm illustrating a neuron responding to uncorrelated (upper panel) or correlated (lower panel) random spike trains. A Poisson train (labelled in red) is copied into all input neurons to generate correlations among independent spike trains. (b) The distribution of the input signal when Inline graphicHz when the common spike train is copied into a portion of randomly selected input synapses (see explanation in the text). (c) The response curve between input and output rate. Simulation results (symbols) are compared to theoretical predictions (curves). The red and blue colours are for uncorrelated (Inline graphic) and correlated (Inline graphic) input cases.

Hence, the correlation in spikes injected from different recurrent synapses improves the responsiveness of neurons. With input correlation and response sensitivity, each neuron can maintain spike generation when the overall firing rate is low.

Mean-field theory of single module dynamics

To further understand the dynamic mechanism underlying the transition of dynamic modes together with the reduction in firing rate, we derive the equations of average neural activity in each module and the interaction among the modules by a novel mean-field technique [19] (see Method and Supplementary Notes II. Neural Dynamics for details). The field equations of a single separate module with connection density Inline graphic receiving excitatory Poisson background input trains with rate Inline graphic are

graphic file with name TM0088.gif (2)

where Inline graphic are the average E, I voltages, Inline graphic is the average firing rate of Inline graphic neurons, Inline graphic are the average excitatory, inhibitory synaptic time courses received by the neuron and Inline graphic. Inline graphic are GWN terms. Inline graphic are effective parameters to construct the voltage-dependent mean population firing rate (see Method for more details). The strong complexity of COB IF dynamics challenges an analytical (self-consistent) estimation of the effective parameters Inline graphic [19]. Taking different fixed Inline graphic, the field equations can qualitatively predict the decay of the rate with connection density Inline graphic (Fig. 6(b)). To achieve the best prediction, we numerically estimate the effective parameters Inline graphic through the formula

graphic file with name TM0100.gif (3)

from simulations of the single module IF spiking network. The simulations can numerically obtain the steady-state mean voltage Inline graphic and mean firing rate Inline graphic of Inline graphic neurons. The results of Inline graphic from modules with different densities Inline graphic are shown in the bottom panel of Fig. 4(a). Under this setting, the field equations can well quantitatively predict the decrease in firing rate as Inline graphic increases (Fig. 6(b)). Note that qualitative prediction can already be achieved by fixing the effective parameters Inline graphic value in Equation (2) (Fig. 6(b)). Importantly, the field equations reveal that the change in dynamics is associated with a (supercritical) Hopf bifurcation. The dominant eigenvalue of the equilibrium in Equation (2) is complex, and its real part approaches zero as Inline graphic increases (Fig. 6(b)). Thus, the firing rate oscillation emerges by approaching the Hopf bifurcation under noise perturbation, which induces critical avalanches [19]. However, the finite-size effect in a small module (500 neurons) hinders the precision of a mean-field theory. Thus, in the spiking IF model, the MSD achieves a minimum at Inline graphic (Fig. 4(a)), whereas the field equations do not reach the Hopf bifurcation point, and the dynamic is perturbed to bifurcation by noise. Note that a Hopf bifurcation indicates that a periodic motion emerges from zero amplitude, corresponding to the continuous increase of the synchrony in the spiking network (Fig. 4(a)). The CV of activity, measured by the firing rate series of field equations, grows as the Inline graphic increases (Fig. 6(b)). Finally, the response size of the voltage computed from the field equations (Fig. 6(b)) also qualitatively predicts the increase in response sensitivity for denser modules (examples of Inline graphic and Inline graphic are shown in Fig. 6(a), compared to Fig. 3(c)). This is because when approaching a bifurcation point, the system will respond more sensitively and will take a longer time to damp back to the fixed point after perturbation, a phenomenon called critical slowing down [42].

Mean-field theory of the MN

The above investigation of separated modules with various connection densities under weak external background driving provides an understanding of the change in dynamic modes and firing rates with respect to the rewiring probability Inline graphic in the original MN (Fig. 1). First, there is a correspondence between the density in a module Inline graphic and the rewiring probability Inline graphic:

graphic file with name TM0116.gif (4)

(refer to Equation (S1.5)), as shown in Figs 6(c) and 2(b). Furthermore, the field equations of the whole MN can be written as (see Method and Supplementary Notes II. Neural Dynamics for details)

graphic file with name TM0117.gif (5)

with Inline graphic corresponding to the quantities of Inline graphic neurons in the k-th module (see Method for more details). Thus, the whole MN can be considered as Inline graphic coupled identical neural oscillators. During the rewiring process, the coupling strength between different modules (∼Inline graphic) decreases, whereas the self-coupling strength (∼Inline graphic) increases. In this process, although different modules become less affected by each other, the increase in their internal density significantly shapes the dynamic properties of each module, as revealed in separated modules (Fig. 4). Here, the effective parameters Inline graphic to construct Inline graphic in Equation (5) depend on the rewiring probability Inline graphic through their optimal dependence on Inline graphic in separated modules shown in the bottom panel of Fig. 4(a) and the relationship between Inline graphic and Inline graphic (Equation (4)). The numerical results in Fig. 6(c) show that the coupled field equations qualitatively predict the decrease in firing rate and increase in CV of activity and response sensitivity to transient stimuli for increasing rewiring probability Inline graphic, as observed in the spiking neural model in Fig. 2. Furthermore, Equation (5) with Inline graphic also predicts a decrease in the nonzero firing rate (Fig. 6(c)), which is a qualitative prediction of self-sustainability during rewiring (Fig. 2(a)). Note, however, that the mean-field analysis here does not capture the effect of changes in input patterns (e.g. increased input correlation) of a module during the rewiring process of the MN. This is a source of prediction errors that lead to the difference between single-module field equations and a module in the coupled-modules field equations when the latter is constructed by the Inline graphic parameters of the former (refer to Fig. S8). An improvement in the future may be made by assuming oscillatory input in the coupled-modules field model where the oscillatory amplitude increases with rewiring. To conclude, the mean-field theory predicts the dynamical transition (approaching a Hopf bifurcation) of a module with increasing internal density, and this emergent behaviour is maintained for the whole MN with mutually coupled modules when rewiring the inter-modular links to intra-modular links.

CONCLUSION AND DISCUSSION

In this study, we have unveiled the principle of neural networks allowing cost-efficient optimization in both structure and dynamics simultaneously. There have been many studies on either side, considering the optimization of brain network structure [7–9] or energy-efficient neural dynamics [22–26]. For example, energy-efficient cortical action potentials are facilitated by body temperature [25], and cellular ion channel expression is optimized to achieve function while minimizing the metabolic cost of action potentials [31]. However, most previous studies considered the efficiency of the network structure (wiring cost) or the efficiency of dynamics (running cost) separately. Here, considering both structure and dynamics at the circuit level, we show that a wiring-economical MN can support response-sensitive critical dynamics with a much lower running cost while maintaining self-sustainability. This is a notable counterintuitive ‘less-is-more’ result because we obtained greatly enhanced functional values with significant decreases in cost rather than a trade-off between them.

In our model, the efficiency of activity is achieved by critical avalanche states. Different from the traditional critical branching region, the critical dynamics in the synchronous transition region simultaneously achieve greater response sensitivity and a lower firing rate. Previous studies have shown that critical avalanches can appear under various network topologies, for example, scale-free networks with small-world features [43]. Here, we show that a locally dense but globally sparse MN is an efficient organization of the network structure that enables both a low global wiring cost and response-sensitive critical dynamics with a low running cost. It would be interesting to further explore its dynamic advantages on specific cognitive tasks such as working memory recall and decision making.

The origin and mechanism of functionally sensitive critical dynamics in neural systems [13–16,28–30] is a long-standing, challenging and controversial topic. Considering the physical mechanism that supports such a co-optimization of structure and dynamics, here we reveal that with increasing topological correlation in the E–I balanced network, the spike correlation increases, and so does the fluctuation of the inputs received by neurons. In this case, neurons can be activated by a lower firing rate, and the network has higher sensitivity. From the perspective of nonlinear dynamics, these features are captured by a novel mean-field analysis, which reduces the whole MN into coupled oscillators describing the macroscopic dynamics of each module. We elucidate the dynamic mechanism for producing avalanches in proximity to a Hopf bifurcation in the mean field. Close to the bifurcation point, the resulting synchronized spikes in each module are temporally organized as critical avalanches. This stronger collective firing rate variability allows greater computation and coding power [44]. In the highly (yet not totally) rewired MN, the sparse inter-modular connections can provide weak external input to a module from other modules. Meanwhile, as modules are dense enough to be around the response-sensitive critical dynamic states, these weak inputs are sufficient to maintain the whole MN in self-sustained states with low rates.

In principle, the analytical theory for treating biologically plausible COB IF neuronal networks is still an open question [45]. Our approximation semi-analytical mean-field technique serves as an effective theory to study the macroscopic dynamics of such realistic networks. It is important to stress that our work put several important features of neural systems into an integrated framework. Spatial embedding of neural circuits under the wiring cost constraint gives rise to local dense connections and modular organization [7–9]. The E–I balance is a fundamental property of neural circuits [32,33]. Collective activities such as critical avalanches and oscillations are pronounced dynamic features of neural networks [13–16,28–30]. Our modelling and theoretical analysis framework reveals intricate interactions among wiring and running costs, MN topology, critical avalanche dynamic modes and sensitive responses to weak stimulations. Thus, it provides an integrative principle for structural-dynamic cost-efficient optimization in neural systems. Our integrative studies with generic network manipulation and novel mean-field theory with realistic neural dynamics can be extended to coupled cortical areas to offer an understanding of critical dynamics across the whole brain [46,47] based on a hierarchical modular connectome [48]. Furthermore, spatial networks with connections to only the nearest neighbours can exhibit propagating waves with critical dynamic properties [49]. It would be interesting to generalize our model to such nearest-neighbour coupling scenarios and explore the effect of extrasparse long-range connections in such models. This type of model may share similar principles revealed in this study, as in our extended model where short-distance links are dominant (see Fig. S6). The physical principles revealed in our work can guide further development of brain-inspired efficient computing [50].

METHOD

Mean-field theory of IF neural dynamics

In this section, we present the outline of the mean-field theory for deriving the field equations Equation (2) and Equation (5). More details are provided in Supplementary Notes II. Neural Dynamics. In our model, for the i-th neuron in the k-th module, we denote its spiking train as Inline graphic, its Inline graphic (E or I) neighbours in the l-th module as Inline graphic, its voltage as Inline graphic, its input conductance received from recurrent excitatory, recurrent inhibitory neurons as Inline graphic, its external input spike trains (with rate Inline graphic) and input conductance from external neurons as Inline graphic and Inline graphic (if there are external inputs). Then, the network dynamic equation Equation (1) can be written in the following more specific form:

graphic file with name TM0140.gif (6)

where Inline graphic.

Denote Inline graphic, Inline graphic, Inline graphicand Inline graphic. We first adopt a diffusion approximation that Inline graphic, with Inline graphicbeing independent standard GWNs. Thus, Inline graphic, with Inline graphicbeing independent standard GWNs. Then, taking the average Inline graphic to the first equation of Equation (6), with the decoupling approximation Inline graphic, we obtain the first equation of Equation (5). Next, the firing rate of the Inline graphic neurons in the k-th module can be approximated as Inline graphic [19]. This form essentially captures the sub- and supra-threshold microscopic dynamics of a spiking network, that is, Inline graphic represents the proportion of Inline graphic type neurons that spike between Inline graphic and Inline graphic (Inline graphic is an infinitely small quantity) as well as the mean firing rate of Inline graphic type neurons at time Inline graphic with unit per ms. Here, Inline graphic are effective parameters to construct the voltage-dependent mean population firing rate. Note that this approximation scheme based only on the first-order statistics neglects several factors that affect the accurate firing rate, including higher-order statistics, noise correlation and refractory time. Thus, it does not have an analytical form, and Inline graphic should be estimated numerically.

Under the mean-field approximation, we have Inline graphic, where Inline graphic is the average number of Inline graphic neighbours in the l-th module of a neuron in the k-th module. Thus, Inline graphic, where Inline graphic is the connection probability from module Inline graphic to module Inline graphic so that

graphic file with name TM0170.gif (7)

Taking Inline graphic or Inline graphic to the second and third equations of Equation (6), we get the second equation of Equation (5), which finishes the construction of the coupled field equations.

In the limit of Inline graphic (all rewired), modules are almost separated. Let Inline graphic in Equation (5), and we obtain the field equations corresponding to one separate module with additional external excitatory inputs, i.e. Equation (2), with Inline graphic being the connection density of the module.

Supplementary Material

nwab102_Supplemental_File

Contributor Information

Junhao Liang, Department of Physics, Centre for Nonlinear Studies and Beijing-Hong Kong-Singapore Joint Centre for Nonlinear and Complex Systems (Hong Kong), Institute of Computational and Theoretical Studies, Hong Kong Baptist University, Hong Kong, China.

Sheng-Jun Wang, Department of Physics, Shaanxi Normal University, Xi’An 710119, China.

Changsong Zhou, Department of Physics, Centre for Nonlinear Studies and Beijing-Hong Kong-Singapore Joint Centre for Nonlinear and Complex Systems (Hong Kong), Institute of Computational and Theoretical Studies, Hong Kong Baptist University, Hong Kong, China; Department of Physics, Zhejiang University, Hangzhou 310027, China; Beijing Computational Science Research Center, Beijing 100193, China.

Funding

This work was supported by the Hong Kong Baptist University (HKBU) Strategic Development Fund, the Hong Kong Research Grant Council (GRF12200217 and GRF12200620), the HKBU Research Committee and Interdisciplinary Research Clusters Matching Scheme 2018/19 (RC-IRCMs/18-19/SCI01) and the National Natural Science Foundation of China (11975194 and 11675096). This research was conducted using the resources of the High-Performance Computing Cluster Centre at HKBU, which receives funding from the Hong Kong Research Grant Council and the HKBU.

Author Contributions

S.-J.W. and C.Z. conceived the study. S.-J.W. and J.L. performed the numerical simulation and theoretical analysis. J.L., S.-J.W. and C.Z. wrote the manuscript. C.Z. supervised the project. J.L. and S.-J.W. contributed equally to this work.

Conflict of interest statement. None declared.

References

  • 1. Herculano-Houzel S. The human brain in numbers: a linearly scaled-up primate brain. Front Hum Neurosci 2009; 3: 31. 10.3389/neuro.09.031.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Kaas JH. Evolution of columns, modules, and domains in the neocortex of primates. Proc Natl Acad Sci USA 2012; 109: 10655–60. 10.1073/pnas.1201892109 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Roberts JA, Perry A, Lord ARet al. The contribution of geometry to the human connectome. Neuroimage 2016; 124: 379–93. 10.1016/j.neuroimage.2015.09.009 [DOI] [PubMed] [Google Scholar]
  • 4. Hilgetag CC, Kaiser M.. Clustered organization of cortical connectivity. Neuroinformatics 2004; 2: 353–60. 10.1385/NI:2:3:353 [DOI] [PubMed] [Google Scholar]
  • 5. Meunier D, Lambiotte R, Bullmore ET.. Modular and hierarchically modular organization of brain networks. Front Neurosci 2010; 4: 200. 10.3389/fnins.2010.00200 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Gollo LL, Roberts JA, Cropley VLet al. Fragility and volatility of structural hubs in the human connectome. Nat Neurosci 2018; 21: 1107–16. 10.1038/s41593-018-0188-z [DOI] [PubMed] [Google Scholar]
  • 7. Cherniak C, Mokhtarzada Z, Rodriguez-Esteban Ret al. Global optimization of cerebral cortex layout. Proc Natl Acad Sci USA 2004; 101: 1081–6. 10.1073/pnas.0305212101 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Bullmore E, Sporns O.. The economy of brain network organization. Nat Rev Neurosci 2012; 13: 336–49. 10.1038/nrn3214 [DOI] [PubMed] [Google Scholar]
  • 9. Chen Y, Wang S, Hilgetag CCet al. Trade-off between multiple constraints enables simultaneous formation of modules and hubs in neural systems. PLoS Comput Biol 2013; 9: e1002937. 10.1371/journal.pcbi.1002937 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Chen Y, Wang S, Hilgetag CCet al. Features of spatial and functional segregation and integration of the primate connectome revealed by trade-off between wiring cost and efficiency. PLoS Comput Biol 2017; 13: e1005776. 10.1371/journal.pcbi.1005776 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Merolla PA, Arthur JV, Alvarez-Icaza Ret al. A million spiking-neuron integrated circuit with a scalable communication network and interface. Science 2014; 345: 668–73. 10.1126/science.1254642 [DOI] [PubMed] [Google Scholar]
  • 12. Softky WR, Koch C.. The highly irregular firing of cortical cells is inconsistent with temporal integration of random EPSPs. J Neurosci 1993; 13: 334–50. 10.1523/JNEUROSCI.13-01-00334.1993 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Beggs JM, Plenz D. Neuronal avalanches in neocortical circuits. J Neurosci 2003; 23: 11167–77. 10.1523/JNEUROSCI.23-35-11167.2003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Fontenele AJ, de Vasconcelos NAP, Feliciano Tet al. Criticality between cortical states. Phys Rev Lett 2019; 122: 208101. 10.1103/PhysRevLett.122.208101 [DOI] [PubMed] [Google Scholar]
  • 15. Kaiser M, Goerner M, Hilgetag CC.. Criticality of spreading dynamics in hierarchical cluster networks without inhibition. New J Phys 2007; 9: 110. 10.1088/1367-2630/9/5/110 [DOI] [Google Scholar]
  • 16. Wu S, Zhang Y, Cui Yet al. Heterogeneity of synaptic input connectivity regulates spike-based neuronal avalanches. Neural Networks 2019; 110: 91–103. 10.1016/j.neunet.2018.10.017 [DOI] [PubMed] [Google Scholar]
  • 17. Hahn G, Ponce-Alvarez A, Monier Cet al. Spontaneous cortical activity is transiently poised close to criticality. PLoS Comput Biol 2017; 13: e1005543. 10.1371/journal.pcbi.1005543 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Dalla Porta L, Copelli M. Modeling neuronal avalanches and long-range temporal correlations at the emergence of collective oscillations: continuously varying exponents mimic M/EEG results. PLoS Comput Biol 2019; 15: e1006924. 10.1371/journal.pcbi.1006924 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Liang J, Zhou T, Zhou C.. Hopf bifurcation in mean field explains critical avalanches in excitation-inhibition balanced neuronal networks: a mechanism for multiscale variability. Front Syst Neurosci 2020; 14: 580011. 10.3389/fnsys.2020.580011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Sethna JP, Dahmen KA, Myers CR.. Crackling noise. Nature 2001; 410: 242–50. 10.1038/35065675 [DOI] [PubMed] [Google Scholar]
  • 21. Yang D-P, Zhou H-J, Zhou C.. Co-emergence of multi-scale cortical activities of irregular firing, oscillations and avalanches achieves cost-efficient information capacity. PLoS Comput Biol 2017; 13: e1005384. 10.1371/journal.pcbi.1005384 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Wang R, Zhu Y.. Can the activities of the large scale cortical network be expressed by neural energy? A brief review. Cogn Neurodyn 2016; 10: 1–5. 10.1007/s11571-015-9354-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Laughlin SB, van Steveninck RR de, Anderson R.. The metabolic cost of neural information. Nat Neurosci 1998; 1: 36–41. 10.1038/236 [DOI] [PubMed] [Google Scholar]
  • 24. Zhu F, Wang R, Pan Xet al. Energy expenditure computation of a single bursting neuron. Cogn Neurodyn 2019; 13: 75–87. 10.1007/s11571-018-9503-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Yu Y, Hill AP, McCormick DA.. Warm body temperature facilitates energy efficient cortical action potentials. PLoS Comput Biol 2012; 8: e1002456. 10.1371/journal.pcbi.1002456 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Yu L, Yu Y. Energy-efficient neural information processing in individual neurons and neuronal networks. J Neurosci Res 2017; 95: 2253–66. 10.1002/jnr.24131 [DOI] [PubMed] [Google Scholar]
  • 27. Raichle ME, Gusnard DA.. Appraising the brain's energy budget. Proc Natl Acad Sci USA 2002; 99: 10237–9. 10.1073/pnas.172399499 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Shew WL, Yang H, Petermann Tet al. Neuronal avalanches imply maximum dynamic range in cortical networks at criticality. J Neurosci 2009; 29: 15595–600. 10.1523/JNEUROSCI.3864-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Kinouchi O, Copelli M.. Optimal dynamical range of excitable networks at criticality. Nat Phys 2006; 2: 348–51. 10.1038/nphys289 [DOI] [Google Scholar]
  • 30. Cocchi L, Gollo LL, Zalesky Aet al. Criticality in the brain: a synthesis of neurobiology, models and cognition. Prog Neurobiol 2017; 158: 132–52. 10.1016/j.pneurobio.2017.07.002 [DOI] [PubMed] [Google Scholar]
  • 31. Hasenstaub A, Otte S, Callaway Eet al. Metabolic cost as a unifying principle governing neuronal biophysics. Proc Natl Acad Sci USA 2010; 107: 12329–34. 10.1073/pnas.0914886107 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Van Vreeswijk C, Sompolinsky H.. Chaos in neuronal networks with balanced excitatory and inhibitory activity. Science 1996; 274: 1724–6. 10.1126/science.274.5293.1724 [DOI] [PubMed] [Google Scholar]
  • 33. Shu Y, Hasenstaub A, McCormick DA.. Turning on and off recurrent balanced cortical activity. Nature 2003; 423: 288–93. 10.1038/nature01616 [DOI] [PubMed] [Google Scholar]
  • 34. Sporns O, Honey CJ.. Small worlds inside big brains. Proc Natl Acad Sci USA 2006; 103: 19219–20. 10.1073/pnas.0609523103 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Wang S-J, Ouyang G, Guang Jet al. Stochastic oscillation in self-organized critical states of small systems: sensitive resting state in neural systems. Phys Rev Lett 2016; 116: 018101. 10.1103/PhysRevLett.116.018101 [DOI] [PubMed] [Google Scholar]
  • 36. Yu Y, Herman P, Rothman DLet al. Evaluating the gray and white matter energy budgets of human brain function. J Cereb Blood Flow Metab 2018; 38: 1339–53. 10.1177/0271678X17708691 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Lennie P. The cost of cortical computation. Curr Biol 2003; 13: 493–7. 10.1016/S0960-9822(03)00135-0 [DOI] [PubMed] [Google Scholar]
  • 38. Guo D, Li C.. Self-sustained irregular activity in 2-D small-world networks of excitatory and inhibitory neurons. IEEE Trans Neural Networks 2010; 21: 895–905. [DOI] [PubMed] [Google Scholar]
  • 39. Marshall N, Timme NM, Bennett Net al. Analysis of power laws, shape collapses, and neural complexity: new techniques and MATLAB support via the NCC toolbox. Front Physiol 2016; 7: 250. 10.3389/fphys.2016.00250 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40. Luck SJ. An Introduction to the Event-Related Potential Technique. Cambridge: MIT Press, 2014. [Google Scholar]
  • 41. Kuhn A, Aertsen A, Rotter S.. Higher-order statistics of input ensembles and the response of simple model neurons. Neural Comput 2003; 15: 67–101. 10.1162/089976603321043702 [DOI] [PubMed] [Google Scholar]
  • 42. Scheffer M, Bascompte J, Brock WAet al. Early-warning signals for critical transitions. Nature 2009; 461: 53–9. 10.1038/nature08227 [DOI] [PubMed] [Google Scholar]
  • 43. Massobrio P, Pasquale V, Martinoia S. Self-organized criticality in cortical assemblies occurs in concurrent scale-free and small-world networks. Sci Rep 2015; 5: 10578. 10.1038/srep10578 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44. Ostojic S. Two types of asynchronous activity in networks of excitatory and inhibitory spiking neurons. Nat Neurosci 2014; 17: 594–600. 10.1038/nn.3658 [DOI] [PubMed] [Google Scholar]
  • 45. Renart A, Brunel N, Wang X-J. Mean-field theory of irregularly spiking neuronal populations and working memory in recurrent cortical networks. In: Feng J (ed.). Computational Neuroscience: A Comprehensive Approach. London: CRC Press, 2004, 431–90. [Google Scholar]
  • 46. Tagliazucchi E, Balenzuela P, Fraiman Det al. Criticality in large-scale brain fMRI dynamics unveiled by a novel point process analysis. Front Physio 2012; 3: 15. 10.3389/fphys.2012.00015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Haimovici A, Tagliazucchi E, Balenzuela Pet al. Brain organization into resting state networks emerges at criticality on a model of the human connectome. Phys Rev Lett 2013; 110: 178101. 10.1103/PhysRevLett.110.178101 [DOI] [PubMed] [Google Scholar]
  • 48. Wang R, Lin P, Liu Met al. Hierarchical connectome modes and critical state jointly maximize human brain functional diversity. Phys Rev Lett 2019; 123: 038301. 10.1103/PhysRevLett.123.038301 [DOI] [PubMed] [Google Scholar]
  • 49. Chen G, Gong P.. Computing by modulating spontaneous cortical activity patterns as a mechanism of active visual processing. Nat Commun 2019; 10: 4915. 10.1038/s41467-019-12918-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Fan D, Sharad M, Sengupta Aet al. Hierarchical temporal memory based on spin-neurons and resistive memory for energy-efficient brain-inspired computing. IEEE Trans Neural Netw Learning Syst 2015; 27: 1907–19. 10.1109/TNNLS.2015.2462731 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

nwab102_Supplemental_File

Articles from National Science Review are provided here courtesy of Oxford University Press

RESOURCES