Abstract
Pattern formation, i.e., the generation of an inhomogeneous spatial activity distribution in a dynamical system with translation invariant structure, is a well-studied phenomenon in neuronal network dynamics, specifically in neural field models. These are population models to describe the spatio-temporal dynamics of large groups of neurons in terms of macroscopic variables such as population firing rates. Though neural field models are often deduced from and equipped with biophysically meaningful properties, a direct mapping to simulations of individual spiking neuron populations is rarely considered. Neurons have a distinct identity defined by their action on their postsynaptic targets. In its simplest form they act either excitatorily or inhibitorily. When the distribution of neuron identities is assumed to be periodic, pattern formation can be observed, given the coupling strength is supracritical, i.e., larger than a critical weight. We find that this critical weight is strongly dependent on the characteristics of the neuronal input, i.e., depends on whether neurons are mean- or fluctuation driven, and different limits in linearizing the full non-linear system apply in order to assess stability. In particular, if neurons are mean-driven, the linearization has a very simple form and becomes independent of both the fixed point firing rate and the variance of the input current, while in the very strongly fluctuation-driven regime the fixed point rate, as well as the input mean and variance are important parameters in the determination of the critical weight. We demonstrate that interestingly even in “intermediate” regimes, when the system is technically fluctuation-driven, the simple linearization neglecting the variance of the input can yield the better prediction of the critical coupling strength. We moreover analyze the effects of structural randomness by rewiring individual synapses or redistributing weights, as well as coarse-graining on the formation of inhomogeneous activity patterns.
Keywords: pattern formation, spiking neurons, linear model, mean-driven, fluctuation driven, ring networks, small-world networks
1. Introduction
Understanding the dynamics of neuronal networks and its dependence on connection topology, weight distribution or input statistics is essential to understand network function. One very successful field in uncovering structure-dynamics relationships are so-called neural field models (see e.g., Beurle, 1956; Wilson and Cowan, 1972; Amari, 1977; Ermentrout and Cowan, 1979a; Ben-Yishai et al., 1995; Ermentrout, 1998; Coombes, 2005). In these models the positions of neurons in space are substituted by densities and their respective coupling is described by coupling kernels depending on pairwise spatial distance. Given certain symmetries such as translation invariance, and homogeneous input the resulting non-linear dynamics is often reduced to a low-dimensional system that can be studied analytically. Such systems can produce various spatio-temporal phenomena, such as traveling waves, activity bumps and formation of periodic patterns, which can be linked to activity in biological systems, e.g., visual hallucinations, feature selectivity, short term memory, or EEG rhythms (see e.g., Ermentrout, 1998; Coombes, 2005 and references therein).
However, when thinking of actual neuronal tissue, the symmetry requirements and continuity assumptions apply on a rather macroscopic scale, when neuron densities are high and heterogeneities are negligible. Moreover, such rate-based models cannot fully resolve the statistics of synaptic input currents, but it is well-known that neurons and neuronal populations react quite strongly to changes in the statistics of incoming currents in terms of mean, auto- and cross-covariance (see e.g., Mainen and Sejnowsky, 1996; Silberberg et al., 2004; Fourcaud-Trocmé and Brunel, 2005; De la Rocha et al., 2007; Boucsein et al., 2009; Tchumatchenko and Wolf, 2011). Finally, it is often unclear how to quantitatively interpret the parameters of abstract field models, especially when compared to spiking network simulations.
Along these lines, Usher et al. (1995) demonstrated numerically how periodic pattern-formation takes place in a two-dimensional toroidal network of excitatory and inhibitory pulse-coupled integrate-and-fire oscillators. The bifurcation occurs when the relative surround inhibition reaches a certain threshold that is determined from the Fourier-transform of the effective coupling matrix (dispersion relation). In their study, Usher et al. (1995) moreover observed that the dynamics of the emerging pattern depends on the average external input current, with slowly diffusing hexagonally arranged activity bumps for low input current, over stationary bumps for intermediate input strength, to coherently moving bumps for strong external input.
In a later study, Bressloff and Coombes (1998, 2000) showed rigorously how periodic orbits in networks of pulse-coupled inhibitory and excitatory neurons with Mexican-hat topology indeed undergo a discrete Turing-Hopf-bifurcation, if the system is coupled strongly enough.
Roxin et al. (2005) studied a neuronal field model with particular emphasis on the role of delays. Similar to earlier studies of neuronal field models, (see e.g., Ben-Yishai et al., 1995; Compte et al., 2000; Shriki et al., 2003) neurons were distributed along a ring and coupling was distance-dependent. In particular, the coupling kernel was of the form J(|x − y|) = J0 + J1 cos(x − y), where x and y are neuron positions in space and J0 and J1 are coupling strength coefficients. Depending on the choice of J0, J1 and the delay the system can show a plethora of activity states such as traveling waves, stationary or oscillatory bumps, and also aperiodic behavior. However, this choice of interaction between two neurons is rather qualitative. It is usually symmetric and changes sign with distance, properties that are not in line with the biology of individual neurons and synapses.
Thus, to compare the field-model results to simulations of networks of excitatory and inhibitory spiking neurons, Roxin et al. (2005) set up a model where the same number of excitatory and inhibitory neurons were uniformly distributed across two rings and were coupled by sinusoidally modulated connection probability. This set-up resulted in an effectively one-dimensional model where at each position excitation and inhibition combine to a net-coupling comparable to the field model, and spatio-temporal patterns qualitatively matched the field-model predictions. The parameters in the spiking model were chosen with regard to biophysically plausible values, and the non-linearity of the current-to-rate transduction in the field model was chosen as threshold linear in an ad-hoc manner, however a direct quantitative mapping between neuronal network and field model was not tried.
Here, we investigate the dynamics of identical leaky integrate-and-fire neurons that are arranged on ring and grid networks where both inhibitory and excitatory neurons are uniquely assigned positions in space rather than representing a density. No two neurons can be at the same position, and this is thus a step toward spatially embedded networks of spiking neurons instead of neuronal populations with possibly ambiguous coupling weights. Instead, individual neurons have a distinctive identity in that they act either excitatorily or inhibitorily on all their post-synaptic targets (Dale's principle) and the coupling between any two neurons is usually not symmetrical. This is important since neglecting the identity of a neuron can lead to dramatically different dynamics (Kriener et al., 2008, 2009). If the system is translation-invariant, a Turing-instability occurs for a critical synaptic coupling strength Jc, such that there is a bistability between the spatially homogeneous firing rate distribution and an inhomogeneous periodic spatial pattern, similar to what was described before (Usher et al., 1995; Bressloff and Coombes, 1998, 2000).
We find that the value of Jc depends strongly on the statistics of the neuronal input: if neurons are mean-driven, i.e., receive suprathreshold input, the instability occurs for much smaller coupling strength than when neurons are in the strongly fluctuation-driven regime, where the mean input is subthreshold and spikes are evoked by spontaneous suprathreshold fluctuations. The latter is the key mechanism in explaining asynchronous-irregular firing in balanced neuronal networks.
In assessing stability of an invariant network state to perturbations the usual modus operandi is to linearize around the fixed point (see e.g., Ermentrout and Cowan, 1979a,b, 1980; Usher et al., 1995; Bressloff and Coombes, 1998, 2000; Coombes, 2005), implying crucially the derivative of the neuronal input-output function. In the fluctuation-driven regime the input-output function depends on both the mean and variance of the input current and thus both respective derivatives are expected to influence the quantitative prediction of linear fixed point stability (Amit and Brunel, 1997; Tetzlaff et al., 2012; Helias et al., 2013). Indeed, in the strongly fluctuation-driven regime, such that the standard-deviation of the input current is much larger than the average distance to firing threshold, the correct critical coupling strength is derived by linearization with respect to both mean and variance around the fixed point. In the mean-driven regime, on the other hand, the linearization becomes independent on the input current variance, and the derivative with respect to the mean is constant over wide ranges of working points implying independence on the exact location of the rate fixed point. In this regime pattern formation typically occurs for considerably smaller coupling strength than in the fluctuation-driven regime. Interestingly, if the input current is in an intermediate regime with both moderate subthreshold input mean, yet considerable variance the linearization independent of the input variance can give a better estimate than the one taking it into account. We will see that usually, the true critical coupling strength will lie inbetween both estimates, with Jc generally closer to the mean-driven regime prediction.
Moreover, in order to relate our ring model to the earlier findings of pattern formation in the classical Ermentrout-Cowan networks (Ermentrout and Cowan, 1979a,b, 1980) we also introduce a coarse-grained network that maps to a two-dimensional system which is formally identical to Ermentrout-Cowan networks, but lacks some of the details of the full, effectively five-dimensional ring network considered here.
Finally, we study the role of robustness of pattern formation to randomness: first, structural randomness is caused by the introduction of short-cuts which move the ring topology into the so-called small-world regime (Watts and Strogatz, 1998); secondly, we also discuss the effect of randomness in the weight distribution introduced by lifting the constraint that neurons are either fully excitatory in their action or fully inhibitory (Dale's principle).
2. Materials and methods
2.1. Neuron and network model
We study ring networks of N leaky integrate-and-fire (LIF) neurons with current-based synapses. NE = βN, β ∈ [0, 1], of all neurons are excitatory, the residual NI = N − NE neurons are inhibitory. The membrane potential Vi(t) of neuron i is governed by the differential equation
(1) |
where the membrane potential Vi(t) is reset to Vres whenever it reaches the threshold potential Vthr and a spike is emitted. The neuron is then refractory for a time τref. τm denotes the membrane time constant, , where Is is the input current received from the local network, and sj(t) = ∑k δ(t − tj, k) is the spike train produced by some neuron j. d is the transmission delay, and W is the coupling matrix, with entries that are either zero (no synapse present), JE = J (excitatory synapse) or JI = −gJ (inhibitory synapse). To keep spiking activity from entering a high rate state, we assume that the recurrent input Is is net inhibitory by choosing g > NE/NI.
Ix denotes the external input current that is modeled as stationary Poisson noise injected via current-based synapses of strength Jx (for details, see below), and R is the membrane resistance. Such leaky integrate-and-fire neurons were extensively studied in various input (see e.g., Amit and Tsodyks, 1991; Amit and Brunel, 1997; Lindner, 2004; Fourcaud-Trocmé and Brunel, 2005; De la Rocha et al., 2007; Vilela and Lindner, 2009; Helias et al., 2010) and network settings, especially in the context of balanced random networks (Brunel, 2000; Tetzlaff et al., 2012). Here, we consider ring networks of LIF neurons, such that each fifth neuron is inhibitory and the coupling matrix W is given by (cf. Figure 3A)
(2) |
Thus, each neuron is connected to its κ nearest neighbors, where we assume κ = 0.1N. We note that the network layout chosen here is motivated by an actual embedding of individual neurons into some space, in contrast to earlier studies of pattern formation in spiking neuron networks where inhibitory and excitatory neurons are often distributed in equal numbers along two identical rings, and where the main difference lies in the spatial interaction ranges. These systems are usually motivated by population-density descriptions and can often be reduced to a two-dimensional model (see e.g., Ben-Yishai et al., 1995; Compte et al., 2000; Shriki et al., 2003; Roxin et al., 2005). As we will demonstrate in section 3.2, the network studied here leads to an effectively five-dimensional model. In section 3.4 we will discuss a reduction to a two-dimensional model that however already deviates noticeably in its details from the full five-dimensional network.
2.2. Constant external input in the mean-driven and intermediate regime
In the case of the mean-driven (μ[RI] ≥ Vthr − Vres) and intermediate regimes, which will be discussed later, the external current is purely excitatory, such that RIx(t) = τmJxsx(t) with Jx > 0 and rate νX = E[sx(t)], where E[.] denotes the expectation value. To quantify the effective strength of the external input, we express νX = ηθ/Jxτm, i.e., the parameter η gives the strength of the external drive in terms of the excitatory rate necessary to reach threshold. Assuming stationary spiking activity with rate νo for both the excitatory and inhibitory population, i.e., νE = E[sE(t)] = νo and νI = E[sI(t)] = νo respectively, the mean and variance of the total input RI = R(Is + Ix) are thus given by
(3) |
For stronger local coupling strength J the negative mean recurrent contribution μ[RIs] to the total input is stronger, while the variance σ2[RIs] increases. For the networks we consider here, for small J the network activity is thus typically driven by the mean μ[RI] of the total input, while for larger J the variance σ2[RI] has a notable impact. Moreover, the network firing rate νo depends on J.
2.3. External input adjusted to ensure constant total current in the strongly fluctuation-driven regime
The strongly fluctuation-driven regime is characterized by subthreshold mean total input, μ[RI] < Vthr − Vres, but high variance σ2[RI], such that spikes are initiated by transient input fluctuations. To guarantee that we are in the high variance regime for all J, we inject excitatory and inhibitory external currents with rates νEx, νIx, such that the total input current mean and variance, i.e., μ[RI] and σ2[RI], stay the same when varying the local coupling strength J. This is achieved by adapting μ[RIx] = μ[RI] − μ[RIs] and σ2[RIx] = σ2[RI] − σ2[RIs] as a function of νEx, νIx. The external current injected into each local neuron by current-based synapses of strengths JEx = Jx and JIx = −gJx is thus given by
(4a) |
(4b) |
(4c) |
Because the variance of the recurrent input σ2[RIs] increases quickly with increasing J, the total variance σ2[RI]− in order to keep it fixed for all J− usually needs to be chosen quite large in order to avoid non-sensical rates νIx < 0. This moreover implies that this approach of choosing the external input is not useful for the mean-driven scenario.
For all neurons we set θ = Vthr − Vres = 20 mV, τm = 20 ms, and R = 80 MΩ. The excitatory synaptic coupling strengths J, Jx, the relative strength of inhibition g, as well as τref and d will be specified individually. The model and parameters are listed in Table A1 in Appendix A1. All simulations were carried out with NEST (Gewaltig and Diesmann, 2007).
3. Results
For small coupling strength J and large mean current input μ[RI] = E[R(Ix + Is)] the spiking dynamics of the ring networks is characterized by locally clustered activity on the spatial scale of the footprint κ, cf. Figures 1A,G (Kriener et al., 2009). For increasing J, however, a clear spatial pattern emerges, cf. Figures 1C,F. Since all neurons receive by construction statistically identical input, all neurons should fire at the same rate. But there is a distinct critical coupling strength Jc beyond which the homogeneous distribution of firing rates becomes unstable and the system enters a spatially inhomogeneous state after sufficient perturbation (Bressloff and Coombes, 1998). We note that the spatio-temporal spiking activity can be quite rich in networks with translation-invariant symmetry, as discussed for example in (Ben-Yishai et al., 1995; Usher et al., 1995; Bressloff and Coombes, 1998, 2000; Shriki et al., 2003; Roxin et al., 2005; Marti and Rinzel, 2013). Here, we do not systematically analyze the temporal aspects of the activity in terms of traveling waves or oscillations, but focus on the onset of periodic pattern formation in space as a function of coupling strength.
What determines this critical coupling strength Jc where the bistability of the homogeneous spatial rate distribution and the spatially modulated distribution occurs? We find that Jc depends strongly on the input regime the neurons operate in: if neurons are driven predominantly by the mean of the input current, i.e., if μ[RI] ≥ θ, we find that Jc is much smaller than if neurons are strongly fluctuation-driven, i.e., μ[RI] < θ with large enough variance of the input current to occasionally drive the membrane potential to threshold, cf. Figure 1.
In the following we present two different linear rate models of the integrate-and-fire dynamics that explain and predict the respective occurrence of the pattern formation.
3.1. Two linearizations for the self-consistent rate in networks of LIF-neurons
The analysis of the stability of spatially homogeneous activity dynamics to spatial perturbations follows a simple concept: first, the stationary, homogeneous state is determined in a mean-field approximation. In particular, in the diffusion limit, i.e., under the assumption that the coupling strength J is small compared to the distance between reset and threshold θ and that all neurons receive statistically identical input, the stationary firing rate νo of the neurons can be derived self-consistently as a function of mean and variance of the input current (a solution of the mean first-passage-time problem first derived by Siegert (Siegert, 1951), but also see e.g., Amit and Tsodyks, 1991; Brunel, 2000), such that
(5) |
where μo = ∑i ∈ {E,I,X} Jiνi, oτm and σ2o = ∑i ∈ {E,I,X} J2iνi, oτm are the self-consistent input mean and variance. Because all neurons receive statistically identical input, in absence of symmetry-breaking the firing rate distribution is spatially homogeneous, i.e., νo, i ≡ νo for all i. To assess simple stability of the homogeneous state to spatial perturbations, we need to perform linear perturbation analysis. The critical eigenvector of the resulting linear stability operator then determines the spatial pattern that develops.
We will show that the exact nature of the appropriate linearization depends critically on the input current regime, and derive the respective linearization in the limits of subthreshold mean input with fluctuation-driven spiking in section 3.1.1, and of dominant suprathreshold mean total input current μ[R(Ix + Is)] > θ in section 3.1.2. In the fluctuation-driven regime a quantitative approximation is obtained from the modulation of both mean and fluctuations of the synaptic input to the neurons, while in the mean-driven regime we recover the intuitive result that the system is governed by the noise-free input-output relation given by the f-I-curve of the individual neuron. In general, the full spiking dynamics will lie somewhere in between these two limiting cases, the case we call “intermediate.”
3.1.1. Linearization of the input-output relation in the fluctuation-driven regime
We first discuss the case when the system is fluctuation-driven, i.e., the average input current is subthreshold (μ[RI] < θ) and spikes are induced by membrane-potential fluctuations.
To assess linear stability in this regime we need to derive a linear model that self-consistently describes the dynamics of the fluctuations ui(t) = 〈si(t) − νo〉 = νi(t) − νo of the activity of neuron i around its mean firing rate or working point νo. Treating the response of integrate-and-fire neurons by first order perturbation theory is a well-established method first introduced in Abbott and van Vreeswijk (1993) for the perfect integrator and in Amit and Brunel (1997); Brunel and Hakim (1999) for the leaky integrate-and-fire neuron.
The dynamics of u(t) can then be written in the form
(6) |
where h(t) is a linear filter and x(t) is a white noise imitating the spiking nature of the signal. In the Fourier-domain Equation (6) becomes the simple product U(ω) = H(ω)[U(ω) + X(ω)], where the capitalizations indicate the Fourier-transforms of h(t), u(t) and x(t). Pattern formation is typically a slow process and the transfer function H(ω) of leaky integrate-and-fire neurons is dominated by low-pass behavior1. In order to assess stability, we therefore estimate H(ω) in the low-frequency limit, i.e., . With and as well as , we obtain (see also Amit and Brunel, 1997; Tetzlaff et al., 2012; Helias et al., 2013 for details) the effective coupling matrix of the system as the local derivative of the input-output-curve given by Equation (5):
(7a) |
(7b) |
(7c) |
with
(8) |
The effective coupling matrix (Wij) is thus structurally identical to W given by Equation (2), with ij = (J) if j is excitatory, and ij = (−gJ) if j is inhibitory, yielding an effective relative inhibitory strength := |(−gJ)/(J)|. The zero frequency mode U(0) therefore fullfills the equation
(9) |
We expect linearity to break down if has any eigenvalue with real part larger than unity. The critical coupling strength Jfdc for which this is the case can be determined implicitly from Equation (7).
3.1.2. Linearization in the low input current noise limit
Here, we discuss the linearization in the case that neurons are mean-driven, i.e., when the mean of the input current μ[RI] is suprathreshold, while the variance of the input is negligible [see also Amit and Tsodyks (1991) for analogous arguments]. In this case, the input-output-relation of the neuron becomes effectively linear in μ[RI] until it reaches saturation at νmax = 1/τref (Amit and Tsodyks, 1991; Salinas and Sejnowsky, 2000). This can be seen by considering the input-output relation of an individual neuron in response to a constant current, i.e., the f-I-curve
(10) |
For RI » θ, i.e., and negligible τref, it holds (see Appendix A2 for details)
(11) |
This linear-threshold-type relationship fits that observed in simulations very well, cf. Figures 2A,B.
To derive a rate equation for the full spiking network dynamics, we first define the instantaneous rate of neuron i as
(12) |
where E[.] is the average over realizations.
Substituting the spike trains by their respective firing rates, the input to neuron i takes the form RIi = τm ∑j Wij νj + RIx. If we only consider networks with rates νi ≥ 0 and approximate the firing rate of neuron i by the linear expression (11), we arrive at the linear equation
(13) |
Indeed, if τref is negligible, Equation (1) can be rewritten as (Kriener et al., 2008)
(14) |
Equation (13) can then be derived analogously by temporal averaging, assuming stationarity d EΔt[V(t)]/dt ≡ 0, i.e., EΔt[V(t)] ≡ const, and thus EΔt[ν(t)] = const = νo. For high rates in response to large μ[RI] we expect E[V(t)] ≈ (Vthr − Vres)/2 = θ/2, as the voltage passes through the interval between reset and threshold with approximately constant velocity, leading to
(15) |
which is identical to (13). We note, that the self-consistent solution of Equation (15) yields the correct quantitative rates over a wide range of relative input magnitude, if instead of θ/2 the actual E[Vi(t)]-values measured in simulations are inserted. If all neurons are identical and receive statistically identical input, this mean value can be obtained from ∀i E[Vi(t)] ≡ V := ∫θ−∞ VP(V) dV, where P(V) denotes the stationary membrane potential probability density function as e.g., derived in Brunel (2000). All output-rate predictions are compared to the outcome from simulations in Figure 2B. Aside from the prediction that assumes E[V(t)] = θ/2 (dark gray triangles) and underestimates the true rate, all predictions fit the simulation results (blue dots) very well. In particular, the Siegert equation (5) coincides perfectly with the linearized rate Equation (14) if we assume E[V(t)] = V (light gray triangles), while the best fit is obtained with Equation (14) and the actual measured E[Vi(t)] (black asterisks).
The stability of the fixed point rate to small perturbations is determined by the largest real part of all eigenvalues of W/θ. If one eigenvalue λc has real part larger than unity, we expect the corresponding eigenvector vc to grow exponentially, only limited by the non-linearities due to rate-rectification for small rates, and the saturation because of neuronal refractoriness for high rates. This determines the critical coupling strength
(16) |
that in this limit is independent of the exact working point νo > 0. Even though for coupling strengths J > Jmdc the rate instability can lead to pattern formation such that individual neurons fire at quite different rates, the population rate is still captured well by the firing rate predictions assuming homogeneous rates across neurons, see Figure 2B.
Finally, we note that for the constant external drive scenario considered here the mean and variance of the total input current vary with varying J. In particular, for increasing J the system undergoes a cross-over from the mean-driven to the fluctuation-driven regime, see Figure 2C.
3.2. Eigensystem of dale-conform translation-invariant ring networks
In this section we will analytically derive the eigenvalue spectrum of the ring networks under consideration that yields the critical eigenvalue and thus the critical coupling strength Jc with respect to the two linearizations Equation (9) and Equation (13). As described in section 2 we consider ring networks of size N with regular connectivity, in that each neuron is connected to its κ nearest neighbors, κ/2 on each side of the neuron but not to itself. Inhibitory neurons shall be distributed periodically across the network as illustrated in Figure 3A, where the ratio between excitation and inhibition is 4 : 1 (β = 0.8). This is in line with the observed frequencies of excitatory and inhibitory cells in cortex (Schüz and Braitenberg, 1994). The ring can thus be divided into structurally identical elementary cells of four excitatory neurons and one inhibitory neuron. We choose κ as an integer multiple of the size ℓ of an elementary cell (here ℓ = 5, κ mod ℓ = 0), such that it moreover divides the total number of neurons N (N mod κ = 0). The resulting coupling matrix is sketched in Figure 3B, where black squares depict inhibitory weights, white depict excitatory weights and gray denotes the lack of a connection. In this way all neurons receive the exact same amount of inhibition and excitation, and the translation invariant mode (1, 1, ···, 1)ᵀ is an eigenvector of the coupling matrix with eigenvalue μ = κJ(β − g(1 − β)). To keep network activity balanced, inhibitory synapses are assumed to be g-times stronger, with g > 4, cf. Equation (2). Hence, the local network is inhibition dominated.
To compute the eigensystem of the N-dimensional system we make use of its symmetry properties. The coupling matrix commutes with the unitary operator that shifts all neuron indices by multiples of ℓ = 5. Hence, we can diagonalize both in the same eigenbasis and moreover reduce the problem to an effectively five-dimensional one as outlined in Appendix A3. Figure 3C shows the exemplary eigenvalue spectrum of in this case the rescaled coupling matrix W/θ of a network of size N = 2500 with absolute coupling strength J = 1 mV and relative strength of inhibition g = 6. Thus, the eigenvalue λc of W/θ with largest real part will exceed unity, if the amplitude of the absolute coupling strength J exceeds Jmdc = 1/Re[λc], cf. Equation (16), here Jmdc ≈ 0.5.
The corresponding critical weight Jfdc for the fluctuation-driven case is obtained by the identical eigenvalue decomposition of as defined by Equation (7). Since depends non-trivially on the self-consistent working point νo of the system, the critical coupling strength Jfdc is the solution of
(17) |
where λi, i ∈ {1, …, N} are the N eigenvalues of . If J > Jmd/fdc (depending on the regime), the spatially homogeneous rate distribution becomes unstable to perturbations in favor of the fastest growing eigenmode vc, which only depends on the symmetry properties of the matrix as long as g and are larger than four, and it is hence the same irrespective of the input regime, cf. Figures 1C,F,I. The spatial frequency of the emerging spatial pattern (see Figure 3), i.e., the number of maxima of the rate distribution along the ring, is thus given by the wavenumber Λ (see Appendix A3) that corresponds to the respective λc := maxi[Re[λi]]. We note, that the eigenvalue spectra in the fluctuation-driven case look slightly different from those in the mean-driven case (cf. Figure 3C) because of the in general different scaling ≠ g between positive and negative entries in . The maximum of the dispersion relation determining Λc and the eigenvectors are, however, usually the same.
Figure 4 shows the critical eigenvalue λc as a function of J for the linearizations Equation (9) (gray lines) and Equation (13) (red line) in the constant external drive regime, parameterized by η. The critical eigenvalue Jc is given by the intersection with the horizontal line at one. As was pointed out in section 3.1.2, in the noiseless approximation Equation (13) the eigenvalues of the linear operator do not depend on the working point, and thus the critical weight Jmdc is unique for all η such that νo > 0 (here, critical Jmdc = 0.506 indicated by red circle).
For the linearization Equation (9), and taking into account both μ and σ in Equation (7b) (solid lines), the critical weight Jfdc is strongly dependent on the value of η = {3.5, 5, 10, 15, 20, 30, 40} (increasing from dark to light gray), and never comes close to the prediction of the noiseless linearization Jmdc. In particular, for larger η Jfdc increases again.
If we neglect the σ-dependence in Equation (7b), and just take into account dνo/dμo = dνmdo/dRI = 1/τmθ as given by Equation (11), Equation (7) is linear in Wij and we recover the linear noiseless approximation Equation (13) for Equation (9). We might thus expect that Equation (9) in general will give a Jfdc-prediction close to Jmdc if we neglect the σ-dependent quadratic term in Equation (7b). Although the prediction of this Jfd|dνo/dμoc becomes smaller (cf. black triangle for Jfdc = 1.54 mV vs. the gray asterisk for Jfd|dνo/dμoc = 0.89 mV for η = 3.5) it never becomes as small as Jmdc. This is because every change in μo will also influence σo via its impact on νo, and thus dνo(μo, σo)/dμo ≠ dνmdo(RI)/dRI.
We note that the observation Jfdc > Jfd,dνo/dμoc is explained by the fact that (Wij)|dνo/dμo is linear in Wij, and thus the ratio = |(−gJ)/(J)| = g, while for the full Equation (7), (Wij) is quadratic in Wij, yielding a < g. Evaluating Equation (A20) in Appendix A3 shows that the critical eigenvalue λc in both linearizations is a linearly increasing function of g, , respectively, and Jc is thus decreasing as a function of g, .
For comparison, Figure 4 also shows the J-dependence of the real part of the critical eigenvalue λc in the strongly fluctuation-driven regime (μ[RI] = 5 mV, σ[RI] = 60 mV, blue line), with the critical Jc indicated by the crossing of the unity-line (blue triangle).
3.3. Linear stability in dependence of the input current regime
We will now compare the general performance of the linearizations Equations (9) and (13) and in particular the resulting predictions of the critical coupling strength Jc with the actual onset of pattern formation as observed in simulations. Figure 1 demonstrates how the identical ring network structure undergoes pattern formation at very different values of network coupling strength J depending on the input current regime. In particular, pattern formation sets in later if the system is in a fluctuation-driven regime. This can be understood in the two limit-cases that were analyzed in sections 3.1.1 and 3.1.2.
But what determines pattern formation onset in intermediate cases that comprise the most interesting and relevant cases? As we saw in section 3.1.2, the network will even undergo a crossover from the mean-driven to a more fluctuation-driven regime, if the external input is kept constant and not adapted to counterbalance the changes in recurrent input structure for changing J, see Figure 2C. Also, due to the spiking nature of the activity, even in a strongly mean-driven scenario there will always be a considerable amount of variance which might change the onset of pattern formation with respect to Jmdc.
To give more insight into the actual onset of pattern formation, useful reduced measures are the variance and kurtosis of the distribution of firing rates over neurons, cf. Figure 5. Figures 5A–C show the histogram of rates, and the variance and kurtosis of these histograms for various J-values in the mean-driven regime averaged over 10 trials each. Figures 5D–F show the same for the fluctuation-driven case and Figures 5G–I for the intermediate case, all for the same parameters as in Figure 1.
When the system is still well in the linear regime J < Jc the distribution is approximately Gaussian with small variance and a kurtosis close to zero (cf. Figures 5A,D,G, light gray curves), while once the critical weight Jc is surpassed the variance increases strongly (cf. Figures 5B,E,H) and the kurtosis becomes negative (platycurtic), indicating the broadening of the distribution due to the inhomogeneous rate per neuron (cf. Figures 5C,F,G and 5A,D,G, dark gray curves). The respective critical weights are indicated by vertical dashed lines in Figures 5B,E,H and C,F,I for visual guidance (red: Jmdc, black: Jfdc, gray: Jfd|dνo/dμoc).
The onset of pattern formation in the strongly fluctuation-driven regime at Jfdc is more shallow than in the mean-driven regime at Jmdc. This might be due to deviations of the input spike statistics from the asynchronous-irregular activity assumption underlying the linear response derivation of Jfdc: the standard-deviation of the input is extremely large at σ = 60 mV. This implies that the membrane-potential will be hyperpolarized for long times, while at other times it is highly depolarized, and several spikes are emitted in short succession. This typically leads to higher coefficients of variation, i.e., more irregular firing than expected for Poisson spike statistics [see also Brunel (2000) and supplementary material section S1]. Also, even in the highly fluctuation-driven regime there may be residual correlations between spike trains. Lastly, the increased network coupling strength might lead to deviations from the Gaussian white noise approximation of input currents underlying Equation (5) and the linear response theory yielding Equation (7) (see supplementary material section S1 for an analysis). How such deviations of the input statistics can indeed change the spike response behavior of LIF neurons was studied, e.g., in Moreno et al. (2002); Renart et al. (2007); Moreno-Bote et al. (2008); Helias et al. (2010).
However, another contributing factor, and more likely the reason for the shallow transition at Jfdc, is the degeneracy of the maximal eigenvalue, see Figures 3D,E. In the fluctuation-driven regime the system is subject to strong perpetual perturbations that lead to a switching between the two dominant eigenvectors depicted in Figure 3D in the transition regime Jfdc ± ΔJ. Indeed, as can be seen from Figure 1E, the periodic pattern is already clearly distinguishable at Jfdc over several hundred milliseconds, but the average activity depicted in the histogram is still quite flat, yielding smaller variance and kurtosis of the respective rate distribution.
Finally, Figures 5D–F demonstrate clearly how pattern formation occurs for intermediate J-values Jmdc ≲ Jinterc < Jfdc if the system is neither clearly in the mean- nor strongly fluctuation-driven regime, and even changes from one to the other regime with increasing coupling strength, cf. Figure 2C. As we demonstrate in the supplementary material section S1, most input and network settings have pattern formation onsets that usually agree much better with the Jmdc prediction than with Jfdc. This is a very consistent finding, even in cases where neurons are well in the fluctuation-driven regime in terms of subthreshold mean and pronounced input variance.
As we discuss in the supplementary material section S1 the reason for this finding lies in an asymmetry between the excitatory and inhibitory compound input current statistics. The excitatory subpopulation tends to be synchronized, even in the presence of strong balanced external noise, while inhibition actively decorrelates itself. The explanation for this effect lies in the local connectivity of ring networks, leading to a population-specific pattern of correlations already below the critical coupling: excess synchrony of the excitatory population will reinforce further spiking, while spikes from the inhibitory population decreases the instantaneous firing probability [see also Helias et al. (2013) for a related discussion of this effect in the framework of balanced random networks]. Near synchrony is thereby effectively enhanced in the excitatory population and suppressed in the inhibitory one. Volleys of excitatory input spikes act like compound pulses with large amplitude (on the order of up to θ) in an otherwise balanced asynchronous background activity. At these amplitudes the effective gain Equation (7) derived from linear response theory becomes linear in J with a slope proportional to 1/θ, i.e., the slope of the linear model Equation (11), see supplementary material section S1. This explains the fact that Equation (13) has better predictive power also in the intermediate or—comparably weakly—fluctuation-driven scenarios. Moreover, the fluctuation-driven prediction only becomes valid if the additional external noise sufficiently dilutes residual synchrony, such as it is the case for μ[RI] = 5 mV and σ[RI] = 60 mV.
3.4. Coarse-grained ring network
In section 3.2 we saw that the full system can effectively be reduced to a five-dimensional system. However, the computation of eigensystems (cf. Appendix A3) is still quite involved. In this section we study how well a further simplified system that is formally identical to the well-known Ermentrout-Cowan networks [Ermentrout and Cowan (1979a,b, 1980) and supplementary material section S2] predicts the dynamics of the full network.
We coarse-grain the system and combine groups of ℓ = N/NI neurons, such that they contain four excitatory and one inhibitory neuron (ℓ needs to be odd in the following), as indicated in Figure 6. The neurons ic ∈ {0, …, ℓ − 1} within each cell c ∈ {0, …, C − 1}, C = N/ℓ, are connected to every other neuron within their cell c but not to themselves. Moreover, all neurons within one local cell c are connected to all neurons within the K < C/2 neighboring cells to the left and to the right, i.e., {(c − K) mod C, …, (c − 1) mod C, (c + 1) mod C, …, (c + K) mod C}. The computation of the eigensystem is outlined in Appendix A4. Due to the two mirror symmetries corresponding to exchange of neurons within the unit cell (with corresponding eigenvalues r, s ∈ {−1, 1}) and the translational symmetry with the eigenvalue determined by the wavenumber α, the system can be reduced to effectively two dimensions.
3.4.1. Coarse-grained network dynamics
In the coarse-grained system we only need knowledge about two elements of the ring, one excitatory and one inhibitory neuron within the same cell (arbitrarily chosen as v0, v(ℓ − 1)/2 in c = 0, without loss of generality). For a fixed set of eigenvalues r, s, α, the activities of all remaining neurons are then uniquely determined. The homogeneous mode νo(t) couples to the symmetric subspace r, s = 1 only and we can write νo(t) thus as a linear combination of the two eigenvectors vIα0 and vE1,1,α0, where α0 = 0 (cf. Appendix A4).
To analyze the stability of this mode with respect to spatial perturbations we write the population activity as a linear combination of eigenvectors with non-zero wavenumbers αi, vE1,1,αi, vIαi, such that
(18) |
Then, for all neurons n ∈ {0, …, N − 1} the input, (W ν(t)) [n], in the reduced system becomes
with [n] := ⌊n/ℓ⌋. In the first identity we made use of the operators R, S and Tℓ to express the input-connectivity of the network. The second identity follows from the linearity of the sum over eigenmodes and the respective eigenvalues of the operators R, S and Tℓ. We excluded synaptic self-coupling, so we need to subtract that input of the cell to itself, i.e.
Leaving out one inhibitory input if the receiving neuron is inhibitory, and one excitatory input if the receiver is excitatory leads to an asymmetry in the total input of excitatory and inhibitory neurons. This implies that the spatially homogeneous mode is not an eigenmode of the system anymore. To compensate for this the weights are rescaled such that the input per neuron equals that in the full (non-coarse-grained) system, i.e., to μ = κJ(β − g(1 − β)). Hence,
Note, that if we do not exclude synaptic self-coupling in the coarse-grained system the eigensystem looks dramatically different from that of the full system. In particular the eigenvalues are nearly fully real and the critical eigenvalue is much smaller implying a higher critical weight.
In the homogeneous mode the input to each neuron n ∈ {0, …, N − 1} is completely identical, all neurons fire at the same rate νo given by the self-consistent solution
where . The eigenvalues of the reduced rescaled coupling matrix
then determine for which parameters g, J the spatially homogeneous state becomes unstable and which wavenumber Λi(αi) will grow fastest and define the spatial pattern. The critical eigenvalue for the coarse-grained system is thus given by the eigenvalue of with largest real part that will cross unity from below first, i.e.
In the fluctuation-driven regime J and g in Equation (21) need to be substituted by : = (J) and , such that the critical eigenvalue is given by λfdc(,).
On a mesoscopic level the coarse-grained and full system give rise to very similar activity patterns—differences become apparent in the fine-scale features, e.g., the full analysis of a network of 2500 neurons with 10% connectivity and g = 6 predicts destabilization of wavenumber 14 in both mean- and fluctuation-driven regime, while the coarse-grained model predicts 13, see also Figure 7. We note that the coarse-grained model presented in this section is formally identical to the linearization derived in Ermentrout and Cowan (1980) in the context of a neural field ring model if one assumes a boxcar-footprint. We summarize the respective derivation in the supplementary material section S2.
3.5. Sensitivity to randomness
To conclude, we study how robust the findings of the previous sections are to effects of randomness in either structure or weight distribution. Though we again only show the eigenvalue spectra for the synaptic coupling matrix W/θ, that is the relevant linear operator in the mean-driven case, all results map straight-forwardly to the fluctuation-driven regime with appropriate rescaling of J and g.
3.5.1. Small-world networks: structural noise
First, we will study the effect of the introduction of structural randomness by rewiring individual connections with probability pr [for details see Kriener et al. (2009)] such that the input composition from the network to every neuron stays constant at μ. The random rewiring procedure means that each realization of a network will be different and moreover lacks invariance to Tℓ such that the eigensystems of individual coupling matrices need to be computed numerically. For low pr the rate model for the translation-invariant network still gives very good results and the patterns that form for large J are indeed strongly correlated with the critical eigenvector vc, see Figure 8. For higher pr predictions become worse and more than one vector can contribute to the pattern (e.g., Figures 8C2,C3). For low rewiring probability pr < 0.02 we can still analytically estimate the critical eigenvalue by employing a mean field model (Grabow et al., 2012) for small-world networks in which we rewire “on average,” cf. Appendix A5 and Figures 8A2–C2,A4–C4. Hence, also for small-world networks of excitatory and inhibitory spiking neurons, rates and linear stability can be estimated from the translation-invariant network rate model if pr is small. The results of the mean-field model are slightly worse than for unweighted networks (cf. Grabow et al., 2012) because of the two qualitatively different types of edges in the networks considered here (excitatory and inhibitory). For the small number of edges κ we assume in Figure 8 it can occur that for very small pr no inhibitory edges are rewired at all, while the mean-field model still reassigns inhibitory connection density. We find the mean field to work better for denser or larger networks, and if the excitatory and inhibitory input per neuron is kept the same during rewiring (not shown).
If the constant input per neuron constraint is dropped, the homogeneous mode (1, …, 1)ᵀ is not an eigenmode anymore and also the prediction of stability fails in large parts. On the one hand there are activity inhomogeneities forming out simply due to the fact that some areas on the ring will by chance get more or less net inhibitory input, and hence some neuronal subpopulation will be driven to zero rates, while others—lacking net inhibitory input from these rectified neurons—will have very high rates. Similar effects arise when the distribution of inhibitory neurons is not periodic anymore, but, e.g., distributed randomly along the ring.
3.5.2. Ring networks not conform with Dale's principle
Finally, if Dale's principle—i.e., the biological fact that each neuron can only either depolarize or hyperpolarize all its postsynaptic targets, but never both at the same time—is violated, both the eigensystem (Figure 9A), as well as the spiking dynamics looks akin to that of a random network, even if the underlying topology is a ring, cf. Figures 9B,C. Hence, the identity of neurons in terms of being excitatory or inhibitory is a necessary condition for the formation of structured patterns in the case of identical footprints for inhibition and excitation considered here.
Even though in all cases pattern formation is either strongly or completely impaired, the prediction Equation (14) of individual rates in the low coupling regime is excellent, also in inhomogeneous networks, where different neurons receive very different inputs, given the individual mean membrane potentials 〈Vi(t)〉t are used in Equation (15).
4. Discussion
In this paper we analyzed the dynamics of pattern formation in networks of excitatory and inhibitory spiking neurons that are embedded on a ring architecture. In particular, we studied the dependence of the underlying bifurcation on the input current statistics and structural noise, and derived a coarse-grained system that is formally analogous to the classical Ermentrout-Cowan networks. To conclude, we will summarize and discuss our findings.
4.1. Impact of input current statistics
The very nature of spiking neuron networks leads to considerable input current fluctuations, and hence in general both mean μ[RI] and standard deviation σ[RI] of the input will shape the output rate of the neuron. For leaky integrate-and-fire neurons the output rate as a function of μ[RI] and σ[RI], assuming stationary Poisson-like uncorrelated input spike trains and weak synaptic coupling J, is given by the Siegert equation (5) (Amit and Tsodyks, 1991; Amit and Brunel, 1997). Thus, in order to analyze the linear stability of this self-consistent solution, in general the derivatives with respect to both μ[RI] and σ[RI] need to be taken into account, cf. Equation (7).
We analyzed two different ways to drive the networks: in the first all neurons in the network were driven with the same excitatory Poisson-noise νX, independent on the recurrent coupling strength J. In this case the total input RI the neurons receive, i.e., the external input RIx together with the recurrent input RIs from the network, changes with increasing J, such that the mean μ[RI] decreases, while the standard deviation σ[RI] increases.
In the second input scenario, we corrected for the change in recurrent input contributions with changing J by administering compensating inhibitory and excitatory external Poisson input νIx, νEx. Thus, μ[RI] and σ[RI], and hence also νo are approximately constant with increasing J.
We find that in the first input scenario the critical Jc as observed in simulations deviates strongly from the one predicted by Equations (7),(9), which are derived from the self-consistent solution Equation (5) by linear response theory. In particular, for the mean-driven regime, where μ[RI] > Vthr, Jmdc becomes basically independent of the rate νo as well as of σ[RI], irrespective of its yet considerable magnitude. Instead, a very simple linearization derived from the noiseless f-I-curve Equation (10) applies. The Jmdc derived from this linearization is always considerably smaller than the prediction by Equations (7), (9). For intermediate cases where neurons are neither mean- nor strongly fluctuation-driven, the critical coupling strength lies in between the two predictions, and often closer to Jmdc.
Possible reasons for the failure of the prediction of Equations (7), (9) in the constant external drive scenario (parameterized by η, cf. section 2) are the break-down of the weak-coupling assumption, as well as the uncorrelated Poisson assumption for the input spike trains in the recurrent contribution. In line with earlier findings (Tetzlaff et al., 2012), we could not identify deviations with respect to coupling strength, however, we observed clear deviations from the Poisson assumption.
If spiking is Poissonian we expect the interspike intervals to be distributed exponentially, and the coefficient of variation (CV) of the interspike intervals (ISI) to be unity. A smaller CV indicates more regular, a larger one more irregular spiking. As can be seen from Figures 1A–C and 1G–I, activity is spatio-temporally structured, with locally clustered spiking that can induce significant pairwise correlations, and low coefficients of variation of the interspike intervals for small J (cf. also supplementary material section S1, Figures S1B,C). In effect, these deviations from uncorrelated Poisson spiking lead to deviations from the Gaussian white noise approximation underlying the derivation of Equations (5) and (9), especially for increasing J.
We observe that in particular the excitatory subpopulation tends to synchronize because of the highly recurrent local structure and positive feedback, while the inhibitory population actively desynchronizes itself due to negative feedback (see supplementary material section S1, Figures S3, S4, and Helias et al. (2013) for a related discussion). Thus, high amplitude volleys of excitatory input in a balanced background comprised of asynchronous inhibition and external Poisson-drive move the network in a regime that is better captured by the mean-driven low-noise approximation Equation (11).
Activity in the strongly fluctuation-driven regime, on the other hand, is much more asynchronous and irregular by construction, cf. Figures 1D–F with CV[ISI] of unity and larger (see supplementary material, section S1, Figures S1B,C). When in this scenario σ[RI] is very large, while μ[RI] is sufficiently subthreshold, the prediction for the critical coupling strength Jfdc from Equations (7), (9) for linearity to break down indeed appears to be correct, or even seems to underestimate the actual onset of pattern formation.
This shallow transition around Jfdc can be explained by slow noise-induced transitions between the two degenerate eigenvectors which grow quickest once pattern formation takes place. However, the fact that the CV is larger than unity will also lead to deviations from the prediction. In a series of papers Moreno et al. (2002); Renart et al. (2007) and Moreno-Bote et al. (2008) studied the impact of deviations of the Poisson-assumption of input spike trains on the firing rates of neurons. They studied the effect of positive and negative spike train auto- and cross-correlations parametrized by the Fano-factor F = σ2[counts]/μ[counts] of spike counts on the output firing rate of current-based LIF neurons in both the fluctuation- and the mean-driven regime. For renewal processes the CV of interspike intervals is directly related to the Fano-factor of counts by F = CV2 (Cox, 1962). In Moreno et al. (2002); Renart et al. (2007); Moreno-Bote et al. (2008) it was shown that in the fluctuation-driven regime output rates are quite sensitive to input noise correlations (parameterized by the Fano factor), while in the mean-driven regime sensitivity is less pronounced.
In simulations of ring networks in the strongly fluctuation-driven case we observe that rates decrease with increasing correlations, while in otherwise equivalent random networks, rates increase (not shown). So the different recurrent input structure plays a crucial role for the effective input current and thus network dynamics properties. These differences are often, as here for ring and grid networks, a direct consequence of complex network topology, and it is thus important to understand how connectivity structure translates to single neuron activity, as well as collective network states. A more thorough analysis of such effects in the context of firing rate stability in random and spatially structured networks will be subject of subsequent research.
4.2. Impact of finite time scales
We assumed throughout the manuscript that the system has a negligible refractory time constant, small delays and instantaneous post-synaptic currents (δ-synapses). If these assumptions are dropped we observe differences especially for the mean-driven case: for larger delays d in the order of several milli-seconds, pronounced oscillations can occur that tend to synchronize the system locally on the order of the footprint (Kriener et al., 2009), or for very strong external drive also on a population-wide level (Brunel, 2000). A systematic study of the complex effects of delays on pattern formation in neural-field-type ring networks of spiking neurons was presented in Roxin et al. (2005). In the network we analyzed here, noisy delay oscillations mainly perturb developing patterns and can thus shift the onset of clear pattern formation to larger Jc.
Similar effects arise in the case of finite refractory time τref. A larger τref, especially in the presence of oscillations, i.e., highly coincidental input spiking, decreases the effective network input, in particular if τref > d, and thus also shifts the input regime. This can be compensated for by explicitly taking into account τref in the derivation of the noiseless linearization.
Finally, if synaptic time constants are finite, while total input currents stay the same, delay oscillations can be smoothened out, and the effect of absolute refractoriness is decreased. Increasing synaptic time constants can thus counteract the effect of finite τref and d.
4.3. Spiking networks vs. neural field models
Pattern formation, such as activity bump formation or periodic patterns, is a well-studied phenomenon in ring and toroidal networks (see e.g., Ermentrout and Cowan, 1979a,b; Ben-Yishai et al., 1995; Usher et al., 1995; Bressloff and Coombes, 1998, 2000; Shriki et al., 2003; Marti and Rinzel, 2013). Earlier studies of periodic pattern formation in spiking neuron networks already showed that ring- and grid-networks of mean-driven neurons can undergo a Turing-bifurcation (Usher et al., 1995; Bressloff and Coombes, 1998, 2000). In particular, Usher et al. (1995) observed a dependence on the input level. For weaker, yet suprathreshold mean input current, activity patches diffuse chaotically across the spatial extend of the system, while for stronger inputs a stable spatial pattern develops that relates to the critical eigenmode. We observe similar dynamics in the networks considered here as well, if the system is in the intermediate regime. This can be understood as follows: if the noise component is relatively strong, it shifts the actual critical coupling strength Jmdc to higher values Jinterc inbetween Jmdc and Jfdc. If the system is thus effectively slightly sub-critical in terms of Jinterc, input-noise can transiently excite eigenmodes close to effective criticality, due to Hebbian (Dayan and Abbott, 2001) or non-normal amplification (Murphy and Miller, 2009).
To compare our results to the classical work on neural field models on ring topologies (Ermentrout and Cowan, 1979a,b, 1980) we moreover studied a coarse-grained model that reduces the N-dimensional system to an effectively two-dimensional one that is structurally equivalent to the linearized two-population model presented, e.g., in (Ermentrout and Cowan, 1980) (see supplementary material section S2) which, however, allows for a direct quantitative mapping to the spiking network dynamics.
Because of the coarse-graining some of the connectivity details get lost and thus the critical eigenmode is not identical with that of the original ring network. Moreover, we observe the importance of excluding self-coupling in the coarse-grained model: if that is not excluded the resulting eigensystem looks dramatically different with nearly completely real-valued eigenvalues.
4.4. Impact of structural noise
For the translation-invariant distribution of inhibitory neurons across a ring topology the eigensystem, and hence the respective linear stability, can be computed analytically. If inhibitory neurons are however randomly distributed, some parts of the ring will have a higher, others a lower density of inhibitory neurons, such that these networks form an ensemble of possible realizations, most of which are not invariant to translations. The resulting inhomogeneous rate per neuron in the subcritical regime can still be predicted well from the linear rate model. However, even when knowing the dominant eigenvector, it will usually not be periodic, and due to the inhomogeneous input per neuron different neurons will be at quite different working points. There is interference between the rate distribution and the dominant mode. Thus, the activity pattern that evolves in the supracritical regime is not straightforward to predict.
If the weights are distributed fully randomly across the whole ring topology irrespective of the identities of the neurons, i.e., in violation of Dale's principle, the resulting eigensystems look very similar to those of corresponding networks with random topologies. Only a few surviving singular eigenvalues serve to distinguish the underlying ring from the random topology. This directly translates to the spiking dynamics: the activity is much more asynchronous than that of Dale-conform networks and most of all pattern formation can not take place due to the effective lack of lateral inhibition.
4.5. Impact of network size
If κ/N is kept constant when varying the network size N the critical coupling strength in both the mean- and the fluctuation-driven regime decreases. For example, for N = 10000 and κ = 1000 the critical coupling strength becomes Jmdc ≈ 0.2 mV for the mean-driven, and Jfdc ≈ 0.32 mV for the fluctuation-driven regime. If the number of synapses per neuron κ is fixed when N increases, both Jmdc and Jfdc hardly change because the maximal eigenvalue is approximately unaffected by network dilution.
We emphasize that the analysis presented here can be extended to two-dimensional grid networks in a straightforward manner, see supplementary material section S3. We conclude by remarking that the rate model Equation (13) is also a valuable tool to study the rate dynamics of spatially embedded spiking neuron networks with e.g., inhomogeneous neuron distribution and distance-dependent connectivity in the mean-driven regime, allowing for a treatment of more general networks than commonly studied random networks or neural field models. The presented model is in conclusion a further step in the efforts to find and understand appropriate descriptions of the high-dimensional spiking activity in structured networks.
Conflict of interest statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
We want to thank the reviewers for their helpful comments. We gratefully acknowledge funding by the eScience program of the Research Council of Norway under grant 178892/V30 (eNeuro), the Helmholtz Association: HASB and portfolio theme SMHB, the Next-Generation Supercomputer Project of MEXT, EU Grant 269921 (BrainScaleS), and by the German Federal Ministry for Education and Research (BMBF, grant 01GQ0420 to BCCN Freiburg, and grant 01GQ0830 to BFNT Freiburg*Tübingen). All network simulations were carried out with NEST (http://www.nest-initiative.org).
A. Appendix
A1. Model and simulation description
Table A1.
A | MODEL SUMMARY |
---|---|
Populations | Three: excitatory, inhibitory, external input |
Connectivity | Ring connectivity with excitatory and inhibitory neurons arranged on the same ring, such that each fifth neuron is inhibitory; same connection footprint for both populations |
Neuron model | Leaky integrate-and-fire (LIF), fixed voltage threshold precise spike timing (Hanuschkin et al., 2010) |
Synapse model | δ-current inputs (discontinuous voltage jumps) |
Input | Independent Poisson spike trains |
B | POPULATIONS | |
---|---|---|
Name | Elements | Size |
E | LIF neuron | NE = 4NI |
I | LIF neuron | NI |
Ex | excitatory Poisson input generator | one realization per neuron |
Ix | inhibitory Poisson input generator (only for the fluctuation-driven case) | one realization per neuron |
C | CONNECTIVITY | ||
---|---|---|---|
Name | Source | Target | Pattern |
EE | E | E | connect to all excitatory neurons within a given distance on the ring, weight J, delay d |
IE | E | I | connect to all excitatory neurons within a given distance on the ring, weight J, delay d |
EI | I | E | connect to all excitatory neurons within a given distance on the ring, weight −gJ, delay d |
II | I | I | connect to all excitatory neurons within a given distance on the ring, weight −gJ, delay d |
Ex | Ex | E ∪ I | independent Poisson spike trains, weight Jx, rate νEx |
Ix | Ix | E ∪ I | independent Poisson spike trains, weight −gJx, rate νIx |
D | NEURON AND SYNAPSE MODEL |
---|---|
Name | LIF neuron |
Type | Leaky integrate-and-fire, δ-current input |
Subthreshold dynamics | |
Spiking | If V(t−) < θ ∧ V(t+) ≥ θ |
1. set t* = t | |
2. emit spike with time-stamp t* |
E | INPUT |
---|---|
Type | Description |
Poisson generators | Rate νEx, νIx, each generator projects independent realizations to all neurons |
A2. Approximation of the siegert formula by an affine-linear input-output relation
For strong mean drive μ[RI] » θ we can approximate the input-output relation (10) by an affine linear relationship. In the following we assume the firing rate 0 < ν « 1/τref, which is the case for all relevant regimes, so we can neglect the effect of the refractory time τref. We use the Ansatz
(A1) |
By taking the derivative with respect to RI on both sides of the last expression we obtain
(A2) |
Equating the terms in proportion to (RI) and in proportion to (RI)2 in the denominator fixes the parameters and , leading to the approximation
(A3) |
Indeed, to show that for θ » RI
(A4) |
we introduce the auxiliary variable θ/RI =: (1 − x) and compute the limit
(A5) |
Both numerator and denominator of this expression tend to zero, thus we apply L'Hôpital's rule and obtain for (A5)
(A6) |
Again, numerator and denominator tend to zero, thus we apply L'Hôpital's rule again and obtain for (A6)
(A7) |
This yields (A4).
As a heuristic observation, we note moreover, that in the large-σo-limit, i.e., σo » μo, the integrand of Equation (5) becomes basically unity, and the integral thus scales as . In particular we observe that the input-output-relation Equation (5) for Vres = 0 takes the following linear form in σo:
(A8) |
The input-output-relation of the LIF is thus also linear in σo for σo » μo, see supplementary material section S4.
A3. Eigenspectrum of the ring graph with equidistant inhibitory neurons
We consider graphs, in which each neuron i is connected to neurons i − κ/2 to i + κ/2, κ even, but not to itself. To keep analysis simple, we assume that each ℓ-th neuron, N mod ℓ = 0 is inhibitory.
Define Tℓ as the translation operator that shifts all neuron indices by ℓ such that for every vector v we have
(A9) |
i.e., for the node index n
(A10) |
Tℓ is a unitary operator and has eigenvalues φl and eigenvectors wl, such that
(A11) |
with
(A12) |
Hence the N/ℓ ℓ-fold degenerated eigenvalues of Tℓ are φl = ei2πlℓ/N, l ∈ {0, …, N/ℓ − 1}. The eigenvectors wl are discrete harmonics and fullfil
(A13) |
We define
(A14) |
and the map η that embeds the ℂℓ as a subspace into the ℂN
(A15) |
where ei, i ∈ {0, …, N − 1} are the N canonical basis-vectors of the ℝN and i ∈ ℂ is the i-th vector component of . We moreover define the operator family Pl, l ∈ {0, …, N/ℓ − 1}
(A16) |
We see that (Pl ◦ η) indeed projects into the eigenspaces and is an isomorphism
(A17) |
as can be readily checked:
We see that for each l ∈ {0, …, N/ℓ − 1}
(A18) |
where in the equation marked by (*) we performed an index shift and used (e−iαTℓ)N/ℓ = (e−iαTℓ)0.
Linearity:
(A19) |
Injectivity:
Surjectivity:
Now, we can calculate the spectrum of W. Due to [W, Tℓ] = 0, i.e., W and Tℓ commute, we can diagonalize both matrices, such that they have a common system of eigenvectors. Hence, it is sufficient to reduce the problem to the ℓ-dimensional vector-space embedded in ℂN, spanned by eigenvectors of Eig(Tℓ, eiφl)|ℂℓ, i.e., l, l ∈ {0, …, N/ℓ − 1}. To obtain the eigenspaces Eig(W, λli) we only need to solve the diagonalization problem for the corresponding (ℓ × ℓ)-matrices η−1 ◦ W ◦ Pl ◦ η = (Pl W)|ℂℓ × ℓ:
(A20) |
Once we got the ℓ eigenvalues and eigenvectors {λl, i, l, i} of W in the respective sub-spaces, the operator Pl ◦ η generates the corresponding N-dimensional eigenvector wl.
A4. Eigensystem of the reduced ring system
We now consider a coarse-grained network model where connectivity is defined between C = N/ℓ cells of ℓ neurons each as depicted in Figure 6A. Each neuron is connected to all other ℓ − 1 neurons within the cell it resides in, as well as to all neurons in the neighboring 2K = pC cells, yielding block-wise connectivity matrices as indicated in Figure 7B. We introduce a new global neuron labeling such that
(A21) |
Symmetries:
The system in invariant to shifts about elementary cells, i.e., for the cell index n we have again invariance to the operation
(A22) |
Moreover, there are two mirror symmetries R and S within each cell (cf. Figures 6B,C), such that
(A23) |
(A24) |
R and S are operators, such that for the eigenvalues rj and sj, and for the eigenvectors vj and wj we have R2vj = r2jvj = vj and S2wj = s2jwj = wj. Hence, rj, sj ∈ {−1, 1}. We have three commuting operators Tℓ, R, S, i.e.,
So, we can find a common basis of eigenvectors to span the complete system. For each wavenumber Λi = αi N/2πℓ one eigenvector is obviously given by
(A25) |
where e(ℓ − 1)/2 is the (ℓ − 1)/2-th canonical basisvector of the ℝN and (ℓ − 1)/2 is the index of the first inhibitory neuron on the ring enumerated as in Figure 6, since this dimension does not couple to either R or S. To get the residual N − NI eigenvectors, we choose
(A26) |
with first canonical basis vector e0.
Indeed, we see again that for each c ∈ {0, …, N/ℓ − 1}
(A27) |
and with x ∈ {r, s} and ∈ {R, S}, so vr, s, α are eigenvectors with eigenvalues r, s, α. Evaluation of the eigenvectors for the two possible eigenvalues of R, S each yields
(A29) |
A5. Mean-field approximation of small-world networks of excitatory and inhibitory neurons
Analogous to the mean-field approach developed in (Grabow et al., 2012) for unweighted networks we will now compute the eigensystem of the mean-field small world network derived from the ring network with equidistant inhibition by random rewiring of edges (Kriener et al., 2009). In this case instead of assigning explicit weights Wij ∈ {0, J, −gJ} to individual entries of the coupling matrix, all possible connections carry their expected weight as a function of the rewiring probability pr such that all weights within the original ring neighborhood are scaled by p1(pr), and all weights that were 0 in the original ring are set to Wij = p2(pr) J if j is excitatory, and Wij = − p2(pr) gJ if j is inhibitory, where (cf. Kriener et al., 2009; Grabow et al., 2012)
(A30) |
The resulting average coupling matrix is again invariant with respect to Tℓ and the eigensystem can be determined exactly as described in Appendix A3.
Footnotes
1Note, that for exponential response kernels h(t) = τ exp[−t/τ] the response is structurally equivalent to a linear first order differential equation, though it has been demonstrated that in ring networks this response kernel can be substantially different from exponential and even contain oscillatory components (Lindner and Schimansky-Geier, 2001; Pernice et al., 2012).
Supplementary material
The Supplementary Material for this article can be found online at: http://www.frontiersin.org/journal/10.3389/fncom.2013.00187/abstract
References
- Abbott L. F., van Vreeswijk C. (1993). Asynchronous states in networks of pulse-coupled oscillators. Phys. Rev. E 48, 1483–1490 10.1103/PhysRevE.48.1483 [DOI] [PubMed] [Google Scholar]
- Amari S.-i. (1977). Dynamics of pattern formation in lateral-inhibition type neural fields. Biol. Cybern. 27, 77–87 10.1007/BF00337259 [DOI] [PubMed] [Google Scholar]
- Amit D., Brunel N. (1997). Model of global spontaneous activity and local structured activity during delay periods in the cerebral cortex. Cereb. Cortex 7, 237–252 10.1093/cercor/7.3.237 [DOI] [PubMed] [Google Scholar]
- Amit D. J., Tsodyks M. V. (1991). Quantitative study of attractor neural network retrieving at low spike rates I: substrate – spikes, rates and neuronal gain. Network 2, 259 10.1088/0954-898X/2/3/003 [DOI] [Google Scholar]
- Ben-Yishai R., Bar-Or R. L., Sompolinsky H. (1995). Theory of orientation tuning in visual cortex. Proc. Natl. Acad. Sci. U.S.A. 92, 3844–3848 10.1073/pnas.92.9.3844 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Beurle R. L. (1956). Properties of a mass of cells capable of regenerating pulses. Philos. Trans. R. Soc. Lond. B240, 55–94 10.1098/rstb.1956.0012 [DOI] [Google Scholar]
- Boucsein C., Tetzlaff T., Meier R., Aertsen A., Naundorf B. (2009). Dynamical response properties of neocortical neuron ensembles: multiplicative versus additive noise. J. Neurosci. 29 1006–1010 10.1523/JNEUROSCI.3424-08.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bressloff P. C., Coombes S. (1998). Spike train dynamics underlying pattern formation in integrate-and-fire oscillator networks. Phys. Rev. Lett. 81, 2384–2387 10.1103/PhysRevLett.81.2384 [DOI] [Google Scholar]
- Bressloff P. C., Coombes S. (2000). Dynamics of strongly coupled spiking neurons. Neural Comput. 12, 91–129 10.1162/089976600300015907 [DOI] [PubMed] [Google Scholar]
- Brunel N. (2000). Dynamics of sparsely connected networks of excitatory and inhibitory spiking neurons. J. Comput. Neurosci. 8, 183–208 10.1023/A:1008925309027 [DOI] [PubMed] [Google Scholar]
- Brunel N., Hakim V. (1999). Fast global oscillations in networks of integrate-and-fire neurons with low firing rates. Neural Comput. 11, 1621–1671 10.1162/089976699300016179 [DOI] [PubMed] [Google Scholar]
- Compte A., Brunel N., Goldman-Rakic P. S., Wang X. J. (2000). Synaptic mechanisms and network dynamics underlying spatial working memory in a cortical network model. Cereb. Cortex 10, 910–923 10.1093/cercor/10.9.910 [DOI] [PubMed] [Google Scholar]
- Coombes S. (2005). Waves, bumps, and patterns in neural field theories. Biol. Cybern. 93, 91–108 10.1007/s00422-005-0574-y [DOI] [PubMed] [Google Scholar]
- Cox D. R. (1962). Renewal Theory. (Science Paperbacks). London: Chapman and Hall [Google Scholar]
- Dayan P., Abbott L. F. (2001) Theoretical Neuroscience. Cambridge, MA: MIT Press [Google Scholar]
- De la Rocha J., Doiron B., Shea-Brown E., Kresimir J., Reyes A. (2007). Correlation between neural spike trains increases with firing rate. Nature 448 802–807 10.1038/nature06028 [DOI] [PubMed] [Google Scholar]
- Ermentrout B. (1998). Neural networks as spatio-temporal pattern-forming systems. Rep. Prog. Phys. 61, 353–430 10.1088/0034-4885/61/4/002 [DOI] [Google Scholar]
- Ermentrout G. B., Cowan J. D. (1979a). A mathematical theory of visual hallucination patterns. Biol. Cybern. 34, 137–150 10.1007/BF00336965 [DOI] [PubMed] [Google Scholar]
- Ermentrout G. B., Cowan J. D. (1979b). Temporal oscillations in neuronal nets. J. Math. Biol. 7, 265–280 10.1007/BF00275728 [DOI] [PubMed] [Google Scholar]
- Ermentrout B., Cowan J. (1980). Large scale spatially organized activity in neural nets. SIAM J. Appl. Math. 38, 1–21 10.1137/0138001 [DOI] [Google Scholar]
- Fourcaud-Trocmé N., Brunel N. (2005). Dynamics of the instantaneous firing rate in response to changes in input statistics. J. Comput. Neurosci. 18, 311–321 10.1007/s10827-005-0337-8 [DOI] [PubMed] [Google Scholar]
- Gewaltig M.-O., Diesmann M. (2007) NEST (NEural Simulation Tool). Scholarpedia 2, 1430 [Google Scholar]
- Girko V. (1984). Circular law. Theory Probab. Appl. 29, 694–706 10.1137/1129095 [DOI] [Google Scholar]
- Grabow C., Grosskinsky S., Timme M. (2012). Small-world spectra in mean field theory. Phys. Rev. Lett. 106, 218701 10.1103/PhysRevLett.108.218701 [DOI] [PubMed] [Google Scholar]
- Hanuschkin A., Kunkel S., Helias M., Morrison A., Diesmann M. (2010). A general and efficient method for incorporating precise spike times in globally time-driven simulations. Front. Neuroinform. 4:113 10.3389/fninf.2010.00113 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Helias M., Deger M., Rotter S., Diesmann M. (2010). Instantaneous non-linear processing by pulse-coupled threshold units. PLoS Comput. Biol. 6:e1000929 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Helias M., Tetzlaff T., Diesmann M. (2013). Echoes in correlated neural systems. New J. Phys. 15, 023002 10.1088/1367-2630/15/2/023002 [DOI] [Google Scholar]
- Kriener B., Helias M., Aertsen A., Rotter S. (2009). Correlations in spiking neuronal networks with distance dependent connections. J. Comput. Neurosci. 27, 177–200 10.1007/s10827-008-0135-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kriener B., Tetzlaff T., Aertsen A., Diesmann M., Rotter S. (2008). Correlations and population dynamics in cortical networks. Neural Comput. 20, 2185–2226 10.1162/neco.2008.02-07-474 [DOI] [PubMed] [Google Scholar]
- Lindner B. (2004). Interspike interval statistics of neurons driven by colored noise. Phys. Rev. E 69, 0229011–0229014 10.1103/PhysRevE.69.022901 [DOI] [PubMed] [Google Scholar]
- Lindner B., Schimansky-Geier L. (2001). Transmission of noise coded versus additive signals through a neuronal ensemble. Phys. Rev. Lett. 86, 2934–2937 10.1103/PhysRevLett.86.2934 [DOI] [PubMed] [Google Scholar]
- Mainen Z. F., Sejnowski T. J. (1996). Influence of dendritic structure on firing pattern in model neocortical neurons. Nature 382, 363–366 10.1038/382363a0 [DOI] [PubMed] [Google Scholar]
- Marti D., Rinzel J. (2013). Dynamics of feature categorization. Neural Comput. 25, 1–45 10.1162/NECO_a_00383 [DOI] [PubMed] [Google Scholar]
- Moreno R., de la Rocha J., Renart A., Parga N. (2002). Response of spiking neurons to correlated inputs. Phys. Rev. Lett. 89, 288101 10.1103/PhysRevLett.89.288101 [DOI] [PubMed] [Google Scholar]
- Moreno-Bote R., Renart A., Parga N. (2008). Theory of input spike auto- and cross-correlations and their effect on the response of spiking neurons. Neural Comput. 20, 1651–1705 10.1162/neco.2008.03-07-497 [DOI] [PubMed] [Google Scholar]
- Murphy B. K., Miller K. D. (2009). Balanced amplification: a new mechanism of selective amplification of neural activity patterns. Neuron 61, 635–648 10.1016/j.neuron.2009.02.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pernice V., Staude B., Cardanobile S., Rotter S. (2012). Recurrent interactions in spiking networks with arbitrary topology. Phys. Rev. E 85, 031916 10.1103/PhysRevE.85.031916 [DOI] [PubMed] [Google Scholar]
- Renart A., Moreno-Bote R., Wang X. J., Parga N. (2007). Mean-driven and fluctuation-driven persistent activty in recurrent networks. Neural Comput. 19, 1–46 10.1162/neco.2007.19.1.1 [DOI] [PubMed] [Google Scholar]
- Roxin A., Brunel N., Hansel D. (2005). The role of delays in shaping spatio-temporal dynamics of neuronal activity in large networks. Phys. Rev. Lett. 94, 238103 10.1103/PhysRevLett.94.238103 [DOI] [PubMed] [Google Scholar]
- Salinas E., Sejnowski T. J. (2000). Impact of correlated synaptic input on output firing rate and variability in simple neuronal models. J. Neurosci. 20, 6193–6209 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schüz A., Braitenberg V. (1994). Constraints to a random plan of cortical connectivity, in Structural and Functional Organization of the Neocortex, eds Albowitz B., Albus K., Kuhnt U., Nothdurft H.-C., Wahle P. (Heidelberg: Springer-Verlag; ), 161–169 [Google Scholar]
- Shriki O., Hansel D., Sompolinsky H. (2003). Rate models for conductance-based cortical neuronal networks. Neural Comput. 15, 1809–1841 10.1162/08997660360675053 [DOI] [PubMed] [Google Scholar]
- Siegert A. J. (1951). On the first passage time probability problem. Phys. Rev. 81, 617–623 10.1103/PhysRev.81.617 [DOI] [Google Scholar]
- Silberberg G., Bethge M., Markram H., Pawelzik K., Tsodyks M. (2004). Dynamics of population codes in ensembles of neocortical neurons. J. Neurophysiol. 91, 704–709 10.1152/jn.00415.2003 [DOI] [PubMed] [Google Scholar]
- Tchumatchenko T., Wolf F. (2011). Representation of dynamical stimuli in populations of threshold neurons. PLoS Comput. Biol. 7:e1002239 10.1371/journal.pcbi.1002239 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tetzlaff T., Helias M., Einevoll G., Diesmann M. (2012). Decorrelation of neural-network activity by inhibitory feedback. PLoS Comput. Biol. 8:e1002596 10.1371/journal.pcbi.1002596 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Usher M., Stemmler M., Olami Z. (1995). Dynamic pattern formation leads to 1/f noise in neural populations. Phys. Rev. Lett. 74, 326–329 10.1103/PhysRevLett.74.326 [DOI] [PubMed] [Google Scholar]
- Vilela R. D., Lindner B. (2009). Comparative study of different integrate-and-fire neurons: spontaneous activity, dynamical response, and stimulus-induced correlation. Phys. Rev. E 80, 031909 10.1103/PhysRevE.80.031909 [DOI] [PubMed] [Google Scholar]
- Watts D. J., Strogatz S. H. (1998). Collective dynamics of small-world networks. Nature 393, 440–442 10.1038/30918 [DOI] [PubMed] [Google Scholar]
- Wilson H. R., Cowan J. D. (1972). Excitatory and inhibitory interactions in localized populations of model neurons. Biophys. J. 12, 1–24 10.1016/S0006-3495(72)86068-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.