Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2018 May 16;98(4):846–860.e5. doi: 10.1016/j.neuron.2018.04.017

The Dynamical Regime of Sensory Cortex: Stable Dynamics around a Single Stimulus-Tuned Attractor Account for Patterns of Noise Variability

Guillaume Hennequin 1,10,, Yashar Ahmadian 2,3,4,5,9, Daniel B Rubin 2,6,9, Máté Lengyel 1,7,8, Kenneth D Miller 2,3,8
PMCID: PMC5971207  PMID: 29772203

Summary

Correlated variability in cortical activity is ubiquitously quenched following stimulus onset, in a stimulus-dependent manner. These modulations have been attributed to circuit dynamics involving either multiple stable states (“attractors”) or chaotic activity. Here we show that a qualitatively different dynamical regime, involving fluctuations about a single, stimulus-driven attractor in a loosely balanced excitatory-inhibitory network (the stochastic “stabilized supralinear network”), best explains these modulations. Given the supralinear input/output functions of cortical neurons, increased stimulus drive strengthens effective network connectivity. This shifts the balance from interactions that amplify variability to suppressive inhibitory feedback, quenching correlated variability around more strongly driven steady states. Comparing to previously published and original data analyses, we show that this mechanism, unlike previous proposals, uniquely accounts for the spatial patterns and fast temporal dynamics of variability suppression. Specifying the cortical operating regime is key to understanding the computations underlying perception.

Keywords: cortical variability, circuit dynamics, noise correlations, variability quenching, V1, MT, theoretical neuroscience

Highlights

  • A simple network model explains stimulus-tuning of cortical variability suppression

  • Inhibition stabilizes recurrently interacting neurons with supralinear I/O functions

  • Stimuli strengthen inhibitory stabilization around a stable state, quenching variability

  • Single-trial V1 data are compatible with this model and rules out competing proposals


Stimuli suppress cortical correlated variability. Hennequin et al. show that a cortical operating regime of inhibitory stabilization around a single stable state—the “stabilized supralinear network”—explains this suppression’s tuning and timing, while alternative proposed regimes do not.

Introduction

Neuronal activity throughout cerebral cortex is variable, both temporally during epochs of stationary dynamics and across repeated trials despite constant stimulus or task conditions (Softky and Koch, 1993, Churchland et al., 2010). Moreover, variability is modulated by a variety of factors, most notably by external sensory stimuli (Churchland et al., 2010, Kohn and Smith, 2005, Ponce-Alvarez et al., 2013), planning and execution of limb movements (Churchland et al., 2006, Churchland et al., 2010), and attention (Cohen and Maunsell, 2009, Mitchell et al., 2009). Modulation of variability occurs at the level of single-neuron activity, e.g., membrane potentials or spike counts (Finn et al., 2007, Poulet and Petersen, 2008, Cardin et al., 2008, Gentet et al., 2010, Churchland et al., 2010, Tan et al., 2014), as well as in the patterns of joint activity across populations, as seen in multiunit activity or the local field potential (LFP) (Tan et al., 2014, Chen et al., 2014, Lin et al., 2015). Variability modulation shows stereotypical patterns. First, the onset of a stimulus quenches variability overall and, in particular, correlated variability in firing rates that is “shared” across many neurons (Lin et al., 2015, Goris et al., 2014, Ecker et al., 2014, Ecker et al., 2016, Churchland et al., 2010). Moreover, the degree of variability reduction can depend systematically on the tuning of individual cells. For example, in area MT, variability is quenched more strongly in cells that respond best to the stimulus, and correlations decrease more among neurons with similar stimulus preferences (Ponce-Alvarez et al., 2013, Lombardo et al., 2015). Although these patterned modulations of variability are increasingly included in quantitative analyses of neural recordings (Renart and Machens, 2014, Orbán et al., 2016), it is still unclear what they imply about the dynamical regime in which the cortex operates.

There have been two dynamical mechanisms proposed to explain selected aspects of the modulation of cortical variability by stimuli. In “multi-attractor” models, the network operates in a multi-stable regime in the absence of a stimulus, such that it noisily wanders among multiple possible stable states (“attractors”). This wandering among attractors occurs in a concerted way across the population, resulting in substantial shared variability (Figure 1A, top). Stimuli then suppress this shared variability by pinning fluctuations to the vicinity of one particular attractor (Figure 1A, bottom; Blumenfeld et al., 2006, Litwin-Kumar and Doiron, 2012, Deco and Hugues, 2012, Burak and Fiete, 2012, Ponce-Alvarez et al., 2013, Doiron and Litwin-Kumar, 2014, Mochol et al., 2015). In chaotic network models (Sompolinsky et al., 1988), firing rates exhibit strong chaotic fluctuations, and certain types of stimuli can suppress chaos by forcing the dynamical state of the network to follow a specific trajectory, thus quenching across-trial variability (Figure 1B; Molgedey et al., 1992, Bertschinger and Natschläger, 2004, Sussillo and Abbott, 2009, Rajan et al., 2010). While both the multi-attractor and the chaotic mechanisms can explain the general phenomenon of stimulus-induced reduction of variability, only the former has been proposed to explain the stimulus-tuning of variability reduction. However, even in that case, a considerable fine-tuning of parameters or very strong noise was needed to keep the network in the regime with multiple attractors, such that the system stays near attractors, yet noise can move the system between them (Ponce-Alvarez et al., 2013).

Figure 1.

Figure 1

Three Different Dynamical Regimes that Could Explain Variability Modulation by Stimuli

(A–C) Two schematic neural trajectories (red and green) corresponding to two separate trials are plotted for each dynamical regime, before (top) and after (bottom) stimulus onset. Spontaneous activity is redrawn in gray beneath evoked activity to allow comparison of variability. Dotted ellipses outline activity covariances around the fixed point(s) of the dynamics (if any exist).

(A) Multi-attractor dynamics: spontaneous activity wanders stochastically between a set of attractor states (three shown), resulting in large trial-by-trial variability (top). Stimulus onset constrains fluctuations to the vicinity of a single attractor, reducing variability across both time and trials (bottom).

(B) Chaos suppression: chaos yields large across-trial variability in spontaneous dynamics (top), which is suppressed by the stimulus, leading to a reduction of variability across trials but not necessarily across time (bottom).

(C) Stochastic SSN: both spontaneous and evoked dynamics are stable with a single fixed point, but the stimulus can shrink the effective size of the basin of attraction of the fixed point (as well as shifting its location), resulting in a reduction of both across-time and across-trial variability.

Here, we explore a qualitatively different regime of cortical dynamics. We describe activity fluctuations as being driven by noise but shaped by nonlinear, recurrent interactions. In contrast to previous models, our network operates around a single stable point that depends on the stimulus (Figure 1C). Crucially, individual neurons have supralinear (expansive) input/output functions. This causes the gains of neurons, and thus the effective synaptic strengths in the network, to increase with network activation. This is a stochastic generalization of the stabilized supralinear network (SSN) model that has successfully accounted for a range of phenomena related to the stimulus dependence of trial-averaged responses in visual cortex (Ahmadian et al., 2013, Rubin et al., 2015). Introducing stochasticity allows us to model the variability of responses and thus use data on neural variability to identify hallmarks of this regime and distinguish it from previous proposals.

In our network, stimulus-dependent changes in effective connectivity shape the magnitude and structure of activity fluctuations in the network. Specifically, stimuli change the balance of two opposing effects of recurrent network dynamics on variability: hidden feedforward interactions (“balanced amplification”; Murphy and Miller, 2009, Hennequin et al., 2014) and recurrent excitation, which amplify variability and dominate for very weak (spontaneous) inputs; and stabilizing inhibitory feedback, which quenches variability (Renart et al., 2010, Tetzlaff et al., 2012) and dominates for stronger inputs.

By studying this network mechanism in a progression of recurrent architectures with increasingly detailed structure, we find that it naturally and robustly explains the modulation of shared cortical variability by stimuli, including its tuning dependence. We first analyze variability in the simplest instantiation of the model, with two unstructured populations of excitatory (E) and inhibitory (I) cells, and find that an external stimulus can strongly modulate the variability of population activities. In particular, the model predicts stimulus-induced quenching of variability, as well as a reduction of the low-temporal-frequency coherence between local population activity and single-cell responses, as found experimentally (Poulet and Petersen, 2008, Churchland et al., 2010, Chen et al., 2014, Tan et al., 2014). Next, we extend our analysis to a more detailed architecture with structured connectivity to account for the tuning-dependent modulations of Fano factors and noise correlations by stimuli. Critically, these results reveal robust qualitative differences between the predictions of our model and those of previously proposed network mechanisms, based on multi-attractor or chaotic dynamics, for both the spatial patterns and temporal dynamics of variability suppression. We tested these predictions against experimental data and found the SSN model to be the most consistent with previously analyzed data from primary visual cortex (V1) and MT (Churchland et al., 2010, Ponce-Alvarez et al., 2013) as well as with our novel analyses of published V1 recordings in the awake monkey (Ecker et al., 2010). Such comparisons of different models are crucial for guiding future experiments that can make targeted measurements to fully resolve the dynamical regime in which the cortex operates—a key first step in identifying the computational strategies underlying perception.

Results

We used a standard model to study the dynamical evolution of momentary firing rates in a recurrently coupled network of excitatory and inhibitory neurons (Figure 2A; Dayan and Abbott, 2001; see also STAR Methods). In this model, neurons integrate their external and recurrent inputs linearly in their membrane potentials, Vm, but their output firing rates, r, are a nonlinear function of the voltage: r=f(Vm) (Figure 2B). Crucially, we studied variants of this model in which the nonlinearity f is an expansive (supralinear) function (Figure 2B) and in which inhibition was both sufficiently fast and strong and appropriately structured to stabilize the network in the face of recurrent excitation and the supralinear input/output function. This is the stabilized supralinear network (SSN) model (Ahmadian et al., 2013). In order to study response variability, we added to this model a stochastic component (slow noise) in the membrane potential dynamics of all cells. Stabilization meant that the network operated around a single steady state, albeit a stimulus-dependent one.

Figure 2.

Figure 2

Activity Variability in a Reduced, Two-Population Stochastic SSN

(A) The network is composed of two recurrently connected units, summarizing the activity of two populations of excitatory (red) and inhibitory (blue) neurons. Both units receive private input noise and a common constant input h.

(B) Threshold-quadratic neural input/output function determining the relationship between membrane potential and momentary firing rate of model neurons (Equation 2).

(C) Sample VE/I traces for the two units (top), as the input is increased in steps from h=0 to 2 mV to 15 mV (bottom).

(D) Dependence of population activity statistics on stimulus strength h. Top: mean E (red) and I (blue) firing rates; middle: mean VE/I; bottom: standard deviation of VE/I fluctuations. The comparison with a purely feedforward network (W=0) receiving the same input h is shown in gray. Dots are based on numerical simulations of 500 trials. Solid lines show analytical approximations (Hennequin and Lengyel, 2016).

Real neurons, of course, have an input/output function that ultimately saturates. We focus on an expansive, non-saturating input/output function because V1 cortical neurons show such a relationship between mean voltage and firing rate across their full dynamic range, without saturation even for the strongest visual stimuli (Priebe and Ferster, 2008). Thus, saturation does not appear to play a role in stabilizing cortical activity, a fact that we capture by using a non-saturating input/output function. Such an expansive input/output function arises in spiking neurons when their firing is driven by voltage fluctuations, with the mean voltage sub- or peri-threshold (Hansel and van Vreeswijk, 2002, Miller and Troyer, 2002), a firing regime that produces the highly variable spiking seen in cortical neurons (Troyer and Miller, 1997, Amit and Brunel, 1997). We assume that the voltage fluctuations giving rise to the expansive input/output function are fast compared to the timescales of variability studied here and do not explicitly model them.

We focused on analyzing how the intrinsic dynamics of the network shaped fixed input noise to give rise to stimulus-dependent patterns of response variability. We studied a progression of connectivity architectures of increasing complexity, all involving two separate populations of excitatory and inhibitory neurons. We also validated our results in large-scale simulations of spiking neuronal networks.

Variability of Population Activity: Modulation by External Input

We first considered a simple circuit motif: an excitatory (E) unit and an inhibitory (I) unit, recurrently coupled and receiving the same mean external input h as well as their own independent noise (Figure 2A). In this reduced model, the two units represent two randomly connected populations of E and I neurons, a canonical model of cortical networks (Amit and Brunel, 1997, Vogels et al., 2005). Thus, their time-varying activity, VE(t) and VI(t), represents the momentary population-average membrane potential of all the E and I cells, respectively. Despite its simplicity, this architecture accounted well for the overall population response properties in the larger networks, with more detailed connectivity patterns, that we analyzed later.

Activity in the network exhibited temporal variability due to the stochastic component of the dynamics. We found that this (correlated) variability of VE and VI fluctuations, together with their means, V¯E/I, was strongly modulated by the external steady input h (Figures 2C and 2D). When h=0, there was no input to drive the network, and VE and VI both hovered around Vrest=70 mV, fluctuating virtually independently, with standard deviations essentially matching those that would arise without recurrent connections (gray line in Figure 2D, bottom). For a somewhat larger input, h=2 mV, both E and I populations fired at moderate rates (3–4 Hz) (Figure 2D, top), but now also exhibited large and synchronous population Vm fluctuations (Figure 2C, black circle mark). For yet larger inputs (h=15 mV), fluctuations remained highly correlated, but their magnitude was strongly quenched (Figure 2C, green circle mark).

Figure 2D shows how the temporal (or, equivalently, the across-trial) mean and variability of activities varied over a broad range of input strengths. We observed that population mean Vm increased monotonically with growing external input, first linearly or supralinearly for small inputs, but strongly sublinearly for larger inputs, with V¯I growing faster than V¯E (Figure 2D, middle; Ahmadian et al., 2013, Rubin et al., 2015). In contrast, variability in both VE and VI typically increased for small inputs, peaking around this transition between supralinear and sublinear growth, and then decreased with increasing input (Figure 2D, bottom). Importantly, input modulation of variability required recurrent network interactions. This was revealed by comparing our network to a purely feedforward circuit that exhibited qualitatively different behavior (Figure 2D, gray). In the feedforward circuit, mean Vm remained linear in h, so that mean rates rose quadratically with Vm or h (reflecting the input/output nonlinearity; Figure 2B), and fluctuations in Vm no longer depended on the input strength.

Variability Suppression with a Single Stable State Is a Robust Phenomenon

In order to demonstrate that the overall dynamical regime of the stabilized supralinear network, rather than just a particular instantiation of our model, underlies variability modulation, we used a combination of numerical simulations and analytical results to confirm the robustness of our findings.

We simulated 1,000 model networks with random parameter values within wide brackets. We found that variability suppression was robust over a broad range of network parameters (connection weights, input strengths and correlations, and the exponent and scale of the firing-rate nonlinearity), as long as they ensured dynamical stability even for strong inputs (Figures S1 and S2). Although the precise amplitude and position of the peak of Vm variance depended on network parameters, the overall non-monotonic shape of variability modulation was largely conserved. In particular, we could show analytically that variability suppression occurs earlier (for smaller input h) in networks with strong connections or, for fixed overall connection strength, in networks that are more dominated by feedback inhibition (Methods S3). More generally, we found that the firing rates at the peak of variability are typically low (2.5 Hz on average over a thousand randomly parameterized stable networks and below 6 Hz for 90% of them; cf. Methods S2). As these rates are comparable to cortical spontaneous firing rates, this predicts that increased sensory drive should generally result in variability quenching in cortical LFPs.

In order to better understand the robustness of variability suppression in the model, we took advantage of the fact that our network was characterized by a single attractor at each level of the input, h, and analyzed the dynamics of small activity fluctuations, δV, around this stable state (such that V=V¯(h)+δV, where V¯(h) is the mean activity in the stable state; STAR Methods). These dynamics are governed by a set of effective connection weights, Weff, that quantify the impact of a small momentary change in the Vm of the presynaptic neuron on the total input to its postsynaptic partner. The dependence of the effective connection weights on the stable state and thus on the external input, h, that determines the stable state is simply given by:

Wijeff(h)Wijf'[V¯j(h)] (Equation 1)

where Wij is the strength of the “biophysical” connection from unit j to unit i, and f' is the slope of the single-neuron firing-rate nonlinearity at the stable state. Importantly, f' increases with increasing V¯(h), because f is an expansive, convex nonlinearity (Figure 2B). Thus, in general, effective connectivity increases with increasing h, reflecting the growth of V¯(h) (Figure 2D, middle).

An increase in effective connectivity can have conflicting effects: it can increase excitatory or driving effects that amplify fluctuations and increase variability (Murphy and Miller, 2009, Hennequin et al., 2014), but it can also increase inhibitory feedback, suppressing fluctuations and decreasing variability (Renart et al., 2010, Tetzlaff et al., 2012). Thus, understanding how changes in effective connectivity translate into changes in variability required further analysis (Methods S3). We found that the net behavior of the network indeed included a combination of both effects (Figure S3). As the input grew from zero, variability first rapidly increased, due primarily to the growth of effective feedforward weights (“balanced amplification”; Murphy and Miller, 2009) but also of recurrent excitatory loops. Then, beginning at firing rates comparable to spontaneous activity as described above, variability steadily decreased with increasing stimulus strength due to increasingly strong inhibitory feedback (Figure 2D, bottom).

Crucially, we were able to show analytically that variability quenching effects must ultimately dominate, leading to progressively stronger quenching of variability as the input increases. This is due to the faster growth of I activity relative to E activity in the network, which is a robust outcome of dynamic stabilization by feedback inhibition (Figure S1; Ahmadian et al., 2013, Rubin et al., 2015) and which has been observed in rodent S1 (Shao et al., 2013) and V1 (Adesnik, 2017). We also found that ignoring the variability-increasing effects, which are characteristic of excitatory-inhibitory dynamics (Kriener et al., 2008, Murphy and Miller, 2009) and thus largely absent from models that do not include separate excitatory and inhibitory populations, can fail to capture the full extent of variability modulation and lead to an underestimation of the level of spontaneous variability obtained at zero-to-weak input levels (Figure S4).

Variability Quenching and Synchronization in Single Neurons

In order to study variability in single neurons and at the level of spike counts, we implemented the two-population architecture of Figure 2A in a network of spiking neurons (Figure 3; STAR Methods). The network consisted of 4,000 E neurons and 1,000 I neurons, randomly connected with low probability and with synaptic weights chosen such that the overall connectivity matched that of the reduced model. Each neuron emitted action potentials stochastically with an instantaneous rate given by Equation 3 (this additional stochasticity accounted for the effects of unmodelled fluctuations in synaptic inputs that occur on timescales faster than the 30 ms effective time resolution of our model; Methods S4). The external input to the network again included a constant term, h, and a noise term that was temporally correlated on a 50 ms timescale with uniform spatial correlations of strength 0.2.

Figure 3.

Figure 3

Modulation of Variability in a Randomly Connected Stochastic Spiking SSN

(A) Top: raster plot of spiking activity, for 40 (out of 4,000) excitatory neurons (red) and 10 (out of 1,000) inhibitory neurons (blue). Upper middle: momentary E and I population firing rates. Lower middle: LFP (momentary population-averaged Vm). Bottom: Vm of two randomly chosen excitatory neurons. The dashed vertical line marks the onset of stimulus, when h switches from 2 mV to 15 mV. Population firing rates, LFP, and Vm traces were smoothed with a Gaussian kernel of 50 ms width.

(B) Top, normalized LFP power in spontaneous (black) and evoked (green) conditions; bottom, average (± SEM) spectral coherence between single-cell Vm and the LFP; left, model; right, data from V1 of the awake monkey, reproduced from Tan et al. (2014).

At the population level, the network behaved as predicted by the reduced model. Neurons fired irregularly (Figure 3A, top), with firing rates that grew superlinearly with small input h but sublinearly with stronger input (Figure S5). Moreover, fluctuations in E and I population activities were strongly synchronized (Figure 3A, upper middle), and LFP variability decreased with increasing h (Figure 3A, lower middle). Importantly, variability quenching also occurred at the level of individual neurons’ Vm, accompanied by a reduction of pairwise correlations (Figure 3A, bottom; these required that single neurons shared part of their input noise; Methods S3).

The model primarily suppressed shared rather than private (to individual neurons) variability (Figure S5), as in experiments (Churchland et al., 2010). This was because the spatially uniform average connectivity of the network meant that its dynamics were only significantly coupled to patterns of uniform activity across E or across I cells. These patterns were thus the ones affected by stimulus-induced changes in effective connectivity (Figure S3). Correlated noise drove such uniform patterns so that they carried significant variability. Thus, these shared excitatory and inhibitory activity patterns behaved as the activity of the individual units of the previous reduced two-population model, and so variability suppression in the reduced model implied the suppression specifically of shared variability in this more detailed model.

Our model also accounted for the stimulus-induced modulation of the power spectrum and cross-coherence of LFP and single-cell Vm fluctuations, as observed in V1 of the awake monkey (Figure 3B; Tan et al., 2014). Strong external input reduced the LFP power at low frequencies, due to enhanced effects of feedback inhibition; increased it at intermediate frequencies, due to the faster timescales associated with relatively enhanced inhibition; and also increased it at high frequencies, due to the larger firing rates, which contributed additional, high-frequency fluctuations in synaptic drive (Figure 3B, top left). This asymmetric modulation of LFP power at low and high frequencies is also seen in experiments (Figure 3B, top right). Moreover, as increasing inputs suppressed variability at the population level, the private noise in the activity of each neuron had a proportionately larger contribution to its overall variability, leading to a drop in pairwise correlations (Figure 3A) and Vm-LFP coherence specifically at low frequencies where the suppression of population variability occurred, as seen in experiments (Figure 3B, bottom).

Stimulus-Tuning of Variability Suppression in V1

Neuronal recordings in visual areas have shown that Fano factors drop at the onset of the stimulus (drifting gratings or plaids) in almost every neuron, which was well accounted for by the randomly connected network we studied above. However, in the experiments, variability did not drop uniformly across cells, but exhibited systematic dependencies on stimulus tuning (Ponce-Alvarez et al., 2013, Lombardo et al., 2015, Lin et al., 2015). This could not be explained by randomly connected architectures, so we extended our model to include tuning dependence in connectivity and input noise correlations.

We studied an architecture in which the preferred stimulus of E/I neuron pairs varied systematically around a “ring” representing an angular stimulus variable, such as stimulus orientation in V1 or motion direction in MT (Figure 4A; STAR Methods). We describe the case in which the variable is orientation, which ranges from 0 to 180o; identical results describe direction if all angles are doubled. The average input to a cell (either E or I) was composed of a constant baseline, which drove spontaneous activity in the network, and a term that depended on the angular distance between the stimulus orientation and the preferred orientation (PO) of the cell, and that scaled with image contrast, c (Figure 4C). Input noise correlations depended on tuning differences (STAR Methods): cells with more similar tuning received more strongly correlated inputs. The strength of recurrent connections depended on the difference in preferred orientation between pre- and postsynaptic neurons and whether they were excitatory or inhibitory (Figure 4B).

Figure 4.

Figure 4

Modulation of Variability in a Stochastic SSN with a Ring Architecture

(A) Schematics of the ring architecture. Excitatory (red) and inhibitory neurons (blue) are arranged on a ring, their angular position indicating their preferred stimulus (expressed here as preferred stimulus orientation, PO). The stimulus is presented at 0.

(B) Synaptic connectivities all follow the same circular Gaussian profiles with peak strengths that depend on the type of pre- and post-synaptic populations (excitatory, E, or inhibitory, I).

(C) Each neuron receives a constant input with a baseline (black line, 0% contrast), which drives spontaneous activity, and a tuned component with a bell-shaped dependence on the neuron’s preferred orientation and proportional to contrast, c (dark and light green, 50% and 100% contrast, respectively). Neurons also receive spatially and temporally correlated noise, with spatial correlations that decrease with tuning difference (see Figure 5D).

(D) Single-trial network activity (E cells), before and after the onset of the stimulus (100% contrast). Neurons are arranged on the y axis according to their preferred stimuli.

(E) Reduction in membrane potential variability across trials: membrane potential traces in 5 independent trials (top) and Fano factors (bottom) for an E cell tuned to the stimulus orientation (left) or tuned to the orthogonal orientation (right). For Vm, orange and brown lines and shading show (analytical approximation of) across-trial mean ± SD.

(F) Reduction of average spike count Fano factor in the population following stimulus onset in the model (top) and experimental data (bottom). Spikes were counted in 100 ms time windows centered on the corresponding time points.

(G) Mean firing rates (top), Fano factors (middle), and std. of voltage fluctuations (bottom) at different contrast levels as a function of the neuron’s preferred stimulus in the model (left) and, for rate and Fano factor, experimental data (right, averaged across 99 neurons). Colors indicate different contrast levels (model: colors as in C; data: black, spontaneous, green, 100% contrast).

(H) Shared variability in normalized spike counts, as estimated via factor analysis (STAR Methods; Churchland et al., 2010), before (spontaneous, black) and after stimulus onset (evoked, green) in the model (left) and experimental data (right). Dots in (F) and (G) are based on numerical simulations of 500 trials. For the model, colored lines and shaded areas in (E) and solid lines in (F) and (G) show analytical approximations (Hennequin and Lengyel, 2016). Experimental data analyzed in (F)–(H) are from awake monkey V1 (Ecker et al., 2010), with error bars denoting 95% CI.

The bump of stimulus-driven input drove a similar, but narrower, bump of network response (Figures 4D and 4G). Although this architecture appears similar to a form of multi-attractor model that has a continuum of attractors—a bump of activity that (in the absence of stimuli) can be centered at any location (the so-called “ring attractor model”; Goldberg et al., 2004, Ben-Yishai et al., 1995, Ponce-Alvarez et al., 2013)—our model was actually quite different. While multi-attractor networks show a bump of sustained activity even once the stimulus is removed (leaving only non-specific background excitation), in our network the bump of activity depends on the similar bump of stimulus-driven input. When the stimulus is removed, our network returns within a single membrane time constant to a homogeneous level of baseline activity, driven by the homogeneous baseline input (Figure 4D). As we show below, this dynamical regime is also characterized by fundamentally different patterns of response variability than multi-attractor dynamics.

We applied this model to study the stimulus dependence of variability quenching in V1 and compared our results to a new analysis we performed of previously published recordings in V1 of the awake monkey (Ecker et al., 2010). In the absence of visual input (0% contrast), the network exhibited spatially patterned fluctuations in momentary firing rates around a few Hz (Figure 4D) with large across-trial variability in single-cell Vm (Figure 4E). In evoked conditions, the input drove a hill of network activity around the stimulus orientation as in the data (Figures 4D and 4G), resulting in approximately contrast-invariant tuning curves (Priebe and Ferster, 2008). At stimulus offset, activity rapidly decayed back to spontaneous levels with the cellular time constant (Figure 4D), as observed in cortex when thalamic input is silenced (Reinhold et al., 2015, Guo et al., 2017).

The fluctuating firing rates in spontaneous activity implied super-Poisson variability in spike counts—Fano factors greater than 1 (Figure 4F, top) —given the stochastic spiking mechanism described above (Figure 3). This was consistent with the high level of spontaneous variability in the data (Figure 4F, bottom). Both the model and the data exhibited a pronounced drop in Fano factor following stimulus onset (Figure 4F) and displayed a U-shaped tuning of variability suppression with stimulus orientation (Figure 4G, middle), such that variability suppression was stronger for cells whose preferred orientation was close to the stimulus. The model made similar predictions for variability in membrane potentials: a U-shaped profile of Vm variance suppression in stimulus-evoked conditions relative to spontaneous fluctuations (Figure 4G, bottom).

Notably, for similar reasons as in the randomly connected network (Figure 3; Figure S5), it was primarily the shared and not the private part of variability that was quenched by stimuli in the model (Figure 4H, left), and this required some degree of spatial correlations in the input noise (Figure S6). This was because the spatially smooth nature of the connectivity meant that only spatially smooth patterns of activity were strongly coupled to the network dynamics. A substantial suppression of shared variability at stimulus onset has been observed across many cortical areas (Churchland et al., 2010) as well as in our analysis of the V1 data (Figure 4H, right; Ecker et al., 2010; see also Lombardo et al., 2015).

We again explored a broad range of parameters to show that the tuning of variability suppression was a robust outcome of the model. We found that Fano factor and Vm variance were always most strongly suppressed in the neurons that were most strongly driven by the stimulus (the “dip” of the U shape) consistent with the V1 data (see above). Interestingly, there were some cases when neurons tuned to the opposite stimulus also showed a strong reduction of Fano factor (though not of membrane potential variance; Figure S7)—consistent with recent findings of an M-shaped modulation of Fano factors (and spike count correlations of similarly tuned cells) in area MT of the awake macaque (Figure S7; Ponce-Alvarez et al., 2013). However, while such an M-shaped modulation was previously attributed to marginally stable multi-attractor dynamics (Ben-Yishai et al., 1995, Ponce-Alvarez et al., 2013), our model still produced this with a single stable attractor: the spike count variability of oppositely tuned cells dropped when input tuning in the model was as narrow as, or narrower than, the tuning of recurrent connections. In this configuration, oppositely tuned cells received so small a net input on average that their membrane potential fluctuations barely crossed the threshold of the firing rate nonlinearity, thus producing very little spiking variability. In turn, this loss of firing rate variance even overcame the effect of dividing by very small firing rates in computing Fano factors for these neurons. Under the same conditions, a similar M shape was apparent for spike count correlations between similarly tuned neurons, as a function of their (common) preferred orientation (Figure S7).

Patterns of Noise Variability Arise from Low-Dimensional Bump Kinetics

Next, we analyzed the origin and mechanism of the stimulus-tuning of noise variability in the ring architecture. As mentioned above, for a fixed stimulus, the most prominent feature of population activity was a “bump” of high Vm in the cells with preferred orientations near the stimulus orientation and a lower baseline of activity in the surround (Figure 5A, left and middle). In general, variability in the bump and the baseline captured most of the network’s variance and its suppression with increasing stimulus strength (Figure S8). Here and in the next section we specifically focus on the structure of the quenched noise variability after stimulus onset.

Figure 5.

Figure 5

Low-Dimensional Bump Kinetics Explain Noise Variability in the Ring SSN

(A) Sample of Vm fluctuations across the network in the evoked condition (left, “true activity,” 100% contrast), to which we fitted a circular-Gaussian function (bump) Vi(t)=a(t)exp[(cos(θiμ(t))1)/σ2(t)] across the excitatory population in each time step (center), parametrized by its location, μ, and width, σ. The amplitude of the bump, a, was chosen in each time step so as to keep total population firing rate constant. Fluctuations in location and width were independent, and the fit captured 87% of the variability in Vm (right).

(B) The two principal modes of bump kinetics: small changes (red arrows) in location (top) and width (bottom) of the activity bump result in the hill of network activity deviating from the prototypical bump (gray shadings). Plots on the right show how the activity of each neuron changes due to these modes of bump kinetics.

(C) Time series of μ and σ extracted from the fit.

(D) Ongoing fluctuations in each bump parameter contribute a template matrix of Vm covariances (color maps show covariances between cells with preferred orientation [PO] indicated on the axes of the “full” matrix, bottom right), obtained from (the outer product of) the differential patterns on the right of (B). Insets show Vm covariance implied by each template for pairs of identically tuned cells (orange, PO difference 0) and orthogonally tuned cells (gray, PO difference =90), as a function of stimulus orientation relative to the average PO of the two cells. The two templates sum up to a total covariance matrix (“bump kinetics”), which captures the key qualitative features of the full Vm covariance matrix (“full”). The covariance matrix of the input noise (“input”) is also shown above for reference. The stimulus is at 0 throughout.

After stimulus onset, most of the shared variability (87%; Figure S8) arose from variability in the location, μ, and width, σ, of the bump of activity (Figure 5A, middle and right). Notably, fluctuations in bump amplitude and width scaled inversely with one another, as the nonlinear interactions among neurons in our network resulted in strong normalization (Ahmadian et al., 2013, Rubin et al., 2015), preserving overall activity. Each of these small transformations resulted in a characteristic pattern of momentary deviation of network activity from the mean bump (Figure 5B). In turn, these two patterns of momentary fluctuations (Figure 5C) contributed two distinct spatial covariance templates (Figure 5D). For example, sideways motion of the bump increased the firing rates of all the cells with preferred orientations on one side of the stimulus orientation and decreased firing rates for all cells on the other side (Figure 5B, top). This resulted in positive covariances between cells with preferred orientations on the same side of the stimulus orientation and negative covariances for cells on opposite sides (Figure 5D, top: μ-template; Moreno-Bote et al., 2014). Conversely, an increase in bump width (and thus a decrease in amplitude) increased the activities of cells on the flanks of the bump, tuned away from the stimulus, while decreasing the activity of cells near the peak, tuned for the stimulus (Figure 5B, bottom). This generated positive covariances within each of these groups and negative covariances between the two groups (Figure 5D, bottom: σ-template).

Taken together, the ongoing jitter in bump location and width contributed a highly structured pattern of response covariances, which accounted for most of the structure in the full covariance matrix of the network (Figure 5D, compare “bump kinetics” with “full”). In particular, bump kinetics correctly predicted the Vm variances of cells (given by the diagonal of the full covariance matrix indicated by the filled arrow in Figure 5D), showing less variance for cells tuned to the stimulus orientation of 0o than for cells tuned to orthogonal orientations (see Figure 4G, bottom, green), and hence explained the U-shaped modulation of Fano factors (Figure 4G, middle, green). Moreover, the recurrent dynamics generated negative correlations in the Vm fluctuations of cells with orthogonal tuning, despite such pairs receiving positively correlated inputs (Figure 5D, “input” versus “bump kinetics,” secondary diagonal with open arrow).

Experimental Predictions: Stimulus Tuning

For a direct comparison of the dynamical regime of the SSN with previously proposed mechanisms for variability modulation, based on marginally stable or chaotic dynamics, we first studied the predictions of the models for the spatial patterns of spike count noise correlations. Chaotic models have not (Rajan et al., 2010), and probably can not, predict the tuning of mean responses, let alone that of variability suppression, so we focused on a comparison with a multi-attractor ring model. This model has been suggested to account for stimulus-modulated changes in variability in area MT (Ponce-Alvarez et al., 2013). We matched it to our model such that it produced similar tuning curves and overall levels of variability (Figure S9).

While there were several differences apparent in the detailed correlations predicted by the two models (Figures 6A and 6C), many of these could be explained away by trivial factors that neither model captured fully. For example, the average correlation was substantially larger in the SSN than in the attractor network—but this difference could be eliminated by invoking, in the attractor model, an additional (potentially extrinsic) mechanism that adds a single source of shared variability across neurons, resulting in a uniform (possibly stimulus strength-dependent) positive offset to all correlations (Lin et al., 2015). As another example, the attractor network always exhibited an M-shaped modulation of correlations, whereas, just as for Fano factors (see above), the SSN mostly showed a U-shaped modulation but could show an M shape for particular parameters (Figure S7).

Figure 6.

Figure 6

Stimulus Tuning of Spike Count Correlations in the Ring SSN versus the Multi-attractor Ring Model

(A) Spike count correlation matrix in the ring SSN during evoked activity (100% contrast). Color map shows correlations between cells with preferred orientation (PO) indicated on the axes, relative to stimulus orientation at 0. Arrows indicate axes along which cell pairs are similarly (orange) or orthogonally tuned (gray). Spike count correlations along the diagonal show correlation for identically tuned cells, rather than for identical cells, and are thus less than one due to private spiking noise.

(B) Average spike count correlations in the SSN, for pairs of similarly tuned cells (orange, PO difference less than 45) and orthogonally tuned cells (gray, PO difference greater than 45), as a function of stimulus orientation relative to the average PO of the two cells.

(C and D) Same as (A) and (B), for the multi-attractor ring network.

(E) Same as (B) and (D), for data from awake monkey V1 (Ecker et al., 2010). Data were symmetrized for negative and positive stimulus orientations. Shaded regions denote 95% CI. SSN simulations in this figure used the same parameters as in Figures 4 and 5.

Therefore, we focused on distinctions that were robust to model details and followed from a fundamental difference of bump kinetics in the two models: in contrast to the richer patterns of variability generated by the SSN, multi-attractor dynamics showed a more limited repertoire, dominated by sideways motion of the bump with barely any fluctuations in bump width (Figure S9; Burak and Fiete, 2012). As fluctuations in bump location and width had opposite effects on the correlations between orthogonally tuned cells in the SSN model (Figure 5D insets, gray), their cancellation made these correlations only very weakly modulated by the stimulus (Figure 6A, gray arrow; Figure 6B, gray). In particular, this modulation was much shallower than that for similarly tuned cells (Figure 6A, orange arrow; Figure 6B, orange), for which variability in bump location and width had congruent effects (Figure 5D insets, orange) that added to rather than cancelled each other. In contrast, in the attractor model, there was no such cancellation even for orthogonally tuned cells due to the absence of fluctuations in bump width (Figure S9). This meant that correlations between orthogonally tuned cells were just as deeply modulated as those between similarly tuned cells (Figures 6C and 6D).

Previous reports on the stimulus-tuning of noise correlations examined only similarly tuned cells and reported mostly M-shaped modulation, which does not distinguish between the models. Therefore, we conducted our own analyses of a previously published dataset of V1 responses in the awake monkey (Ecker et al., 2010) (Figure 6E). The modulation of these correlations by the stimulus could only be accounted for by the SSN. First, we found that correlations between similarly tuned cells were significantly modulated by the stimulus (Figure 6E, orange; repeated-measures ANOVA F(2,274)=5.29, p=0.006), and this modulation had a U rather than an M shape. More critically, also in agreement with the predictions of the SSN but not of the attractor model, correlations between orthogonally tuned cells were unaffected by the stimulus (Figure 6E, gray; repeated-measures ANOVA F(2,274)=0.04, p=0.961). While the magnitude of correlations in either model was overall larger than in the data, this simply reflected the relatively small number of neurons in the models (model correlations could be decreased without affecting the shape and extent of their stimulus tuning by substituting each model unit by several neurons with independent spiking noise).

Experimental Predictions: Temporal Dynamics of Variability Modulation

We hypothesized that the fundamentally different mechanisms responsible for variability modulation in the SSN, the multi-attractor, and the chaotic dynamical regimes (Figure 1) should be revealed in the dynamics of variability suppression at stimulus onset and of variability recovery at stimulus offset. In order to test this, we used the same models for the SSN and multi-attractor models as above, and we implemented the classical chaotic model of Rajan et al. (2010) (STAR Methods), in which variability suppression had previously been shown to occur. We then measured the across-trial variability (averaged across neurons) following the onset and offset of a step stimulus in each model (Figures 7A–7C, shaded areas), as we parametrically varied the amplitude of the stimulus and therefore the degree of variability suppression (Figures 7A–7C, dark to light colors).

Figure 7.

Figure 7

Temporal Dynamics of Variability Modulation in the SSN versus Other Models

(A) Time course of variability reduction and recovery in the ring SSN in response to a step input (shaded area, 500 ms duration) of increasing amplitude (dark to light). Variability is quantified by the population-averaged across-trial Vm variance.

(B) Same as (A), for chaotic network dynamics (Rajan et al., 2010).

(C) Same as (A), for a continuous, multi-attractor network (Ponce-Alvarez et al., 2013). The stimulus is twice as long as in (A) and (B), so that variability suppression can be observed following the characteristic transient increase.

(D) Timescale of variability suppression (time to reach half of the total suppression) as a function of the percentage of variance suppression in the three models, extracted from their corresponding variability trajectories (colors as in A–C).

(E) Same as (D), for recovery timescales (time to recover half of the total suppression). In both (D) and (E), open yellow squares indicate V1 data from anesthetized cat (estimated from Figure S4 in Churchland et al., 2010); yellow circles show data from anesthetized monkey MT (Figure S4 in Churchland et al., 2010); dotted yellow circles show our analysis of the awake monkey V1 data of Ecker et al., 2010. Variability refers to the above-Poisson part of spike count variability (i.e., population-averaged Fano factor minus one), and time constants discard latencies in data. In the data of Ecker et al., 2010, the Fano factor dropped below one, effectively resulting in >100% variance suppression with our definition (right-pointing arrows). All results regarding the SSN and the multi-attractor model shown in this figure were obtained by using the same parameters as in previous figures (Figures 4, 5, and 6).

In the SSN, the timescales on which both suppression and recovery of variability occurred were nearly as fast as the single-neuron time constant (20 ms in these simulations; Figures 7D and 7E, green). In contrast, in chaotic networks, both these timescales were several (4–15) times longer than the single-neuron membrane time constant (Figures 7D and 7E, blue). More importantly, recovery times were much longer than suppression times in the chaotic network and increased with increasing stimulus strength and thus increasing amount of variability suppression during the stimulus period, neither of which was the case in the SSN. In the multi-attractor network, both the dynamics of the network activity and those of variability were much slower than in the SSN (Figures 7D and 7E, red). Moreover, we found that, unlike in the SSN or the chaotic model, variability increased transiently immediately following stimulus onset (before eventually decreasing to its new steady state; Figure 7C). The cause of this behavior was the slow rotation of the activity bump from its random position at the time of stimulus onset to the location where cells’ preferred orientation matched the stimulus orientation (Figure S10). Thus, we expect this behavior to be generic at least to the subclass of multi-attractor models that have a continuous ring of attractors and thus show such rotational response, which likely include those that can address the orientation- or direction-tuning of variability reduction in V1 and MT.

The timescales of variability suppression and recovery found experimentally in anaesthetized cat V1 and awake monkey MT (Figures 7D and 7E, open square and circle; Churchland et al., 2010) and by our own analysis of awake monkey V1 data (Figures 7D and 7E, dotted square; Ecker et al., 2010) were short and nearly identical. Moreover, recovery times showed little dependence on the amount of variability suppression (comparing across areas), and there was no transient increase in variability at stimulus onset (Figure 4F; Churchland et al., 2010). These results confirm the predictions of the SSN and are at odds with the dynamics of variability modulation as predicted by the multi-attractor and chaotic regimes.

Discussion

We studied the modulation of variability in a stochastic, nonlinear model of cortical circuit dynamics. We focused on a simple circuit motif that captured the essence of cortical networks: noisy excitatory and inhibitory populations interacting in a recurrent but stable way despite expansive single-neuron nonlinearities. This stochastic stabilized supralinear network (SSN) reproduced key aspects of variability in the cortex. During spontaneous activity, i.e., for weak external inputs, model neurons showed large and relatively slow synchronous fluctuations in their membrane potentials. These fluctuations were considerably amplified by the network relative to that expected from the input alone and were quickly quenched and decorrelated by stimuli. The model thus explains and unifies a large body of experimental observations made in diverse systems under various conditions (Churchland et al., 2006, Churchland et al., 2010, Finn et al., 2007, Poulet and Petersen, 2008, Gentet et al., 2010, Poulet et al., 2012, Tan et al., 2014, Chen et al., 2014). Moreover, the drop in variability was tuned to specific stimulus features in a model of V1/MT, also capturing recent experimental findings (Ponce-Alvarez et al., 2013, Lin et al., 2015, Lombardo et al., 2015) as well as our own analyses of a previously published dataset (Ecker et al., 2010).

The main insight of our analysis was that in a network of nonlinear neurons with an expansive firing rate nonlinearity, increasing the input increases the effective connection strengths of the network, which in turn modulates the variability of responses. We identified two opposing effects of increasing effective connectivity on variability: the amplification of variability by excitatory-inhibitory interactions (balanced amplification), which dominates at very low (spontaneous) levels of input, and the quenching of variability by increased inhibitory feedback, which dominates for stimulus-driven input. Critically, these network effects preferentially act on smooth patterns of activity that are aligned with the anatomical connectivity of the network, so that it is the shared component of variability that is suppressed and modulated by stimuli. Taken together, we showed that these mechanisms robustly produced experimentally observed spatial and temporal patterns of variability quenching and modulation, whereas the dynamics of the network always remained in the vicinity of a single attractor state, unlike previously proposed mechanisms based on multi-attractor or chaotic dynamical regimes.

Sources and Effects of Stochasticity

We focused on how the network shapes variability and assumed that the variability originates in correlated noise input to the network; such input correlations could arise due to upstream areas already exhibiting noise correlations (e.g., thalamic input to V1, Sadagopan and Ferster, 2012) and/or because of feedforward connectivity implying shared inputs (e.g., Kanitscheider et al., 2015). In contrast, other models have focused on how circuits intrinsically generate slow correlated variability (Litwin-Kumar and Doiron, 2012, Stringer et al., 2016). Nevertheless, our model also points to an important mechanism for creating shared variability, namely the strong amplification of the input noise by balanced amplification (see also Kriener et al., 2008, Murphy and Miller, 2009, Hennequin et al., 2014).

Although most of our analyses were based on rates, rather than spikes, the effect of fast fluctuations resulting from spiking noise were not ignored, but were incorporated implicitly in the power-law input/output nonlinearity of neurons in the model (Equation 3) and in the stochastic spike-generation mechanism used in our spiking network simulations (Figure 3, STAR Methods). Theoretical work (Miller and Troyer, 2002, Hansel and van Vreeswijk, 2002) shows that these fast fluctuations are the key factor causing momentary firing rates (on the 30–50 ms timescale of Vm fluctuations considered here) to be a supralinear, power-law function of mean voltages, a critical feature of our model. As experiments, as well as our model, show that only the shared but not the private part of variability is modulated by stimuli (Churchland et al., 2010), we expect our assumption that the exponent of the threshold power-law nonlinearity can be considered constant (implying that fast private spiking fluctuations are not affected by stimuli) to be valid to a good approximation. We also expect that a more detailed model explicitly including these fast fluctuations would allow a more systematic study of the effects of stimuli on high-frequency (gamma) oscillations (Ray and Maunsell, 2010), which our current model could only partially account for (Figure 3B).

Tight versus Loose E-I Balance

While we focused on the sources and modulation of slower, correlated fluctuations, a classical model of cortical variability, the “balanced network” (van Vreeswijk and Sompolinsky, 1998), focused on the origin of fast fluctuations from spiking noise. In that model, very large external and recurrent inputs cancel or “balance” to yield a much smaller net input. This mechanism can self-consistently generate the voltage variability to generate irregular spiking. However, the very strong, very fast inhibitory feedback in the balanced network suppresses correlated rate fluctuations away from the stable state (van Vreeswijk and Sompolinsky, 1998, Renart et al., 2010, Tetzlaff et al., 2012), leaving only fast, private variability due to irregular spiking (though “breaking balance” can restore correlated variability; Litwin-Kumar and Doiron, 2012, Rosenbaum et al., 2017). Because the shared variability is already eliminated, stimuli cannot modulate that variability.

As opposed to the “tight balance” between excitation and inhibition in the classical balanced network model, the SSN in the stimulus-driven regime is “loosely balanced”: the same mathematical cancellation of external and recurrent input occurs, but in a regime in which inputs are not large and the net input after cancellation is comparable in size to the factors that cancel (Ahmadian et al., 2013). This regime is supported by observations that external input is comparable to, rather than very much larger than, the net input received by cortical cells (Ferster et al., 1996, Chung and Ferster, 1998, Lien and Scanziani, 2013, Li et al., 2013). This loose balance allows correlated variability to persist and be modulated by stimuli. Variability quenching in the stochastic SSN robustly occurred as the input pushed the dynamics to stronger and stronger inhibitory dominance. Consistent with this, with increasing strength of external input, the ratio of inhibition to recurrent excitation received by neurons in the network increases (Rubin et al., 2015), as observed in layers 2/3 of mouse S1 (Shao et al., 2013) and V1 (Adesnik, 2017). In the balanced network, the ratio of inhibitory to excitatory activity would be fixed regardless of the strength of activation. The balanced network also only yields responses that are linear functions of the input (though see Mongillo et al., 2012), whereas the loosely balanced regime replicates many nonlinear cortical response properties (Rubin et al., 2015), including the profound dependence of correlated variability on stimuli. Although our model does not focus on the origins of fast spiking variability, spiking models in the loosely balanced SSN regime can, given noisy inputs (e.g., Sadagopan and Ferster, 2012), yield the irregular spiking characteristic of cortex (unpublished data).

Further Factors Modulating Variability

We analyzed variability modulation solely as arising from intrinsic network interactions, but other factors may also contribute (Doiron et al., 2016). External inputs may be modulated; for example, the drop with contrast in Fano factors in the lateral geniculate nucleus (LGN) has been argued to underlie Vm variability decreases in V1 simple cells (Sadagopan and Ferster, 2012; but see Malina et al., 2016). However, since high-contrast stimuli also cause firing rates to increase in LGN, the total variance of LGN-to-V1 inputs (scaling with the product of the LGN Fano factor and mean rate) is modulated far less by contrast. This provides some justification for our model choice that input variance did not scale with contrast. Changes in input correlations have also been suggested as a potential mechanism underlying variability modulation (Bujan et al., 2015). However, the proposed mechanism would require a stimulus to specifically increase the correlations of the different inputs onto individual cells (and this increase should be tuned to the stimulus) while leaving the correlation of inputs between cells unchanged. This seems difficult to achieve in cortex, where nearby cells are likely to share a significant amount of input and correlations are generally observed to decrease, rather than increase, with stimulus strength (Churchland et al., 2010).

One particular form of external input modulation, that involving changes in brain state, has been proposed to directly contribute to correlated variability in both awake (Poulet and Petersen, 2008, Ecker et al., 2016) and anesthetized cortex (Ecker et al., 2014, Goris et al., 2014, Lin et al., 2015, Mochol et al., 2015), so that a reduction of state switching would underlie the reduction of shared variability (Mochol et al., 2015, Ecker et al., 2016). To the extent that correlated noise in the input to our model is aligned with a uniform activity pattern, this input can also be regarded as having a single scalar “brain state”-like component that is changing in time. However, our analysis suggests that the variability of this component needs not be modulated directly by the stimulus to account for variability quenching in network responses. Instead, our network used its intrinsic mechanisms to quench variability in response to a stimulus. Importantly, these intrinsic mechanisms not only quenched this uniform component of variability (Figure S8), but also produced more complex patterns of variability modulation via “bump” kinetics that a single brain state-dependent mechanism could not account for.

Cellular factors may also modulate variability. For example, inhibitory reversal potential or spike threshold may set boundaries limiting voltage fluctuations, which would more strongly limit voltage fluctuations in more hyperpolarized or more depolarized states, respectively; conductance increases will reduce voltage fluctuations; and dendritic spikes may contribute more to voltage fluctuations in some states than others (Stuart and Spruston, 2015). A joint treatment of external input, cellular, and recurrent effects may be needed to explain, for example, why Vm variability appears strongest near the preferred stimulus in anaesthetized cat V1 (Finn et al., 2007) or why overall Vm variability grows with visual stimulation in some neurons of awake macaque V1 (Tan et al., 2014).

Cellular properties may themselves be subject to change over time, thereby causing changes in variability. For example, various mechanisms (e.g., attention, intrinsic and synaptic plasticity, neuromodulators, anesthetics) can change the input/output gain of single neurons and the synaptic efficacies of the network. As all these changes eventually lead to changes in effective connectivity, our work offers a principled approach to study their effects on variability and is thus complementary to previous studies that focused on the consequences of different anatomical connectivity patterns on correlations (Kriener et al., 2008, Tetzlaff et al., 2012, Ostojic, 2014, Hennequin et al., 2014).

Effects of Normalization on Variability

The nonlinear response properties of our network were crucial for the modulation of variability by stimuli. These nonlinearities had been shown to capture ubiquitous phenomena involving nonlinear response summation to multiple stimuli, including normalization, surround suppression, and their dependencies on stimulus contrast (Rubin et al., 2015, Ahmadian et al., 2013). As such, the SSN reproduces much of the phenomenology of the “normalization model” of cortical responses (Carandini and Heeger, 2011) and provides a circuit substrate for it.

However, while response normalization has previously been studied for deterministic steady-state responses, our results can be interpreted as showing that it also plays a role in the suppression of ongoing variability by stimuli, as well as shaping the structure of stimulus-evoked noise correlations. Specifically, in the deterministic SSN, steady-state responses to multiple stimuli add sublinearly, and as one stimulus becomes stronger than another, the response to their simultaneous presentation becomes “winner take all,” i.e., dominated by the response to the stronger stimulus alone (Rubin et al., 2015). This provides an alternative conceptual explanation of why, in the stochastic SSN, a stronger mean input drive relative to the noise input leads to greater suppression of the noise’s contribution to the total network response, thus quenching variability.

Our results go beyond what could be predicted based on this simple qualitative link between steady-state normalization and variability quenching. First, we found a specific quantitative form of normalization in our network: an approximate conservation of the integrated activity across a bump of activity that forms around cells tuned to the stimulus orientation, despite fluctuations in its width. In turn, this predicted a specific pattern of noise correlations that we found contributed substantially to noise variability in V1 of the awake monkey (Figure 6). Second, we were able to study the dynamics with which variability was suppressed following stimulus onset and recovered following stimulus offset and found a good match to experimental data (Figure 7).

The Origin and Role of Inhibitory Dominance

We found that an increase in inhibitory dominance was necessary for the suppression of variability and correlations in the SSN. In line with that, Stringer et al. (2016) studied rodent A1 and V1 in various awake and anesthetized brain states and found that desynchronized states with weaker correlations were accompanied by enhanced activity of putative fast-spiking inhibitory neurons. By fitting a recurrent spiking E-I network model to the data, they found that enhanced inhibitory feedback was the key factor capturing the suppression of correlations. However, the enhanced dominance of inhibition with increasing network activation, which suppresses correlations, was artificially incorporated into the model by making the inhibitory conductance an exponential function of the inhibitory spike count. In contrast, our model provides a dynamical mechanism by which inhibition becomes increasingly dominant with increasing network activation.

Kanashiro et al. (2017) proposed a mechanism similar to ours for the top-down suppression of correlated variability by attention, rather than bottom-up suppression by a stimulus. They also proposed that this arises from enhanced inhibitory feedback resulting from increased effective connectivity due to expansive input/output functions. However, their conclusions differed significantly from ours. They found that, for attention to suppress variability, attentional input had to be directed dominantly to inhibitory cells, while for attention to increase the gain of response to stimuli, stimuli had to give input dominantly to excitatory cells. Note that this implies that stimuli would not suppress variability. We have found that neither of these conditions are necessary (Methods S4) and that stimuli robustly suppress variability. In particular, increasing input strength decreased variability across a wide range of relative strengths of input to excitatory versus inhibitory cells (Figure S2). The main reason for these differences in conclusions is the special, non-generic parametrization of the model studied by Kanashiro et al. (2017) in which a neuron’s projections to excitatory and to inhibitory neurons were statistically identical, which precluded the SSN regime (Methods S2).

The Dynamical Regime of Cortical Activity

Two proposals have been made previously to explain quenching of variability by a stimulus: a stimulus may quench multi-attractor dynamics to create single-attractor dynamics (Blumenfeld et al., 2006, Litwin-Kumar and Doiron, 2012, Deco and Hugues, 2012, Ponce-Alvarez et al., 2013, Doiron and Litwin-Kumar, 2014, Mochol et al., 2015), and a stimulus may quench chaotic dynamics to produce non-chaotic dynamics (Molgedey et al., 1992, Bertschinger and Natschläger, 2004, Sussillo and Abbott, 2009, Rajan et al., 2010, Laje and Buonomano, 2013). Our results propose a very different dynamical regime underlying variability quenching, which can be distinguished from the multi-attractor or chaos-suppression models.

Conceptually, the stochastic SSN differs from previous models of stimulus-driven quenching of shared variability in exhibiting a single stable state in all conditions—spontaneous, weakly driven, strongly driven—whereas the others show this only when strongly driven. Furthermore, quenching of variability and correlations in the SSN is highly robust, arising from two basic properties of cortical circuits: inhibitory stabilization of strong excitatory feedback (Tsodyks et al., 1997, Ozeki et al., 2009) and supralinear input/output functions in single neurons (Priebe and Ferster, 2008). In contrast, models of multi-attractor or chaotic dynamics can either account only for the modulation of average pairwise correlations (Mochol et al., 2015) or else require considerable fine tuning of connections (Litwin-Kumar and Doiron, 2012, Ponce-Alvarez et al., 2013) to account for more detailed correlation patterns. Moreover, as studied thus far (Rajan et al., 2010, Ponce-Alvarez et al., 2013, Mochol et al., 2015; but see Harish and Hansel, 2015, Kadmon and Sompolinsky, 2015, Mastrogiuseppe and Ostojic, 2017), they typically ignore Dale’s law (the separation of E and I neurons) and its consequences for variability, e.g., balanced amplification. These differences between the SSN and previous models also lead to two main experimentally testable features that we used to distinguish their respective dynamical regimes: the tuning and the timing of variability modulation.

With respect to the stimulus tuning of spike count Fano factors and noise correlations, we found that multi-attractor networks could only predict an M-shaped modulation while the SSN could produce either M- or U-shaped modulations depending on the tuning width of inputs relative to that of connectivity. Indeed, while most types of stimuli in MT were found to result in an M-shaped modulation (Ponce-Alvarez et al., 2013), coherent plaids (Ponce-Alvarez et al., 2013) and random moving dots (Lombardo et al., 2015) in the macaque as well as moving dot fields and drifting gratings in the marmoset (Selina Solomon, personal communication; Solomon et al., 2015) result in a pronounced U-shaped modulation of Fano factors in MT, and our own analyses of V1 data also revealed a U-shaped modulation. Interestingly, our results also suggested that irrespective of the precise shape of the modulation of spike count statistics, membrane potential variability in the SSN should always exhibit a U-shaped profile (Figure 4), which could be tested in future experiments. Critically, we also identified a rarely analyzed aspect of spatial correlation patterns that could most clearly distinguish between different models: the modulation of correlations between orthogonally tuned cells. The SSN predicted only very weak modulation for such cell pairs, while multi-attractor dynamics resulted in modulations that were as deep as for pairs of similarly tuned cells. We found that data from awake macaque V1 supported the SSN.

Another distinctive feature of the SSN regime is the speed of its dynamics, and in particular the speed with which variability is modulated as the stimulus is changed. In contrast to multi-attractor and chaotic dynamics, in which variability modulation happens on timescales that are considerably slower than the single neuron time constant, the SSN produces fast variability modulation on a timescale comparable to the neural time constant. The timescales of variability modulation we extracted from data recorded in monkey visual cortical areas (Churchland et al., 2010, Ecker et al., 2010) were fast, on the order of 20–50 ms, providing further support to the SSN.

In summary, the SSN robustly captures multiple aspects of stimulus modulation of correlated variability and suggests a dynamical regime that uniquely captures a wide array of behaviors of sensory cortex. In turn, our work suggests a principled approach to use data on cortical variability to identify the dynamical regime in which the cortex operates.

STAR★Methods

Key Resources Table

REAGENT or RESOURCE SOURCE IDENTIFIER
Deposited Data

Awake monkey V1 dataset Ecker et al., 2010 http://bethgelab.org/datasets/v1gratings

Software and Algorithms

OCaml (for all simulations) Open source http://www.ocaml.org
Sqlite3 (for V1 data analysis) Sqlite Consortium https://sqlite.org/index.html
Mathematica Wolfram https://www.wolfram.com/mathematica
Gnuplot Open source http://www.gnuplot.info

Contact for Reagent and Resource Sharing

As Lead Contact, Guillaume Hennequin is responsible for all reagent and resource requests. Please contact Guillaume Hennequin at g.hennequin@eng.cam.ac.uk with requests and inquiries.

Method Details

The values of all the parameters mentioned below are listed in the tables below. All differential equations detailed below were integrated using a simple Euler scheme with time step 0.1 ms.

Parameters Used in the SSN Simulations

Symbol Figure 2 Figure 3 Figures 4, 5, 6, and 7 Unit Description
NE 1 4,000 50 Number of excitatory units
NI 1 1,000 50 Number of inhibitory units
τE 20 ms Membrane time constant (E neurons)
τI 10 ms Membrane time constant (I neurons)
Vrest −70 mV Resting membrane potential
V0 −70 mV Rectification threshold potential
k 0.3 mV−n ⋅ s−1 Nonlinearity gain
n 2 Nonlinearity exponent
WEE 1.25 mV ⋅ s E→E connection weight (or sum thereof)
WIE 1.2 mV ⋅ s E→I connection weight (or sum thereof)
WEI 0.65 mV ⋅ s I→E connection weight (or sum thereof)
WII 0.5 mV ⋅ s I→I connection weight (or sum thereof)
τnoise 50 ms Noise correlation time constant
σ0,E 0.2 1 mV Noise standard deviation (E neurons)
σ0,I 0.1 0.5 mV Noise standard deviation (I neurons)
pE 0.1 Outgoing connection probability (E neurons)
pI 0.4 Outgoing connection probability (I neurons)
τsyn 2 ms Synaptic time constants
Δ 0.5 ms Axonal delay
syn 45 deg. Connectivity length scale
noise 60 deg. Noise correlation length scale
stim 60 deg. Stimulus tuning length scale of the input
b 2 mV Input baseline
Amax 20 mV Maximum input modulation (100% contrast)
θstim 0 deg. Stimulus direction

Parameters Used in the Multi-attractor Network Simulations

Symbol Figures 6 and 7 Unit Description
N 100 Number of units
τm 10 ms Membrane time constant
k 0.1 mV−1 Nonlinearity gain
gmax 100 ms−1 ⋅ mV−1 Maximal firing rate
W −40/gmax mV ⋅ s Average connection weight
WΔ 33/gmax mV ⋅ s Tuning-dependent modulation of connection weight
τnoise 50 ms Noise correlation time constant
σ0 0.15 mV Noise standard deviation
noise 60 deg. Noise correlation length scale
stim 60 deg. Stimulus tuning length scale of the input
b 2 mV Input baseline
A 0.1 mV Depth of input tuning
θstim 0 deg. Stimulus direction

Parameters Used in the Chaotic Network Simulations

Symbol Figure 7 Unit Description
N 2,000 Number of units
τm 10 ms Membrane time constant
σW 2 Standard deviation of connection weights

SSN model

Our rate-based networks contained NE excitatory and NI inhibitory units, yielding a total N=NE+NI units. The circuit dynamics were governed by (see also Methods S1):

τidVidt=Vi+Vrest+hi(t)+ηi(t)+jE cellsWijr(Vj)jI cellsWijr(Vj), (Equation 2)

where Vi denotes the Vm of neuron i, τi is its membrane time constant, Vrest is a resting potential, Wij is the (positive or zero) strength of the synaptic connection from neuron j to neuron i, and hi(t) is the potentially time-varying but deterministic component (the mean) of external input to which a noise term ηi(t) is added (see below, “Input noise”). The momentary firing rate of cell j was given by a threshold-powerlaw function of its membrane potential:

r(Vj)=kVjV0+n. (Equation 3)

Experiments support Equation 3 when both membrane potentials and spike counts are averaged in 30 ms time bins (Priebe and Ferster, 2008). Accordingly, Vi in Equation 2 can be understood as the coarse-grained (low-pass filtered) version of the raw somatic membrane potential; in particular it does not incorporate the action potentials themselves. Thus the effective time resolution of our model was around 30 ms which allowed studying the effects of inputs that did not change significantly on timescales shorter than that. Accordingly, in Equation 2 we assumed that external noise had a time constant τnoise=50 ms, in line with membrane potential and spike count autocorrelation timescales found across the cortex (Azouz and Gray, 1999, Berkes et al., 2011, Murray et al., 2014).

Equations 2 and 3 together define the stabilized supralinear network model studied in Ahmadian et al. (2013) and Rubin et al. (2015), but formulated with voltages rather than rates as the dynamical variables (the two formulations are mathematically equivalent when all neurons have the same time constant, Miller and Fumarola, 2012) and with the crucial addition of noise that enables us to study variability. In all the figures of the main text, the exponent of the power-law nonlinearity was set to n=2 (but see Figure S2 for n > 2). Methods S2 explores more general scenarios.

Mean external input

In the reduced rate model of Figure 2, each unit received the same constant mean input h. In the ring model, the mean input to neuron i was the sum of two components,

hi(θstim)=b+cAmaxexp(cos(θiθstim)1stim2). (Equation 4)

The first term b=2 mV is a constant baseline which drives spontaneous activity. The second term models the presence of a stimulus with orientation θstim in the visual field as a circular-Gaussian input bump of “half width” stim centered around θstim and scaled by a factor c (increasing c represents increasing stimulus contrast), taking values from 0 to 1, times a maximum amplitude Amax. We assumed for simplicity that E and I cells are driven equally strongly by the stimulus, though this could be relaxed.

Input noise

The input noise term ηi(t) in Equation 2 was modeled as a multivariate Ornstein-Uhlenbeck process:

τnoisedη=ηdt+2τnoiseΣnoisedξ, (Equation 5)

where dξ is a collection of N independent Wiener processes and Σnoise is an N×N input covariance matrix (see below). Note that Equation 5 implies ηi(t)ηj(t+τ)t=Σijnoisee|τ|/τnoise.

In the reduced two-population model (Figure 2), noise terms were chosen to be uncorrelated, i.e., Σijnoise=σα(i)2δij (where δij=1 if i=j and 0 otherwise), α(i){E,I} is the E/I type of neuron i, and σα2 is the variance of noise fed to population α (see Equation 7 below). In the spiking two-population model (Figure 3), input noise covariance was uniform, such that Σijnoise=σnoise2[δij(1ρ)+ρ], with the pairwise correlation coefficient set to ρ=0.2 (see Figure S5 for the dependence of our results on ρ). In the ring model (Figures 4, 5, 6, and 7), the noise had spatial structure, with correlations among neurons decreasing with the difference in their preferred directions following a circular-Gaussian:

Σijnoise=σα(i)σα(j)exp(cos(θiθj)1noise2), (Equation 6)

where θi and θj are the preferred orientations of neurons i and j (exc. or inh.), and noise is the correlation length (see table “Parameters Used in the SSN Simulations”). The noise amplitude has the natural scaling

σα=σ0,α1+τατnoise(α{E,I}) (Equation 7)

such that, in the absence of recurrent connectivity (W=0), the input noise alone would drive Vm fluctuations of standard deviation σ0,E or σ0,I, measured in mV, in the E or I cells, respectively. We chose values of σ0,E that yielded spontaneous Fano factors in the range 1.3-1.5 where appropriate, and chose σ0,I=σ0,E/2 to make up for the difference in membrane time constants between E and I cells (see table “Parameters Used in the SSN Simulations”).

Connectivity

The synaptic weight matrix in the reduced model was given by

W=(WEEWEIWIEWII), (Equation 8)

where WAB is the magnitude of the connection from the unit of type B (E or I) to that of type A (see table “Parameters Used in the SSN Simulations” for parameter values). In the ring model, connectivity fell off with angular distance on the ring, following a circular-Gaussian profile:

Wijexp(cos(θiθj)1syn2), (Equation 9)

where θi and θj are the preferred orientations of neurons i and j (exc. or inh.), and syn sets the length scale over which synaptic weights decay (see table “Parameters Used in the SSN Simulations”). The connectivity matrix W was further rescaled in each row and in each quadrant, such that the sum of incoming E and I weights onto each E and I neuron (4 cases) matched the values of WEE, WIE, WEI and WII in the reduced model. Thus, all connectivity matrices used in the SSN model obeyed Dale’s law.

Simulated spike counts

To relate the firing rate model to spiking data in Figures 4 and 6, we assumed that action potentials were emitted as inhomogeneous (doubly stochastic) Poisson processes with time-varying rate r(Vm) given by Equation 3. Unlike in the full spiking model (see below), spikes did not “re-enter” the dynamics of Equation 2, according to which neurons influence each other through their firing rates. Spikes were counted in 100 ms time bins and spike count statistics such as Fano factors and pairwise correlations were computed as standard.

Spiking SSN model

Dynamics

In the spiking model (Figure 3), neuron i emitted spikes stochastically with an instantaneous probability equal to dtr(Vi), with time-varying rate r(Vi) given by Equation 3, consistent with how (hypothetical) spikes were modeled in the rate-based case (cf. above). Presynaptic spikes were filtered by synaptic dynamics into exponentially decaying postsynaptic currents (E or I):

dajdt=ajτsyn+tjδ(ttjΔ), (Equation 10)

where the tj’s are the firing times of neuron j, τsyn=2 ms is the synaptic time constant, and Δ=0.5 ms is a small axonal transmission delay (which enabled the distribution of the simulations onto multiple compute cores; Morrison et al., 2005). Synaptic currents then contributed to membrane potential dynamics according to

τidVidt=Vi+Vrest+hi(t)+ηi(t)+jEcellsJijaj(t)jIcellsJijaj(t), (Equation 11)

where the synaptic efficacies Jij are described below, and the noise term ηi was modeled exactly as described above.

Connectivity

For each neuron i, we drew pENE excitatory and pINI inhibitory presynaptic partners, uniformly at random. Connection probabilities were set to pE=0.1 and pI=0.4 respectively. The corresponding synaptic weights took on values Jij=Wαβ/(τsynpβNβ) where {α,β}{E,I} denote the populations to which neuron i and j belong respectively, and Wαβ are the connections in the reduced model (see table “Parameters Used in the SSN Simulations”). This choice was such that, for a given set of mean firing rates in the E and I populations, average E and I synaptic inputs to E and I cells matched the corresponding recurrent inputs in the rate-based model. Synapses that were not drawn were obviously set to Jij=0.

Local field potential

As a proxy for LFP in Figure 3, we took the momentary population-averaged Vm (Mazzoni et al., 2015 simulated various proxies and, although some proxies were more accurate, they found the average Vm to be reasonably accurate).

Multi-attractor model

We compared our ring SSN model to a version of the ring attractor model published by Ponce-Alvarez et al. (2013). The ring attractor model had a single population with a similar ring topology, and—using the same notation as above—the connectivity took the form (cf. Equation 9)

Wij=W¯+WΔNcos(θiθj), (Equation 12)

where N=100 is the number of neurons, and W¯ and WΔ are two parameters that control the average connection strength and modulation with tuning dissimilarity, respectively. Note that, in general, this connectivity matrix could violate Dale’s law but with the specific parameters used here it did not (see table “Parameters Used in the Multi-attractor Network Simulations”). Instead, all connections were inhibitory to keep the system in the marginally stable regime (as in Ponce-Alvarez et al., 2013). The dynamics of the network obeyed a similar stochastic differential equation as for the ring SSN (Equation 2), namely

τmdVidt=Vi+hi(t)+ηi(t)+jWijr(Vj), (Equation 13)

with the momentary firing rate of cell j given by a rectified saturating firing rate nonlinearity (cf. Equation 3):

r(Vj)=gmaxtanh(kVj+), (Equation 14)

and a noise process η identical to the one we used in the SSN (same spatial and temporal correlations, Equations 5 and 6), with a variance adjusted so as to obtain Fano factors of about 1.5 during spontaneous activity (Figure S9B, black). The external input had both a constant baseline, b, and a contrast-dependent modulated component (cf. Equation 4):

hi=b+c(1A+Acos(θiθstim)), (Equation 15)

where A controlled the depth of the modulation, and c represents stimulus strength.

Note that although the phenomenology and dynamical regime of this model was consistent with that of Ponce-Alvarez et al. (2013) (Figure S9), the model differed from their original implementation in some of the details: our dynamics were written in voltage form, not in rate form, we had only one unit at each location on the ring (as opposed to small pools of neurons), and our input noise process had spatial correlations to allow for a more direct and consistent comparison with the ring SSN.

Our analysis of variability in this ring attractor network is presented in Figure S9 in a format identical to that of Figure 5, and shows that shared variability is entirely dominated by the fluctuations in the location of an otherwise very stable bump of activity.

Chaos suppression model

We also implemented a chaotic rate network of size N=2,000 with the following (deterministic) dynamics (cf. Equations 2 and 13):

τmdVidt=Vi+hi(t)+jWijr(Vj), (Equation 16)

with an (unrectified) saturating firing rate nonlinearity (cf. Equations 3 and 14)

r(Vj)=tanh(Vj) (Equation 17)

(which could thus go negative as well as positive). Elements of the synaptic weight matrix were sampled i.i.d. from a normal distribution (thus violating Dale’s law, cf. Equations 9 and 12):

WijN(0,σW2/N), (Equation 18)

with σW=2, which placed the network in the chaotic regime (Sompolinsky et al., 1988). The external input was a constant input vector of the form (cf. Equations 4 and 15)

hi=ccos(ϕi), (Equation 19)

where ϕi is a phase sampled i.i.d. from a uniform distribution between 0 and 2π, and c represents stimulus strength. See table “Parameters Used in the Chaotic Network Simulations” for all parameter values. As shown in Rajan et al. (2010), chaos is suppressed for large enough c.

Quantification and Statistical Analysis

Dataset

We analyzed neural recordings from the V1 of two awake monkeys (Figures 4, 6, and 7). A full description of the experimental protocol and recordings can be found in the original publication (Ecker et al., 2010). We discarded all cells that were poorly isolated (contamination >5%), leaving us with 330 cells to analyze. The stimuli consisted of moving gratings of various orientations, all at 100% contrast. We fitted orientation tuning curves (Figure S11; average firing rate in the first 500 ms following stimulus onset, as a function of stimulus orientation) of the form f(θ)f0+fmexp[κ(cos(2(θθpref))1)], where θ is the stimulus orientation (thus, we neglected the direction of motion, which could be in either of the two directions orthogonal to the orientation of the grating). The fit was achieved using nonlinear least-squares regression.

For each neuron, we calculated an orientation tuning index (OTI), defined based on the fitted tuning curve as

OTI=f(θpref)f(θorth)f(θpref)+f(θorth), (Equation 20)

where θorth=θpref+π/2. As the ring architecture we studied in Figures 4, 5, 6, and 7 only applied to neurons with well-defined tuning curves, we excluded cells that had OTI<0.75 as well as average evoked rates (measured during the stimulus period) below 1 spike/sec. This left us with 99 well-tuned cells to analyze.

Our analysis of the stimulus tuning of Fano factors and pairwise spike-count correlations was based on a time window of 100 ms starting at stimulus onset.

Factor analysis

We performed factor analysis of spike counts, either for a single stimulus condition in the model (the model had a natural rotational symmetry), or separately for each stimulus condition (direction) in the V1 dataset, subsequently averaging the reported quantities across conditions. We worked with normalized spike counts, defined as c˜ik=cik/cikk where cik is the spike count of neuron i in trial k and k denotes averaging across trials. Note that the variances of these normalized spike counts are exactly the Fano factors, i.e., the usual measure of spike count variability. This prevented the normalized spike count covariance matrix C˜ from being contaminated by a rank-1 pattern of network covariance merely reflecting the tuning of single-neuron firing rates (the “Poisson” part of variability, which tends to scale with the mean count). Factor analysis decomposes C˜ as the sum of a rank-k covariance matrix C˜shared representing k modes of network covariances, and a diagonal matrix C˜private. In the rate model, we could near-perfectly estimate the spike count covariance matrix, so we performed factor analysis by direct eigendecomposition of C˜, thus defining C˜shared=i=1kλivivi whereby the top k eigenvectors v1,,vk of C˜ contributed to shared variability in proportion of the corresponding eigenvalues λi. For factor analysis of the monkey V1 data, we performed direct maximization of the data likelihood (Cunningham and Ghahramani, 2015), also keeping k factors. In Figure 4, we set k=3, but we observed quenching of shared variability irrespective of k (Figure S12).

Data and Software Availability

The code used for model simulations and data analysis is available from the Lead Contact, Dr Guillaume Hennequin, upon request.

Acknowledgments

This work was supported by NIH grant R01-EY11001 (K.D.M.); NSF award DBI-1707398 (K.D.M.); the Gatsby Charitable Foundation (K.D.M.); Medical Scientist Training Program grant 5 T32 GM007367-36 (D.B.R.); the Swartz Program in Computational Neuroscience at Columbia University (Y.A.); the Postdoc Program of École des Neurosciences, Paris, France (Y.A.); Swiss National Science Foundation Advanced Postdoctoral Fellowship P300P3.154636 (G.H.); Wellcome Trust Investigator Award 095621/Z/11/Z (M.L., G.H.); and Wellcome Trust Seed Award 202111/Z/16/Z (G.H.). We are grateful to A. Ecker, P. Berens, M. Bethge, and A. Tolias for making their data publicly available. We thank L. Abbott, A. Tan, S. Solomon, A. Renart, and P. Latham for helpful discussions. Y.A. would like to thank D. Hansel and the Centre de Neurophysique, Physiologie, et Pathologie, Paris, for their hospitality.

Author Contributions

G.H. performed the model simulations and data analysis and made the figures. G.H., Y.A., D.B.R., M.L., and K.D.M. conceived the study. G.H., Y.A., M.L., and K.D.M. performed mathematical analysis, discussed and planned approaches throughout the research, and wrote the manuscript.

Declaration of Interests

The authors declare no competing interests.

Published: May 16, 2018

Footnotes

Supplemental Information includes 12 figures and additional methods and can be found with this article online at https://doi.org/10.1016/j.neuron.2018.04.017.

Supplemental Information

Document S1. Methods and Figures S1–S12
mmc1.pdf (1.9MB, pdf)
Document S2. Article plus Supplemental Information
mmc2.pdf (5.9MB, pdf)

References

  1. Adesnik H. Synaptic Mechanisms of Feature Coding in the Visual Cortex of Awake Mice. Neuron. 2017;95:1147–1159.e4. doi: 10.1016/j.neuron.2017.08.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Ahmadian Y., Rubin D.B., Miller K.D. Analysis of the stabilized supralinear network. Neural Comput. 2013;25:1994–2037. doi: 10.1162/NECO_a_00472. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Amit D.J., Brunel N. Dynamics of a recurrent network of spiking neurons before and following learning. Network: Computation in Neural Systems. 1997;8:373–404. [Google Scholar]
  4. Azouz R., Gray C.M. Cellular mechanisms contributing to response variability of cortical neurons in vivo. J. Neurosci. 1999;19:2209–2223. doi: 10.1523/JNEUROSCI.19-06-02209.1999. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Ben-Yishai R., Bar-Or R.L., Sompolinsky H. Theory of orientation tuning in visual cortex. Proc. Natl. Acad. Sci. USA. 1995;92:3844–3848. doi: 10.1073/pnas.92.9.3844. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Berkes P., Orbán G., Lengyel M., Fiser J. Spontaneous cortical activity reveals hallmarks of an optimal internal model of the environment. Science. 2011;331:83–87. doi: 10.1126/science.1195870. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bertschinger N., Natschläger T. Real-time computation at the edge of chaos in recurrent neural networks. Neural Comput. 2004;16:1413–1436. doi: 10.1162/089976604323057443. [DOI] [PubMed] [Google Scholar]
  8. Blumenfeld B., Bibitchkov D., Tsodyks M. Neural network model of the primary visual cortex: from functional architecture to lateral connectivity and back. J. Comput. Neurosci. 2006;20:219–241. doi: 10.1007/s10827-006-6307-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Bujan A.F., Aertsen A., Kumar A. Role of input correlations in shaping the variability and noise correlations of evoked activity in the neocortex. J. Neurosci. 2015;35:8611–8625. doi: 10.1523/JNEUROSCI.4536-14.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Burak Y., Fiete I.R. Fundamental limits on persistent activity in networks of noisy neurons. Proc. Natl. Acad. Sci. USA. 2012;109:17645–17650. doi: 10.1073/pnas.1117386109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Carandini M., Heeger D.J. Normalization as a canonical neural computation. Nat. Rev. Neurosci. 2011;13:51–62. doi: 10.1038/nrn3136. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Cardin J.A., Palmer L.A., Contreras D. Cellular mechanisms underlying stimulus-dependent gain modulation in primary visual cortex neurons in vivo. Neuron. 2008;59:150–160. doi: 10.1016/j.neuron.2008.05.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Chen M., Wei L., Liu Y. Motor preparation attenuates neural variability and beta-band LFP in parietal cortex. Sci. Rep. 2014;4:6809. doi: 10.1038/srep06809. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Chung S., Ferster D. Strength and orientation tuning of the thalamic input to simple cells revealed by electrically evoked cortical suppression. Neuron. 1998;20:1177–1189. doi: 10.1016/s0896-6273(00)80498-5. [DOI] [PubMed] [Google Scholar]
  15. Churchland M.M., Afshar A., Shenoy K.V. A central source of movement variability. Neuron. 2006;52:1085–1096. doi: 10.1016/j.neuron.2006.10.034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Churchland M.M., Yu B.M., Cunningham J.P., Sugrue L.P., Cohen M.R., Corrado G.S., Newsome W.T., Clark A.M., Hosseini P., Scott B.B. Stimulus onset quenches neural variability: a widespread cortical phenomenon. Nat. Neurosci. 2010;13:369–378. doi: 10.1038/nn.2501. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Cohen M.R., Maunsell J.H.R. Attention improves performance primarily by reducing interneuronal correlations. Nat. Neurosci. 2009;12:1594–1600. doi: 10.1038/nn.2439. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Cunningham J.P., Ghahramani Z. Linear dimensionality reduction: survey, insights, and generalizations. Journal of Machine Learning Research. 2015;16:2859–2900. [Google Scholar]
  19. Dayan P., Abbott L.F. MIT Press; Cambridge, MA: 2001. Theoretical neuroscience. [Google Scholar]
  20. Deco G., Hugues E. Neural network mechanisms underlying stimulus driven variability reduction. PLoS Comput. Biol. 2012;8:e1002395. doi: 10.1371/journal.pcbi.1002395. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Doiron B., Litwin-Kumar A. Balanced neural architecture and the idling brain. Front. Comput. Neurosci. 2014;8:56. doi: 10.3389/fncom.2014.00056. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Doiron B., Litwin-Kumar A., Rosenbaum R., Ocker G.K., Josić K. The mechanics of state-dependent neural correlations. Nat. Neurosci. 2016;19:383–393. doi: 10.1038/nn.4242. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Ecker A.S., Berens P., Keliris G.A., Bethge M., Logothetis N.K., Tolias A.S. Decorrelated neuronal firing in cortical microcircuits. Science. 2010;327:584–587. doi: 10.1126/science.1179867. [DOI] [PubMed] [Google Scholar]
  24. Ecker A.S., Berens P., Cotton R.J., Subramaniyan M., Denfield G.H., Cadwell C.R., Smirnakis S.M., Bethge M., Tolias A.S. State dependence of noise correlations in macaque primary visual cortex. Neuron. 2014;82:235–248. doi: 10.1016/j.neuron.2014.02.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Ecker A.S., Denfield G.H., Bethge M., Tolias A.S. On the structure of neuronal population activity under fluctuations in attentional state. J. Neurosci. 2016;36:1775–1789. doi: 10.1523/JNEUROSCI.2044-15.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Ferster D., Chung S., Wheat H. Orientation selectivity of thalamic input to simple cells of cat visual cortex. Nature. 1996;380:249–252. doi: 10.1038/380249a0. [DOI] [PubMed] [Google Scholar]
  27. Finn I.M., Priebe N.J., Ferster D. The emergence of contrast-invariant orientation tuning in simple cells of cat visual cortex. Neuron. 2007;54:137–152. doi: 10.1016/j.neuron.2007.02.029. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Gentet L.J., Avermann M., Matyas F., Staiger J.F., Petersen C.C.H. Membrane potential dynamics of GABAergic neurons in the barrel cortex of behaving mice. Neuron. 2010;65:422–435. doi: 10.1016/j.neuron.2010.01.006. [DOI] [PubMed] [Google Scholar]
  29. Goldberg J.A., Rokni U., Sompolinsky H. Patterns of ongoing activity and the functional architecture of the primary visual cortex. Neuron. 2004;42:489–500. doi: 10.1016/s0896-6273(04)00197-7. [DOI] [PubMed] [Google Scholar]
  30. Goris R.L.T., Movshon J.A., Simoncelli E.P. Partitioning neuronal variability. Nat. Neurosci. 2014;17:858–865. doi: 10.1038/nn.3711. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Guo Z.V., Inagaki H.K., Daie K., Druckmann S., Gerfen C.R., Svoboda K. Maintenance of persistent activity in a frontal thalamocortical loop. Nature. 2017;545:181–186. doi: 10.1038/nature22324. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Hansel D., van Vreeswijk C. How noise contributes to contrast invariance of orientation tuning in cat visual cortex. J. Neurosci. 2002;22:5118–5128. doi: 10.1523/JNEUROSCI.22-12-05118.2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Harish O., Hansel D. Asynchronous rate chaos in spiking neuronal circuits. PLoS Comput. Biol. 2015;11:e1004266. doi: 10.1371/journal.pcbi.1004266. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Hennequin, G., and Lengyel, M. (2016). Characterizing variability in nonlinear recurrent neuronal networks. arXiv, arXiv:1610.03110, https://arxiv.org/abs/1610.03110.
  35. Hennequin G., Vogels T.P., Gerstner W. Optimal control of transient dynamics in balanced networks supports generation of complex movements. Neuron. 2014;82:1394–1406. doi: 10.1016/j.neuron.2014.04.045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Kadmon J., Sompolinsky H. Transition to chaos in random neuronal networks. Phys. Rev. X. 2015;5:041030. [Google Scholar]
  37. Kanashiro T., Ocker G.K., Cohen M.R., Doiron B. Attentional modulation of neuronal variability in circuit models of cortex. eLife. 2017;6:e23978. doi: 10.7554/eLife.23978. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Kanitscheider I., Coen-Cagli R., Pouget A. Origin of information-limiting noise correlations. Proc. Natl. Acad. Sci. USA. 2015;112:E6973–E6982. doi: 10.1073/pnas.1508738112. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Kohn A., Smith M.A. Stimulus dependence of neuronal correlation in primary visual cortex of the macaque. J. Neurosci. 2005;25:3661–3673. doi: 10.1523/JNEUROSCI.5106-04.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Kriener B., Tetzlaff T., Aertsen A., Diesmann M., Rotter S. Correlations and population dynamics in cortical networks. Neural Comput. 2008;20:2185–2226. doi: 10.1162/neco.2008.02-07-474. [DOI] [PubMed] [Google Scholar]
  41. Laje R., Buonomano D.V. Robust timing and motor patterns by taming chaos in recurrent neural networks. Nat. Neurosci. 2013;16:925–933. doi: 10.1038/nn.3405. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Li Y.T., Ibrahim L.A., Liu B.H., Zhang L.I., Tao H.W. Linear transformation of thalamocortical input by intracortical excitation. Nat. Neurosci. 2013;16:1324–1330. doi: 10.1038/nn.3494. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Lien A.D., Scanziani M. Tuned thalamic excitation is amplified by visual cortical circuits. Nat. Neurosci. 2013;16:1315–1323. doi: 10.1038/nn.3488. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Lin I.-C., Okun M., Carandini M., Harris K.D. The nature of shared cortical variability. Neuron. 2015;87:644–656. doi: 10.1016/j.neuron.2015.06.035. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Litwin-Kumar A., Doiron B. Slow dynamics and high variability in balanced cortical networks with clustered connections. Nat. Neurosci. 2012;15:1498–1505. doi: 10.1038/nn.3220. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Lombardo J., Macellaio M., Liu B., Osborne L.C., Palmer S.E. 2015 Neuroscience Meeting Planner (online) Society for Neuroscience; Washington, DC: 2015. Direction tuning of response variability in populations of MT neurons is different in awake versus anesthetized recordings. [Google Scholar]
  47. Malina K.C.-K., Mohar B., Rappaport A.N., Lampl I. Local and thalamic origins of ongoing and sensory evoked cortical correlations. bioRxiv. 2016:058727. doi: 10.1038/ncomms12740. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Mastrogiuseppe F., Ostojic S. Intrinsically-generated fluctuating activity in excitatory-inhibitory networks. PLoS Comput. Biol. 2017;13:e1005498. doi: 10.1371/journal.pcbi.1005498. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Mazzoni A., Lindén H., Cuntz H., Lansner A., Panzeri S., Einevoll G.T. Computing the local field potential (LFP) from integrate-and-fire network models. PLoS Comput. Biol. 2015;11:e1004584. doi: 10.1371/journal.pcbi.1004584. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Miller K.D., Fumarola F. Mathematical equivalence of two common forms of firing rate models of neural networks. Neural Comput. 2012;24:25–31. doi: 10.1162/NECO_a_00221. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Miller K.D., Troyer T.W. Neural noise can explain expansive, power-law nonlinearities in neural response functions. J. Neurophysiol. 2002;87:653–659. doi: 10.1152/jn.00425.2001. [DOI] [PubMed] [Google Scholar]
  52. Mitchell J.F., Sundberg K.A., Reynolds J.H. Spatial attention decorrelates intrinsic activity fluctuations in macaque area V4. Neuron. 2009;63:879–888. doi: 10.1016/j.neuron.2009.09.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Mochol G., Hermoso-Mendizabal A., Sakata S., Harris K.D., de la Rocha J. Stochastic transitions into silence cause noise correlations in cortical circuits. Proc. Natl. Acad. Sci. USA. 2015;112:3529–3534. doi: 10.1073/pnas.1410509112. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Molgedey L., Schuchhardt J., Schuster H.G. Suppressing chaos in neural networks by noise. Phys. Rev. Lett. 1992;69:3717–3719. doi: 10.1103/PhysRevLett.69.3717. [DOI] [PubMed] [Google Scholar]
  55. Mongillo G., Hansel D., van Vreeswijk C. Bistability and spatiotemporal irregularity in neuronal networks with nonlinear synaptic transmission. Phys. Rev. Lett. 2012;108:158101. doi: 10.1103/PhysRevLett.108.158101. [DOI] [PubMed] [Google Scholar]
  56. Moreno-Bote R., Beck J., Kanitscheider I., Pitkow X., Latham P., Pouget A. Information-limiting correlations. Nat. Neurosci. 2014;17:1410–1417. doi: 10.1038/nn.3807. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Morrison A., Mehring C., Geisel T., Aertsen A.D., Diesmann M. Advancing the boundaries of high-connectivity network simulation with distributed computing. Neural Comput. 2005;17:1776–1801. doi: 10.1162/0899766054026648. [DOI] [PubMed] [Google Scholar]
  58. Murphy B.K., Miller K.D. Balanced amplification: a new mechanism of selective amplification of neural activity patterns. Neuron. 2009;61:635–648. doi: 10.1016/j.neuron.2009.02.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Murray J.D., Bernacchia A., Freedman D.J., Romo R., Wallis J.D., Cai X., Padoa-Schioppa C., Pasternak T., Seo H., Lee D., Wang X.-J. A hierarchy of intrinsic timescales across primate cortex. Nat. Neurosci. 2014;17:1661–1663. doi: 10.1038/nn.3862. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Orbán G., Berkes P., Fiser J., Lengyel M. Neural variability and sampling-based probabilistic representations in the visual cortex. Neuron. 2016;92:530–543. doi: 10.1016/j.neuron.2016.09.038. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Ostojic S. Two types of asynchronous activity in networks of excitatory and inhibitory spiking neurons. Nat. Neurosci. 2014;17:594–600. doi: 10.1038/nn.3658. [DOI] [PubMed] [Google Scholar]
  62. Ozeki H., Finn I.M., Schaffer E.S., Miller K.D., Ferster D. Inhibitory stabilization of the cortical network underlies visual surround suppression. Neuron. 2009;62:578–592. doi: 10.1016/j.neuron.2009.03.028. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Ponce-Alvarez A., Thiele A., Albright T.D., Stoner G.R., Deco G. Stimulus-dependent variability and noise correlations in cortical MT neurons. Proc. Natl. Acad. Sci. USA. 2013;110:13162–13167. doi: 10.1073/pnas.1300098110. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Poulet J.F.A., Petersen C.C.H. Internal brain state regulates membrane potential synchrony in barrel cortex of behaving mice. Nature. 2008;454:881–885. doi: 10.1038/nature07150. [DOI] [PubMed] [Google Scholar]
  65. Poulet J.F.A., Fernandez L.M.J., Crochet S., Petersen C.C.H. Thalamic control of cortical states. Nat. Neurosci. 2012;15:370–372. doi: 10.1038/nn.3035. [DOI] [PubMed] [Google Scholar]
  66. Priebe N.J., Ferster D. Inhibition, spike threshold, and stimulus selectivity in primary visual cortex. Neuron. 2008;57:482–497. doi: 10.1016/j.neuron.2008.02.005. [DOI] [PubMed] [Google Scholar]
  67. Rajan K., Abbott L.F., Sompolinsky H. Stimulus-dependent suppression of chaos in recurrent neural networks. Phys. Rev. E Stat. Nonlin. Soft Matter Phys. 2010;82:011903. doi: 10.1103/PhysRevE.82.011903. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Ray S., Maunsell J.H.R. Differences in gamma frequencies across visual cortex restrict their possible use in computation. Neuron. 2010;67:885–896. doi: 10.1016/j.neuron.2010.08.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Reinhold K., Lien A.D., Scanziani M. Distinct recurrent versus afferent dynamics in cortical visual processing. Nat. Neurosci. 2015;18:1789–1797. doi: 10.1038/nn.4153. [DOI] [PubMed] [Google Scholar]
  70. Renart A., Machens C.K. Variability in neural activity and behavior. Curr. Opin. Neurobiol. 2014;25:211–220. doi: 10.1016/j.conb.2014.02.013. [DOI] [PubMed] [Google Scholar]
  71. Renart A., de la Rocha J., Bartho P., Hollender L., Parga N., Reyes A., Harris K.D. The asynchronous state in cortical circuits. Science. 2010;327:587–590. doi: 10.1126/science.1179850. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Rosenbaum R., Smith M.A., Kohn A., Rubin J.E., Doiron B. The spatial structure of correlated neuronal variability. Nat. Neurosci. 2017;20:107–114. doi: 10.1038/nn.4433. [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Rubin D.B., Van Hooser S.D., Miller K.D. The stabilized supralinear network: a unifying circuit motif underlying multi-input integration in sensory cortex. Neuron. 2015;85:402–417. doi: 10.1016/j.neuron.2014.12.026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Sadagopan S., Ferster D. Feedforward origins of response variability underlying contrast invariant orientation tuning in cat visual cortex. Neuron. 2012;74:911–923. doi: 10.1016/j.neuron.2012.05.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Shao Y.R., Isett B.R., Miyashita T., Chung J., Pourzia O., Gasperini R.J., Feldman D.E. Plasticity of recurrent l2/3 inhibition and gamma oscillations by whisker experience. Neuron. 2013;80:210–222. doi: 10.1016/j.neuron.2013.07.026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Softky W.R., Koch C. The highly irregular firing of cortical cells is inconsistent with temporal integration of random EPSPs. J. Neurosci. 1993;13:334–350. doi: 10.1523/JNEUROSCI.13-01-00334.1993. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Sompolinsky H., Crisanti A., Sommers H.J. Chaos in random neural networks. Phys. Rev. Lett. 1988;61:259–262. doi: 10.1103/PhysRevLett.61.259. [DOI] [PubMed] [Google Scholar]
  78. Solomon S.S., Chen S.C., Morley J.W., Solomon S.G. Local and global correlations between neurons in the middle temporal area of primate visual cortex. Cereb. Cortex. 2015;25:3182–3196. doi: 10.1093/cercor/bhu111. [DOI] [PubMed] [Google Scholar]
  79. Stringer C., Pachitariu M., Steinmetz N.A., Okun M., Bartho P., Harris K.D., Sahani M., Lesica N.A. Inhibitory control of correlated intrinsic variability in cortical networks. eLife. 2016;5:e19695. doi: 10.7554/eLife.19695. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Stuart G.J., Spruston N. Dendritic integration: 60 years of progress. Nat. Neurosci. 2015;18:1713–1721. doi: 10.1038/nn.4157. [DOI] [PubMed] [Google Scholar]
  81. Sussillo D., Abbott L.F. Generating coherent patterns of activity from chaotic neural networks. Neuron. 2009;63:544–557. doi: 10.1016/j.neuron.2009.07.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Tan A.Y.Y., Chen Y., Scholl B., Seidemann E., Priebe N.J. Sensory stimulation shifts visual cortex from synchronous to asynchronous states. Nature. 2014;509:226–229. doi: 10.1038/nature13159. [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Tetzlaff T., Helias M., Einevoll G.T., Diesmann M. Decorrelation of neural-network activity by inhibitory feedback. PLoS Comput. Biol. 2012;8:e1002596. doi: 10.1371/journal.pcbi.1002596. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Troyer T.W., Miller K.D. Physiological gain leads to high ISI variability in a simple model of a cortical regular spiking cell. Neural Comput. 1997;9:971–983. doi: 10.1162/neco.1997.9.5.971. [DOI] [PubMed] [Google Scholar]
  85. Tsodyks M.V., Skaggs W.E., Sejnowski T.J., McNaughton B.L. Paradoxical effects of external modulation of inhibitory interneurons. J. Neurosci. 1997;17:4382–4388. doi: 10.1523/JNEUROSCI.17-11-04382.1997. [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. van Vreeswijk C., Sompolinsky H. Chaotic balanced state in a model of cortical circuits. Neural Comput. 1998;10:1321–1371. doi: 10.1162/089976698300017214. [DOI] [PubMed] [Google Scholar]
  87. Vogels T.P., Rajan K., Abbott L.F. Neural network dynamics. Annu. Rev. Neurosci. 2005;28:357–376. doi: 10.1146/annurev.neuro.28.061604.135637. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Document S1. Methods and Figures S1–S12
mmc1.pdf (1.9MB, pdf)
Document S2. Article plus Supplemental Information
mmc2.pdf (5.9MB, pdf)

RESOURCES