Significance
Cortical networks have the remarkable ability to self-assemble into dynamic regimes in which excitatory positive feedback is balanced by recurrent inhibition. This inhibition-stabilized regime is increasingly viewed as the default dynamic regime of the cortex, but how it emerges in an unsupervised manner remains unknown. We prove that classic forms of homeostatic plasticity are unable to drive recurrent networks to an inhibition-stabilized regime due to the well-known paradoxical effect. We next derive a novel family of cross-homeostatic rules that lead to the unsupervised emergence of inhibition-stabilized networks. These rules shed new light on how the brain may reach its default dynamic state and provide a valuable tool to self-assemble artificial neural networks into ideal computational regimes.
Keywords: inhibition-stabilized networks, paradoxical effect, homeostatic plasticity
Abstract
Self-sustained neural activity maintained through local recurrent connections is of fundamental importance to cortical function. Converging theoretical and experimental evidence indicates that cortical circuits generating self-sustained dynamics operate in an inhibition-stabilized regime. Theoretical work has established that four sets of weights (WE←E, WE←I, WI←E, and WI←I) must obey specific relationships to produce inhibition-stabilized dynamics, but it is not known how the brain can appropriately set the values of all four weight classes in an unsupervised manner to be in the inhibition-stabilized regime. We prove that standard homeostatic plasticity rules are generally unable to generate inhibition-stabilized dynamics and that their instability is caused by a signature property of inhibition-stabilized networks: the paradoxical effect. In contrast, we show that a family of “cross-homeostatic” rules overcome the paradoxical effect and robustly lead to the emergence of stable dynamics. This work provides a model of how—beginning from a silent network—self-sustained inhibition-stabilized dynamics can emerge from learning rules governing all four synaptic weight classes in an orchestrated manner.
Self-sustained patterns of neural activity maintained by local recurrent excitation underlie many cortical computations and dynamic regimes, including the persistent activity associated with working memory (1–3), motor control (4, 5), asynchronous states associated with the default cortical dynamic regime (6–10), and up-states (8, 11). Recurrent excitation also has the potential to drive pathological and epileptiform regimes (12, 13). Converging theoretical and experimental evidence indicates that cortical circuits that generate self-sustained dynamics operate in an inhibition-stabilized regime, in which positive feedback is held in check by recurrent inhibition (7, 14–20). There is also evidence that inhibition-stabilized regimes may comprise the default awake cortical dynamic regime (8, 9, 20).
At the computational level inhibition-stabilized networks are often modeled as a simplified circuit composed of interconnected excitatory (E) and inhibitory (I) neural populations with four classes of synaptic weights: WE←E, WE←I, WI←E, and WI←I. In the inhibition-stabilized regime, recurrent excitation produces positive feedback, which is held in check by rapid inhibition. The dynamics settles into a stable fixed-point attractor and instantiates an inhibition-stabilized network. A signature of the inhibition-stabilized regime is the presence of the paradoxical effect, in which an increase in excitatory drive to inhibitory neurons produces a net decrease in the firing rate of those same inhibitory neurons (14–16, 18), a phenomenon that has been observed in the awake resting cortex (Fig. 1A) (20). Analytical and numerical studies have demonstrated that in order to support inhibition-stabilized dynamics, the four weight classes must obey certain “balanced” relationships; for example, if excitation is too strong, runaway (or saturated) excitation occurs, whereas if inhibition is too strong the activity falls into a quiescent fixed point (6, 7, 14–16, 19). However, in most computational models the set of four weights is determined analytically or through numerical searches—or in a few cases by allowing one or two weights to be plastic while appropriately hardwiring the others. In contrast, experimental studies both in vitro and in vivo have shown that self-sustained activity emerges autonomously during development (21–26), indicating that synaptic plasticity rules are in place to orchestrate the unsupervised emergence of inhibition-stabilized dynamics. Additionally, because the four weight classes have been observed to undergo synaptic plasticity in experimental studies (27–32), here we ask how inhibition-stabilized dynamics might emerge in a self-organizing manner.
One possibility is that standard homeostatic forms of plasticity underlie the emergence of inhibition-stabilized dynamics. Homeostatic plasticity rules generally assume that excitatory weights are regulated in a manner proportional to the difference between some ontogenetically determined setpoint and average neural activity (for both excitatory and inhibitory neurons)—and conversely that inhibitory weights onto excitatory neurons are regulated in the opposite direction (33–38). However, it remains an open question whether homeostatic rules can lead to the self-organized emergence of inhibition-stabilized networks. Here we use computational models and analytical methods to explore families of homeostatic plasticity rules that operate in parallel in all four synapse classes and lead to inhibition-stabilized dynamics. We show that when driving the network toward inhibition-stabilized regimes, standard forms of homeostatic plasticity are stable only in a narrow region of parameter space. Here we prove that this instability arises from the paradoxical effect. Indeed, it can be seen that like increasing external input to an inhibitory neuron, increasing the excitatory weights onto an inhibitory neuron firing below its setpoint produces a paradoxical (and anti-homeostatic) decrease in inhibitory activity (Fig. 1 B and C). While inhibition-stabilized regimes can operate in the presence or absence of external input, here we focus primarily on “fully self-sustained” activity in the absence of any external tonic input; however, we show that our results also apply when external inputs are present.
We conclude that homeostatic manipulations in the inhibitory population lead to paradoxical outcomes, making the rules unstable in this context. Therefore, homeostatic plasticity rules that aim to bring the network to the relevant dynamic regime of the cerebral cortex must work in paradoxical conditions. We developed a family of homeostatic plasticity rules that include “cross-homeostatic” influences and lead to the unsupervised emergence of fully self-sustained dynamics in the inhibition-stabilized regime in a robust manner. These rules are consistent with experimental data and generate explicit predictions regarding the effects of manipulations of excitatory and inhibitory neurons on synaptic plasticity.
Results
Standard Homeostatic Plasticity Rules Cannot Account for the Emergence of Stable Self-Sustained Activity.
We first ask, when starting from a network in a silent regime similar to cortical circuits early in development, whether standard homeostatic rules can drive networks to stable self-sustained dynamics in the absence of any external input. Based on experimental studies we assume that both excitatory and inhibitory neurons exhibit ontogenetically programmed firing rate setpoints (38–42) and ask whether homeostatic plasticity of excitatory and inhibitory weights can drive neurons to these setpoints. Homeostatic plasticity rules are traditionally defined by changes in synaptic weights that are proportional to an “error term” defined by the difference between the setpoint and the neurons’ average activity levels (34–37, 39, 43, 44), for example, ΔWE←E ∝ Eset − Eavg, where any departure of the excitatory activity Eavg from the setpoint Eset would lead to a compensatory correction in the value of the weight WE←E.
We first examined whether stable self-sustained dynamics can emerge in the standard two-population model (Fig. 1B) (19) through homeostatic mechanisms. We initialized the four weights (WE←E, WE←I, WI←E, and WI←I) of the model to small values and applied a standard family of homeostatic plasticity rules to all four weight classes (Fig. 2A). It is well established that PV+-inhibitory neurons have higher firing rates than pyramidal neurons during periods of self-sustained activity (45, 46); thus, based on previous data using intracellular recordings, we set the setpoints for the E and I populations to 5 and 14 Hz, respectively (46). We first asked whether the family of four standard homeostatic plasticity rules can lead to a stable self-sustained dynamic regime in response to a brief external input. Since in the absence of external input (or intrinsic spontaneous activity) networks capable of self-sustained activity have a trivial stable silent (down-state) regime, at the beginning of each trial we administered a brief external input to engage the network (low levels of noise were used to avoid fluctuation-induced transitions). Although the rules are homeostatic in nature (e.g., if I is below Iset, an increase in WI←E and a decrease in WI←I would be induced), in the example shown in Fig. 2 B and C the network failed to converge to a stable self-sustained regime (Fig. 2 B and C). Initially (trial 1) an external input to the excitatory population does not engage recurrent activity because WE←E is too weak. By trial 200 the weights have evolved and the brief external input triggers self-sustained activity, but activities E and I do not match the corresponding setpoints; the network is in a nonbiologically observed regime in which E > I, so the weights keep evolving. By trial 600 E = Eset but I < Iset, and rather than converging to Iset, the network returns to a regime without self-sustained activity by trial 1,000. At that point both setpoint error terms have increased, leading to continued weight changes (Fig. 2C). Results across 100 simulations with different weight initializations (SI Appendix, Supplementary Methods) further indicate that the standard homeostatic rules are ineffective at driving E and I toward their respective setpoints and generating stable self-sustained dynamics (Fig. 2D).
To gain insights into why a family of homeostatic plasticity rules that might intuitively converge fails to do so, we can consider the case in which a network is initialized to a set of weights that already match Eset and Iset (Fig. 2E). Although the neural subsystem alone is stable at this condition (trial 1), small fluctuations in E and I cause the homeostatic rules to drive the weight values and the average activity of the network away from the setpoints (trial 500). It is possible to understand this instability by performing an analytical stability analysis. Specifically, a two-population network in which the weights undergo plasticity can be characterized as a dynamic system composed of two subsystems: the neural subsystem, composed of the two differential equations that define E and I dynamics, and the synaptic homeostatic plasticity rule subsystem, defined by the four plasticity rules (SI Appendix, Section 2.1). We use the two very different time scales of the neural (fast) and plasticity rule (slow) subsystems to perform a quasi-steady-state approximation of the neural subsystem; then we compute the eigenvalues of the four-dimensional plasticity rule subsystem and finally get an analytical expression for the stability condition of the plasticity rules (SI Appendix, Section 2.3). For the entire system to be stable, both the neural and plasticity rule subsystems have to be stable. For the results presented in Fig. 2 B–E we assumed the learning rates driving plasticity onto the excitatory (αE) and inhibitory neurons (αI) to be equal. Under these conditions, the standard homeostatic rules are mostly unstable for biologically meaningful parameter values in which the neural system is stable. The regions of stability can be seen in Fig. 2F. Critically, Fig. 2F shows that the stability region of the neural subsystem, that is, an inhibition-stabilized network (15, 19), is almost entirely within the region where the homeostatic plasticity rule system is unstable. The only region where a stable self-sustained dynamics can exist is the small triangle where the neural and synaptic stability regions overlap (SI Appendix, Fig. S1). Only when plasticity onto the excitatory neuron is significantly faster (αE ≫ αI, resulting in very slow convergence to the inhibitory setpoints) is there a substantial region of overlap between the stability of the neural and plasticity rule subsystems (Fig. 2G; SI Appendix, Section 1.1).
Because inhibitory neurons seem to undergo homeostatic plasticity as quickly as or more quickly than excitatory neurons (40–42, 47, 48) we conclude that standard homeostatic rules by themselves do not account for the emergence of stable self-sustained and inhibition-stabilized dynamics. Similarly, a combination of analytical and numerical methods also indicates that variants of these homeostatic rules, such as synaptic scaling, are also stable only in a narrow region of parameter space (SI Appendix, Section 1.5). We next show that the inherent instability of standard homeostatic plasticity rules is related to the paradoxical effect.
The Paradoxical Effect Hampers the Ability of Homeostatic Rules to Lead to Self-Sustained Activity.
The inability of the homeostatic plasticity rules to generate stable self-sustained activity is in part a consequence of the paradoxical effect, a counterintuitive yet well described property of two-population models of inhibition-stabilized networks (14, 15). Specifically, if during self-sustained activity one increases the excitatory drive to the inhibitory population, the net result is a decrease in the firing rate of the inhibitory units. This paradoxical effect can be understood in terms of the I→E→I loop: The increased inhibitory drive leads to a lower steady-state rate for E, but this new steady-state value requires a decrease in the I firing rate to maintain an appropriate E/I balance (in effect, the decrease in E decreases the drive to I by more than the external increase to I). This paradoxical effect has profound consequences for plasticity rules that attempt to drive excitatory and inhibitory weights to an activity setpoint.
The relationship of the paradoxical effect and the homeostatic rule performance is presented in Fig. 2H. The region of stability for the homeostatic plasticity rules is shown in a parameter regime where inhibitory plasticity is much faster (αE ≪ αI). Contrary to when excitatory plasticity dominates, the region of stability is small, and there is no overlap with the region of stability of the neural subsystem. Crucially, the boundary of the stability region of the plasticity rule coincides with the condition for the paradoxical effect to be present (right of the blue line in Fig. 2H, SI Appendix, Sections 2.2.4 and 2.3.6). Under these conditions, the rules can be stable only when the network is not in an inhibition-stabilized regime. If a network regime with nonzero E is forced to exist in that region (e.g., via a tonic external current, Fig. 2I), it would be stable only in the nonparadoxical region with the plasticity rules in place (SI Appendix, Section 2.5). Note that in the absence of a sufficiently strong external input it is not possible to have stable self-sustained activity that is not inhibition stabilized.
To understand the impact of the paradoxical effect on homeostatic plasticity rules, consider a network state in which the I rate falls significantly below its setpoint and the E rate is close to its setpoint (Fig. 3A). In order to reach the I setpoint, homeostatic plasticity in the inhibitory neuron would intuitively result in an increase of WI←E. However, because of the paradoxical effect, an increase in WI←E actually makes I decrease (Fig. 3B), thus increasing the error term Iset − I. To increase the steady-state inhibitory rate, we can “anti-homeostatically” decrease the excitatory weight onto the inhibitory neurons (Fig. 3C). (Note that the converse is true for the WI←I weight.) This simple example shows the complexity of designing a coherent set of rules in a strongly coupled system (an analysis of the paradoxical effect is in SI Appendix, Section 2.2.4). This analysis also explains why homeostatic plasticity rules can lead to self-sustained activity at the appropriate setpoints when αE ≫ αI. Essentially, by allowing plasticity onto the E population to be faster, one overcomes the counterproductive homeostatic plasticity associated with the paradoxical effect.
The interaction between the paradoxical effect and homeostatic plasticity in inhibitory neurons leads to the question of whether anti-homeostatic plasticity rules may be more effective that standard homeostatic rules; for example, ΔWI←E ∝ −(Iset − Iavg). Thus, we also examined a number of hybrid families of plasticity rules with different combinations of homeostatic and anti-homeostatic rules. Indeed, some hybrid families exhibited large degrees of overlap between the stable regions of the network and plasticity rule subsystems. However, numerical simulations revealed that these rules were mostly ineffective in driving networks to self-sustained activity at the target setpoints (SI Appendix, Section 1.2 and Fig. S2). These two results are not inconsistent because the stability analysis speaks to cases when the network is initialized to weights that satisfy Eset and Iset, not whether the rules will drive network activity into these stable areas from any initial state, including an early developmental state. Thus, we interpret these results as meaning that while anti-homeostatic plasticity can contribute to stability of this dual dynamic system, anti-homeostatic plasticity is ineffective at driving the dynamics toward setpoints (in other words, that anti-homeostatic plasticity might allow for stable inhibition-stabilized dynamics but does not necessarily generate sizable basins of attraction around the fixed point).
Cross-Homeostatic Rule Robustly Leads to the Emergence of Self-Sustained Dynamics.
Given that a standard family of homeostatic plasticity rules did not robustly lead to stable dynamics, we explored alternative learning rules. By defining a loss function based on the sum of the excitatory and inhibitory errors, we analytically derived a set of learning rules using gradient descent (SI Appendix, Section 3). This approach led to mathematically complex and biologically implausible rules; however, approximations and simulations inspired a simple class of learning rules that we will refer to as cross-homeostatic (see Methods). The main characteristic of this set of rules is that the homeostatic setpoints are “crossed” (Fig. 4A). Specifically, the weights onto the excitatory neuron (WE←E and WE←I) are updated to minimize the inhibitory error, while weights into the inhibitory neuron (WI←E and WI←I) change to minimize the excitatory error. Although apparently nonlocal, from the perspective of an excitatory neuron these rules can be interpreted as cells having a setpoint for the total inhibitory input current onto the cell. Such inputs could be read by a cell as the activation of metabotropic receptors (e.g., gamma-aminobutyric acid B [GABAB] and metabotropic glutamate; see Discussion). Indeed, a similar cross-homeostatic rule has been recently derived for WI←E weights (49).
An example of the performance of the cross-homeostatic rules is shown in Fig. 4 B and C. After an initial phase with no self-sustained firing (trial 1), recurrent activity reaches stable self-sustained dynamics (trial 20), whose average rate continues to converge toward its defined setpoints (trial 100) until the learning rule system reaches steady state (trial 500). The average E and I rates of the network evolve asymptotically toward the defined setpoints, as the weights evolve and converge (Fig. 4C). Across different weight initializations the rules proved effective in driving the mean activity of the network to the target E and I setpoints and led to balanced, inhibition-stabilized dynamics (Fig. 4 D and E). The weight trajectory from its initial value to its final one is shown for 100 different simulations (Fig. 4D). Each line corresponds to individual experiments with different initializations (for visualization purposes weights were initialized around the final stable weights; SI Appendix, Fig. S3 includes broader initialization conditions). Circles indicate the final values of the weights. Independently of the initial conditions, the weights converge to a line attractor (actually a two-dimensional plane attractor in four-dimensional weight space; SI Appendix, Section 2.1). Note that this attractor refers to the sets of weights that generate self-sustained dynamics where E and I activity matches Eset and Iset, respectively. That is, for a given pair of setpoints (Eset, Iset) the final values of the weights WE←I and WI←I are linear functions of the “free” weights WE←E and WI←E, respectively. This is a direct consequence of the steady-state conditions for the nontrivial fixed-point of the two-population model (14, 15), where the slope of the line is defined by the setpoints Eset/Iset (see Methods). For example, to satisfy dE/dt = 0 at the neural activity fixed point, the net excitation and inhibition must obey a specific balance, meaning that once WE←E or WE←I is determined, the other weight is analytically constrained for a given set of setpoints and parameters. Once the weights reach this specific relationship, the E and I rates reach their corresponding Eset and Iset values (Fig. 4E). Numerical simulations confirm that the cross-homeostatic rule robustly guides self-sustained activity to different Eset and Iset setpoints (Fig. 4F), whose ratios define the slopes of the final relationship between the weights (Fig. 4G). The above implementations were trial-based, that is, the weights were updated at the end of every trial. An “online” implementation, in which weights were continuously updated, also led to convergence to the setpoints (SI Appendix, Fig. S4).
To further validate the effectiveness and stability of the cross-homeostatic rule, we again used analytic methods to determine the eigenvalues of the four-dimensional dynamic system describing the family of four cross-homeostatic rules. As above, stability is determined by the sign of the real part of the eigenvalues of the system. It can be shown (SI Appendix, Section 1.3) that these learning rules are stable for any set of parameter values, provided that the stability conditions of the neural subsystem are satisfied (Fig. 4H). Importantly, these results demonstrate that cross-homeostatic rules work in both paradoxical and nonparadoxical conditions. Furthermore, the stability of the rules is independent of the absence or presence of external input (SI Appendix, Section 2.5). Therefore, it is possible to formally establish that cross-homeostatic learning rules are inherently stable and can robustly account for the emergence and maintenance of self-sustained, inhibition-stabilized dynamics in the two-population model.
Cross-Homeostatic Rules Drive Average Activity to Setpoints in a Multiunit Model.
The previous results demonstrate the robustness of the cross-homeostatic family of rules in driving a two-subpopulation rate model to a stable self-sustained, inhibition-stabilized regime. We next examined whether these rules are also effective for a multiunit model in which there are many excitatory and inhibitory units. The firing-rate model was composed of 80 excitatory and 20 inhibitory recurrently connected neurons (Fig. 5A). In this case, individual neurons adjust their weights to minimize the average error of their presynaptic partners (SI Appendix, Supplementary Methods). Starting with normally distributed weights, the network reaches stable self-sustained dynamics (Fig. 5 B and C). However, individual units converge to different final rate values, satisfying the defined setpoints only as an average (green and red thick lines of Fig. 5B). This is a result of the nature of the cross-homeostatic rules: Neurons adjust their weights to minimize the error of the mean activity of its presynaptic partners. For this reason, although the network is globally balanced, single units do not converge to the same balanced E–I line attractor (Fig. 5 D and E). After cross-homeostatic plasticity, some differential structure is visible among the various weight classes (Fig. 5F). Simulations across 400 different initialization conditions demonstrate that the rules lead the average excitatory and inhibitory population activity to Eset and Iset, respectively (Fig. 5 G and H). The cross-homeostatic rules are thus capable of driving a multiunit model to a stable self-sustained regime, but they do not guide individual units to local setpoints. Similar results are obtained when the network weights are initialized with log-normal distributions (SI Appendix, Fig. S5).
Learning Rule with Cross-Homeostatic and Homeostatic Terms Leads to Local Convergence to Setpoints.
The above results demonstrate a potential limitation of the cross-homeostatic family of rules: The target setpoints are reached only at the population level. An additional and potentially more serious limitation is that cross-homeostatic rules predict that artificially altering the activity of a small number of excitatory neurons within a large network would not directly produce homeostatic plasticity in these neurons but would directly produce plasticity in their postsynaptic inhibitory neurons. This prediction seems to conflict with homeostatic plasticity experiments that have targeted specific cell types rather than globally altered activity through pharmacological means (50, 51). We therefore assessed the scenario in which both cross-homeostatic and homeostatic rules operate in parallel, resulting in a two-term cross-homeostatic family of rules. Interestingly, this family of rules can be obtained from an approximation of a gradient descent derivation on a loss function that includes the difference between E and I and their respective setpoints (SI Appendix, Section 3). In a two-population model, we first confirmed that this two-term cross-homeostatic family is stable, assuming that the learning rate of the homeostatic term does not dominate (SI Appendix, Section 1.4).
Simulations with the same multiunit model as in Fig. 5 show that with the two-term cross-homeostatic rule all individual units converge to their respective Eset and Iset (Fig. 6 A–C). Importantly, in contrast to the single-term cross-homeostatic rule, the total excitatory and inhibitory weight of each individual unit converged to the E–I balance of the line attractor predicted by the network equations (Fig. 6 D and E), while some structure in the different weight classes is also observed in the connectivity matrices (Fig. 6F). The convergence to the setpoints was stable across a wide range of initial states (Fig. 6 G and H). Thus, a hybrid family of plasticity rules that includes both cross-homeostatic and homeostatic forces provides global network stability while also locally driving each unit to their setpoint and a balanced E–I regime.
Spiking Neural Network Model with Sparse Connectivity Converges to an Inhibition-Stabilized Regime at the Setpoints.
The previous results demonstrate the ability of the two-term cross-homeostatic plasticity rule to guide firing-rate-based models to inhibition-stabilized regimes at the target setpoints. We next examined the effectiveness of this family of learning rules in a sparsely connected spiking neural network (Fig. 7A). In a sparsely connected network, the cross-homeostatic component of the learning rule can be implemented globally (e.g., excitatory plasticity onto an excitatory unit is based on the mean error of the entire population of inhibitory units) or locally (e.g., excitatory plasticity onto an excitatory neuron is based on the mean error of its presynaptic inhibitory partners). Here we used a local implementation, in which, for example, an excitatory neuron has a setpoint interpreted as a target for the total amount of GABAB receptor activation it should receive (see Discussion).
Starting from a developmental scenario with weak weights (i.e., that do not support any self-sustaining activity), the rules successfully drive the network to stable self-sustained and asynchronous spiking activity near the setpoints (Fig. 7 B–E; CVISI∼1). As seen in the firing rate multiunit model, the weights self-organize from an unstructured initial condition (Fig. 7 F and G) to a state where a balance of excitation and inhibition emerges (Fig. 7 H–K). After training, the firing rates of the excitatory and inhibitory neurons distribute around their setpoints (Fig. 7H). As expected, the emergent network dynamics exhibits the paradoxical effect; specifically, when inhibitory neurons are transiently activated with an external current, a net decrease in the mean firing rate is observed (Fig. 7I). Finally, convergence holds across networks initialized with different weights (Fig. 7L). These results demonstrate that the theoretical and computational results in rate networks translate to more complex and biologically realistic scenarios.
Discussion
Elucidating the learning rules that govern the connectivity within neural circuits is a fundamental goal in neuroscience, in part because learning rules establish unifying principles that span molecular, cellular, systems, and computational levels of analyses. Elucidation of Hebbian associative synaptic learning, for example, linked simple computations at the level of single proteins (the N-methyl-d-aspartate receptor) with higher-order computations at the system and computational levels (52–56). However, it remains the case that because most studies have focused on learning rules at one or two synapse classes, little is known about the learning rules that give rise to complex neural dynamic regimes. Here we have taken steps toward exploring families of learning rules that operate in parallel at four different synapse classes, and starting from a silent state, capture the experimentally observed emergence of self-sustained, inhibition-stabilized dynamics in cortical networks.
We first explored whether standard formulations of homeostatic plasticity can account for the unsupervised emergence of self-sustained, inhibition-stabilized regimes. Based on experimental data we assumed that both excitatory and inhibitory neurons have an ontogenetically programmed activity setpoint during self-sustained activity and that plasticity in the four weight classes is driven by standard formulations of homeostatic plasticity. Numerical simulations and analytical stability analyses revealed that while some initial conditions and parameter regimes led to self-sustained dynamics, they occupied a narrow region of parameter space, when the rate of synaptic plasticity onto inhibitory neurons is much lower than that onto excitatory neurons (Fig. 2G and SI Appendix). When the rates of inhibitory and excitatory plasticity are comparable, analytical stability analyses confirmed that the region of stability of the network dynamics overlapped only in a narrow region. Such a narrow stability area seems incompatible with the robustness necessary in biological systems and with experimental data showing that inhibitory neurons exhibit homeostatic plasticity as fast as or faster than excitatory neurons (40–42, 47, 48). We thus conclude that a family of standard homeostatic plasticity rules operating in all four synapse classes is not sufficient to account for the experimentally observed emergence of self-sustained dynamics in cortical circuits.
Cross-Homeostatic Plasticity.
Analyses of approximations of a gradient-descent-derived learning rule suggested, somewhat counterintuitively, that adjusting the E population based on the error of the I population (and vice versa) may prove to be an effective family of learning rules. Indeed, numerical simulations and analytical stability analyses revealed that this cross-homeostatic rule was robustly stable (Fig. 4). However, the convergence to the excitatory and inhibitory setpoints in a multiunit network occurred only at the population level, not at the level of individual units. This observation is not inconsistent with experimental data, which show that in vivo neurons do exhibit a wide range of variability in their apparent setpoints (57, 58). However, a significant concern with this single-term cross-homeostatic rule is that it predicts that selectively increasing activity in a subpopulation of excitatory neurons would first induce plasticity in inhibitory neurons (WI←E and WI←I), which could in turn lead to plasticity in the manipulated excitatory neurons (WE←E and WE←I). Most homeostatic plasticity studies do not speak to this prediction because they have used pharmacological manipulations of both excitatory and inhibitory neurons. However, some studies have used cell-specific manipulations—such as cell-specific overexpression of potassium channels (50, 51)—that strongly support the notion that synaptic plasticity is guided at least in part by their own deviation from setpoint.
In our opinion, and although we have explored alternative rules (SI Appendix, Section 1.6), the most biologically plausible set of plasticity rules that lead to stable self-sustained dynamics comprises a hybrid rule that includes both standard homeostatic and cross-homeostatic terms. Such a two-term cross-homeostatic rule robustly led to a self-sustained, inhibition-stabilized network, with all units converging to their setpoints, and is directly consistent with current experimental data.
Biological Plausibility of Cross-Homeostatic Plasticity.
While the neural mechanisms underlying homeostatic plasticity remain to be elucidated, it is generally assumed that an individual neuron can maintain a running average of their firing rate over the course of hours as a result of Ca2+-activated sensors. Based on the deviation of this value from an ontogenetically determined setpoint, neurons up- or down-regulate the density of postsynaptic receptors accordingly (38, 58–60). Two-term cross-homeostatic plasticity would require additional, and apparently nonlocal information about the error in a given neuron’s presynaptic partners. Importantly however, this rule can be implemented locally because any postsynaptic neuron has access to the mean activity of its presynaptic partners simply as a result of its postsynaptic receptor activation. Indeed, a plasticity rule for WI←E weights with a similar cross-homeostatic error term has also been recently proposed and implemented based on the mean activation of postsynaptic receptors—more specifically the net postsynaptic currents, which provide a coupled measure of average presynaptic firing and synaptic weights (49).
Here we propose that cross-homeostatic plasticity could be implemented through postsynaptic metabotropic receptors (e.g., mGlu and GABAB). Such receptors would provide a mechanism for postsynaptic neurons to maintain a running average of the activity of its presynaptic partners that is decoupled from the synaptic weights. Metabotropic receptors are G protein coupled receptors that provide a low-pass filtered measure of presynaptic activity and are involved in a large number of incompletely understood neuromodulatory roles (61, 62). Since metabotropic receptors appear to undergo less homeostatic and associative plasticity, they provide a measure of presynaptic activity that is naturally decoupled from the ionotropic receptors (e.g., AMPA and GABAA) that are being up- and down-regulated.
Further support for the notion that individual neurons have access to global network activity emerges from studies suggesting that neurons might not homeostatically regulate activity at the individual neuron level but rather at the global population level (63). Such a global-level homeostasis could be achieved by nonsynaptic paracrine transmission. Indeed, retrograde messenger systems are ideally suited for this role, as they have already been implicated in signaling mean activity levels to local capillaries, driving the activity-dependent vasodilation that underlies functional MRI (64).
Paradoxical Effect and Standard Homeostatic Rules.
The paradoxical effect is one of the defining features of inhibition-stabilized networks, and a growing body of evidence suggests that the cortex operates in this particular dynamic regime (20, 65–67). As recent work has begun to hint (68), here we formally prove that the paradoxical effect applies important constraints to the potential learning rules that lead to the emergence of inhibition-stabilized networks. In the simplified case in which there is only homeostatic plasticity onto the inhibitory neurons, we can immediately see why the paradoxical effect renders standard homeostatic rules ineffective. If the I population is below its setpoint, standard homeostatic rules would increase WI←E, which paradoxically would further decrease I (Fig. 3), thus further increasing the error instead of decreasing it (Fig. 3). This reasoning is related to why, when using the standard family of homeostatic rules, the rate of plasticity onto the inhibitory neurons has to be much smaller—in effect making the “paradoxical homeostatic plasticity effect” much slower. Furthermore, our analytical stability analyses show that in the limit of vanishingly small excitatory learning rates (αEE,EI ≪ αIE,II) the stability region of the weight subsystem is bounded by the paradoxical condition. This means that the only allowed stable states with nonzero E activity will occur in the nonparadoxical regime, if any, and they will not be proper inhibition-stabilized, self-sustained regimes.
Future Directions and Experimental Predictions.
While we have taken the approach of implementing homeostatic plasticity rules at all four synapse classes in our model, it is important to stress that we have omitted other well-characterized forms of synaptic plasticity. In particular, we did not include associative long-term potentiation or spike-timing dependent plasticity. These forms of plasticity are generally considered to capture the correlation structure in networks that are driven by structured inputs. Arguably, because there is evidence that self-sustained forms of activity such as up-states develop in the absence of any structured external input (23, 25, 26) and because all excitatory and inhibitory neurons synchronously shift between quiescent and active states, associative forms of plasticity may not contribute significantly to these regimes. Nevertheless, we envision the cross-homeostatic rules we propose working hand in hand with associative forms of plasticity that impose high-dimensional structure on the top of the inhibition-stabilized dynamics. In fact, preliminary observations reveal that cross-homeostatic rules are capable of stabilizing recurrent networks while preserving imposed Hebbian-like structure in the weight matrix (SI Appendix, Fig. S6). Future work should explore the computational advantage of cross-homeostatic plasticity in models performing complex computational tasks, such as working memory, sensory timing, or motor control (5, 69–72).
An important implication of our results is that neuronal and network properties can operate in fundamentally different modes. That is, while homeostatic plasticity can lead to single neurons to reach their target setpoints in simple feedforward circuits, those same rules can be highly unstable when the neurons are placed even in the simplest of recurrent excitatory/inhibitory circuits with emergent dynamics. Furthermore, because emergent neural dynamic regimes are highly nonlinear, and in particular that stable self-sustained dynamic regimes exhibit a paradoxical effect, it is likely that the brain exhibits paradoxical or counterintuitive learning rules to generate self-sustained dynamic regimes.
Methods
Computational Model.
A two-population firing-rate model was implemented based on a previous inhibition-stabilized network model (19). The firing rate of the excitatory (E) and inhibitory (I) population obeyed Wilson and Cowan dynamics (73).
[1] |
[2] |
where WXY represents the weight between the presynaptic unit Y and postsynaptic unit X. The parameters τX and ηX represent a time constant and an independent noise term, respectively. The time constants were set to τE = 10 ms for the excitatory and τI = 2 ms for the inhibitory subpopulations. The noise term was an Ornstein–Uhlenbeck process with mean μx = 0, a time constant 1/Θx = 1 ms, and a sigma parameter of σx = 10. To elicit self-sustained activity, a step current was injected at the beginning of each trial on the excitatory population.
The function fY(x) represents the intrinsic excitability of the neurons, and it is modeled as a threshold-linear function with threshold θY and gain gY.
[3] |
The thresholds were set to θE = 4.8 and θI = 25, and the gains to gE = 1 and gI = 4. The higher thresholds in PV neurons are consistent with experimental findings (46).
The linear relationship between excitatory and inhibitory weights (Fig. 4) corresponds to the steady-state solution of the neural subsystem when the inhibitory and excitatory rates are at their target setpoints. The solution can be obtained by setting the left side of Eqs. 1 and 2 to zero and substituting the steady-state and values with Eset and Iset:
[4] |
[5] |
Thus, the slope of the E/I balance line in Fig. 4 corresponds to Eset/Iset. We chose WE←E and WI←E as the “free” weights. See details and analytical results in SI Appendix, Section 2.2.
Synaptic Plasticity.
Plasticity at all four weight classes (WE←E, WE←I, WI←E, and WI←I) was governed by different families of homeostatic-based plasticity rules, all driven by the deviation of the actual excitatory and inhibitory rates from their target setpoints (Eset and Iset). Three different learning rules are presented in the main text of this article.
Standard homeostatic family of rules.
[6] |
where αE and αI are the learning rates onto the excitatory and inhibitory units, respectively. The setpoints were based on empirically measured values in ex vivo cortical circuits (46), Eset = 5 and Iset = 14 Hz and follow a classic homeostatic formulation (33, 34, 43, 44). As outlined in SI Appendix, Section 1.5, we also examined variants of these rules, such as standard synaptic scaling (which includes the weight as a factor).
We prove that these rules are stable only in a narrow parameter regime: when excitatory plasticity dominates (SI Appendix, Section 2).
Cross-homeostatic family of rules.
[7] |
These rules differ from the standard homeostatic formulation in that the setpoints are “crossed,” meaning that the weights onto the excitatory (inhibitory) population change in order to minimize the inhibitory (excitatory) error. We prove that these rules are stable for any set of parameters (SI Appendix, Section 1.3).
Two-term cross-homeostatic family of rules.
[8] |
The two-term rules combine homeostatic and cross-homeostatic terms. This exact formulation can be obtained after an approximation of a gradient descent derivation on the following loss function:
[9] |
The mathematical derivation can be found in the SI Appendix, Section 3. We prove that these rules are stable for a biologically meaningful set of parameter values, as long as the homeostatic part does not dominate (SI Appendix, Section 1.4).
For all other methods, including the implementation of the multiunit firing rate and spiking models, numerical and analytical methods, proofs, and derivation of the two-term cross-homeostatic rule, see the SI Appendix.
Supplementary Material
Acknowledgments
We thank Ben Liu, Shanglin Zhou, Juan Romero-Sosa, and Helen Motanis for technical assistance and helpful discussions. We thank Joana Soldado-Magraner for comments on the manuscript. D.V.B. and M.J.S. were supported by the National Institutes of Health grant no. NS116589. R.L. was supported by Universidad Nacional de Quilmes, CONICET, and The Pew Charitable Trusts grant no. 2009-000360-006. S.S.-M. was supported by the Swiss National Science Foundation grant nos. P2ZHP3-187943 and P500PB-203133.
Footnotes
The authors declare no competing interest.
This article is a PNAS Direct Submission.
This article contains supporting information online at https://www.pnas.org/lookup/suppl/doi:10.1073/pnas.2200621119/-/DCSupplemental.
Data, Materials, and Software Availability
Software necessary to reproduce all data shown throughout this manuscript is fully available in the public repository GitHub (see links below). Computational analysis for the rate-based model was performed in custom-written MATLAB R2020a software (https://www.mathworks.com/). Brian2 was used to simulate the spiking model (https://briansimulator.org/). SageMath was used for the analytical proofs (https://www.sagemath.org/) (SI Appendix). The MATLAB source code that reproduces Figs. 2, 4–6 and SI Appendix, Fig. S4 is available at https://github.com/saraysoldado/Paradoxical2022 (74). The spiking model can be found at https://github.com/mikejseay/spiking-upstates/tree/Paradoxical2022 (75). The Jupyter notebooks with SageMath code to reproduce all analytical results are available at https://github.com/SMDynamicsLab/Paradoxical2022 (76).
References
- 1.Fuster J. M., Jervey J. P., Inferotemporal neurons distinguish and retain behaviorally relevant features of visual stimuli. Science 212, 952–955 (1981). [DOI] [PubMed] [Google Scholar]
- 2.Goldman-Rakic P. S., Cellular basis of working memory. Neuron 14, 477–485 (1995). [DOI] [PubMed] [Google Scholar]
- 3.Wang X.-J., Synaptic reverberation underlying mnemonic persistent activity. Trends Neurosci. 24, 455–463 (2001). [DOI] [PubMed] [Google Scholar]
- 4.Churchland M. M., et al. , Neural population dynamics during reaching. Nature 487, 51–56 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Hennequin G., Vogels T. P., Gerstner W., Optimal control of transient dynamics in balanced networks supports generation of complex movements. Neuron 82, 1394–1406 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.van Vreeswijk C., Sompolinsky H., Chaotic balanced state in a model of cortical circuits. Neural Comput. 10, 1321–1371 (1998). [DOI] [PubMed] [Google Scholar]
- 7.Brunel N., Dynamics of sparsely connected networks of excitatory and inhibitory spiking neurons. J. Comput. Neurosci. 8, 183–208 (2000). [DOI] [PubMed] [Google Scholar]
- 8.Destexhe A., Hughes S. W., Rudolph M., Crunelli V., Are corticothalamic ‘up’ states fragments of wakefulness? Trends Neurosci. 30, 334–342 (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Renart A., et al. , The asynchronous state in cortical circuits. Science 327, 587–590 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Ostojic S., Two types of asynchronous activity in networks of excitatory and inhibitory spiking neurons. Nat. Neurosci. 17, 594–600 (2014). [DOI] [PubMed] [Google Scholar]
- 11.Steriade M., McCormick D. A., Sejnowski T. J., Thalamocortical oscillations in the sleeping and aroused brain. Science 262, 679–685 (1993). [DOI] [PubMed] [Google Scholar]
- 12.McCormick D. A., GABA as an inhibitory neurotransmitter in human cerebral cortex. J. Neurophysiol. 62, 1018–1027 (1989). [DOI] [PubMed] [Google Scholar]
- 13.Steriade M., Contreras D., Spike-wave complexes and fast components of cortically generated seizures. I. Role of neocortex and thalamus. J. Neurophysiol. 80, 1439–1455 (1998). [DOI] [PubMed] [Google Scholar]
- 14.Tsodyks M. V., Skaggs W. E., Sejnowski T. J., McNaughton B. L., Paradoxical effects of external modulation of inhibitory interneurons. J. Neurosci. 17, 4382–4388 (1997). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Ozeki H., Finn I. M., Schaffer E. S., Miller K. D., Ferster D., Inhibitory stabilization of the cortical network underlies visual surround suppression. Neuron 62, 578–592 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Rubin D. B., Van Hooser S. D., Miller K. D., The stabilized supralinear network: A unifying circuit motif underlying multi-input integration in sensory cortex. Neuron 85, 402–417 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Rutishauser U., Slotine J.-J., Douglas R., Computation in dynamically bounded asymmetric systems. PLOS Comput. Biol. 11, e1004039 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Litwin-Kumar A., Rosenbaum R., Doiron B., Inhibitory stabilization and visual coding in cortical circuits with multiple interneuron subtypes. J. Neurophysiol. 115, 1399–1409 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Jercog D., et al. , UP-DOWN cortical dynamics reflect state transitions in a bistable network. eLife 6, e22425 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Sanzeni A., et al. , Inhibition stabilization is a widespread property of cortical networks. eLife 9, e54875 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Plenz D., Kitai S. T., Up and down states in striatal medium spiny neurons simultaneously recorded with spontaneous activity in fast-spiking interneurons studied in cortex-striatum-substantia nigra organotypic cultures. J. Neurosci. 18, 266–283 (1998). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Seamans J. K., Nogueira L., Lavin A., Synaptic basis of persistent activity in prefrontal cortex in vivo and in organotypic cultures. Cereb. Cortex 13, 1242–1250 (2003). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Johnson H. A., Buonomano D. V., Development and plasticity of spontaneous activity and Up states in cortical organotypic slices. J. Neurosci. 27, 5915–5925 (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Golshani P., et al. , Internally mediated developmental desynchronization of neocortical network activity. J. Neurosci. 29, 10890–10899 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Motanis H., Buonomano D., Delayed in vitro development of Up states but normal network plasticity in Fragile X circuits. Eur. J. Neurosci. 42, 2312–2321 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Motanis H., Buonomano D., Decreased reproducibility and abnormal experience-dependent plasticity of network dynamics in Fragile X circuits. Sci. Rep. 10, 14535 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Donato F., Rompani S. B., Caroni P., Parvalbumin-expressing basket-cell network plasticity induced by experience regulates adult learning. Nature 504, 272–276 (2013). [DOI] [PubMed] [Google Scholar]
- 28.Froemke R. C., Plasticity of cortical excitatory-inhibitory balance. Annu. Rev. Neurosci. 38, 195–219 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Hennequin G., Agnes E. J., Vogels T. P., Inhibitory plasticity: Balance, control, and codependence. Annu. Rev. Neurosci. 40, 557–579 (2017). [DOI] [PubMed] [Google Scholar]
- 30.Ahmadian Y., Miller K. D., (2019) What is the dynamical regime of cerebral cortex? arXiv:1908.10101 [q-bio]. [DOI] [PMC free article] [PubMed]
- 31.Chen Y.-H., et al. , PV network plasticity mediated by neuregulin1-ErbB4 signalling controls fear extinction. Mol. Psychiatry 27, 896–906 (2021). [DOI] [PubMed] [Google Scholar]
- 32.He X., et al. , Gating of hippocampal rhythms and memory by synaptic plasticity in inhibitory interneurons. Neuron 109, 1013–1028.e1019 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Turrigiano G. G., Leslie K. R., Desai N. S., Rutherford L. C., Nelson S. B., Activity-dependent scaling of quantal amplitude in neocortical neurons. Nature 391, 892–896 (1998). [DOI] [PubMed] [Google Scholar]
- 34.van Rossum M. C., Bi G. Q., Turrigiano G. G., Stable Hebbian learning from spike timing-dependent plasticity. J. Neurosci. 20, 8812–8821 (2000). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Kilman V., van Rossum M. C., Turrigiano G. G., Activity deprivation reduces miniature IPSC amplitude by decreasing the number of postsynaptic GABA(A) receptors clustered at neocortical synapses. J. Neurosci. 22, 1328–1337 (2002). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Turrigiano G. G., Nelson S. B., Homeostatic plasticity in the developing nervous system. Nat. Rev. Neurosci. 5, 97–107 (2004). [DOI] [PubMed] [Google Scholar]
- 37.Peng Y.-R., et al. , Postsynaptic spiking homeostatically induces cell-autonomous regulation of inhibitory inputs via retrograde signaling. J. Neurosci. 30, 16220–16231 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Pozo K., Goda Y., Unraveling mechanisms of homeostatic synaptic plasticity. Neuron 66, 337–351 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Turrigiano G. G., The self-tuning neuron: Synaptic scaling of excitatory synapses. Cell 135, 422–435 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Hengen K. B., Lambo M. E., Van Hooser S. D., Katz D. B., Turrigiano G. G., Firing rate homeostasis in visual cortex of freely behaving rodents. Neuron 80, 335–342 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Ma Z., Turrigiano G. G., Wessel R., Hengen K. B., Cortical circuit dynamics are homeostatically tuned to criticality in vivo. Neuron 104, 655–664.e654 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Kuhlman S. J., et al. , A disinhibitory microcircuit initiates critical-period plasticity in the visual cortex. Nature 501, 543–546 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Liu J. K., Buonomano D. V., Embedding multiple trajectories in simulated recurrent neural networks in a self-organizing manner. J. Neurosci. 29, 13172–13181 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Vogels T. P., Sprekeler H., Zenke F., Clopath C., Gerstner W., Inhibitory plasticity balances excitation and inhibition in sensory pathways and memory networks. Science 334, 1569–1573 (2011). [DOI] [PubMed] [Google Scholar]
- 45.Neske G. T., Patrick S. L., Connors B. W., Contributions of diverse excitatory and inhibitory neurons to recurrent network activity in cerebral cortex. J. Neurosci. 35, 1089–1105 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Romero-Sosa J. L., Motanis H., Buonomano D. V., Differential excitability of PV and SST neurons results in distinct functional roles in inhibition stabilization of up states. J. Neurosci. 41, 7182–7196 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Keck T., et al. , Loss of sensory input causes rapid structural changes of inhibitory neurons in adult mouse visual cortex. Neuron 71, 869–882 (2011). [DOI] [PubMed] [Google Scholar]
- 48.Gainey M. A., Aman J. W., Feldman D. E., Rapid disinhibition by adjustment of PV intrinsic excitability during whisker map plasticity in mouse S1. J. Neurosci. 38, 4749–4761 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Mackwood O., Naumann L. B., Sprekeler H., Learning excitatory-inhibitory neuronal assemblies in recurrent networks. eLife 10, e59715 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Burrone J., O’Byrne M., Murthy V. N., Multiple forms of synaptic plasticity triggered by selective suppression of activity in individual neurons. Nature 420, 414–418 (2002). [DOI] [PubMed] [Google Scholar]
- 51.Xue M., Atallah B. V., Scanziani M., Equalizing excitation-inhibition ratios across visual cortical neurons. Nature 511, 596–600 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Hebb D. O., The Organisation of Behaviour: A Neuropsychological Theory (Science Editions, New York, 1949). [Google Scholar]
- 53.Miller K. D., Keller J. B., Stryker M. P., Ocular dominance column development: Analysis and simulation. Science 245, 605–615 (1989). [DOI] [PubMed] [Google Scholar]
- 54.Buonomano D. V., Merzenich M. M., Cortical plasticity: From synapses to maps. Annu. Rev. Neurosci. 21, 149–186 (1998). [DOI] [PubMed] [Google Scholar]
- 55.Martin S. J., Grimwood P. D., Morris R. G., Synaptic plasticity and memory: An evaluation of the hypothesis. Annu. Rev. Neurosci. 23, 649–711 (2000). [DOI] [PubMed] [Google Scholar]
- 56.Song S., Miller K. D., Abbott L. F., Competitive Hebbian learning through spike-timing-dependent synaptic plasticity. Nat. Neurosci. 3, 919–926 (2000). [DOI] [PubMed] [Google Scholar]
- 57.Hengen K. B., Torrado Pacheco A., McGregor J. N., Van Hooser S. D., Turrigiano G. G., Neuronal firing rate homeostasis is inhibited by sleep and promoted by wake. Cell 165, 180–191 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Trojanowski N. F., Bottorff J., Turrigiano G. G., Activity labeling in vivo using CaMPARI2 reveals intrinsic and synaptic differences between neurons with high and low firing rate set points. Neuron 109, 663–676.e665 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Liu Z., Golowasch J., Marder E., Abbott L. F., A model neuron with activity-dependent conductances regulated by multiple calcium sensors. J. Neurosci. 18, 2309–2320 (1998). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Joseph A., Turrigiano G. G., All for one but not one for all: Excitatory synaptic scaling and intrinsic excitability are coregulated by CaMKIV, whereas inhibitory synaptic scaling is under independent control. J. Neurosci. 37, 6778–6785 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Blein S., Hawrot E., Barlow P., The metabotropic GABA receptor: Molecular insights and their functional consequences. Cell. Mol. Life Sci. 57, 635–650 (2000). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Niswender C. M., Conn P. J., Metabotropic glutamate receptors: Physiology, pharmacology, and disease. Annu. Rev. Pharmacol. Toxicol. 50, 295–322 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Slomowitz E., et al. , Interplay between population firing stability and single neuron dynamics in hippocampal networks. eLife 4, e04378 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Drew P. J., Vascular and neural basis of the BOLD signal. Curr. Opin. Neurobiol. 58, 61–69 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Zucca S., et al. , An inhibitory gate for state transition in cortex. eLife 6, e26177 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Mahrach A., Chen G., Li N., van Vreeswijk C., Hansel D., Mechanisms underlying the response of mouse cortical networks to optogenetic manipulation. eLife 9, e49967 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Sadeh S., Clopath C., Inhibitory stabilization and cortical computation. Nat. Rev. Neurosci. 22, 21–37 (2021). [DOI] [PubMed] [Google Scholar]
- 68.Sadeh S., Clopath C., Excitatory-inhibitory balance modulates the formation and dynamics of neuronal assemblies in cortical networks. Sci. Adv. 7, eabg8411 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Sussillo D., Abbott L. F., Generating coherent patterns of activity from chaotic neural networks. Neuron 63, 544–557 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Laje R., Buonomano D. V., Robust timing and motor patterns by taming chaos in recurrent neural networks. Nat. Neurosci. 16, 925–933 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Goudar V., Buonomano D. V., Encoding sensory and motor patterns as time-invariant trajectories in recurrent neural networks. eLife 7, e31134 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Cueva C. J., et al. , Low-dimensional dynamics for working memory and time encoding. Proc. Natl. Acad. Sci. U.S.A. 117, 23021–23032 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.Wilson H. R., Cowan J. D., Excitatory and inhibitory interactions in localized populations of model neurons. Biophys. J. 12, 1–24 (1972). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.S. Soldado-Magraner, Paradoxical2022. GitHub. https://github.com/saraysoldado/Paradoxical2022. Deposited 7 July 2022. [Google Scholar]
- 75.M. J. Seay, Spiking-upstates. GitHub. https://github.com/mikejseay/spiking-upstates/tree/Paradoxical2022. Deposited 29 June 2022. [Google Scholar]
- 76.R. Laje, Paradoxical2022. GitHub. https://github.com/SMDynamicsLab/Paradoxical2022. Deposited 30 June 2022. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Software necessary to reproduce all data shown throughout this manuscript is fully available in the public repository GitHub (see links below). Computational analysis for the rate-based model was performed in custom-written MATLAB R2020a software (https://www.mathworks.com/). Brian2 was used to simulate the spiking model (https://briansimulator.org/). SageMath was used for the analytical proofs (https://www.sagemath.org/) (SI Appendix). The MATLAB source code that reproduces Figs. 2, 4–6 and SI Appendix, Fig. S4 is available at https://github.com/saraysoldado/Paradoxical2022 (74). The spiking model can be found at https://github.com/mikejseay/spiking-upstates/tree/Paradoxical2022 (75). The Jupyter notebooks with SageMath code to reproduce all analytical results are available at https://github.com/SMDynamicsLab/Paradoxical2022 (76).