Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2020 Aug 7;16(8):e1007659. doi: 10.1371/journal.pcbi.1007659

Achieving stable dynamics in neural circuits

Leo Kozachkov 1,2,3,, Mikael Lundqvist 1,2,4,, Jean-Jacques Slotine 1,3,, Earl K Miller 1,2,‡,*
Editor: Adrian M Haith5
PMCID: PMC7446801  PMID: 32764745

Abstract

The brain consists of many interconnected networks with time-varying, partially autonomous activity. There are multiple sources of noise and variation yet activity has to eventually converge to a stable, reproducible state (or sequence of states) for its computations to make sense. We approached this problem from a control-theory perspective by applying contraction analysis to recurrent neural networks. This allowed us to find mechanisms for achieving stability in multiple connected networks with biologically realistic dynamics, including synaptic plasticity and time-varying inputs. These mechanisms included inhibitory Hebbian plasticity, excitatory anti-Hebbian plasticity, synaptic sparsity and excitatory-inhibitory balance. Our findings shed light on how stable computations might be achieved despite biological complexity. Crucially, our analysis is not limited to analyzing the stability of fixed geometric objects in state space (e.g points, lines, planes), but rather the stability of state trajectories which may be complex and time-varying.

Author summary

Stability is essential for any complex system including, and perhaps especially, the brain. The brain’s neural networks are highly dynamic and noisy. Activity fluctuates from moment to moment and can be highly variable. Yet it is critical that these networks reach a consistent state (or sequence of states) for their computations to make sense. Failures in stability have consequences ranging from mild (e.g incorrect decisions) to severe (disease states). In this paper we use tools from control theory and dynamical systems theory to find mechanisms which produce stability in recurrent neural networks (RNNs). We show that a kind of “unlearning” (inhibitory Hebbian and excitatory anti-Hebbian plasticity), balance of excitation and inhibition, and sparse anatomical connectivity all lead to stability. Crucially, we focus on the stability of neural trajectories. This is different from traditional studies of stability of fixed points or planes. We do not assess what trajectories our networks will follow but, rather, when these trajectories will all converge towards each other to achieve stability.

Introduction

Behavior emerges from complex neural dynamics unfolding over time in multi-area brain networks. Even in tightly controlled experimental settings, these neural dynamics often vary between identical trials [1,2]. This can be due to a variety of factors including variability in membrane potentials, inputs, plastic changes due to recent experience and so on. Yet, in spite of these fluctuations, brain networks must achieve computational stability: despite being “knocked around” by plasticity and noise, the behavioral output of the brain on two experimentally identical trials needs to be similar. How is this stability achieved?

Stability has played a central role in computational neuroscience since the 1980’s, with the advent of models of associative memory that stored neural activation patterns as stable point attractors [37], although researchers were thinking about the brain’s stability since as early as the 1950’s [8]. The vast majority of this work is concerned with the stability of activity around points, lines, or planes in neural state space [9,10]. However, recent neurophysiological studies have revealed that in many cases, single-trial neural activity is highly dynamic, and therefore potentially inconsistent with a static attractor viewpoint [1,11]. Consequently, there has been a number of recent studies—both computational and experimental—which focus more broadly on the stability of neural trajectories [12,13], which may be complex and time-varying.

While these studies provide important empirical results and intuitions, they do not offer analytical insight into mechanisms for achieving stable trajectories in recurrent neural networks. Nor do they offer insights into achieving such stability in plastic (or multi-modal) networks. Here we focus on finding conditions that guarantee stable trajectories in recurrent neural networks and thus shed light onto how stable trajectories might be achieved in vivo.

To do so, we used contraction analysis, a concept developed in control theory [14]. Unlike a chaotic system where perturbations and distortions can be amplified over time, the population activity of a contracting network will converge towards the same trajectory, thus achieving stable dynamics (Fig 1). One way to understand contraction is to represent the state of a network at a given time as a point in the network’s ‘state-space’, for instance the space spanned by the possible firing rates of all the networks’ neurons. This state-space has the same number of dimensions as the number of units n in the network. A particular pattern of neural firing rates corresponds to a point in this state-space. This point moves in the n dimensions as the firing rates change and traces out a trajectory over time.

Fig 1. Cartoon demonstrating the contraction property.

Fig 1

In a network with N neural units and S dynamic synaptic weights, the network activity can be described a trajectory over time in an (N + S)-dimensional space. In a contracting system all such trajectories will converge exponentially in some metric towards each other over time, regardless of initial conditions. In other words, the distance between any two trajectories shrinks to zero—potentially after transient divergence (as shown).

In a contracting network, all such trajectories converge. These contracting dynamics have previously been used in several applications, including neural networks with winner take all dynamics [15,16], in a model of action-selection in the basal ganglia [17], and to explain how neural synchronization can protect from noise [18]. Here, we instead explore how contraction can be achieved generally in more complex recurrent neural networks (RNNs) including those with plastic weights. We used RNNs that received arbitrary time-varying inputs and had synapses that changed on biologically relevant timescales [1921]. Our analysis reveals several novel classes of mechanisms that produced contraction including inhibitory Hebbian plasticity, excitatory anti-Hebbian plasticity, excitatory-inhibitory balance, and sparse connectivity. For the first two parts of the Results section, we focus on contraction of both neural activity and components of the weight matrix (Fig 1). For the remaining parts of the Results section, we hold the weights fixed (i.e they become parameters, not variables) and focus on contraction of neural activity alone.

Results

The main tool we used to characterize contraction was the logarithmic norm (also known as a matrix measure). The formal definition of the logarithmic norm is as follows (from [22] section 2.2.2): let A be a matrix in Cn×n and ∥⋅∥i be an induced matrix norm on Cn×n. Then the corresponding logarithmic norm is the function μ():Cn×nR defined by

μ(A)=limϵ0+I+ϵAi1ϵ

In the same way that different vector norms induce different matrix norms, different vector norms also induce different logarithmic norms. Two important logarithmic norms which we use throughout the paper are those induced by the vector 1-norm and the vector 2-norm:

μ1(A)=maxj[ajj+ijn|aij|]μ2(A)=λmax(A*+A2)

Where λmax denotes the largest eigenvalue. To study the contraction properties of RNNs, we applied the logarithmic norm to the RNN’s Jacobians. The Jacobian of a dynamical system is a matrix essentially describing the local ‘traffic laws’ of nearby trajectories of the system in its state space. More formally, it is the matrix of partial derivatives describing how a change in any system variable impacts the rate of change of every other variable in the system. It was shown in [14] that if the logarithmic norm of the Jacobian is negative then all nearby trajectories are funneled towards one another (see S1A Text Section 1.2 for technical review). This, in turn, implies that all trajectories are funneled towards one another at rate called the contraction rate. The contraction rate and the logarithmic norm are related as follows: the maximum value attained by the absolute value of logarithmic norm of the Jacobian along the network’s trajectory is the contraction rate. In other words, if the logarithmic norm of the Jacobian is upper bounded by some negative number −c, where c > 0, then the contraction rate is simply c.

Importantly, the above description can be generalized to different metrics. A metric is a symmetric, positive definite matrix which generalizes the notion of Euclidean distance. Every invertible coordinate transformation y = θx yields a metric M = θTθ. To see this, consider the squared norm of ∥y2 = yTy = xTθTθx = xTMx. Thus, the norm of y is related to the norm of x through the metric M. If one can find metric in which the network is contracting—in the sense that its Jacobian has negative logarithmic norm–this implies contraction for all coordinate systems. This makes contraction analysis useful for analyzing systems where exponential convergence of trajectories is preceded by transient divergence (Fig 1) as in recent models of motor cortex [23,24]. In this case, it is usually possible to find a coordinate system in which the convergence of trajectories is ‘pure’. For example, linear stable systems were recently used in the motor control literature to find initial conditions which produce the most energetic neural response [23] They are ‘purely’ contracting in a metric defined by the eigenvectors of the weight matrix (see Example 5.1 in [14]) but transiently diverging in the identity metric (i.e M = I). Note that the identity metric corresponds to θ = I, which is simply the original, untransformed coordinate system.

Inhibitory hebbian plasticity & excitatory anti-Hebbian plasticity produce contraction

It is known that certain forms of synaptic plasticity can quickly lead to extreme instabilities if left unchecked [9,25]. Thus, the same feature that can aid learning can also yield chaotic neural dynamics if not regulated. It is not known how the brain resolves this dilemma. A growing body of evidence—both experimental and computational—suggests that inhibitory plasticity (that is, the strengthening of inhibitory synapses) can stabilize neural dynamics while simultaneously allowing for learning/training in neural circuits [2628]. By using the Jacobian analysis outlined above, we found that inhibitory Hebbian synaptic plasticity (as well as excitatory anti-Hebbian plasticity) indeed leads to stable dynamics in neural circuits. Specifically, we considered neural networks of the following common form:

x˙i=h(xi)+j=1NWijxj+ui(t) (1)

where the term xi denotes the ‘activation’ of neuron i as a function of time. Here we follow other authors [23] and interpret xi as the deviation from the baseline firing rate of neuron i. Note that this interpretation assumes that the baseline firing rates are positive–thus allowing for x to be negative—and large enough so that baseline + x > 0. The term Wij denotes the weight between neurons i and j the term h(xi) captures the dynamics neuron i would have in the absence of synaptic input, including self-feedback terms arising from the diagonal elements of the weight matrix—in other words, the dynamics neuron i would have if for all i and j, Wij = 0. The term being summed represents the weighted contribution of all the neurons in the network on the activity of neuron i. Finally, the term ui(t) represents external input into neuron i.

We did not constrain the inputs into the RNN (except that they were not infinite) and we did not specify the particular form of h(xi) except that it should be a leak term (i.e has a negative derivative for all x, see S1A Text Section 2.2.4, e.g h(xi) = −xi). Furthermore, we made no assumptions regarding the relative timescales of synaptic and neural activity. Synaptic dynamics were treated on an equal footing as neural dynamics. We considered synaptic plasticity of the following correlational form [29]:

W˙ij=-kijxixj-γ(t)Wij (2)

where the term kij > 0 is the learning rate for each synapse and γ(t) > 0 is a decay factor for each synapse. For technical reasons outlined in the appendix (S1A Text Section 3), we restricted K, the matrix containing the learning rates kij, to be positive semi-definite, symmetric, and have positive entries. A particular example of K satisfying these constraints is to have the learning rates of all synapses to be equal (i.e. kij = k > 0).

Before we show that (2) leads to overall synaptic and neural contraction, it’s useful to spend some time interpreting this plasticity. Since Wij can be positive or negative (corresponding to excitatory and inhibitory synapses, respectively), and xixj can be positive or negative (corresponding to correlated and anticorrelated neurons, respectively), there are four cases to consider. We summarize these cases in Table 1 and discuss them in details below. By Hebbian plasticity we refer to the increase of synaptic efficiency between correlated neurons [30]. In the context of simple neural networks with scalar weights, as we consider here, efficiency refers to the absolute value |w| of a weight. Thus, for excitatory synapses, (2) in fact describes anti-Hebbian plasticity, because the positive synaptic weight becomes less positive (and thus less efficient) between correlated neurons and more positive (thus more efficient) for anticorrelated neurons. For inhibitory synapses, (2) describes Hebbian plasticity because the direction of synaptic weight change is negative between correlated neurons, and thus the synapse becomes more efficient [31,32], while for anticorrelated neurons the direction of synaptic weight change is positive, and thus the synapse becomes less efficient. Plasticity of this form produced contracting neural and synaptic dynamics regardless of the initial values of the weights and neural activity (Figs 2 and 3). The black trace of Fig 3A shows that this is not simply due to the weights decaying to 0. Thus, this plasticity is not only contraction preserving, it is contracting ensuring. Furthermore, we showed that the network is contracting in a non-identity metric (which we derive from the system parameters in K), opening up the possibility of transient divergent dynamics in the identity metric, as seen in the modelling of motor dynamics [23].

Table 1. Summary of the effect of the plasticity described in Eq (2) on excitatory and inhibitory for correlated or anticorrelated pre and post synaptic neurons.

Correlated Neurons Anticorrelated Neurons
xixj > 0 xixj < 0
Excitatory Synapse Less Efficient More Efficient
w > 0 Δ|w| < 0 Δ|w| > 0
Inhibitory Synapse More Efficient Less Efficient
w < 0 Δ|w| > 0 Δ|w| < 0

Fig 2. Contracting dynamics of neural and synaptic activity.

Fig 2

Euclidean distances between synaptic and neural trajectories demonstrate exponential shrinkage over time. The top row of panels shows the activation of a randomly selected neural unit (black) and synapse (blue) across two simulations (dotted and solid line). The bottom row shows the average Euclidean distance in state space for the whole population across simulations with distinct, randomized starting conditions. Leftmost Panel: Simulations of a contracting system where only starting conditions differ over simulations. Center Panel: the same as in Leftmost but with an additional random pulse perturbation in one of the two simulations indicated by a red background shading. Rightmost Panel: same as in Center Panel but with additional sustained noise, unique to each simulation.

Fig 3.

Fig 3

The anti-Hebbian plasticity pushes the weight matrix towards symmetry. (Left) Plotted are the spectral norms (largest singular value) of the overall weight matrix as well as the anti-symmetric part of that matrix. Since every square matrix can be uniquely decomposed as the sum of a symmetric and anti-symmetric component—0.5*(W+W’) and 0.5*(W-W‘), respectively—the teal curve decaying to zero implies that the matrix becomes symmetric. The black trace shows the spectral norm of the overall weight matrix. If this quantity does not decay to zero, it implies that not all the weights have decayed to zero. On the right, we plot the largest eigenvalue of the symmetric part of W. A prerequisite for overall contraction of the network is that this quantity be less than or equal to the ‘leak-rate’ of the individual neurons. The dotted line shows our theoretical upper bound for this quantity, and the solid line shows the actual value of taken from a simulation (see Methods).

To explain how inhibitory Hebbian plasticity and excitatory anti-Hebbian plasticity work to produce contraction across a whole network, we needed to deal with the network in a holistic fashion, not by analyzing the dynamics of single neurons. To do so, we conceptualized RNNs with dynamic synapses as a single system formed by combining two subsystems, a neural subsystem and a synaptic subsystem. We showed that the above plasticity rule led the neural and synaptic subsystems to be independently contracting. Thus contraction analysis of the overall system then boiled down to examining the interactions between these subsystems [33].

We found that this plasticity works like an interface between these systems. It produces two distinct effects that push networks toward contraction. First, it makes the synaptic weight matrix symmetric (Fig 3A, red trace). This means that the weight between neuron i to j is the same as j to i. We showed this by using the fact that every matrix can be written as the sum of a purely symmetric matrix and a purely anti-symmetric matrix. An anti-symmetric matrix is one where the ij element is the negative of the ji element (i.e. Wij = −Wji) and all the diagonal elements are zero. We then showed that anti-Hebbian plasticity shrinks the anti-symmetric part of the weight matrix to zero, implying that the weight matrix becomes symmetric. The symmetry of the weight matrix ‘cancels out’ off-diagonals in the Jacobian matrix (see S1A Text Section 3) of the overall neural-synaptic system. Loosely speaking, off-diagonal terms in the Jacobian represent potentially destabilizing cross-talk between the two subsystems. Furthermore, anti-Hebbian plasticity makes the weight matrix negative semi-definite. This means that all its eigenvalues are less than or equal to zero (Fig 3).

Sparse connectivity pushes networks toward contraction

Synaptic connectivity in the brain is extraordinarily sparse. The adult human brain contains at least 1011 neurons yet each neuron forms and receives on average only 103−104 synaptic connections [34]. If the brain’s neurons were all-to-all connected this number would be on the order of 1011 synaptic connections per neuron (101110111011[synapticconnections][neurons]). Even in local patches of cortex, such as we model here, connectivity is far from all-to-all; cortical circuits are sparse [35]. Our analyses revealed that sparse connectivity helps produce global network contraction for many types of synaptic plasticity.

To account for the possibility that some synapses may have much slower plasticity than others (and can thus be treated as synapses with fixed amplitude), we made a distinction between the total number of synapses and the total number of plastic synapses. These plastic synapses then changed on a similar time-scale as the neural firing rates. By neural dynamics, we mean the change in neural activity as a function of time. We analyzed RNNs with the structure:

x˙i=hi(xi)+j=1NWij(t)r(xj)+ui(t) (3)

Where hi(xi) is a nonlinear leak term (see S1A Text Section 2.2.4), and r(xj) is a nonlinear activation function. The RNNs analyzed in this section are identical to those analyzed in the previous section, with the exception of the r terms, which we constrained to be linear. Under the assumption that the plastic synapses have a ‘forgetting term’, we show in the appendix (S1A Text Section 4) that if the following equation is satisfied for every neuron, then the overall network is contracting:

pi(gmaxwmax+αirmax)<βi (4)

where pi denotes the total number of afferent synapses into neuron i and αi denotes the fraction of afferent plastic synapses into neuron i. The term wmax refers to the maximum possible absolute efficiency of any single synapse. That is, wmax = maxi,j |wij|. Similarly, the term rmax refers to the maximum possible absolute value of r. That is, rmax = maxi,t |ri(t)|. The term βi denotes the contraction rate of the ith isolated neuron. That is, βi=-maxi,thixi(t). Recall from the introduction that the contraction rate measures how quickly the trajectories of a contracting system reconvene after perturbation. Finally, gmax refers to the maximum gain of any neuron in the network. That is, gmax=maxi,t|rixi|(t). Note that because βi is a positive number by assumption, it is always possible to decrease pi to the point where (4) is satisfied. Of course, it is possible that the only value of pi that satisfies (4) is the trivial solution pi = 0, which corresponds to removing all interconnections between neurons. Since these neurons are assumed to be contracting in isolation, the network is trivially contracting. However, if the term inside the parentheses of (4) is small enough, or βi is large enough, intermediate value of pi can be found which satisfy the inequality. Because increasing the sparsity of a network corresponds to decreasing pi, we may conclude that increasing the sparsity of connections pushes the system in the direction of contraction. Note that (4) also implies that the faster the individual neurons are contracting (i.e. the larger βi is), the denser you can connect them with other neurons while still preserving overall contraction.

Up to now we have focused our analysis on the case where synaptic weights vary on a timescale comparable to neurons, and must therefore be factored into the stability analysis. For the next two sections, we’ll apply contraction analysis to neural network in the case where the weights may be regarded as fixed relative to the neural dynamics (i.e. there is a separation of timescales).

E-I balance leads to contraction in static RNNs

Apart from making connections sparse, one way to ensure contraction is to make synaptic weights small. This can be seen for the case with static synapses by setting αi = 0 in the section above, where Wmax now has to be small to ensure contraction. Intuitively, this is because very small weights mean that neurons cannot exert much influence on one another. If the neurons are stable before interconnection, they will remain so. Since strong synaptic weights are commonly observed in the brain, we were more interested in studying when contraction can arise irrespective of weight amplitude. Negative and positive synaptic currents are approximately balanced in biology [3638]. We reasoned that such balance might allow much larger weight amplitudes while still preserving contraction since most of the impact of such synapses cancel and the net effect small. This was indeed the case. To show this, we studied the same RNN as in the section above, while assuming additionally that the weights are static. In particular, we show in the appendix (S1A Text Section 5) that contraction can be assessed by studying the eigenvalues of the symmetric part of W (i.e. W+WT2).

Before we discuss the above result in detail, it is useful here to quickly review some facts about the stability of nonlinear systems as compared to the stability of linear systems. In particular, the fact that the eigenvalues of W are only informative for assessing contraction in regions where the dynamics may be regarded as linear. This is because in linear time-variant (LTI) systems (i.e. x˙=Ax) stability is completely characterized in terms of the eigenvalues of A. However, this is not true for nonlinear systems, even those of the linear time-varying form x˙=A(t)x. To see this, consider the following counter-example (from [39], section 4.2.2):

[x˙y˙]=[-1e2t0-1][xy] (5)

The eigenvalues of A(t) are (−1, 1) for all time, however one can verify by direct evalution that the solution of this system satisfies y = y(0)et, x˙=-x+y(0)et which is unstable along x. However, it can be shown straightforwardly that if the eigenvalues of the symmetric part of A(t) are all negative, then the system is stable [39]. This fact underlies our analysis, and highlights the reason why the eigenvalues of the symmetric part of W are important for stability.

Returning to our results, we show that if excitatory to inhibitory connections are of equal amplitude (and opposite sign) as inhibitory to excitatory connections, they will not interfere negatively with stability—regardless of amplitude (see S1A Text Section 5). This is because connections between inhibitory and excitatory units will be in the off-diagonal of the overall weight matrix and get cancelled out when computing the symmetric part. As an intuitive example, consider a two-neuron circuit made of one excitatory neuron and one inhibitory neuron connected recurrently (as in [40], Fig 1A). Assume that the overall weight matrix has the following structure:

W=(w-ww-w)

When taking the symmetric part of this matrix, the off-diagonal elements cancel out, leaving only the diagonal elements to consider. Since the eigenvalues of a diagonal matrix are simply its diagonal elements, we can conclude that if the excitatory and inhibitory subpopulations are independently contracting (w is less than the contraction rate of an isolated neuron), then overall contraction is guaranteed. It is straightforward to generalize this simple two-neuron example to circuits achieving E-I balance through interacting populations (see S1A Text Section 5). It is also straightforward to generalize to the case where E-I and I-E connections do not cancel out exactly neuron by neuron, but rather they cancel out in a statistical sense where the mean amplitudes are matched. Another way to view this E-I balance is in the framework of combinations of contracting systems (Fig 4). It is known that combining independently contracting systems in negative feedback preserves contraction [14]. We show that E-I balance actually translates to this negative feedback and thus can preserve contraction.

Fig 4. Cartoon illustrating the combination properties of contracting systems.

Fig 4

A) Two isolated, contracting systems. The Jacobian of the overall system is block diagonal, with all zeros on the off-diagonal—corresponding to the fact that the systems are not connected. B) If one of the systems is connected to the other in a feedforward manner, the overall Jacobian is changed by the presence of non-zero terms on the bottom left block—corresponding to the connections going from the ‘top’ system to the ‘bottom’ system. This Jacobian may not be negative definite. However, it is known that a coordinate change exists which will make it negative definite. Thus, hierarchically connected contracting systems are contracting. C) If the systems are reciprocally connected, the system may lose its contracting properties (for example in the case of positive feedback). However, it is known that if the feedforward connections (blue) are ‘equal and opposite’ to the feedback connections (green) then the overall system is contracting. We use this property in the main text to prove that inhibitory Hebbian plasticity and excitatory anti-Hebbian plasticity lead to contracting neural circuits.

Relation to other models with fading memory

As can be seen in Fig 2, contracting systems have ‘fading memories’. This means that past events will affect the current state, but that the impact of a transient perturbation gradually decays over time. Consider the transient input in Fig 2 (red panel) presented on only one of the two trials to the network. Because the input is only present on one trial and not the other we call it a perturbation. When this perturbation occurs, the trajectories of the two trials become separated. However, after the disturbance is removed, the distance between the network’s trajectories starts shrinking back to zero again. Thus, the network does not hold onto the memory of the perturbation indefinitely—the memory fades away. A similar property has been used in Echo State Networks (ESNs) and liquid state machines (LSMs) to perform useful brain-inspired computations [41,42]. These networks are an alternative to classical attractor models in which neural computations are performed by entering stable states rather than by ‘fading memories’ of external inputs [43].

While there are several distinctions between the networks described above and ESNs (e.g. ESNs are typically discrete time dynamical systems, rather than continuous), we show in the appendix (S1A Text Section 5.1) that they are a special case of the networks considered here. We show this for ESNs as opposed to LSMs because LSMs are typically implemented on integrate and fire neurons which, because of the spike reset, have a sharp discontinuity in their dynamics—making them unamenable to contraction analysis.

By highlighting the link between contraction and ESNs, we demonstrate that the contracting neural networks considered here are in principle capable of performing useful and interesting neural computations. In other words, the strong stability properties of contracting neural networks do not automatically prohibit them from doing interesting computations. By working within the framework of contraction analysis we were able to study networks both with dynamic synapses and non-identity metrics—a much broader model space than allowed by the standard ESN framework.

Discussion

We studied a fundamental question in neuroscience: How do neural circuits maintain stable dynamics in the presence of disturbance, noisy inputs and plastic change? We approached this problem from the perspective of dynamical systems theory, in light of the recent successes of understand neural circuits as dynamical systems [44]. We focused on contracting dynamical systems, which are yet largely unexplored in neuroscience, as a solution to the problem outlined above. We did so for three reasons:

  1. Contracting systems can be input-driven. This is important because neural circuits are typically bombarded with time-varying inputs either from the environment or from other brain areas. Previous stability analyses have focused primarily on the stability of RNNs without time-varying input. These analyses are most insightful in situations where the input into a circuit can be approximated as either absent or constant. However, naturalistic stimuli tend to be highly time-varying and complex [45].

  2. Contracting systems are robust to noise and disturbances. Perturbations to a contracting system are forgotten at the rate of the contraction and noise therefore does not stack up over time. Importantly, the rate of forgetting (i.e the contraction rate) does not change with the size of the perturbation. Thus dynamic stability can co-exist with high trial-to-trial variability in contracting neural networks, as observed in biology.

  3. Contracting systems can be combined with one another in ways that preserve contraction (Fig 4). This is not true of most dynamical systems which can easily ‘blow up’ when connected in feedback with one another [8]. This combination property is important as it is increasingly clear that cognitive functions such as working memory or attention are distributed in multiple cortical and sub-cortical regions [46,47]. In particular, prefrontal cortex has been suggested as a hub that can reconfigure the cortical effective network based on task demands [48]. Brain networks must therefore be able to effectively reconfigure themselves on a fast time-scale without loss of stability. Most attempts in modelling cognition, for instance working memory, tend to utilize single and often autonomous networks. Contracting networks display a combination of input-driven and autonomous dynamics, and thus have key features necessary for combining modules into flexible and distributed networks.

To understand what mechanisms lead to contraction in neural circuits, we applied contraction analysis to RNNs. For RNNs with static weights, we found that the well- known Echo State Networks are a special case of a contracting network. Since realistic synapses are complex dynamical systems in their own right, we went one step further and asked when neural circuits with dynamic synapses would be contracting. We found that inhibitory Hebbian plasticity as well as excitatory anti-Hebbian plasticity and synaptic sparsity all lead to contraction in a broad class of RNNs.

Inhibitory plasticity has recently been the focus of many experimental and computational studies due to its stabilizing nature as well as its capacity for facilitating nontrivial computations in neural circuits [27,28,49]. It is known to give rise to excitatory-inhibitory balance and has been implicated as the mechanism behind many experimental findings such as sparse firing rates in cortex [28]. Similarly, anti-Hebbian plasticity exists across many brain areas and species, such as salamander and rabbit retina [31], rat hippocampus [50,51], electric fish electrosensory lobe [52] and mouse prefrontal cortex [53]. Anti-Hebbian dynamics can give rise to sparse neural codes which decrease correlations between neural activity and increase overall stimulus representation in the network [54]. Because of this on-line decorrelation property, anti-Hebbian plasticity has also been implicated in predictive coding [31,52]. Our findings suggest that it also increase the stability of networks.

For more general forms of synaptic dynamics, we showed that synaptic sparsity pushes RNNs towards being contracting. This aligns well with the experimental observation that synaptic connectivity is typically extremely sparse in the brain. Our results suggest that sparsity may be one factor pushing the brain towards dynamical stability. It is therefore interesting that synapses are regulated by homeostatic processes where synapses neighboring an upregulated synapse are immediately downregulated [55]. On the same note, we also observed that balancing the connections between excitatory and inhibitory populations leads to contraction. Balance between excitatory and inhibitory synaptic inputs are often observed in biology [3638], and could thus serve contractive stability purposes. Related computational work on spiking networks has suggested that balanced synaptic currents leads to fast response properties, efficient coding, increased robustness of function and can support complex dynamics related to movements [21,5658].

A main advantage to our approach is that it provides provable certificates of global contractive stability for nonlinear, time-varying RNNs with synaptic plasticity. This distinguishes it from previous works where—while very interesting and useful—stability is experimentally observed, but not proven [12]. In some cases [23,24], linear stability around the origin is proven (which implies that there is a contraction region around the origin) but the size of this region is neither established nor sought after. Indeed, one future direction we are pursuing is the question of: given an RNN, can one provide a certificate of contractive stability in a region? An answer to this question would shed light on the stability properties of known RNN models in the literature (e.g. trained RNNs, biologically-detailed spiking models, etc.).

Experimental neuroscience is moving in the direction of studying many interacting neural circuits simultaneously. This is fueled by the expanding capabilities of recording multiple areas simultaneously in vivo and study their interactions. This increases the need for multi-modal cognitive models. We therefore anticipate that the presented work can provide a useful foundation for how cognition in noisy and distributed computational networks can be understood.

Materials and methods

In the interested of space and cohesion, we’ve placed all the detailed proofs of main results into the appendix. The appendix was written to be self-contained, and thus also contains additional definitions of mathematical objects used throughout the text. Simulations (Figs 2 and 3) were performed in Python. Code to reproduce the figures is available at [https://github.com/kozleo/stable_dynamics]. Numerical integrating was performed using sdeint, an open-source collection of numerical algorithms for integrating stochastic ordinary differential equations.

Fig 2 details:

All parameters and time constants in Eqs (1) and (2) were set to one. The integration step-size, dt, was set to 1e-2.

Initial conditions for both neural and synaptic activation were drawn uniformly between -1 and 1. Inputs into the network were generated by drawing N frequencies uniformly between dt and 100dt, phases between 0 and 2π, amplitudes between 0 and 20 and generating an N x Time vector of sinusoids with the above parameters.

The perturbations of the network was achieved by adding a vector of all 10s (i.e an additive vector input into the network, with each network of the element equal to 10) to the above input on one of the trials for 100 time steps in the middle of the simulation.

The noise was generated by driving each neural unit with an independent Weiner process (sigma = .2).

Fig 3 details:

The weight matrix used was the same as in Fig 2, leftmost panel (without perturbation, without noise).

Supporting information

S1 Text. The supplementary appendix file contains extensive mathematical proofs of the results stated above.

We kept the appendix self-contained by restating the basic results of contraction analysis and linear algebra which we used often in our proofs.

(PDF)

Acknowledgments

We thank Pawel Herman for comments on an earlier version of this manuscript. We thank Michael Happ and all members of the Miller Lab for helpful discussions and suggestions.

Data Availability

All detailed proofs of main results are found in the appendix. Simulations (Figs 2 and 3) were performed in Python. Code to reproduce the figures is available at [https://github.com/kozleo/stable_dynamics]. Numerical integration was performed using sdeint, an open-source collection of numerical algorithms for performing integrations of stochastic ordinary differential equations.

Funding Statement

This work was supported by NIMH R37MH087027, The MIT Picower Institute Innovation Fund, ONR MURI N00014-16-1-2832, and Swedish Research Council Starting Grant 2018-04197. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Lundqvist M, Rose J, Herman P, Brincat SL, Buschman TJ, Miller EK. Gamma and Beta Bursts Underlie Working Memory. Neuron. 2016;90: 152–164. 10.1016/j.neuron.2016.02.028 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Churchland MM, Yu BM, Cunningham JP, Sugrue LP, Cohen MR, Corrado GS, et al. Stimulus onset quenches neural variability: A widespread cortical phenomenon. Nat Neurosci. 2010;13: 369–378. 10.1038/nn.2501 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Hopfield JJ. Neural networks and physical systems with emergent collective computational abilities. Proc Natl Acad Sci. 1982;79: 2554–2558. 10.1073/pnas.79.8.2554 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Hirsch MW. Convergent activation dynamics in continuous time networks. Neural Networks. 1989. pp. 331–349. 10.1016/0893-6080(89)90018-X [DOI] [Google Scholar]
  • 5.Cohen MA, Grossberg S. Absolute Stability of Global Pattern Formation and Parallel Memory Storage by Competitive Neural Networks. IEEE Trans Syst Man Cybern. 1983;SMC-13: 815–826. 10.1109/TSMC.1983.6313075 [DOI] [Google Scholar]
  • 6.Lundqvist M, Herman P, Lansner A. Theta and gamma power increases and alpha/beta power decreases with memory load in an attractor network model. J Cogn Neurosci. 2011;23: 3008–3020. 10.1162/jocn_a_00029 [DOI] [PubMed] [Google Scholar]
  • 7.Lansner A, Ekeberg O. Reliability and Speed of Recall in an Associative Network. IEEE Trans Pattern Anal Mach Intell. 1985;PAMI-7: 490–498. 10.1109/tpami.1985.4767688 [DOI] [PubMed] [Google Scholar]
  • 8.Ashby W. Design for a brain: The origin of adaptive behaviour. Chapman & Hall Ltd; 1952. [Google Scholar]
  • 9.Dayan P, Abbot LF. Theoretical Neuroscience Computational Neuroscience. The MIT press; 2005. 10.1016/j.neuron.2008.10.019 [DOI] [Google Scholar]
  • 10.Zhang H, Wang Z, Liu D. A comprehensive review of stability analysis of continuous-time recurrent neural networks. IEEE Trans Neural Networks Learn Syst. 2014;25: 1229–1262. 10.1109/TNNLS.2014.2317880 [DOI] [Google Scholar]
  • 11.Spaak E, Watanabe K, Funahashi S, Stokes MG. Stable and dynamic coding for working memory in primate prefrontal cortex. J Neurosci. 2017;37: 6503–6516. 10.1523/JNEUROSCI.3364-16.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Laje R, Buonomano DV. Robust timing and motor patterns by taming chaos in recurrent neural networks. Nat Neurosci. 2013;16: 925–933. 10.1038/nn.3405 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Chaisangmongkon W, Swaminathan SK, Freedman DJ, Wang XJ. Computing by Robust Transience: How the Fronto-Parietal Network Performs Sequential, Category-Based Decisions. Neuron. 2017;93: 1504–1517.e4. 10.1016/j.neuron.2017.03.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Lohmiller W, Slotine J-JE. On Contraction Analysis for Non-linear Systems. Automatica. 1998;34: 683–696. 10.1016/S0005-1098(98)00019-3 [DOI] [Google Scholar]
  • 15.Rutishauser U, Douglas RJ, Slotine J-J. Collective stability of networks of winner-take-all circuits*. 2018. [cited 31 Oct 2019]. [DOI] [PubMed] [Google Scholar]
  • 16.Rutishauser U, Slotine J-J, Douglas R. Computation in Dynamically Bounded Asymmetric Systems. PLoS Comput Biol. 2015;11: 1004039 10.1371/journal.pcbi.1004039 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Girard B, Tabareau N, Pham QC, Berthoz A, Slotine J-J. Where neuroscience and dynamic system theory meet autonomous robotics: A contracting basal ganglia model for action selection. Neural Networks. 2008;21: 628–641. 10.1016/j.neunet.2008.03.009 [DOI] [PubMed] [Google Scholar]
  • 18.Tabareau N, Slotine JJ, Pham QC. How synchronization protects from noise. PLoS Comput Biol. 2010;6: 1–9. 10.1371/journal.pcbi.1000637 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Orhan AE, Ma WJ. A diverse range of factors affect the nature of neural representations underlying short-term memory. Nat Neurosci. 2019;22: 275–283. 10.1038/s41593-018-0314-y [DOI] [PubMed] [Google Scholar]
  • 20.Mongillo G, Barak O, Tsodyks M. Synaptic Theory of Working Memory. Science (80-). 2008;319: 1543 10.1126/science.1150769 [DOI] [PubMed] [Google Scholar]
  • 21.Lundqvist M, Compte A, Lansner A. Bistable, Irregular Firing and Population Oscillations in a Modular Attractor Memory Network. Morrison A, editor. PLoS Comput Biol. 2010;6: e1000803 10.1371/journal.pcbi.1000803 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Vidyasagar M. Nonlinear systems analysis. 2002. [Google Scholar]
  • 23.Hennequin G, Vogels TP, Gerstner W. Optimal control of transient dynamics in balanced networks supports generation of complex movements. Neuron. 2014;82: 1394–1406. 10.1016/j.neuron.2014.04.045 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Stroud JP, Porter MA, Hennequin G, Vogels TP. Motor primitives in space and time via targeted gain modulation in cortical networks. Nat Neurosci. 2018;21: 1774–1783. 10.1038/s41593-018-0276-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Zenke F, Gerstner W, Ganguli S. The temporal paradox of Hebbian learning and homeostatic plasticity. Current Opinion in Neurobiology. 2017. pp. 166–176. 10.1016/j.conb.2017.03.015 [DOI] [PubMed] [Google Scholar]
  • 26.Vogelsy TP, Froemkey RC, Doyon N, Gilson M, Haas JS, Liu R, et al. Inhibitory synaptic plasticity: Spike timing-dependence and putative network function. Frontiers in Neural Circuits. 2013. 10.3389/fncir.2013.00119 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Hennequin G, Agnes EJ, Vogels TP. Inhibitory Plasticity: Balance, Control, and Codependence. Annu Rev Neurosci. 2017;40: 557–579. 10.1146/annurev-neuro-072116-031005 [DOI] [PubMed] [Google Scholar]
  • 28.Vogels TP, Sprekeler H, Zenke F, Clopath C, Gerstner W. Inhibitory Plasticity Balances Excitation and Inhibition in Sensory Pathways and Memory Networks. Science (80-). 2011;334: 1569–1573. 10.1126/science.1211095 [DOI] [PubMed] [Google Scholar]
  • 29.Gerstner W, Kistler WM. Mathematical formulations of Hebbian learning. Biol Cybern. 2002;87: 404–415. 10.1007/s00422-002-0353-y [DOI] [PubMed] [Google Scholar]
  • 30.Gerstner W, Kistler WM. Mathematical formulations of Hebbian learning. Biol Cybern. 2002;87: 404–415. 10.1007/s00422-002-0353-y [DOI] [PubMed] [Google Scholar]
  • 31.Hosoya T, Baccus SA, Meister M. Dynamic predictive coding by the retina. Nature. 2005;436: 71 Available: 10.1038/nature03689 [DOI] [PubMed] [Google Scholar]
  • 32.Gerstner W, Kistler WM. Mathematical formulations of Hebbian learning. Biol Cybern. 2002;87: 404–415. 10.1007/s00422-002-0353-y [DOI] [PubMed] [Google Scholar]
  • 33.Slotine JJE. Modular stability tools for distributed computation and control. Int J Adapt Control Signal Process. 2003;17: 397–416. 10.1002/acs.754 [DOI] [Google Scholar]
  • 34.Kandel ER, Schwartz JH, Jessell TM, Jessell D of B and MBT, Siegelbaum S, Hudspeth AJ. Principles of neural science. McGraw-hill New York; 2000.
  • 35.Song S, Sjöström PJ, Reigl M, Nelson S, Chklovskii DB. Highly nonrandom features of synaptic connectivity in local cortical circuits. PLoS Biol. 2005;3: 0507–0519. 10.1371/journal.pbio.0030068 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Mariño J, Schummers J, Lyon DC, Schwabe L, Beck O, Wiesing P, et al. Invariant computations in local cortical networks with balanced excitation and inhibition. Nat Neurosci. 2005;8: 194–201. 10.1038/nn1391 [DOI] [PubMed] [Google Scholar]
  • 37.Wehr M, Zador AM. Balanced inhibition underlies tuning and sharpens spike timing in auditory cortex. Nature. 2003;426: 442–446. 10.1038/nature02116 [DOI] [PubMed] [Google Scholar]
  • 38.Shu Y, Hasenstaub A, McCormick DA. Turning on and off recurrent balanced cortical activity. Nature. 2003;423: 288–293. 10.1038/nature01616 [DOI] [PubMed] [Google Scholar]
  • 39.Slotine J-JE, Li W. Applied nonlinear control. Prentice hall; Englewood Cliffs, NJ; 1991. [Google Scholar]
  • 40.Murphy BK, Miller KD. Balanced Amplification: A New Mechanism of Selective Amplification of Neural Activity Patterns. Neuron. 2009;61: 635–648. 10.1016/j.neuron.2009.02.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Jaeger H. The “echo state” approach to analysing and training recurrent neural networks-with an erratum note. Bonn, Ger Ger Natl Res Cent Inf Technol GMD Tech Rep. 2001;148: 13. [Google Scholar]
  • 42.Pascanu R, Jaeger H. A Neurodynamical Model for Working Memory. www.reservoir-computing.org/organic [DOI] [PubMed]
  • 43.Buonomano DV, Maass W. State-dependent computations: spatiotemporal processing in cortical networks. 2009. [cited 11 Mar 2019]. 10.1038/nrn2558 [DOI] [PubMed] [Google Scholar]
  • 44.Sussillo D. Neural circuits as computational dynamical systems. Curr Opin Neurobiol. 2014;25: 156–163. 10.1016/j.conb.2014.01.008 [DOI] [PubMed] [Google Scholar]
  • 45.Van Steveninck RRDR, Lewen GD, Strong SP, Koberle R, Bialek W. Reproducibility and Variability in Neural Spike Trains. 1997;275. [DOI] [PubMed] [Google Scholar]
  • 46.Chatham CH, Badre D. Multiple gates on working memory. Curr Opin Behav Sci. 2015;1: 23–31. 10.1016/j.cobeha.2014.08.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Halassa MM, Kastner S. Thalamic functions in distributed cognitive control. Nat Neurosci. 2017;20: 1669–1679. 10.1038/s41593-017-0020-1 [DOI] [PubMed] [Google Scholar]
  • 48.Miller EK, Cohen JD. An Integrative Theory of Prefrontal Cortex Function. Annu Rev Neurosci. 2001;24: 167–202. 10.1146/annurev.neuro.24.1.167 [DOI] [PubMed] [Google Scholar]
  • 49.Vogelsy TP, Froemkey RC, Doyon N, Gilson M, Haas JS, Liu R, et al. Inhibitory synaptic plasticity: Spike timing-dependence and putative network function. Front Neural Circuits. 2013;7: 1–11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Lisman J. A mechanism for the Hebb and the anti-Hebb processes underlying learning and memory. Proc Natl Acad Sci. 1989;86: 9574–9578. 10.1073/pnas.86.23.9574 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Kullmann DM, Lamsa KP. Long-term synaptic plasticity in hippocampal interneurons. Nat Rev Neurosci. 2007;8: 687–699. 10.1038/nrn2207 [DOI] [PubMed] [Google Scholar]
  • 52.Enikolopov AG, Abbott L, Sawtell NB. Internally Generated Predictions Enhance Neural and Behavioral Detection of Sensory Stimuli in an Electric Fish. 2018. [cited 1 Mar 2019]. 10.1016/j.neuron.2018.06.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Ruan H, Saur T, Yao W-D. Dopamine-enabled anti-Hebbian timing-dependent plasticity in prefrontal circuitry. Front Neural Circuits. 2014;8: 38 10.3389/fncir.2014.00038 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Földiák P. Forming sparse representations by local anti-Hebbian learning. Biol Cybern. 1990;64: 165–170. 10.1007/BF02331346 [DOI] [PubMed] [Google Scholar]
  • 55.El-Boustani S, Ip JPK, Breton-Provencher V, Knott GW, Okuno H, Bito H, et al. Locally coordinated synaptic plasticity of visual cortex neurons in vivo. Science (80-). 2018;360: 1349–1354. 10.1126/science.aao0862 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Denève S, Machens CK. Efficient codes and balanced networks. Nat Neurosci. 2016;19: 375–382. 10.1038/nn.4243 [DOI] [PubMed] [Google Scholar]
  • 57.Hennequin G, Vogels TP, Gerstner W. Optimal control of transient dynamics in balanced networks supports generation of complex movements. Neuron. 2014;82: 1394–1406. 10.1016/j.neuron.2014.04.045 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Brunel N. Dynamics of Sparsely Connected Networks of Excitatory and Inhibitory Spiking Neurons. J Comput Neurosci. 2000. Available: https://web.stanford.edu/group/brainsinsilicon/documents/BrunelSparselyConnectedNets.pdf [DOI] [PubMed] [Google Scholar]
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007659.r001

Decision Letter 0

Lyle J Graham, Adrian M Haith

21 Apr 2020

Dear Dr. Miller,

Thank you very much for submitting your manuscript "Achieving stable dynamics in neural circuits" for consideration at PLOS Computational Biology.

As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by two independent reviewers. We apologize for the lengthy review process for this manuscript which was due to one reviewer being unable to complete their review.

The reviewers appreciated the novel perspective on an important problem, but did raise a number of concerns and felt that several parts of the paper required further clarification. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.

We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.

Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Adrian M Haith

Associate Editor

PLOS Computational Biology

Lyle Graham

Deputy Editor

PLOS Computational Biology

***********************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: The authors have some interesting results that would be useful for theorists/modelers in computational neuroscience to be aware of. The usefulness would be enhanced if some extra work is done to clarify the findings and state the limits on their generality or whether they could be expanded. I point out the areas where this can be achieved below:

l.57 “all such trajectories converge”: this suggests the focus is on networks with a single fixed point. Are these really the interesting ones? Is there some generalization to networks with multiple fixed points where one can discuss regions of contraction? Such networks may be more versatile and useful.

Eq. 1: The description “in the absence of input from other neurons” suggests that the diagonal part of the connection matrix, W_ij is contained within the first term, h(x). Is this true? Or are you only considering W_ij without diagonal self-connection terms? If not, then “other” neurons should be modified to include self-feedback.

Eq. 1: It is important to indicate the range of “x” as it appears to enter the expression as if it is a firing rate which can only be non-negative. Are you allowing negative neural outputs?

Eq.2: Following the above comment, if “x” represents firing rate so can only be non-negative, then the equation for changes in connection strengths prevent any stable positive connections from arising (the initial term is negative or zero and the second term has opposite sign to W). This seems wrong. Similarly if x can be negative some discussion should be made about the meaning of negative rates and how the change in synaptic strength should depend on such rates. Perhaps zero “x” corresponds to a threshold for the rate-dependent switch between synaptic depression and potentiation and the formulae would all work with shifted “x” and threshold subtracted, to represent non-negative firing rates?

l. 130: I think the logic is correct but the explanation a little too brief:

According to Hebb’s rule if a neuron helps cause another neuron to fire then a change in connection enhances the likelihood of such a causal event occurring again. i.e. Hebb’s rule is all about positive feedback (the synapse changes to enhance existing correlations).

For excitatory neurons that means strengthening following correlated spikes.

For an inhibitory neuron, it can help a downstream neuron fire by not spiking at the same time.

Also, a decrease in the inhibitory synapse will help the postsynaptic neuron fire.

So, firing at different times (anticorrelated) should lead to a decrease in inhibitory strength (Wij more positive) as it is then more likely that the inhibitory cell helps the excitatory cell fire -- or conversely, firing at the same time leads to an increase in inhibitory strength (Wij more negative as stated here). Perhaps a short-hand could be that a “positive feedback” process is Hebbian and a negative feedback process is anti-Hebbian? However, I am still confused by the values of “x” and their sign being perhaps negative. It may help to describe this as negative x meaning low-rate and positive x meaning high-rate. But then this rule would allow for strong plasticity when both pre- and post-synaptic rates are minimal (x very negative) in contradiction to known plasticity mechanisms. These issues should be discussed.

Figure 3: Please define “Norm of Matrix Component” via an equation

l. 146 typo (“of”)

l. 173-174 The calculation is incorrect. You are comparing connections per neuron with total connections in the brain. The numerator should be multiplied by number of neurons (10^11).

l. 197: It seems that there are two separate (but linked) systems operating on very different time scales. On the fast time scale, for a given weight matrix, one can ask if the neural firing rates of the system are contracting. At times (especially when Echo State networks are compared or the requirements for weight matrices) it seems like the goal is to generate contracting networks in terms of neural firing rates. However, l. 197 suggests that the discussions of contraction are in the components of the weight matrix. The authors should clearly separate discussions and results concerning weight matrices that contract (such that the network produces a reliable/reproducible – perhaps in this case temporally variable – set of synaptic strengths) versus whether the firing rates of neurons follow relatively stable trajectories.

Eq. 4: The terms here need a little more description and explanation of how they enter. While the main point is trivial (decreasing p can lead to stability) the terms like “g_max” appear to be related to the maximum derivative of connection strength with respect to neural activation (Eq. 4.0.3) rather than what is generally termed “gain” of a neuron (gradient of f-I curve). Also, does w_max refer to maximum strength of plastic synapses (in which case why is it not in the “alpha” term) or is it simply the maximum value of the assigned fixed synapses? Perhaps explain why increasing the fraction of static to plastic synapses allows for a larger w_max.

l. 211: “has” not “have”

l. 218 “impact of such synapses cancel” but there are time-differences in practice that commonly lead to oscillations. I think “cancel” is an overstatement.

l.224-225 “will not interfere with stability”: in fact the antisymmetric part can enhance stability (this is easy to see from the determinant of a 2x2 matrix). So, while the intent is correct, the point could be made a little more clearly. That is, if the symmetric part is stable the complete matrix will be, however, if the symmetric part is not stable the antisymmetric part could be sufficient to generate stability overall.

l. 232 type “that” should be “the”

l. 232 “the off-diagonal elements cancel out”. No! They impact the determinant of the matrix and hence can impact the values (and signs) of the eigenvalues, therefore impacting stability. But again, I think only in the sense of enhancing stability.

Fig. 4: legend is pretty opaque and stops half-way through a sentence in my version.

Section 3.4 on Echo State networks: this seems a little strange and overly expounded, especially when discussing the discrete-time nature of Echo State networks when there are almost identical Liquid State networks that gained prominence at the same time (sometimes in joint reviews) with the major difference being the Liquid State networks operate in continuous time. Why focus on Echo State not Liquid State given the latter is more similar to the current paper?

Discussion: I think papers that train some chaotic networks, for example Laje and Buonomano 2012, should be mentioned at a minimum and, even better, discussed in terms of contracting trajectories given that is what they appear to do. How is their plasticity rule or network similar or dissimilar to any mentioned in this paper that lead to contraction? Or is there system not contracting? Or only contracting in a portion of phase space? It would be nice to see this sort of analysis (even if only descriptive) applied to a few prior networks that appear to achieve contraction

l. 346 “increases” not “increase”

Reviewer #2: Overview: The authors analyze the mechanisms that neural circuits employ in order to maintain stable dynamics in face of perturbations and plasticity through the lens of a control theoretic perspective, namely, contraction analysis. The study indicates that mechanisms such as inhibitory Hebbian plasticity, excitatory anti-Hebbian plasticity, synaptic sparsity and E-I balance can all lead to stability within the considered networks.

Comments:

• Contribution: The authors point out that stability of neural circuits has been considered in the literature widely, but what is less explored is how despite trial to trial variability, the brain manages to produce robust computations and therefore robust responses. The method considered here i.e., contraction analysis, states that the trajectories produced by population activity converge quickly following perturbation and therefore are robust in the face of small perturbations.

The paper brings together certain established analytical methods (contraction theory) to treat a long standing problem in computational neuroscience. While some of the factors discussed (inhibitory plasticity, E-I balance) have been studied in the context of achieving stability, they are brought together here under a common analytical framework. In this sense, the paper could constitute a useful contribution to the literature. However, the technical novelty is at time ambiguous and the presentation could be more precise in this regard. I elaborate on specific concerns below:

• Main concerns.

o The authors frequently refer to computations in the Appendix section to narrate their findings. Important quantifications (not the extensive mathematical derivations) should be introduced in line with the claims made in the Result section to make the narrative more cohesive. For instance, in line 80 the authors introduce a matrix measure called the ‘logarithmic norm of the Jacobian’ which happens to be an important metric for the discussions that follow, but no concise mathematical form of it is provided to aid the readers.

o The authors have identified two quantitative tools to characterize contraction: contraction rate and the Jacobian of the network. However, in line 85 the authors have expressed one in terms of the other. Does that mean that the quantifications are interchangeable, and the analysis can be interpreted in terms of any one? Some explicit clarifications regarding this might be helpful.

o In line 86, the authors state that the idea of contraction analysis using Jacobian can be generalized to different metrics. What are some of these metrics? What are the metrics that can generate pure convergence of trajectories? What is the identity metric in line 138? A discussion along these lines including the mathematical expression of these metrics will enrich the study.

o The types of perturbations considered for the study are not mentioned explicitly. Does the nature of convergence change depending upon the perturbation inflicted on the network?

o In line 211 the authors state that if alpha_i = 0, then W_max must be reduced to ensure equation 4 i.e., contraction in the network. My question here is in this case do you consider a densely connected network? Since, equation 4 states that if you are sparse enough you can have strong couplings irrespective of whether the synapses are static or dynamic.

o How should Figure 4 in the manuscript be interpreted? I understand that it points to the fact that E-I balance in networks (often employed by achieving E-I balance in subnetworks) facilitate stability and therefore convergence of trajectories under perturbation. But much of the preceding and succeeding discussion assumes prior knowledge of the matter. Introducing some key equations back into the text here will help justify the claims made.

o Are there any neural mechanisms that have been studied experimentally to ensure stability and robustness in computation but cannot be explained solely through contraction analysis?

• Minor Comments:

o Figure 1: Dimensionality of the cartoon state space. The text states that it is a (N+S) dimensional space for a network with N neural units and S dynamic synaptic weights, but the figure does not match up with it or perhaps need to be restructured. What is the index j here?

o Figure 2: Legends are not visible. Consider using larger font size.

o Line 157: There is no red trace in Figure 3A. Is it a typographical error or a missing trace?

o Caption for Figure 4 is incomplete.

o The Materials and Methods section does not add much in this version of the manuscript except where to find the relevant code. Consider expanding this section to include information that are vital to interpretation of the results.

o What is H in the supplementary section (Pg 13). Is it simply H = x_t x_t^T?

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: Yes

Reviewer #2: No: Simulations are specified, but data is not submitted.

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, PLOS recommends that you deposit laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions, please see http://journals.plos.org/compbiol/s/submission-guidelines#loc-materials-and-methods

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007659.r003

Decision Letter 1

Lyle J Graham, Adrian M Haith

27 Jun 2020

Dear Dr. Miller,

We are pleased to inform you that your manuscript 'Achieving stable dynamics in neural circuits' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Adrian M Haith

Associate Editor

PLOS Computational Biology

Lyle Graham

Deputy Editor

PLOS Computational Biology

***********************************************************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: All prior criticisms have been addressed satisfactorily and the paper is clear enough now.

Please just carefully check the grammar before sending it for any publication as there are some typos/non-English forms in the added text (e.g. l.120 "at rate called the contraction rate" should begin "at a rate"). I came across a few of these sorts of issues.

Reviewer #2: The authors have addressed the points raised in my original review. I thank them for their responses and am satisfied with the edits made to the manuscript.

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: Yes

Reviewer #2: None

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Paul Miller

Reviewer #2: No

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007659.r004

Acceptance letter

Lyle J Graham, Adrian M Haith

30 Jul 2020

PCOMPBIOL-D-20-00042R1

Achieving stable dynamics in neural circuits

Dear Dr Miller,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Laura Mallard

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Text. The supplementary appendix file contains extensive mathematical proofs of the results stated above.

    We kept the appendix self-contained by restating the basic results of contraction analysis and linear algebra which we used often in our proofs.

    (PDF)

    Attachment

    Submitted filename: StableDynamicsPLOSReviewResponses3.docx

    Data Availability Statement

    All detailed proofs of main results are found in the appendix. Simulations (Figs 2 and 3) were performed in Python. Code to reproduce the figures is available at [https://github.com/kozleo/stable_dynamics]. Numerical integration was performed using sdeint, an open-source collection of numerical algorithms for performing integrations of stochastic ordinary differential equations.


    Articles from PLoS Computational Biology are provided here courtesy of PLOS

    RESOURCES