Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2021 Apr 30;17(4):e1008673. doi: 10.1371/journal.pcbi.1008673

Efficient and robust coding in heterogeneous recurrent networks

Fleur Zeldenrust 1,*, Boris Gutkin 2,3, Sophie Denéve 2
Editor: Daniel Bush4
PMCID: PMC8115785  PMID: 33930016

Abstract

Cortical networks show a large heterogeneity of neuronal properties. However, traditional coding models have focused on homogeneous populations of excitatory and inhibitory neurons. Here, we analytically derive a class of recurrent networks of spiking neurons that close to optimally track a continuously varying input online, based on two assumptions: 1) every spike is decoded linearly and 2) the network aims to reduce the mean-squared error between the input and the estimate. From this we derive a class of predictive coding networks, that unifies encoding and decoding and in which we can investigate the difference between homogeneous networks and heterogeneous networks, in which each neurons represents different features and has different spike-generating properties. We find that in this framework, ‘type 1’ and ‘type 2’ neurons arise naturally and networks consisting of a heterogeneous population of different neuron types are both more efficient and more robust against correlated noise. We make two experimental predictions: 1) we predict that integrators show strong correlations with other integrators and resonators are correlated with resonators, whereas the correlations are much weaker between neurons with different coding properties and 2) that ‘type 2’ neurons are more coherent with the overall network activity than ‘type 1’ neurons.

Author summary

Neurons in the brain show a large diversity of properties, yet traditionally neural network models have often used homogeneous populations of neurons. In this study, we investigate the effect of including heterogoenous neural populations on the capacity of networks to represent an input stimulus. We use a predictive coding framework, by deriving a class of recurrent filter networks of spiking neurons that close to optimally track a continuously varying input online. We show that if every neuron represents a different filter, these networks can represent the input stimulus more efficiently than if every neuron represents the same filter.

Introduction

It is widely accepted that neurons do not form a homogeneous population, but that there is large variability between neurons. For instance, the intrinsic biophysical properties of neurons, such as the densities and properties of ionic channels, vary from neuron to neuron [15]. Therefore, the way in which individual neurons respond to a stimulus (their ‘encoding’ properties or receptive field) also varies. A classical example is the difference between ‘type 1’ and ‘type 2’ neurons [610]: ‘type 1’ neurons or ‘integrators’ respond with a low firing frequency to constant stimuli, which they increase with the amplitude of the stimulus. ‘Type 2’ neurons or ‘resonators’ on the other hand respond with an almost fixed firing frequency, and are sensitive to stimuli in a limited frequency band. Apart from these intrinsic ‘encoding’ properties of neurons, the ‘decoding’ propertes of neurons also show a large variability (see for instance [11]). Even if we do not take various forms of (short term) plasticity into account, there is a large heterogeneity in the shapes of post-synaptic potentials (PSPs) that converge onto a single neuron (for an overview: [12, 13]), depending on amongst others: the projection site (soma/dendrite: [14]), the number of receptors at the synapse, postsynaptic cell membrane properties ([1517]), the type of neurotransmitter (GABAA, GABAB, glutamate), synapse properties (channel subunits, [18]), the local chloride reversal potential and active properties of dendrites [12]. This heterogeneity results in variability of decay times, amplitudes and overall shapes of PSPs. So neural heterogeneity plays an important role both in encoding and in decoding stimuli. Whereas the study of homogeneous networks has provided us with invaluable insights [1922], the effects of neural heterogeneity on neural coding have only been studied to a limited extent [2325]. Here we show that networks with spiking neurons with heterogeneous encoding and decoding properties can do optimal online stimulus representation. In this framework, neural variability is not a problem that needs to be solved, but it increases the networks’ versatility of coding.

In order to investigate coding properties of neurons and networks, we need to use a framework in which we can assess the encoding and decoding properties and the network properteis. To characterize the relationship between neural stimuli and responses, filter networks (such as the Linear-Nonlinear Poisson (LNP) model [26], [27] and the Generalized Linear Model (GLM) [28, 29], for an overview, see [30], [31]) are widely used. In these models, each neuron in a network compares the input it receives with an ‘input filter’. If the two are similar enough, a spike is fired. In a GLM, unlike the LNP model, this output spike train is filtered and fed back to the neuron, thereby incorporating effectively both the neuron’s receptive field and history-dependent effects such as the refractory period and spike-frequency adaptation. It can be shown, that these types of filter-frameworks describe a maximum-likelihood relation between the input and the output spikes [32], [33]. However, these models are purely descriptive: they only describe how spike trains are generated, not how they should be read out or ‘decoded’. In this paper, we analytically derive a class of recurrent networks of spiking neurons that close to optimally track a continuously varying input online. We start with two very simple assumptions: 1) every spike is decoded linearly and 2) the network aims to perform optimal stimulus representation (i.e. reduces the mean-squared error between the input and the estimate). From this we derive a class of predictive coding networks, that unifies encoding (how a network represents its input in its output spike train) and decoding (how the input can be reconstructed from the output spike train) properties.

We investigate the difference between homogeneous networks, in which all neurons represent similar features in the input and have similar response properties, and heterogeneous networks, in which neurons represent different features and have different spike-generating properties. Firstly, we assess the properties of single-neuron (spike-triggered average, input-output frequency curve and phase-response curve) and show that in this framework, ‘type 1’ or ‘integrator’ [6, 34] neurons and resonators or ‘type 2’ neurons arise quite naturally. We show that the response properties of these neurons (the encoding properties, [35]) and dynamics of the PSPs they send (the decoding properties) are inherently linked, thereby giving a functional interpretation to these classical neuron types. Next, we investigate the effects of these different types of neurons on the network behaviour: we investigate the coding efficiency, robustness and trial-to-trial variability in in-vivo-like simulations. Finally, we predict that 1) integrators show strong correlations with other integrators and resonators are correlated with resonators, whereas the correlations are much weaker between neurons with different coding properties [36] and 2) that ‘type 2’ neurons are more coherent with the overall network activity than ‘type 1’ neurons.

Materials and methods

We analytically derive a recurrent network of spiking neurons that close to optimally tracks a continuously varying input online. We start with two assumptions: 1) every spike is decoded linearly and 2) the network aims to perform optimal stimulus representation (i.e. reduces the mean-squared error between the input and the estimate). We construct a cost function that consists of three terms: 1) the mean-squared error between the stimulus and the estimate, 2) a linear cost that punishes high firing rates and 3) a quadratic cost that promotes distributed firing. Every spike that is fired in the network reduces this cost function. The code for simulating such a network can be found at GitHub: github.com/fleurzeldenrust/Efficient-coding-in-a-spiking-predictive-coding-network.

Derivation of a filter-network that performs stimulus estimation

Suppose we have a set of N neurons j that use filters gj(t) to represent their input. Their spiking will give an estimated input equal to

s^(t)=j=1Ngj(t)*ρjideal(t)=j=1Ni=1njgj(t-tji), (1)

where tji are the spike times i of neuron j (what ‘ideal’ stands for will be explained later). The mean-squared error between the estimate s^ and the stimulus s equals

E(t)=0t(s(t)-s^(t))2dt=0t(s(t)-j=1Ni=1njgj(t-tji))2dt. (2)

Suppose that at time t = T + Δ we want to estimate the difference in error given that there was a spike of neuron m at time T or not. The difference in error is given by

EnospikeatT(T+Δ)-EspikeatT(T+Δ)=0T+Δ(s(t)-j=1Ni=1njgj(t-tji))2dt-0T+Δ(s(t)-j=1Ni=1njgj(t-tji)-gm(t-T))2dt=-0T+Δgm(t-T)2dt+20T+Δgm(t-T)(s(t)-j=1Ni=1njgj(t-tji))dt. (3)

We introduce a greedy spike rule: a spike will only be placed at time T if this reduces the mean squared error at T + Δ (see Fig 1B), so if the expression in Eq (3) is positive. This results in a spike rule:

0T+Δgm(t-T)(s(t)-j=1Ni=1njgj(t-tji))dt>120T+Δgm(t-T)2dt. (4)

Obviously, if the decision is made at time t = T + Δ, a spike cannot be placed in the past (t = T). Therefore, we introduce two spike trains: the real spike train ρ(t), where a spike will be placed at time T + Δ with the spike rule above (Eq (4)). This represents the ‘ideal’ spike train ρideal(t), which is ρ(t) shifted by an amount of Δ, so that a spike at time T + Δ in ρ(t) is equavalent to a spike at time T in spike train ρideal(t). This means that the neurons have to keep track of their own spiking history for at least Δ time and that any prediction of the input is delayed by an amount of Δ (see Fig 1A, j=1Ngj(t)*ρj(t) is an estimate of the input, delayed Δ in time).

Fig 1. Overview of the network model.

Fig 1

A The signal (black), ideal estimate s^ (thin red line), delayed estimate (thick red line), ideal spike trains ρideal (open dots) and real spike trains ρ (filled dots). B Zoom in of the yellow square in A. The decision whether the blue spike (denoted with a question mark) in ρideal will be fired, is made at time t + Δ. The spike is fired, because the estimate with the spike (solid red line) has a smaller error than the estimate without the spike (dashed red line). C Decoding filters g used in this network: neurons 1 and 2 (blue and green) have positive filters, neurons 3 and 4 (pink and yellow) have negative filters.D Each neuron k filters the input s(t) using an input filter gkin. It receives input from other neurons l via lateral filters gkllat and sends input to other neurons l via lateral filters glklat. Finally, the network activity can be decoded linearly using decoding filters gk.

Eq (4) defines a filter-network (Fig 1D). Each neuron m keeps track of a ‘membrane potential’ V

Vm(T+Δ)=0T+Δgm(t-T)s(t)dt-0T+Δgm(t-T)j=1Ni=1njgj(t-tji)dt, (5)

that it compares to a threshold Θm

Θm=120T+Δgm(t-T)2dt+ν=12-TΔgm(x)2dx. (6)

Note that the threshold is not evaluated over the whole filter, but only between t = −T and t = Δ. Only if Δ is larger than the causal part of the filter and the proposed time of the spike T larger than the acausal part, the whole filter is taken into account, and the threshold reduces to Θm=12 for (L2-norm) normalized filters.

The network defined by Eq (4) can spike at any arbitrary frequency, and neurons show no spike-frequency adaptation. Two neurons that have identical filters except for their sign, have lateral filters gmout(0)=2(Θm). Depending on the shape of the filter and the value of Δ, the maximum value of these lateral filters can exceed twice the threshold of the postsynaptic neuron, so a spike in a neuron can induce a spike in a neuron with a filter that is identical except for their sign. This ‘ping-pong effect’, which makes the estimate fluctuate very quicky around the ideal value, can be dampened by introducting a spike cost. So to force the network to choose solutions with realistic firing rates for each neuron, we introduce two additional terms to the threshold:

Θmc(t)=12-TΔgm(x)2dx+ν+μgthreshold(tm-T), (7)

where ν is a spike cost that punishes high firing rates in the network, which makes the code more sparse. μgthreshold(t) is a spike cost kernel that punishes high firing rates in a single neuron. The effect is equivalent to spike-frequency adaptation [37]: every time a neuron fires a spike, it threshold is increased with μ, and decays (with filter gthreshold) back to its original dynamic value. Compared to just adding a constant spike cost (as in citeBoerlin2011,Boerlin2013, Bourdoukan2012, Deneve2017), adding the temporal spike cost makes the network activity more distributed between the neurons: a neuron can fire a few spikes, but by doing that will increase its threshold, and another neuron will take over. In this paper we use an exponential kernel with time constant of 60 ms unless stated differently.

The first term on the right of Eq (5) shows that neuron m is convolving the input s(t) with an input filter gmin(t) that is a flipped and shifted version of the filter neuron m represents:

0T+Δgm(t-T)s(t)dt=0T+Δgmin(T+Δ-t)s(t)dt=(gmin*s)(T+Δ),

where

gmin(t)=gm(Δ-t). (8)

Note that by introducing the time delay Δ between the evaluation time and the spike time, the input filter is now shifted relative to the representing filter. The representing filter can contain a causal part, that consists of the systems estimation on how the stimulus will behave in the near future of the spike, i.e. the systems estimation of the input auto-correlation. Any acausal part of the input filter is not used (since we do not know the future if the input s(t)), and vanishes as long as g(t) = 0 for t > Δ. The optimal value of Δ depends on how much of a prediction the neuron wants to make into the future, but also on how long it is willing to wait with its response.

The second term of Eq (5) denotes the lateral (jm) and output (j = m) filters, that are substracted from the filtered input as a result of the spiking activity of any of the neurons. For well-behaved filters we can write

0T+Δgm(t-T)j=1Ni=1njgj(t-tji)dt=j=1N0T+Δgm(t-T)(gj*ρjideal)(t)dt=j=1N(gmin*gj*ρjideal)(T+Δ), (9)

so that we find an output filter

gmout(t)=-(gmin*gm)(t) (10)

and lateral filters

gmjlat(t)=-(gmin*gj)(t), (11)

where t denotes the time since the spike of neuron j. Note that a spike at T can only influence the decision process at T + Δ, so the part of the filter where 0t-tji<tji+Δ does not influence the spiking process, and is therefore not used.

In summary, we defined network that can perform near-optimal stimulus estimation (Fig 2). Given a set of readout filters gi(t), the membrane potential of each neuron i is defined as

Vi(t)=gi(t-Δ)*(s(t)-j=1Ngj(t)*ρj(t-Δ)). (12)

A spike is fired if this reduces the MSE of the estimate, which is equivalent to when the membrane potential exceeds the threshold Θmc (Eq (7)). This model has two main characteristics: the input, output and lateral filters are defined by the representing filter, and the representation of the input by the output spike train is delayed by an amount Δ in order for the network to use predictions for the stimulus after the spike.

Fig 2. Example network response.

Fig 2

Example of the response of a homogeneous (red) and a heterogeneous (blue) network to a constant input (first column), to a low frequency input (5 Hz, second colum), to a high frequency input (80 Hz, third column) and to a recorded LFP-signal (right). The homogeneous network consists of 50 neurons with a positive representing filter g and 50 neurons with a negative one (see section ‘Derivation of a filter-network that performs stimulus estimation); the heterogeneous network consists of 100 neurons with each a different (but normalized) representing filter g. Network parameters: Δ = 7, 5 ms, ν = μ = 0.5. No noise is added to this network.

We conclude that a classical filter network can perform near-optimal stimulus encoding given a certain relation between the input, output and lateral filters. In the classical framework of LNP and GLM models, acausal filters g are used, and the readout of the model is only used for information theoretic purposes. However, if the readout is being done by a next layer of neurons, or in recurrent networks, one cannot use acausal filters: every spike of a presynaptic neuron m at time tmi, can only influence the membrane potential at its target at t>tmi (i.e. can only cause a post-synaptic potential after the spike). Therefore, in a layered and recurrent networks, a self-consistent code will use only representative filters that are causal

g(t)=0ift<0

This means that for the input filters

gin=0ift>Δ

Note that if Δ = 0, the input filter reduces to a single value at t = 0. If Δ = 0 and g(0) = 0, the input filter and hence the output and lateral filters vanish. Note that the threshold is scaled to the part of g between 0 and Δ, so not to the full integral over g. In this paper, we will normalize the input filters so that the thresholds are the same for all neurons.

Analysis

Generation of neuron filters

In this paper, we choose each neuron filter on the basis of 8 basis functions given by the following Γ-functions:

Γn(t)=tne-t (13)
  • Type 1 neuron The representing filter of a ‘type 1’ neuron (Fig 3A and 3B, blue) is chosen equal to a Γ-function (Eq (13)) with n = 3.

  • Type 2 neuron The representing filter of a ‘type 2’ neuron (Fig 3A and 3B, red) is chosen as gtype 2(t) = Γ3(t)(0.2 − 0.8sin(0.6t)).

  • Off cells We call a neuron with an inverted representing filter an ‘off cell’. So a ‘type 1 off cell’ has representing filter gtype 1,off(t) = −Γ3(t) and a ‘type 2 off cell’ has representing filter gtype 2,off(t) = −Γ3(t)(0.2 − 0.8sin(0.6t)).

  • Homogeneous network The representing filters of a ‘type 1’-network are equal to the ‘type 1’ neuron (half of the neurons) or minus the ‘type 1’ neuron (other half of the neurons).

  • Type 1 & type 2 network A quarter of the representing filters of the neurons in the network are equal to the ‘type 1’ neuron, a quarter to minus the ‘type 1’ neuron, a quarter to the ‘type 2’ neuron and a quarter to minus the ‘type 2’ neuron.

  • Heterogenous network The representing filters in the heterogenous network are products of Γ functions and oscillations. This is in order to ensure that they have realistic properties (i.e. vanishing after tens of miliseconds), while still being able to represent different frequencies. Half of the neurons have a representing filter equal to gn(t) = Γ3(t)(0.2±0.8sin(ψt)), with ψ a random number between 0 and 1.5 and half of these using an addition and half a subtraction. The representing filters of the other half of the population are given by gn(t) = Γ3(t)(0.2±0.8cos(ψt)).

Fig 3. Example of a ‘type 1’ (blue) and a ‘type 2’ (red) neuron.

Fig 3

A Representing filter of a ‘type 1’ (blue) and a ‘type 2’ (red) neuron. B Self-filters for both neuron types. C Instantaneous frequency of the first two spikes in response to step-and-hold inputs of different amplitudes. D Average frequency in response to step-and-hold inputs of different amplitudes. E Phase-response curves for the first and second pairs of spikes, calculated as response to a small pulse (0.1 ms, amplitude 1.5 (‘type 2’) and 3.7 (‘type 2’)) on top of a constant input (amplitude 0.9 (‘type 1’) and 2.2 (‘type 2’)). F Spike-triggered average in response to a white-noise stimulus (amplitude: 0.45 (‘type 1’) and 1.1 (‘type 2’)) filtered with an exponential filter with a time constant of 1 ms.

Spike coincidence factor

The coincidence factor Γ [38, 39] describes how similar two spike trains s1(t) and s2(t) are: it reaches the value 1 for identical spike trains, vanishes for Poissonian spike trains and negative values hint at anti-correlations. It is based on the binning of the spike train in K=Tp bins of binwidth p. The coincidence factor is corrected for the expected amount of coincidences 〈Ncoinc〉 of spike train s1 with a Poissonian spike-train with the same rate ν2 as spike train s2. It gives a measure of 1 for identical spike trains, 0 if all coincidences are accidental and negative values for anti-correlated spike trains. It is defined as

Γ12=Ncoinc-Ncoinc12(N1+N2)1N (14)

in which

Ncoinc=2ν2pN1=2N1N2K

Finally, Γ is normalized by

N=1-2ν2p=1-N2K

so it is bounded by 1. Note that the coincidence factor is not symmetric nor positive, therefore it is not a metric. It is only defined as long as each bin contains at most one event, however, we counted the bins with double spikes as bins containing one spike. Finally, it will in general saturate at a value below one, which can be seen as the spike reliability. The rate as which it reaches this value (for instance as defined by a fit to an exponential function) can be seen as the precision. In Section Heterogeneous networks are more efficient than homogeneous networks we calculated the coincidence factor between the spike-train response to each stimulus presentation for each neuron in the network, and averaged this over neurons to obtain a Γ¯, a measure for the trial-to-trial variability or the degeneracy of the code of the network (Fig 4, left).

Fig 4. eterogeneous networks are more efficient and show less trial-to-trial variability.

Fig 4

Results of two simulations using the same stimulus, but different initial network states, in a homogeneous ‘type 1’ network (first row), a network with ‘type 1’ and ‘type 2’ neurons (second row) and a heterogeneous network (third row). Each data point was simulated ten times and averaged. The network network activity A (first column), spike reliability Γ¯ (second column), network MSE¯ (third column) and efficiency E (fourth column) are shown as a function of the time constant (τ) and amplitude of the stimulus. All networks perform well over a wide range of stimulus amplitudes and frequencies, but the heterogeneous network is more efficient for a wider range of input parameters, in particular for fast-fluctuating low amplitue signals. The bottom row shows the same data presented differently: every data point is the average over 10 trials of a single value of the input stimulus amplitude and τ. This shows how the network activity, error and efficiency depend on the spike reliability Γ¯. Parameters: Δ = 7, 5 ms, ν = μ = 1, 5, N = 100, #trials = 10. The white star denotes the parameter values used in section Heterogeneous networks are more efficient than homogeneous networks.

Mean-squared error, network activity and effciency

The Mean-Squared Error (MSE) is for N measurements in time is defined by

MSE(s,s^)=n=1N(sn-s^n)2

However, this typically increases with the stimulus amplitude and length. To assess the performace of networks independently of stimulus amplitude, we normalized the mean-squared error by deviding it by the mean-squared error between the stimulus and an estimate of a constant zero signal (or equivalently, the MSE between the stimulus and network estimate if the network would be quiescent, MSEno spikes):

MSE¯=MSEMSEnospikes. (15)

An MSE¯ close to zero means a good performance, whereas a value close to one means performance that is comparable to a network that doesn’t show any activity. Given that the goal of this network is to give an approximation of the input signal with the lowest number of spikes as possible, we define the network activity A (in Hz) as

A=#spikesNneuronTsim, (16)

where N is the number of neurons, and T the duration of the stimulus, and the efficiency E (in seconds) of the network as

E=1MSE¯·A. (17)

Alternative measures of efficiency were also calculated (see Figs A and B in Section C of the supplementary S1 Text), but did not significantly alter our conclusions.

Results

In this section, we will discuss the properties of the network derived in section Derivation of a filter-network that performs stimulus estimation. We will start with the general network behaviour, and show that it can track several inputs. Next, we will show that this framework provides a functional interpretation of ‘type 1’ and ‘type 2’ neurons. In the following sections, we will zoom in on the relation between trial-to-trial variability and the degeneracy of the code used, and on the network’s robustness to noise. Finally, we will make experimental predictions based on the network properties.

Network response

In Fig 2, the response of two different networks is shown: a homogeneous network, consisting of 50 neurons with a positive representing filter g (see section Analysis) and 50 neurons with a negative one and a heterogeneous network (bottom), consisting of 100 neurons with each a different (but normalized, based on Γ-fuctions) representing filter g. Both networks can track both constant and fluctuating inputs with different frequencies well. Note that even though there is no noise in the network, the network response is quite irregular, like in in-vivo recordings. Note also that the heterogeneous network is better at tracking fast fluctuations. How well the different types of networks respond to different types of input, what the response properties of the networks are, and what the influence of the type of filter g is, will be investigated in the following sections.

‘Type 1’ and ‘type 2’ neurons

If we create representing filters randomly (see section Analysis), they generally fall into one of two types: unimodal ones (only postitive or only negative) or bimodal ones (both a positive and a negative part). In this section, we will investigate the difference between neurons using these two types of representing filters.

In Fig 3 the response of a single neuron with a unimodal (blue) or multimodal (red) representing filter (both are normalized with respect to the input filter gin, so between t = 0 ms and t = Δ = 7, 5 ms) to different types of input is shown (see section Analysis). Note that these simulations are for single neurons, so there is no network present, like in in-vitro patch-clamp experiments. A neuron with a unimodal representing filter (Fig 3A and 3B, blue) shows a continuous input-frequency curve (Fig 3C and 3D). Such a neuron with a unimodal representing filter has a unimodal Phase Response Curve (PRC) and Spike-Triggered Average (STA) (Fig 3E and 3F). A neuron with a multimodal representing filter (Fig 3A and 3B, red) does initially only respond with a single spike to the switching on of the step-and-hold current. Only for high current amplitudes it starts firing pairs of doublets, due to the interaction between the filtering properties and the spike-frequency adaptation. It has a bimodal PRC within the doublets (Fig 3E, solid red line), but a unimodal PRC between the doublets (Fig 3E, dashed red line). Such a neuron with a multimodal representing filter also has a bimodal Spike-Triggered Average (Fig 3F, red line). The input-frequency curves, PRC and STA together, show that neurons with unimodal representing filters show ‘type-1’ -like behaviour, whereas neurons with multimodal fitlers show ‘type 2’-like behaviour.

Homogeneous and heterogeneous networks

In the previous section, we showed that ‘type 1’ and ‘type 2’ neurons appear naturally in the predictive coding framework we defined in section Derivation of a filter-network that performs stimulus estimation. Even though the single-neuron response properties of ‘type 1’ and ‘type 2’ neurons have been studied extensively, most simulated network consist of a single or a few homogeneous populations of leaky integrate-and-fire (‘type 1’) neurons. Here, we will investigate the effect of heterogeneity in the response properties of single neurons on the network properties and dynamics. We will compare a homogeneous network consisting of ‘type 1’-neurons, an intermediate network consisting of ‘type 1’ and ‘type 2’-neurons, and a heterogeneous network (see section Analysis).

Heterogeneous networks are more efficient than homogeneous networks

The trial-to-trial variability of network responses depends critically on both the network structure and on the input stimuli used. For instance, it has been shown that (sub)cortical responses to stimuli with naturalistic statistics are more reliable than responses to other stimuli [4046]. This suggests that, depending on the cortical area and the input statistics, neural networks can use codes that are highly degenerate or non-degenerate. For clarification, we define here a degenerate code as a code in which the stimulus can be represented with a low error by several different population responses. Therefore, a degenerate code will show a high trial-to-trial variability, or a low spike reliability. We hypothesize that a network consisting of neurons that represent similar features of the common input signal (i.e. several neurons have the same representing filter g) will use more degenerate codes than networks consisting of neurons that represent different features of the input signal (i.e. every neuron has a different representing filter g). So we hypothesize that homogeneous networks will show a higher trial-to-trial variability (i.e. a lower spike reliability). Here, we investigate the relation between trial-to-trial variability, network performance, input statistics and the network heterogeneity. We do this by simulating the response of three different networks with increasing levels of heterogeneity (a network consisting only of identical ‘type 1’ neurons, a mixed network consisting of ‘type 1’ and ‘type 2’ neurons and a heterogeneous network in which each neuron is different, see section Analysis) to input stimuli with different statistical properties (varying the amplitude and the autocorrelation time constant τ).

In Fig 4, we simulated three networks (see also section Analysis): a homogeneous network (first row), a mixed network (second row, consisting for 50% of ‘type 1’ neurons (positive and negative filters) and for 50% of ‘type 2’ neurons (positive and negative filters) and a heterogeneous network (third row). We varied both the amplitude and the time constant of the input signal (by filtering the input forwards and backwards with an exponential filter). To determine the level of degeneracy of the code the network uses, we performed the following simulations: we computed the network response to the same stimulus (T = 2500 ms) twice, but before this stimulus started, we gave the network a 500 ms random start-stimulus. Note that there was no noise in the network except for the different start-stimuli. We calculated four network performance measures:

  • Network Activity A (Hz) We assessed the total network activity (Fig 4, first column), as the average firing frequency per neuron (see section Analysis).

  • Spike Reliability Γ The coincidence factor Γ [38, 39] (see section Analysis) describes how similar two spike trains are: it reaches a value of 1 for identical spike trains, vanishes for Poisson spike trains and negative values hint at anti-correlations. We calculated the coincidence factor between the spike-train response to each stimulus presentation for each neuron in the network, and averaged this over neurons to obtain Γ¯, a measure for the trial-to-trial variability of the network (Fig 4, second column). If the network uses a highly degenerate code, the starting stimulus will put it in a different state just before the start of the stimulus used for comparison, and the trial-to-trial variability will be high (low Γ¯). On the other hand, if the network uses a non-degenerate code, the starting stimulus will have no effect, and the trial-to-trial variability will be low (high Γ¯). Therefore, Γ¯ (Fig 4, left) represents the non-degeneracy of the code: a Γ¯ close to zero corresponds to a high degeneracy and a high trial-to-trial variability, and a Γ¯ close to one a low degeneracy and a low trial-to-trial variability.

  • Error MSE¯ To assess the performace of the network, we calculated the normalized mean-squared error (MSE¯ (Fig 4, third column), see section Analysis), so that a value of MSE¯ close to zero means a good network performance, and a value close to one means performance that is comparable to a network that doesn’t show any activity.

  • Efficiency E (s) Given that the goal of this network is to give an approximation of the input signal with the lowest number of spikes as possible, we define the network efficiency (in seconds) as the inverse of the product of firing rate and the error (see section Analysis), so that the efficiency decreases with both the network activity and the error (MSE¯ (Fig 4, fourth column).

In Fig 4, it is shown that all three networks (homogeneous, type 1 & type 2 and heterogeneous, respective rows) perform well (small error, second column) over a wide range of stimulus amplitudes and frequencies. The performance of the network depends strongly on the heterogeneity of the network and the characteristics of the stimulus: heterogeneous networks show a smaller MSE¯ (second column), especially for fast-fluctuating (small τ) input. However, this comes at the cost of a higher activity A (first column), in particular at larger stimulus amplitudes. If we summarize this by the efficiency E (fourth column), we see that the heterogeneous network is more efficient, in particular for low amplitude and fast fluctuating stimuli. Alternative measures of efficiency were also calculated (see section C of the supplementary S1 Text), but did not significantly alter our conclusions.

The amplitude of the stimulus relative to the amplitudes of the neural filters and the amount of neurons is important. This can be understood using a simplified argument: to represent a high-amplitude stimulus requires all the neurons to fire at the same time, which makes the code highly non-degenerate: Since the filter amplitudes are between 1 and 2, a stimulus with an amplitude (standard deviation) of 20 would need an equivalent number of the N = 100 neurons (the ones with a positive filter) to spike at the same time to reach the amplitude of the peaks. This results in a reliable, non-degenerate code (high Γ¯). Alternatively, when a low-amplitude stimulus is represented by many neurons with relatively high-amplitude filters, the network can ‘choose’ which neuron to use for representing the input, thereby making the code degenerate (low Γ¯). This argument does not take into account the temporal aspect of the representing filters (i.e. that each spike contributes an estimate into the future), the fact that representing filters can have positive and negative parts and the implemented spike-frequency adaptation, but the argument does show that the stimulus amplitude needs to be considerably lower than the sum of the absolute value of all the representing filters in the network.

Note that none of the networks is very good at responding to very low stimulus amplitudes: this is because most fluctuations are smaller than the filter amplitudes. This is reflected in the spike reliability Γ¯ (third column). In a homogeneous network consisting of identical neurons representing a low amplitude input, there is no difference between a spike of neuron A or one of neuron B, making the code highly degenerate (low Γ¯). When the amplitude of the stimulus increases, more neurons are recruited to represent the stimulus, thereby increasing both the network activity A and the spike reliability Γ¯. When the stimulus amplitude becomes too high to represent properly, both the error MSE¯ and Γ¯ increase sharply (bottom left), and the efficiency E decreases (bottom right). Therefore, in our framework, a high trial-to-trial variability, or a low spike reliability Γ¯ is a hallmark of an efficiently coding network, and a strong decrease of the trial-to-trial variability (an increase in Γ¯) is a sign of a network starting to fail to track the stimulus. This happens for lower amplitudes for the heterogeneous network than for the homogeneous network.

In conclusion, all three networks can track stimuli with a wide range of parameters well, but the heterogeneous network performs shows a lower error for an only small increase in activity, therefore the heterogeneous network is more efficient than the homogeneous network. However, the homogeneous network can track higher amplitude stimuli, in particular slowly fluctuating ones. The efficiency of the code is high for intermediate values for the trial-to-trial variability: a very low trial-to-trial veriability (high Γ¯) corresponds with a low efficiency, but so does a very high trial-to-trial variability (Γ¯ close to 0). So in our framework, an high to intermediate trial-to-trial variability is a hallmark of an efficiently coding network.

Heterogeneous networks are more robust against correlated noise than homogeneous networks

In the previous section it was shown that heterogeneous networks are more efficient than homogeneous networks in encoding a wide variety of stimuli. However, these networks did not contain any noise. It has been shown that cortical networks receive quite noisy input, which is believed to be correlated between neurons [4749]. Therefore, we will test in this section how robust homogeneous and heterogeneous networks are against correlated noise.

To test for robustness against noise, we chose a stimulus that all networks responded well to: amplitude = 10, τ = 15 ms (see the white start in Fig 4). Noise was added in a ‘worse case scenario’: it had the same temporal properties as the stimulus (i.e. it was white noise filtered with the same filter with τ = 15 ms) and several neurons received the same noise. Two parameters were varied: the amplitude of the noise, and the correlations between the noise signals that different neurons receive. This was implemented as follows: next to the stimulus, every neuron received a noise signal and we varied the relative amplitude of the noise signal (anoise/asignal). Correlations between the noise signals for different neurons were simulated by only making a limited amount of noise copies, that were distributed among the neurons. So in Fig 5, the top horizontal row of each axis corresponds to the situation in which each neuron receives an independent noise signal, the bottom horizontal row of each axis represents the situation where all neurons receive the same noise signal (and hence there is from the network’s point of view no difference between signal and noise) and the leftmost column of each axis corresponds to the situation without noise.

Fig 5. Heterogeneous networks are more robust against noise.

Fig 5

Network activity A (first column), network error MSE¯ (second column) and efficiency E (third column) of a homogeneous network (top row), a network with ‘type 1’ and ‘type 2’ neurons (middle row) and a heterogeneous network (bottom row). Next to the stimulus, each neuron in the network was presented with a noise input, with a varying relative amplitude (horizontal axis) and number of copies of the noise signal (vertical axis; 1 copy means all neurons receive the same noise, 100 copies means all neurons receive independent noise). Network: Δ = 7, 5 ms, ν = μ = 1, 5, N = 100. Stimulus: amplitude = 10, τ = 15 ms. Noise: τ = 15 ms.

In Fig 5, it is shown that all three networks are very effective in compensating for independent noise by increasing their firing rate. Note that the stimulus and the noise add quadratically, so that a signal with amplitude 10 and noise with amplitude 12 (relative amplitude 1.2) add together to a total input of amplitude 102+122=15.5 for each neuron. However, the firing rate of a network with signal amplitude = 10 and noise amplitude = 12, is much higher than the firing rate of a network that receives a signal with amplitude 15.5 and no noise (compare Figs 4 and 5). So the networks increase their activity both due to the increased amplitude of the input, and in order to compensate for noise. The bottom row of each subplot in Fig 5 corresponds to a simulation where all neurons in the network recieve the same noise signal. In this case, it is impossible for the network to distinguish between signal and noise. In the row above, only two noise copies are present, in the row above that five, and so on. The homogeneneous network can handle higher amplitudes of independent noise (top part of each subplot) before the representation breaks down (MSE¯>.2), but all networks are able to compensate for independent noise amplitudes up to equal to the signal amplitude (signal-to-noise ratio = 1). The heterogeneous network however, is better at dealing with correlated noise (bottom part of each subplot): it shows a lower error and higher efficiency for when there are few copies of the noise signal (alternative measures of efficiency were also calculated (see Fig A in the supplementary S1 Text), but did not significantly alter our conclusions). The type 1’& type 2 network appears to combine the properties of both networks: it has a low error in both representing independent noise and at representing correlated noise. These differences in robustness against noise probably have a relation with ambiguities in degenerate codes, as will be discussed in the Conclusion and discussion.

Predictions for experimental measurements

In the previous sections, we derived a predictive coding framework and assessed the efficiency and robustness of representing a stimulus of homogeneous and heterogeneous networks. In this section, we will simulate often-used experimental paradigms, to assess what can be expected from such measurements with respect to the effect of the predictive coding framework on correlation structures.

Signal and noise correlations

The predictive coding framework that we derived in section Derivation of a filter-network that performs stimulus estimation predicts a specific correlation structure: neurons with similar filters g should show positive signal correlations (because they use similar input filters), but negative noise (also termed spike count) correlations (because they have negative lateral connections). In many experimental papers, noise and signal correlations between neurons are measured [50, 51] (for an overview, see [52]). However, the methods authors use vary strongly: the amount of repetitions of the stimulus varies from tens to hundreds, the windows over which correlations are summed vary from tens of miliseconds to hundreds of miliseconds and the strength of the stimulus (i.e. network response) varies from a few to tens of Hz. All these parameters strongly influence the conclusions one can draw about correlations between (cortical) neurons. In order to be able to compare the correlations this framework predicts with experiments, we performed the following simulation (based on [51], Fig 6): we chose a 20 s. stimulus (exponentially filtered noise, τ = 15 ms), and showed a network of N = 100 neurons 300 repetitions. We chose the stimulus amplitude so that the network response was around 8 Hz. In order to be able to compare neurons with similar filters and neurons with different filters, we used the ‘type 1 & type 2’ network (see section Analysis). To assess the effects of shared noise, each neuron received a noise signal that was the sum of an independent noise signal, and a noise signal that was shared between 10 neurons (amplitude signal = 2.7, amplitude independent noise = 0.5, amplitude correlated noise = 0.5). In this simulation, we can compare neurons that share a noise source, and neurons that don’t.

Fig 6. Signal and noise correlations in a network with ‘type 1’ and ‘type 2’ neurons.

Fig 6

(see section Analysis and Fig 3). Next to the stimulus, each neuron in the network was presented with a noise input (of which half the power was independent, and half was shared with a subset of other neurons). A) Signal correlations between ‘type 1’ on-cells (solid blue line) and between a ‘type 1’ on and off cell (dashed blue line). B) Noise correlations between ‘type 1’ on-cells (solid blue line) and between a ‘type 1’ on and off cell (dashed blue line) receiving independent noise. C) Noise correlations between ‘type 1’ on-cells (solid blue line) and between a ‘type 1’ on and off cell (dashed blue line) receiving shared noise. D) Signal correlations between ‘type 2’ on-cells (solid red line) and between a ‘type 2’ on and off cell (dashed red line). E) Noise correlations between ‘type 2’ on-cells (solid blue line) and between a ‘type 2’ on and off cell (dashed red line) receiving independent noise. F) Noise correlations between ‘type 2’ on-cells (solid red line) and between a ‘type 2’ on and off cell (dashed red line) receiving shared noise. D) Signal correlations between a ‘type 1’ and a ‘type 2’ on-cell (solid purple line) and between a ‘type 1’ on and a ‘type 2’ off cell (dashed purple line). E) Noise correlations between a ‘type 1’ on-cell and a ‘type 2’ on-cell (solid purple line) and between a ‘type 1’ on and a ‘type 2’ off cell (dashed purple line) receiving independent noise. F) Noise correlations between a ‘type 1’ on-cell and a ‘type 2’ on-cell (solid purple line) and between a ‘type 1’ on and a ‘type 2’ off cell (dashed purple line) receiving shared noise. Network parameters: Δ = 7, 5 ms, ν = μ = 1, 5, N = 100, Ntrial = 300. Stimulus: τ = 15 ms, amplitude = 2.7. Noise: τ = 15 ms, amplitude independent noise = 0.5, amplitude shared noise = 0.5.

In Fig 6, the signal and noise correlations between ‘type 1’ and ‘type 2’ and ‘on’ and ‘off’ cells (see section Analysis) and Fig 3) are shown. We ran a simulation consisting of 300 trials, in which the same signal, but a different noise realization was used. We used the same method as [49] (Note that the spike-count correlation is proportional to the area under the noise cross-correlogram in this method):

  • For the signal correlations (Fig 6, first column), we calculated the average spike train over the 300 trials and calculated the cross-correllogram, normalized to the total average number of spikes.

  • For the noise correlations (Fig 6, second column), we subtracted the cross-correlogram of the averaged spike trains (see above) from the cross-correlograms averaged over all trials.

  • For the noise correlations, shared noise condition (Fig 6, third column), we simply calculated the noise correlations as above for two neurons that shared a noise source.

We performed this correlation analysis for two ‘type 1’ neurons (Fig 6, top row, blue), for two ‘type 2’ neurons (Fig 6, middle row, red) and for a ‘type 1’ and a ‘type 2’ neuron (Fig 6, bottom row, purple). We performed the correlation analysis also between two neurons with the same representing filter (solid line) and with a neuron and a neuron with an inverted representing filters (‘off-cells’, dashed lines).

We start by looking at cells with similar filtering properties. In Fig 6A, we show that ‘type 1’ neurons show positive signal correlations with other ‘type 1’ neurons (solid blue line), and negative signal correlations with their off-cells (dashed blue line). As expected, ‘type 1’-neurons show negative noise correlations with other ‘type 1’-neurons (Fig 6B, solid blue line), and positive noise correlations with their off-cells as long as noise is independent (dashed blue line). When neurons share a noise source (Fig 6C), the noise correlations are positive, but show a small negative deflection around zero lag for on cells (solid blue line). For ‘type 2’ neurons, we can show similar conclusions: two on-cells have positive signal correlations (Fig 6D, solid red line), negative noise correlations for independent noise ((Fig 6E, solid red line), and positive noise correlations with a negative peak around zero for shared noise ((Fig 6F, solid red line). An on and an off cell show negative signal correlations (Fig 6D, dashed red line), but hardly any noise correlations for independent noise (Fig 6E, dashed red line) or shared noise (Fig 6F, dashed red line).

We now focus at cells with different filtering properties. ‘Type 1’ and ‘type 2’ on cells show positive signal correlations (Fig 6G), solid purple line). Note that the peak is shifted towards positive lags, meaning that the ‘type 2’ neurons spike earlier in time than ‘type 1’ neurons. ‘Type 1’ and ‘type 2’ on cells show only small noise correlations for independent noise (Fig 6H, solid purple line). When noise is shared (Fig 6I, solid purple line), a small deflection at zero lag can be seen.

Under experimental conditions, correlations are often summed over a window of about 10 miliseconds or more. Therefore, we conclude that even though the predictive coding framework predicts negative noise correlations between similarly tuned neurons, and positive noise correlations between oppositly tuned neurons, these would be very hard to observe experimentally. For shared noise, we expect to see noise correlations with a similar sign as the signal correlations, but with small deflections at small lags, as shown in Fig 6C, 6F and 6I.

‘Type 2’ neurons show more coherence with network activity

To examine how the two different neuron types couple to the network activity in response to a temporally fluctuating stimulus, we study the spike coherence with a simulated Local Field Potential (LFP). In Fig 7, we analyse the network activity of the different types of neurons, using the ‘type 1 & type 2’-network (see section Analysis and Fig 3). Next to the stimulus, each neuron in the network was presented with a noise input (of which half the power was independent, and half was shared with a subset of other neurons). Network: Δ = 7, 5 ms, ν = μ = 1, 5, N = 100. Stimulus: τ = 15 ms, amplitude = 2.7. Noise: τ = 15 ms, amplitude independent noise = 0.5, amplitude shared noise = 0.5. For the network activity, each spike was convolved with a Gaussian kernel (σ = 6 ms). Looking at the network activity (Fig 7A and 7B), it is clear that ‘type 2’ neurons (red line) respond in a much more synchronized manner (high peaks), and ‘type 1’ neurons (blue line) respond in a much more continuous way. We quantified this, by convolving each spike with a Gaussian kernel (σ = 6 ms), and calculating both the correlations between the network activity and that of a single neuron (Fig 7C), and by calculating the ‘spike-triggered network activity’ [53] (Fig 7D). From this it is clear, that ‘type 2’ neurons show a much stronger coupling to the overall network activity (making them ‘chorists’) than ‘type 1’ neurons (making them ‘soloists’). So we predict that ‘type 2’ neurons are more coherent with the network activity/LFP than ‘type 1’ neurons.

Fig 7. Network activity in a ‘type 1 & type 2’ network.

Fig 7

(see section Analysis and Fig 3). A) Average activity of the whole network (black), ‘type 1 on cells’ (blue line) and ‘type 2 on cells’ (red line). B) Spike response of the network. C) Average cross-correllogram between the network activity and ‘type 1, on’ cells (blue line) or ‘type 2, on’ cells (red line). D) Spike-triggered network activity for ‘type 1, on’ cells (blue line) or ‘type 2, on’ cells (red line). Next to the stimulus, each neuron in the network was presented with a noise input (of which half the power was independent, and half was shared with a subset of other neurons). Network: Δ = 7, 5 ms, ν = μ = 1, 5, N = 100. Stimulus: τ = 15 ms, amplitude = 2.7. Noise: τ = 15 ms, amplitude independent noise = 0.5, amplitude shared noise = 0.5. For the network activity, each spike was convolved with a Gaussian kernel (σ = 6 ms).

Conclusion and discussion

Biological data often show a strong heterogeneity: neural properties vary considerably from neuron to neuron, even in neurons from the same network [15], but also see [54]. Theoretical networks, however, often use a single or only a very limited amount of ‘cell types’, where neurons from the same cell type have the same response properties. In order to investigate the effects of network heterogeneity on neural coding, we derive a filter network that efficiently represents its input from first principles. We start with the decoding instead of with the encoding (this is not common, but has been done before [55]), and formulate a spike rule in which a neuron only fires a spike if this reduces the mean-squared error between the received input and a prediction of the input based on the output spike trains of the network, implementing a form of Lewicki’s ‘matching pursuit’ [56]. Linear decoding requires recurrent connectivity, as neurons representing different features in the input should inhibit one another to alllow linear decoding, something that has been shown in experiments [57]. A similar framework has been formulated in homogeneous networks with integrate-and-fire neurons, the so-called ‘Spike-Coding Networks’(SCNs) [5861] and in networks using conductance-based models [62]. Effectively, this network performs a form of coordinate transformation [63]: each neuron represents a particular feature of the input, and only by combining these features the complete stimulus can be reconstructed. This network is related to autoencoders, in that it finds a sparse distributed representation of a stimulus by using an over-complete set of basis functions in the form of a feed-forward neural-network. The homogeneous integrate-and-fire networks in this framework have been shown to operate in a tightly balanced excitatory-inhibitory regime, where a large trial-to-trial variability coexists with a maximally efficient code [64]. Because of their linear read-out and filtering, only linear computations can be performed in this framework, which leaves open the questions whether the results presented here also hold for non-linear network computations [65, 66]. Recently, the SCN framework has been extended to include non-linear computations [67]. It is an interesting question to see whether the conclusions remain the same for non-linear computations.

With the derived filter network, we are able to study both single neuron and network properties. On the single neuron level, we find that the single-neuron response properties are equivalent to those of ‘type 1’ and ‘type 2’ neurons (for an overview see [8, 9, 68]): Neurons using unimodal representing filters showed the same behaviour as ‘type 1’ cells (continuous input-frequency curve [6], unimodal Phase-Response Curve (PRC) [7, 69] and a unimodal Spike-Triggered Average (STA) [70]), whereas neurons using bimodal filters correspond to ‘type 2’ cells (discontinuous input-frequency curve, bimodal PRC and a bimodal STA). This should be the case, as the STA is a result of the filtering properties of the neuron, and proportional to the derivative of the PRC [71]. In this framework, neurons with bimodal representing filters will also send bimodal Post-Synaptic Potentials (PSPs) to other neurons, so excitatory post-synaptic potentials (EPSPs) with an undershoot or inhibitory post-synaptic potentials (IPSPs) with a depolarizing part. This might sound counterintuitive, because we often think of EPSPs as having a purely depolarizing effect on the membrane potential of the postsynaptic neuron, and of IPSPs as having a purely hyperpolarizing effect. However, a post-synaptic potential might have both excitatory and inhibitory parts, depending on the type of synapse and the ion channels present in the membrane. For instance, an undershoot after an EPSP can be observed as an effect of slow potassium channels [17] such as IM [16], IA [15] or IAHP [72]. IPSPs can have direct depolarizing effects when the inhibition is shunting [73], [74], or due to for instance deinactivation of sodium channels, or slow activation of other depolarizing channels such as Ih. The observation that neurons using unimodal representing filters show ‘type 1’ behaviour and neurons using bimodal representing filters show ‘type 1’ behaviour gives a functional interpretation of these classical neuron types: ‘type 1’ cells are more efficient at representing slowly fluctuating inputs, whereas ‘type 2’ cells are made for representing transients and fast-fluctuating input. This can also be observed in the network activity: ‘type 2’ neurons show a much stronger coupling to the overall network activity (making them ‘chorists’ [53]) than ‘type 1’ neurons (making them ‘soloists’). This is expected, as ‘type 1’ neurons are generally harder to entrain [69], whereas ‘type 2’ neurons generally show resonant properties [75]. So we predict that ‘type 2’ neurons are more coherent with the network activity (local field potential) than ‘type 1’ neurons.

We compare both the functional coding properties and the activity of networks with different degrees of heterogeneity. We found that all networks in this framework can respond efficiently and robustly to a large variety of inputs (varying amplitude and fluctuation speed) corrupted with noise with different properties (ampitude, fluctuation speed, correlation between neurons). All networks show a high trial-to-trial variability, that decreases with the network efficiency. So we confirmed that in our framework trial-to-trial variability is not necessarily a result of noise, but is actually a hallmark of efficient coding [7678]. In-vivo recordings typically show strong trial-to-trial variability between spike trains from the same neuron, and spike trains from individual neurons are quite irregular, appearing as if a Poisson process has generated them. In in-vitro recordings on the contrary, neurons show very regular responses to injected input current, especially if this current is fluctuating ([79, 80]. It is often argued that this is due to noise in the system and therefore that the relevant decoding parameter should be the firing rate over a certain time window (as opposed the timing of individual spikes, [81], but see also [82]). Here, we show in noiseless in-vivo-like simulations that the generated spike trains are irregular and show large trial-to-trial variability, even though the precise timing of each spike matters: shifting spike times decreases the performance of the network. Therefore, this model shows how the intuitively contradictory properties of trial-to-trial variability and coding with precise spike times can be combined in a single framework. Trial-to- trial variability is here a sign of degeneracy in the code: the relation between the network size, filter size and homogeneity of the network versus the amplitude determines whether there is strong or almost no trial-to-trial variability. Moreover, we show that the trial-to-trial variability and the coding efficiency depend on the frequency content of the input, as has been shown in several systems [40, 43, 45, 79].

In our framework, heterogeneous networks are more efficient than homogeneous networks, especially in representing fast-fluctuating stimuli: heterogeneous networks represent the input with a smaller error and using fewer spikes, in line with earlier research that found that heterogeneity increases the computational power of a network [23, 24], especially if they match the stimulus statistics [83]. Moreover, heterogeneous networks are not only more efficient, they are also more robust against correlated noise (noise that is shared between neurons) than homogeneous networks, in line with previous results [25, 84, 85]. This might be the result of heterogeneous networks using a less degenerate code: these networks are better at whitening the noise signal, because each neuron projects the noise onto a different filter, thereby effectively decorrelating the noise. Put differently, the heterogeneous network projects the signal and noise into a higher-dimensional space [63], thereby more effectivily projecting noise and signal into different dimensions. Homogeneous networks are better at compensating for independent noise. This is probably due to a combination of two factors: 1) a homogeneous network being better at compensating for erronous spikes (a ‘mistake’ in the estimate is easier to compensate if the same but negative filter exist than if this doesn’t exist) and 2) a homogeneous network is better at representing high-amplitude signals. The optimal mix of neuron types probably depends on the stimulus statistics and remains a topic for further study. The effects of neural diversity on computational properties of networks remains an open question (for a review, see [86]). It has for instance been shown that neural diversity can improve network information transmission [87] and results in efficient and robust encoding of stimuli by a population [4]. Recently, it has been shown that other forms of heterogeneity also result in advantages for network computation: diversity in synaptic or membrane time constants increases stability and robustness in learning [88] and diversity in network connectivity enables networks to track their input more rapidly [89], improves network stability and can lead to better performance [90]. So diversity in network properties has shown to result in computational advantages on several levels and for different functions.

The predictive coding framework predicts a specific correlation structure between neurons: negative noise correlations between neurons with similar tuning, and positive noise correlations between neurons with opposite tuning. This may appear to be contradictory to earlier results [91, 92]. However, as neurons with similar tuning most likely recieve inputs from common sources in previous layers, these neurons will also share noise sources, resulting in correlated noise between neurons. We showed that in this situation, the negative noise correlations are only visible as small deflections of effectively positive correlations. It has been argued that in order to code efficiently and effectively, recurrent connectivity should depend on the statistical structure of the input to the network [93] and that noise correlations should be approximately proportional to the product of the derivatives of the tuning curves [94], although these authors also concluded that these correlations are difficult to measure experimentally. Others suggest that neurons might be actively decorrelated to overcome shared noise [95]. We conclude that even though different (optimal) coding frameworks make predictions about correlation structures between neurons, the opposite is not true: a single correlation structure can correspondd to different coding frameworks. So measuring correlations is not sufficient (although informative) to determine what coding framework is used by a network [35, 96, 97].

Supporting information

S1 Text

Section A: A word on notation. Section B: Technical notes. Section C: Efficiency measures, including Supplemental Figures A and B. Fig A: Comparison of efficiency measures Results of two simulations using the same stimulus, but different initial network states, in a homogeneous ‘type 1’ network (first row), a network with ‘type 1’ and ‘type 2’ neurons (second row) and a heterogeneous network (third row). Three alternative efficiency measures are compared: 1) the network efficiency multiplied by the stimulus amplitude (Ea, see supp. eq. (10), first column), 2) the network efficiency multiplied by the stimulus power (Ea2, see supp/ eq. (11), second column) and 3) the network cost (C, see supp. eq. (12), third column). The bottom row shows how the network efficiency or cost depends on the spike reliability Γ¯. Parameters: Δ = 7, 5 ms, ν = μ = 1, 5, N = 100, #trials = 10. The white star denotes the parameter values used in section ‘Heterogeneous networks are more efficient than homogeneous networks’ of the main text. Fig B: Comparison of efficiency measures for the noise simulations Efficiency normalized by amplitude Ea (first column), amplitude squared Ea2 (second column) and cost (third column) of a homogeneous network (top row), a network with ‘type 1’ and ‘type 2’ neurons (middle row) and a heterogeneous network (bottom row). Next to the stimulus, each neuron in the network was presented with a noise input, with a varying relative amplitude (horizontal axis) and number of copies of the noise signal (vertical axis; 1 copy means all neurons receive the same noise, 100 copies means all neurons receive independent noise). Network: Δ = 7, 5 ms, ν = μ = 1, 5, N = 100. Stimulus: amplitude = 10, τ = 15 ms. Noise: τ = 15 ms.

(PDF)

Data Availability

All code to generate the data can be found at https://github.com/fleurzeldenrust/Efficient-coding-in-a-spiking-predictive-coding-network.

Funding Statement

FZ acknowledges support from the Netherlands Organisation for Scientific Research (Nederlandse Organisatie voor Wetenschappelijk Onderzoek, NWO) Veni grant (863.150.25) and the Radboud University (Christine Mohrmann Foundation), SD acknowledges support from Neuropole Region Île de France (NERF) and ERC consolidator grant “predispike” and BSG acknowledges funding from the Basic Research Program at the National Research University Higher School of Economics (HSE University), ANR-17-EURE- 1553-0017, and ANR-10-IDEX-0001-02. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Destexhe A, Marder E. Plasticity in single neuron and circuit computations. Nature. 2004;431:789–795. 10.1038/nature03011 [DOI] [PubMed] [Google Scholar]
  • 2. Marder E, Taylor AL. Multiple models to capture the variability in biological neurons and networks. Nature neuroscience. 2011;14(2):133–8. 10.1038/nn.2735 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Padmanabhan K, Urban NN. Intrinsic biophysical diversity decorrelates neuronal firing while increasing information content. Nature neuroscience. 2010;13(10):1276–82. 10.1038/nn.2630 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Tripathy SJ, Padmanabhan K, Gerkin RC, Urban NN. Intermediate intrinsic diversity enhances neural population coding. Proceedings of the National Academy of Sciences. 2013;110(20):8248–53. 10.1073/pnas.1221214110 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Gouwens NW, Sorensen SA, Berg J, Lee C, Jarsky T, Ting J, et al. Classification of electrophysiological and morphological neuron types in the mouse visual cortex. Nature Neuroscience. 2019;. 10.1038/s41593-019-0417-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Hodgkin AL. The local electric changes associated with repetitive action in a non-medullated axon. The Journal of Physiology. 1948;107:165–181. 10.1113/jphysiol.1948.sp004260 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Hansel D, Mato G, Meunier C. Synchrony in excitatory neural networks. Neural computation. 1995;7(2):307–337. 10.1162/neco.1995.7.2.307 [DOI] [PubMed] [Google Scholar]
  • 8. Rinzel J, Ermentrout GB. Analysis of neural excitability and oscillations. In: Koch C, Segev I, editors. Methods in Neural Modeling: from synapses to networks. Cambridge, Massachusets: MIT Press; 1989. p. 251–292. [Google Scholar]
  • 9. Izhikevich EM. Neural Excitability, Spiking and Bursting. International Journal of Bifurcation and Chaos in Applied Sciences and Engineering. 2000;10(6):1171–1266. [Google Scholar]
  • 10. Prescott SA, De Koninck Y, Sejnowski TJ. Biophysical Basis for Three Distinct Dynamical Mechanisms of Action Potential Initiation. PLoS Computational Biology. 2008;4(10):e1000198. 10.1371/journal.pcbi.1000198 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Machens CK, Romo R, Brody CD. Functional, but not anatomical, separation of “what” and “when” in prefrontal cortex. The Journal of Neuroscience. 2010;30(1):350–360. 10.1523/JNEUROSCI.3276-09.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Hausser M. Diversity and Dynamics of Dendritic Signaling. Science. 2000;290(5492):739–744. 10.1126/science.290.5492.739 [DOI] [PubMed] [Google Scholar]
  • 13. Mody I, Pearce Ra. Diversity of inhibitory neurotransmission through GABA(A) receptors. Trends in neurosciences. 2004;27(9):569–75. 10.1016/j.tins.2004.07.002 [DOI] [PubMed] [Google Scholar]
  • 14. Miles R, Toth K, Gulyas AI, Hajos N, Freund TF. Differences between somatic and dendritic inhibition in the hippocampus. Neuron. 1996;16(4):815–823. 10.1016/S0896-6273(00)80101-4 [DOI] [PubMed] [Google Scholar]
  • 15. Clements BJD, Nelson PG, Redman SJ. Intracellular tetraethylammonium ions enhance group Ia excitatory post-synaptic potentials evoked in cat motoneurones. The Journal of Physiology. 1986;377:267–282. 10.1113/jphysiol.1986.sp016186 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Tosaka T, Tasaka J, Miyazaki T, Libet B. Hyperpolarization following activation of K+ channels by excitatory postsynaptic potentials. Nature. 1983;305(8):148–150. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Miles R, Wong RKS. Excitatory synaptic interactions between CA3 neurones in the guinea-pig hippocampus. The Journal of Physiology. 1986;373:397–418. 10.1113/jphysiol.1986.sp016055 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Hevers W, Lüddens H. The diversity of GABAA receptors. Pharmacological and electrophysiological properties of GABAA channel subtypes. Molecular neurobiology. 1998;18(1):35–86. [DOI] [PubMed] [Google Scholar]
  • 19. Amit DJ, Brunel N. Model of Global Spontaneous Activity and Local Structured Activity During Delay Periods in the Cerebral Cortex. Cerebral Cortex. 1997;7(3):237–252. 10.1093/cercor/7.3.237 [DOI] [PubMed] [Google Scholar]
  • 20. Brunel N. Dynamics of Sparsely Connected Networks of Excitatory and Inhibitory Spiking Neurons. Journal of Computational Neuroscience. 2000;8. [DOI] [PubMed] [Google Scholar]
  • 21. van Vreeswijk C, Sompolinsky H. Chaos in neuronal networks with balanced excitatory and inhibitory activity. Science. 1996;274:1724–1726. 10.1126/science.274.5293.1724 [DOI] [PubMed] [Google Scholar]
  • 22. van Vreeswijk C, Sompolinsky H. Chaotic Balanced State in a Model Of Cortical Circuits. Neural Computation. 1998;10:1321–1371. 10.1162/089976698300017214 [DOI] [PubMed] [Google Scholar]
  • 23. Hunsberger E, Scott M, Eliasmith C. The competing benefits of noise and heterogeneity in neural coding. Neural computation. 2014;26(8):1600–23. 10.1162/NECO_a_00621 [DOI] [PubMed] [Google Scholar]
  • 24. Duarte R, Morrison A. Leveraging heterogeneity for neural computation with fading memory in layer 2/3 cortical microcircuits. PLoS computational biology. 2019;15(4):e1006781. 10.1371/journal.pcbi.1006781 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Ecker AS, Berens P, Tolias AS, Bethge M. The effect of noise correlations in populations of diversely tuned neurons. The Journal of Neuroscience. 2011;31(40):14272–83. 10.1523/JNEUROSCI.2539-11.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Chichilnisky EJ. A simple white noise analysis of neuronal light. Network: Computation in Neural Systems. 2001;12:199–213. 10.1080/713663221 [DOI] [PubMed] [Google Scholar]
  • 27. Simoncelli EP, Paninski L, Pillow JW, Schwartz O. Characterization of Neural Responses with Stochastic Stimuli. In: Gazzaniga M, editor. The Cognitive Neurosciences. MIT Press; 2004. p. 1385. Available from: http://books.google.com/books?id=ffw6aBE-9ykC. [Google Scholar]
  • 28. Paninski L. Maximum likelihood estimation of cascade point-process neural encoding models. Network: Computation in Neural Systems. 2004;15(4):243–262. 10.1088/0954-898X_15_4_002 [DOI] [PubMed] [Google Scholar]
  • 29. Pillow JW, Shlens J, Paninski L, Sher A, Litke AM, Chichilnisky EJ, et al. Spatio-temporal correlations and visual signalling in a complete neuronal population. Nature. 2008;454(7207):995–9. 10.1038/nature07140 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Pillow JW. Likelihood-Based Approaches to Modeling the Neural Code. In: Doya K, Ishii I, Pouget A, Rao RPN, editors. Bayesian Brain: Probabilistic Approaches to Neural Coding. vol. 70. Cambridge, Massachusets: MIT Press; 2007. p. 53–70. [Google Scholar]
  • 31. Schwartz O, Pillow JW, Rust NC, Simoncelli EP. Spike-triggered neural characterization. Journal of Vision. 2006;6(4):484–507. [DOI] [PubMed] [Google Scholar]
  • 32. Paninski L, Pillow JW, Simoncelli EP. Maximum likelihood estimation of a stochastic integrate-and-fire neural encoding model. Neural Computation. 2004;16(12):2533–61. 10.1162/0899766042321797 [DOI] [PubMed] [Google Scholar]
  • 33. Brenner N, Bialek W, de Ruyter van Steveninck RR. Adaptive rescaling maximizes information transmission. Neuron. 2000;26(3):695–702. 10.1016/S0896-6273(00)81205-2 [DOI] [PubMed] [Google Scholar]
  • 34. Izhikevich EM. Resonate-and-fire neurons. Neural Networks. 2001;14(6-7):883–94. 10.1016/S0893-6080(01)00078-8 [DOI] [PubMed] [Google Scholar]
  • 35. Averbeck BB, Latham PE, Pouget A. Neural correlations, population coding and computation. Nature reviews Neuroscience. 2006;7(5):358–66. 10.1038/nrn1888 [DOI] [PubMed] [Google Scholar]
  • 36. Ko H, Cossell L, Baragli C, Antolik J, Clopath C, Hofer SB, et al. The emergence of functional microcircuits in visual cortex. Nature. 2013;496(7443):96–100. 10.1038/nature12015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Gutkin BS, Zeldenrust F. Spike frequency adaptation. Scholarpedia. 2014;9(2):30643. 10.4249/scholarpedia.30643 [DOI] [Google Scholar]
  • 38. Kistler WM, Gerstner W, van Hemmen JL. Reduction of the Hodgkin-Huxley Equations to a Single-Variable Threshold Model. Neural Computation. 1997;9:1015–1045. [Google Scholar]
  • 39. Jolivet R, Rauch A, Lüscher HR, Gerstner W. Predicting spike timing of neocortical pyramidal neurons by simple threshold models. Journal of Computational Neuroscience. 2006;21:35–49. [DOI] [PubMed] [Google Scholar]
  • 40. Yao H, Shi L, Han F, Gao H, Dan Y. Rapid learning in cortical coding of visual scenes. Nature neuroscience. 2007;10(6):772–8. 10.1038/nn1895 [DOI] [PubMed] [Google Scholar]
  • 41. Rikhye RV, Sur M. Spatial correlations in natural scenes modulate response reliability in mouse visual cortex. Journal of Neuroscience. 2015;35(43):14661–14680. 10.1523/JNEUROSCI.1660-15.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Hasson U, Malach R, Heeger DJ. Reliability of cortical activity during natural stimulation. Trends in Cognitive Sciences. 2010;14(1):40–48. 10.1016/j.tics.2009.10.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Rieke F, Bodnar Da, Bialek W. Naturalistic stimuli increase the rate and efficiency of information transmission by primary auditory afferents. Proceedings of the Royal Society of London Series B. 1995;262(1365):259–65. 10.1098/rspb.1995.0204 [DOI] [PubMed] [Google Scholar]
  • 44. Dan Y, Atick JJ, Reid RC. Efficient coding of natural scenes in the lateral geniculate nucleus: experimental test of a computational theory. The Journal of neuroscience. 1996;16(10):3351–62. 10.1523/JNEUROSCI.16-10-03351.1996 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. Butts DA, Weng C, Jin J, Yeh CII, Lesica Na, Alonso JMM, et al. Temporal precision in the neural code and the timescales of natural vision. Nature. 2007;449(7158):92–96. 10.1038/nature06105 [DOI] [PubMed] [Google Scholar]
  • 46. Desbordes G, Jin J, Alonso JM, Stanley GB. Modulation of temporal precision in thalamic population responses to natural visual stimuli. Frontiers in systems neuroscience. 2010;4(November):151. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Cohen MR, Newsome WT. Context-Dependent Changes in Functional Circuitry in Visual Area MT. Neuron. 2008;60(1):162–173. 10.1016/j.neuron.2008.08.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48. Zohary E, Shadlen MN, Newsome WT. Correlated neuronal discharge rate and its implications for psychophysical performance. Nature. 1994;370(6485):140–143. 10.1038/370140a0 [DOI] [PubMed] [Google Scholar]
  • 49. Bair W, Zohary E, Newsome WT. Correlated firing in macaque visual area MT: time scales and relationship to behavior. The Journal of Neuroscience. 2001;21(5):1676–1697. 10.1523/JNEUROSCI.21-05-01676.2001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. de la Rocha J, Doiron B, Shea-Brown ET, Josic K, Reyes AD. Correlation between neural spike trains increases with firing rate. Nature. 2007;448(7155):802–6. 10.1038/nature06028 [DOI] [PubMed] [Google Scholar]
  • 51. Tkačik G, Marre O, Amodei D, Schneidman E, Bialek W, Berry MJ. Searching for collective behavior in a large network of sensory neurons. PLoS computational biology. 2014;10(1):e1003408. 10.1371/journal.pcbi.1003408 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Cohen MR, Kohn A. Measuring and interpreting neuronal correlations. Nature neuroscience. 2011;14(7):811–9. 10.1038/nn.2842 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Okun M, Steinmetz Na, Cossell L, Iacaruso MF, Ko H, Barthó P, et al. Diverse coupling of neurons to populations in sensory cortex. Nature. 2015;521(7553):511–515. 10.1038/nature14273 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Tripathy SJ, Burton SD, Geramita M, Gerkin RC, Urban NN, Urban NN. Brain-wide analysis of electrophysiological diversity yields novel categorization of mammalian neuron types. Journal of Neurophysiology. 2015;in press. 10.1152/jn.00237.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55. Brockmeier AJ, Choi JS, Kriminger EG, Francis JT. Neural Decoding with Kernel-based Metric Learning. Neural computation. 2014;26:1080–1107. 10.1162/NECO_a_00591 [DOI] [PubMed] [Google Scholar]
  • 56. Smith E, Lewicki MS. Efficient coding of time-relative structure using spikes. Neural Computation. 2005;17(1):19–45. 10.1162/0899766052530839 [DOI] [PubMed] [Google Scholar]
  • 57. Botella-Soler V, Deny S, Martius G, Marre O, Tkačik G. Nonlinear decoding of a complex movie from the mammalian retina. PLOS Computational Biology. 2018;14(5):e1006057. 10.1371/journal.pcbi.1006057 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58. Boerlin M, Denève S. Spike Based Population Coding and Working Memory. PLoS Computational Biology. 2011;7(2):e1001080. 10.1371/journal.pcbi.1001080 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59. Boerlin M, Machens CK, Denève S. Predictive Coding of Dynamical Variables in Balanced Spiking Networks. PLoS Computational Biology. 2013;9(11). 10.1371/journal.pcbi.1003258 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60. Bourdoukan R, Barrett DGT, Machens CK, Denève S. Learning optimal spike-based representations. In: Bartlett P, Pereira FCN, Burges CJC, Bottou L, Weinberger KQ, editors. Advances in Neural Information Processing Systems 25; 2012. p. 2294–2302. [Google Scholar]
  • 61. Denève S, Alemi A, Bourdoukan R. The Brain as an Efficient and Robust Adaptive Learner. Neuron. 2017;94(5):969–977. 10.1016/j.neuron.2017.05.016 [DOI] [PubMed] [Google Scholar]
  • 62. Schwemmer MA, Fairhall AL, Denève S, Shea-Brown ET. Constructing Precisely Computing Networks with Biophysical Spiking Neurons. Journal of Neuroscience. 2015;35(28):10112–10134. 10.1523/JNEUROSCI.4951-14.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63. Pouget A, Snyder LH. Computational approaches to sensorimotor transformations. Nature Neuroscience. 2000;3(11s):1192–1198. [DOI] [PubMed] [Google Scholar]
  • 64. Denève S, Machens CK. Efficient codes and balanced networks. Nature neuroscience. 2016;19(3):375–82. 10.1038/nn.4243 [DOI] [PubMed] [Google Scholar]
  • 65. Eliasmith C. A Unified Approach to Building and Controlling Spiking Attractor Networks. Neural Computation. 2005;17(6):1276–1314. 10.1162/0899766053630332 [DOI] [PubMed] [Google Scholar]
  • 66. Thalmeier D, Uhlmann M, Kappen HJ, Memmesheimer Rm. Learning Universal Computations with Spikes. PLoS Computational Biology. 2016;12(6):1–29. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Nardin M, Phillips JW, Podlaski WF, Keemink SW. Nonlinear computations in spiking neural networks through multiplicative synapses. arXiv. 2020; p. preprint arXiv:2009.03857.
  • 68. Prescott SA, Ratté S, De Koninck Y, Sejnowski TJ. Pyramidal Neurons Switch From Integrators In Vitro to Resonators Under In Vivo-Like Conditions. Journal of Neurophysiology. 2008;100:3030–3042. 10.1152/jn.90634.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69. Ermentrout GB. Type I Membranes, Phase Resetting Curves, and Synchrony. Neural Computation. 1996;8:979–1001. 10.1162/neco.1996.8.5.979 [DOI] [PubMed] [Google Scholar]
  • 70. Gutkin BS, Ermentrout GB, Reyes AD. Phase-response curves give the responses of neurons to transient inputs. Journal of neurophysiology. 2005;94(2):1623–35. 10.1152/jn.00359.2004 [DOI] [PubMed] [Google Scholar]
  • 71. Ermentrout GB, Galan RF, Urban NN. Relating neural dynamics to neural coding. Physical Review Letters. 2007;99(24):248103. 10.1103/PhysRevLett.99.248103 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72. Nicoll RA, Alger BE. Synaptic excitation may activate a calcium-dependent potassium conductance in hippocampal pyramidal cells. Science. 1981;212(4497):957–959. 10.1126/science.6262912 [DOI] [PubMed] [Google Scholar]
  • 73. Gulledge AT, Stuart GJ. Excitatory Actions of GABA in the Cortex. Neuron. 2003;37:299–309. 10.1016/S0896-6273(02)01146-7 [DOI] [PubMed] [Google Scholar]
  • 74. Jonas P, Buzsáki G. Neural Inhibition. Scholarpedia. 2007;2(9):3286. 10.4249/scholarpedia.3286 [DOI] [Google Scholar]
  • 75. Izhikevich EM. Dynamical Systems in Neuroscience: The Geometry of Excitability and Bursting. Cambridge, MA: MIT Press; 2007. [Google Scholar]
  • 76. Denève S, Chalk M. Efficiency turns the table on neural encoding, decoding and noise. Current Opinion in Neurobiology. 2016;37:141–148. 10.1016/j.conb.2016.03.002 [DOI] [PubMed] [Google Scholar]
  • 77. Franke F, Fiscella M, Sevelev M, Roska B, Hierlemann A, Azaredo da Silveira R. Structures of Neural Correlation and How They Favor Coding. Neuron. 2016;89(2):409–422. 10.1016/j.neuron.2015.12.037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78. Montijn JS, Meijer GT, Lansink CS, Pennartz CMA. Population-Level Neural Codes Are Robust to Single-Neuron Variability from a Multidimensional Coding Perspective. Cell Reports. 2016; p. 91–101. [DOI] [PubMed] [Google Scholar]
  • 79. de Ruyter van Steveninck RR, Lewen GD, Strong SP, Koberle R, Bialek W. Reproducibility and Variability in Neural Spike Trains. Science. 1997;275:1805–1808. 10.1126/science.275.5307.1805 [DOI] [PubMed] [Google Scholar]
  • 80. Mainen ZF, Sejnowski TJ. Reliability of Spike Timing in Neocortical Neurons. Science. 1995;268:1503–1506. 10.1126/science.7770778 [DOI] [PubMed] [Google Scholar]
  • 81. Shadlen MN, Newsome WT. Noise, neural codes and cortical organization. Current Opinion in Neurobiology. 1994;4(4):569–79. 10.1016/0959-4388(94)90059-0 [DOI] [PubMed] [Google Scholar]
  • 82. Renart A, Machens CK. Variability in neural activity and behavior. Current Opinion in Neurobiology. 2014;25:211–220. 10.1016/j.conb.2014.02.013 [DOI] [PubMed] [Google Scholar]
  • 83. Mitchell DE, Kwan A, Carriot J, Chacron MJ, Cullen KE. Neuronal variability and tuning are balanced to optimize naturalistic self- motion coding in primate vestibular pathways. eLife. 2018; p. e43019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84. Shamir M, Sompolinsky H. Implications of neuronal diversity on population coding. Neural computation. 2006;18(8):1951–86. 10.1162/neco.2006.18.8.1951 [DOI] [PubMed] [Google Scholar]
  • 85. Sompolinsky H, Yoon H, Kang K, Shamir M. Population coding in neuronal systems with correlated noise. Physical Review E. 2001;64(5):051904. 10.1103/PhysRevE.64.051904 [DOI] [PubMed] [Google Scholar]
  • 86. Gjorgjieva J, Drion G, Marder E. Computational implications of biophysical diversity and multiple timescales in neurons and synapses for circuit performance. Current Opinion in Neurobiology. 2016;37:44–52. 10.1016/j.conb.2015.12.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87. Mejias JF, Longtin A. Optimal Heterogeneity for Coding in Spiking Neural Networks. Physical Review Letters. 2012;108(22):228102. 10.1103/PhysRevLett.108.228102 [DOI] [PubMed] [Google Scholar]
  • 88. Perez-Nieves N, Leung VCH, Dragotti PL, Goodman DFM. Neural heterogeneity promotes robust learning. bioRxiv. 2020; p. 2020.12.18.423468. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89. Tian G, Li S, Huang T, Wu S. Excitation-Inhibition Balanced Neural Networks for Fast Signal Detection. Frontiers in Computational Neuroscience. 2020;14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Kadmon J, Timcheck J, Ganguli S. Predictive coding in balanced neural networks with noise, chaos and delays. arXiv:200614178 [cond-mat, q-bio, stat]. 2020;.
  • 91. Harris KD, Mrsic-Flogel TD. Cortical connectivity and sensory coding. Nature. 2013;503(7474):51–8. 10.1038/nature12654 [DOI] [PubMed] [Google Scholar]
  • 92. Ko H, Hofer SB, Pichler B, Buchanan Ka, Sjöström PJ, Mrsic-Flogel TD. Functional specificity of local synaptic connections in neocortical networks. Nature. 2011;473(7345):87–91. 10.1038/nature09880 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93. Beck JM, Pouget A. Insights from a Simple Expression for Linear Fisher Information in a Recurrently Connected Population of Spiking Neurons. Neural Computation. 2011;23(6):1484–1502. 10.1162/NECO_a_00125 [DOI] [PubMed] [Google Scholar]
  • 94. Moreno-Bote R, Beck JM, Kanitscheider I, Pitkow X, Latham P, Pouget A. Information-limiting correlations. Nature Neuroscience. 2014;17(10):1410–1417. 10.1038/nn.3807 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95. Ecker AS, Berens P, Keliris GA, Bethge M, Logothetis NK, Tolias AS. Decorrelated Neuronal Firing in Cortical Microcircuits. Science. 2010;327:584–587. [DOI] [PubMed] [Google Scholar]
  • 96. Tkačik G, Prentice JS, Balasubramanian V, Schneidman E. Optimal population coding by noisy spiking neurons. Proceedings of the National Academy of Sciences of the United States of America. 2010;107(32):14419–14424. 10.1073/pnas.1004906107 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 97. Schneidman E. Noise, Correlations, and Information in Neural Population Codes. In: Gazzaniga MS, Mangun GR, editors. The Cognitive Neurosciences. 5th ed. MIT PRess; 2014. p. 319–336. [Google Scholar]
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1008673.r001

Decision Letter 0

Daniel Bush, Daniele Marinazzo

12 Feb 2021

Dear Dr Zeldenrust,

Thank you very much for submitting your manuscript "Efficient and robust coding in heterogeneous recurrent networks" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations. In particular, by improving the description of the time-dependent threshold, clarifying the limitations of this approach, and - ideally - elaborating on the experimental predictions of the proposed framework.

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. 

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Daniel Bush

Associate Editor

PLOS Computational Biology

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

***********************

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

[LINK]

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: Overall comments:

This is a really nice paper that shows the advantages of heterogeneity for coding in a particular spiking neuron model. This is a pretty important point because the brain is heterogeneous but most models are not. I particularly liked the observation in the discussion that this is a noiseless in-vivo-like model in which trial-to-trial variability is high but the precise timing of each spike matters.

Major points:

- The argument rests on a model of neurons that produce spikes that perfectly minimise error when doing linear decoding. I think this is an interesting approach, but it might be nice to acknowledge limitations of that approach and discuss to what extent it can be implemented more practically.

- How robust are the results to the choice of efficiency measure? I can see the logic in dividing by number of spikes to get a measure of some sort of accuracy per spike, but 1/MSE as a measure of accuracy doesn't seem totally obvious to me. For example, you might equally argue that there is a cost to each inaccuracy, and a cost for each spike, and you want to minimise the sum of those costs, suggesting that you want to minimise a*num_spikes+b*MSE for some a, b. It would be nice to see that the conclusions about efficiency are robust to the measure chosen here, in the absence of any obvious standard measure.

- Conclusions are a bit too strongly worded, e.g. "high trial-to-trial variability is a hallmark for an efficiently coding network". In the specific case studied here, yes, but I'm not sure you can generalise this.

- On that specific point (on p8, end of second paragraph of right column), it looks like from Fig 4, bottom right, that actually there isn't quite an inverse relation of E with Gamma, but rather that it peaks around Gamma=0.2?

- Fig 4, heterogeneous MSE panel, what is going on at the right hand side of this figure? Why do some values of tau seem to have (randomly) much smaller or larger errors? Is this noise? Is it feasible to re-run the simulations more times to get a smoother figure?

Minor:

- On p8, left column, there's a combinatorial argument that you would expect the lowest reliability for smallest amplitude signals because the network can choose which neuron to use. But in the simple case of all amplitudes being the same, you would have the maximum number of choices for k where n choose k was largest, i.e. k=n/2. I suspect that I'm making a mistake here though.

- p8, bottom right, some sentences/parts of sentences repeated here.

- Section numbers are missing throughout.

Reviewer #2: ### Review summary

This work presents a conceptual novelty that combines heterogeneous neural temporal filters within an efficient coding framework. The analysis of the error and contribution of different properties of noise is interesting and shows unexpected results. The theory's predictive power and its use in future experiments is the weak spot of the work. One critical point that must be addressed is improving the description and implementation of the time-dependent threshold introduced. I recommend publishing this work after some revisions.

### Main contributions

1. This work's main conceptual novelty is considering the effect of using mixed temporal filters for predictive coding. Notably, some recent papers have studied related questions: In [1], the authors considered the benefits of using a population of neurons with heterogeneous time constants on learning. The study in [2] focused on the efficient predictive coding framework with random tuning (random weights) but a uniform temporal profile [2]. This work, for the first time to my knowledge, presents a model for efficient coding in spiking where a temporal filter defines neural tuning. I think this idea can influence future studies beyond the results of the current study.

2. The studies mentioned above demonstrated the benefits of heterogeneity: In [1], different temporal properties of neurons improve the learning of time-dependent data, and in [2], structural disorder improves the stability and can lead to better performances. The current study presents similar nature results, but I find them complementary and not overlapping: a network composed of neurons with different temporal filter can be more efficient.

3. The work provides an interesting analysis of different noise properties' contribution to the network's efficiency and correlation structure. First, they show that a heterogeneous network with mixed temporal profiles is more robust to correlated noise than the homogeneous counterpart. Second, the authors show that correlated noise –which is expected– hides the negative noise correlation predicted by the efficient coding network. The behavior of the model to different noise properties is beneficial for future studies.

### Criticism

1. The authors propose a solution to the known "ping-pong" effect of the efficient coding paradigm by adjusting the threshold. Increasing the threshold is a known solution and was proposed in the early works on efficient coding (ref [58] in the paper). In the current paper, the authors define a time-dependent threshold that acts as a filter. This solution is an interesting new view and seems plausible by considering an adaptation of neurons' firing rates. However, I find the explanation of the model lacking. In particular, it is not clear to me how the temporal kernel of the threshold is implemented. Is the kernel convoluted over the inputs or the membrane potential? The authors need to elaborate on how they implement the temporal-kernel threshold. Does it offer benefit when compared to the solution proposed by ref [58]? It is hard to judge the contribution and the biological plausibility without a clear description of the model.

2. The authors devote a large section to the experimental predictions of their theory. The current state of efficient coding theory lacks testable predictions, which will help advance and disseminate the ideas. Here, the authors describe in detail the results of numerical simulations that use biological values for the parameters. Considering it is a simplified model, I could not understand what the actual biological prediction is. The authors indicate that we should not expect to see the expected negative noise correlations because of correlated noise. While important, it is a postdiction of what we do not see in experiments, rather than a testable prediction. I find the experimental predictions the weakest part of this work.

### Minor issues

1. The authors note that in deep learning, causal filters are needed because of the layered structure and information flow direction. In my view, causal filters are also needed for a recurrent network. Since the authors are proposing a mechanistic model of neural coding –and not a statistical description– causality seems important for recurrent networks.

2. I find some of the definitions slightly confusing:

a. First, the authors define reliability as the similarity between spike trains. Naturally, one expects reliability to depend on the readout. In that case, a network that can produce the same readout with low error for very different spike trains is more reliable. Counterintuitively, the 'reliability' of the network is very prone to noise. I feel that 'consistency' is a more appropriate term. If the authors choose to keep the term 'reliability,' they may want to emphasize that difference.

b. The authors do not normalize the efficiency by the signal amplitude, so high-amplitude signals with high activity give low efficiency (which they notice in the results). It is counterintuitive to the meaning of efficiency, and I think it should be pointed out.

3. I could not understand the bottom-left panel of figure 4. Did the authors change the signal strength in different simulations to get different points? Is it just another way of presenting the top panels? I need a more explicit explanation of what is the take-home message from it.

### Recommendation

This work provides a conceptual novelty using temporal filters and tuning curves together is an efficient coding framework; this alone makes this work publication-worthy. The experimental predictions section is weak, and I would have liked to see some more clearly stated predictions, but I understand it is a lot to ask and would still recommend publication without it. There are some minor corrections to clarity and definitions that could be addressed. The one fault I believe must be addressed before publication is a better explanation of the time-dependent threshold and its implementation, and I recommend publishing the work after resubmission.

### Typos and errors

1. All the section references are missing.

2. The authors use the acronyms EPSP and IPSPS but define them only a few sentences after first use.

3. On page 9, the paragraph below the figure, third row. The reference to figure 5 points to the top row, where it should be the bottom row.

### References

[1] Perez-Nievez et al., Neural heterogeneity promotes robust learning. BioArxiv 2020, (https://www.biorxiv.org/content/10.1101/2020.12.18.423468v2.full)

[2] Kadmon et al., Predictive Coding in balance neural network with noise, chaos and delays, Neurips 2020 (https://papers.nips.cc/paper/2020/hash/c236337b043acf93c7df397fdb9082b3-Abstract.html)

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: Yes

Reviewer #2: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Jonathan Kadmon

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, PLOS recommends that you deposit laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see http://journals.plos.org/ploscompbiol/s/submission-guidelines#loc-materials-and-methods

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1008673.r003

Decision Letter 1

Daniel Bush, Daniele Marinazzo

7 Apr 2021

Dear Dr Zeldenrust,

We are pleased to inform you that your manuscript 'Efficient and robust coding in heterogeneous recurrent networks' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Daniel Bush

Associate Editor

PLOS Computational Biology

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

***********************************************************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: The authors have fully responded to all my points and I'm happy with the paper as it is.

Congratulations on a great paper!

Reviewer #2: The authors have addressed my concerns, and I recommend publishing the paper in its current form.

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1008673.r004

Acceptance letter

Daniel Bush, Daniele Marinazzo

27 Apr 2021

PCOMPBIOL-D-21-00095R1

Efficient and robust coding in heterogeneous recurrent networks

Dear Dr Zeldenrust,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Andrea Szabo

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Text

    Section A: A word on notation. Section B: Technical notes. Section C: Efficiency measures, including Supplemental Figures A and B. Fig A: Comparison of efficiency measures Results of two simulations using the same stimulus, but different initial network states, in a homogeneous ‘type 1’ network (first row), a network with ‘type 1’ and ‘type 2’ neurons (second row) and a heterogeneous network (third row). Three alternative efficiency measures are compared: 1) the network efficiency multiplied by the stimulus amplitude (Ea, see supp. eq. (10), first column), 2) the network efficiency multiplied by the stimulus power (Ea2, see supp/ eq. (11), second column) and 3) the network cost (C, see supp. eq. (12), third column). The bottom row shows how the network efficiency or cost depends on the spike reliability Γ¯. Parameters: Δ = 7, 5 ms, ν = μ = 1, 5, N = 100, #trials = 10. The white star denotes the parameter values used in section ‘Heterogeneous networks are more efficient than homogeneous networks’ of the main text. Fig B: Comparison of efficiency measures for the noise simulations Efficiency normalized by amplitude Ea (first column), amplitude squared Ea2 (second column) and cost (third column) of a homogeneous network (top row), a network with ‘type 1’ and ‘type 2’ neurons (middle row) and a heterogeneous network (bottom row). Next to the stimulus, each neuron in the network was presented with a noise input, with a varying relative amplitude (horizontal axis) and number of copies of the noise signal (vertical axis; 1 copy means all neurons receive the same noise, 100 copies means all neurons receive independent noise). Network: Δ = 7, 5 ms, ν = μ = 1, 5, N = 100. Stimulus: amplitude = 10, τ = 15 ms. Noise: τ = 15 ms.

    (PDF)

    Attachment

    Submitted filename: response reviewers formatted.pdf

    Data Availability Statement

    All code to generate the data can be found at https://github.com/fleurzeldenrust/Efficient-coding-in-a-spiking-predictive-coding-network.


    Articles from PLoS Computational Biology are provided here courtesy of PLOS

    RESOURCES