Abstract
More often than not, action potentials fail to trigger neurotransmitter release. And even when neurotransmitter is released, the resulting change in synaptic conductance is highly variable. Given the energetic cost of generating and propagating action potentials, and the importance of information transmission across synapses, this seems both wasteful and inefficient. However, synaptic noise arising from variable transmission can improve, in certain restricted conditions, information transmission. Under broader conditions, it can improve information transmission per release, a quantity that is relevant given the energetic constraints on computing in the brain. Here we discuss the role, both positive and negative, synaptic noise plays in information transmission and computation in the brain.
Keywords: synaptic noise, information transfer, optimal synapse
Highlights
Synaptic transmission is noisy, partially because neurotransmitter release is unreliable but also because synaptic receptor current varies stochastically.
When synaptic input is below the cell spike threshold, adding noise may push it above the threshold, thus improving information transmission.
Synaptic noise can widen the dynamic range of transmission.
Synaptic noise allows synapses to communicate their degree of uncertainty.
Noise is a ubiquitous feature of brain activity and is likely to be a critical element in any theory of how the brain works.
Neuronal Communication is Noisy
Neurons communicate primarily through chemical synapses, and that communication is critical for proper brain function. However, chemical synaptic transmission appears unreliable: for most synapses, when an action potential arrives at an axon terminal, about half the time, no neurotransmitter is released and so no communication happens. Even strong and reliable synapses, such as the neuromuscular junction or the calyx of Held, operate by using multiple unreliable release sites [1]. Furthermore, when neurotransmitter is released at an individual synaptic release site, the size of the local postsynaptic membrane conductance change is also variable. Given the importance of synapses, the energetic cost of generating action potentials, and the evolutionary timescales over which the brain has been optimized, the high level of synaptic noise seems surprising. Here we ask: how surprised should we be? More to the point: is synaptic noise a bug, or is it a feature? Whilst the role of unreliable release on neural network performance has been explored in multiple studies (see [2., 3., 4.] for informative reviews), here we expand on these views, in particular giving consideration to noisy synaptic conductance (see Glossary).
Noise in Nonlinear Communication Channels
When communicating information, noise would seem to be universally bad. Static on the radio, background noise during a conversation, grainy videos – they all make it harder to extract interesting information. However, for nonlinear systems – especially when a threshold is involved, as in spike generation – noise can be helpful [5., 6., 7.]. Most notably, if a noise-free presynaptic signal is too small, it will never generate a spike and thus will not transmit any information. Add a small amount of noise, and, at least sometimes, a spike will occur. Thus, a subthreshold sensory signal has a better chance of being detected when noise is added [8,9]. This phenomenon is, very broadly, known as stochastic resonance [3,10,11], and it is observable in electrophysiological experiments: add a small amount of noise to subthreshold synaptic currents, and the probability of a spike can go way up [12]. A simple multisynaptic neuron model (Figure 1A) illustrates this point. With no noise in the synaptic conductance, there are very few spikes (Figure 1B), but adding a small amount of variability in the postsynaptic conductance can produce fairly reliable spikes (Figure 1C). Note, however, that decreasing threshold for the generation of an action potential can produce even more reliable spikes (Figure 1D), which has a higher somatic sodium conductance, making the cell more excitable compared with Figure 1B. Although in a wider context we cannot equate more spiking with more information, we will, in the interest of simplicity, assume that is the case here.
Abstractly, the ability to communicate information depends on three things: the range of possible input signals, the noise around those signals, and the transformation from input to output (Box 1 provides a commonly accepted definition of information). For neurons, the transformation from input to output yields either a spike or no spike. In that case, communication is optimal when the integrated input to a neuron is either well above or well below threshold relative to the noise. In such a regime, a spike provides unambiguous information about the input: it tells downstream neurons whether the input was high or low. But typically, the signal and noise cannot be adjusted independently; in particular, the synaptic drive to a neuron is not bimodal – it is graded. So, the simple, optimal, regime is not realized. Consequently, the question of whether synaptic noise is a bug or a feature depends on how the signal and noise interact. For instance, if increasing noise made synaptic drive more bimodal, it could, at least in principle, increase information transmission. And we have already seen that adding noise can increase the reliability of output spikes (Figure 1C). But we have also seen that increasing cell excitability, and thus lowering the threshold for a spike, can do the same thing (Figure 1D). Which one is better, and when?
Box 1. Defining (Mutual) Information.
Spike trains convey information about presynaptic input. Intuitively, this means that if you were to observe spikes from a neuron, you would know more about its synaptic drive than if you did not observe spikes. To quantify this intuition, we need a way to measure knowledge – or, as is much more common, uncertainty. There are many ways to do this, but an especially convenient one is entropy, which is a relatively direct measure of uncertainty; the higher the entropy of a distribution over some variable (i.e., the broader the distribution), the more the uncertainty about that variable (according to Equations (I), (II) as follows). With this choice, the difference in entropy is the Shannon mutual information [59], which indicates how much information about the input signal we can glean from the output signal (Figure I).
The entropy of a continuous distribution P(x), often denoted H(x), is
(I) |
If P(x) is discrete there is an analogous definition; the only difference is that the integral becomes a sum,
(II) |
To compute mutual information, we need to compare the entropy before and after some observation (which for us is a spike train). Let’s use s for the observation, and P(x | s) for the distribution of x (which for us is the synaptic drive) after observing a spike train. In that case, the conditional entropy, denoted H(x | s) – the average entropy after observing a spike train – is
(III) |
with integrals replaced by sums for discrete distributions. The mutual information, I(x,s), between x and s is the difference between the two entropies,
(IV) |
This definition is sufficiently flexible that x could be continuous and s discrete, or vice versa. As is not hard to show, information is symmetric; it can also be written
(V) |
Symmetry is important because it tells us that the higher the entropy of a spike train (which typically translates to higher firing rate), the larger the potential for high information. Importantly, though, increasing the firing rate is not guaranteed to increase information, as that may also increase the second term.
Alt-text: Box 1
Sources of Synaptic Noise
To investigate how synaptic noise affects information transmission in realistic situations, we need to understand the sources of noise. There are two main sources. The first, and best studied, are release failures: when an action potential invades the presynaptic terminal, neurotransmitter may or may not be released [13,14]. The average release probability, Pr – calculated as one minus the probability of failure – ranges from 0.2 to 0.8 at central synapses [15., 16., 17., 18.], with experiments in vivo putting this towards the low end [19]. This generates considerable noise: if ten action potentials arrive at a cell approximately simultaneously, typically somewhere between two and eight of them will cause a release of neurotransmitter, and on any one trial, anywhere from zero to ten will cause neurotransmitter release.
The second source of noise is variability in the synaptic conductance (producing variability in the peak synaptic receptor current) in response to release of one neurotransmitter-filled synaptic vesicle. Fluctuations in the synaptic conductance depend on two main factors. First, synaptic vesicle content can vary depending on prior activity [20,21]. Second, synaptic receptor binding and activation is a stochastic process controlled by conformational changes of receptor proteins [22,23]. The lateral mobility of synaptic receptors [24,25] and, possibly, neurotransmitter release sites [26,27], have also been considered a factor. However, recent super-resolution studies have revealed a structural alignment of presynaptic release sites and postsynaptic receptor clusters [28], which should minimise this source of noise, at least on time scale of minutes [29]. Nonetheless, excitatory postsynaptic currents recorded at individual synapses in response to synaptic vesicle release, in basal conditions, vary significantly. A recent study suggested that the architecture of common excitatory synapses is such that it maximises synaptic conductance variability [30]. If so, it would appear that individual excitatory synaptic connections 'prefer' to be noisy.
An additional source of noise affecting synaptic circuit signalling is stochastic fluctuations of the cell membrane potential. Its multiple origins and its effect on performance have recently been reviewed in detail [31] and are outside the scope of this present opinion article.
How Does Synaptic Noise Affect Information Transmission?
Figure 1 shows that if the spike threshold is set too high, noise from variability in the conductance can improve information transmission. This is not true of release failures: lowering the probability of release decreases the drive to a postsynaptic cell. However, if we conserve transmitter expenditure, by increasing the conductance as we lower the release probability so that the product of conductance and Pr stays constant, the story changes. In this regime, the mean is independent of the release probability and the variance scales as (1– Pr)/Pr. The variance thus increases as Pr decreases – exactly what we need to drive a cell when the threshold is set too high. Consequently, both failures and conductance variability could be beneficial for information transfer.
This is illustrated in Figure 2. Figure 2A shows that reliable synapses (Pr = 1) with constant conductance are perfectly efficient, but only when the overall input is above the threshold (compare left and right). In comparison, unreliable synapses (Pr < 1) with conductance scaled to conserve transmitter expenditure (as described earlier) can transfer more information when below threshold, but their performance is worse above threshold (Figure 2B). Not surprisingly, reliable synapses with variability in conductance will behave in a similar way to unreliable synapses with constant conductance (Figure 2C).
The picture that emerges from this analysis is that when the threshold is set too high, noise increases information transmission; otherwise, noise decreases information transmission (see Figure I in Box 2). This suggests that the optimal strategy is to reduce noise as much as possible, then set the threshold for spike generation to maximize information transmission. As we have seen, increasing release probability and reducing conductance variability both reduce noise. In addition, the nervous system has another option: increase the redundancy of synaptic connections – by allowing individual axons to make multiple synaptic connections on a postsynaptic cell [32]. This averages the noise and so partially compensates for synaptic unreliability [33,34]. This is, in fact, the strategy taken by reliable synapses, such as the neuromuscular junction or the calyx of Held, which are equipped with multiple unreliable release sites [1].
Box 2. Computing (Mutual) Information.
To gain an understanding of how mutual information depends on parameters, consider a simplified scenario in which there is a single presynaptic neuron that either does or does not fire, the output spikes tell us whether or not it fired, time is divided into small bins, say 10 ms, and there are either zero spikes or one spike in each bin.
In this scenario, it is not hard to show that the mutual information (Box 1) is given by
(I) |
where p is the probability that the presynaptic neuron fires, q denotes the probability that the postsynaptic neuron emits a spike given that the presynaptic neuron fired, and h(z) is the entropy of a Bernoulli random variable with probability z,
(II) |
Assume the postsynaptic neuron fires when its input exceeds a threshold, θ, and that the input drive is Gaussian with mean μ and standard deviation σ. In that case, q is a cumulative normal function of (μ – θ)/σ. A plot of mutual information versus μ is shown in Figure I for various values of the noise (the thicker the line, the larger the noise). If the mean synaptic drive is above threshold, noise hurts, because it decreases the probability that a presynaptic spike will cause a postsynaptic one; if it is below threshold, noise helps, because it increases the probability that a presynaptic spike will cause a postsynaptic one. In addition, the higher the firing rate of the presynaptic neuron, the higher the mutual information (compare red and blue curves).
These observations suggest that mutual information is maximized when the input is bimodal [either well above or well below threshold, so that q = 1 according to Equation II, which means h(q) is zero], the noise is small, and the output firing rate is large. However, it is unlikely that these are achievable. For instance, presynaptic spikes are spread out over time. Consequently, the synaptic drive is a continuous function of time, and neurons spike as soon as the drive crosses threshold. Thus, synaptic drive is rarely bimodal. In addition, there are energetic constraints on firing rate, so it can’t be arbitrarily large. Note, though, that in the brain, the input information is much larger than the output information – about 1000 times larger, since each neuron receives about 1000 inputs. Consequently, the information lost due to noise added by failures and variability in the size of postsynaptic currents is likely to be small compared with the information that is thrown away by the finite capacity of neurons, an argument made by Levy and Baxter [37].
Alt-text: Box 2
Taken to its extreme, if neurons can optimise their threshold, zero noise should always be optimal. However, the continual learning that goes on in the brain involves constant adjustment of synaptic strengths and cell excitability. This should make threshold optimisation a continuously evolving process – one that can never be perfectly realised. Indeed, the large range of timescales over which changes in synaptic strength and in cell excitability occur [35,36] suggest that brain circuits cannot rely on permanently optimised thresholds. Thus, it may be advantageous to use synaptic noise to ensure that some spikes are transmitted rather than relying on an optimal threshold. In this context, release probability is not strictly noise: its use-dependent changes may represent a certain 'informative' operation that contributes to the transmitted message.
So far, we have defined optimality purely with respect to information transmission. However, even if a circuit is optimal from an information transmission standpoint, it is unlikely to be optimal once the cost of neurotransmitter release is considered. There is a relatively simple reason for that: spikes have limited bandwidth (sustained firing rates are almost always well below 100 Hz, with only occasional bursts of a few spikes at 1 kHz), so substantial input information, which occurs when release probability is high and/or input activity is intense, is largely wasted. As shown previously [37], for reasonable output firing rates, release probabilities that maximize information per release are in the range 0.2–0.8; exactly what is seen in vivo. Subsequent studies taking into account short-term plasticity [38., 39., 40.] found the same thing: information per release peaked at release probabilities below one. It is easy to see why, at least in the extreme case: if two consecutive presynaptic spikes occur with a small interspike interval, it is more energy efficient to release neurotransmitter on only one of the spikes but double the conductance change than to release neurotransmitter on both spikes without doubling the conductance change. This has the added benefit of increasing slightly the peak conductance, and thus increasing the probability of a postsynaptic spike.
The studies mentioned earlier, on optimal information per release, considered noise only from synaptic failures. However, the same lessons should apply to variability in synaptic conductance but with an important difference: because conductance variability does not affect release probability, if the threshold is set perfectly, adding conductance variability would not increase information per release, and so would seem to be uniformly bad for information transmission per release. However, the fact that excitatory synapses appear to maximise synaptic conductance variability [30] points to the possibility that theory is missing an important role for this kind of noise.
It is also worth mentioning that noise can affect the input–output function of neurons. Modern theories of recurrent spiking networks suggest that most of the time the mean drive to a neuron is subthreshold, and it is temporal fluctuations around the mean that drive activity [31,41., 42., 43., 44.]. Whenever this is the case, increasing the (zero mean) noise – no matter how it is done – increases the probability that a postsynaptic cell will fire. Thus, noise, due either to synaptic failures or variability in peak synaptic conductance, is an energetically cheap way to increase the gain of a neuron (albeit at the cost of larger fluctuations). Noise can also decrease the probability of firing; that happens when the input to a neuron is superthreshold. Thus, the overall effect of noise is to flatten the gain curve (input current to firing rate), and so widen a neuron’s dynamic range. This is something that threshold adjustments cannot do.
A Computational Role for Noise
It is reasonable to assume that information transmission is the main function of synapses. However, there is another potential role, one rooted in the Bayesian approach to neuroscience [45., 46., 47., 48.]. It has become increasingly clear over the last several decades that organisms – from fly larvae to humans – compute and manipulate probabilities [48., 49., 50., 51.]. This is not surprising; it is impossible to make a good decision of any kind without some measure of uncertainty. For instance, every time we cross the street we have to estimate whether the oncoming car will hit us; because it is impossible to get an exact estimate, we must evaluate how uncertain we are about a collision. Without some estimate of uncertainty, we would not know whether we can cross the street reasonably safely. Much of the work in this area has focused on uncertainty in our perception of the outside world. However, there is another source of uncertainty: the computations the brain performs depend on synaptic strengths, but the correct strengths are not known with certainty. In principle, that degree of uncertainty should be communicated to the rest of the brain. Recently it was proposed that synapses compute their uncertainty and then communicate that uncertainty by adding noise to synaptic weights [52]. With this theory, variability in synaptic strength – caused by failures, variability in conductance size, or both – is a feature, not a bug; it is used by the rest of the brain to estimate uncertainty, and, ultimately, to determine confidence. Whether or not synapses do this is currently not known, but it is an avenue worth exploring, we would argue, and one that raises concrete questions for future research.
In our discussion of information transmission, we have operated under the assumption that noise is useful only if the spike threshold is too low. However, noise can be useful in and of itself. For instance, it can boost behavioural variability – an essential feature in most, if not all, adversarial games. It is also useful for exploratory behaviour, one prominent example being identifiable random number generators in birdsongs [53]. In some instances, however, synaptic noise could be unimportant, simply forced on the brain by biological reality. And the cost of noisy synapses may not be very high: chaotic dynamics [54] already introduces a large amount of noise in neural circuits, so adding a little more probably does not cost much in terms of computational capabilities. For instance, the noise associated with failures should have very little effect on the accuracy of computations [55], and the approximations that the brain makes have a much larger effect on performance than the noise found in either synapses or chaotic dynamics [56].
Concluding Remarks
Two key physiological phenomena contribute to what we consider synaptic noise: stochastic neurotransmitter release and variability in postsynaptic conductance given release. Multiple lines of experimental and theoretical evidence suggest that, at least in certain conditions, synaptic noise improves information handling by brain circuits. Here we attempted to discuss, and illustrate, how and when this is the case. Although much progress has been made, we still lack a deep understanding of the role of noise in the nervous system. In particular, it is currently unknown how noise interacts with plasticity at the synaptic level, learning at the behavioural level, or computations at the circuit level. These are all fundamental issues and represent fertile ground for experimental and theoretical exploration. No matter what its role, noise is a ubiquitous feature of brain activity and is likely to be a critical element in any theory of how the brain works (see Outstanding Questions).
Outstanding Questions.
What is the experimental relationship between synaptic noise and synaptic morphology?
How stable are conductance fluctuations at individual synapses over short periods of time?
Is there a relationship between synaptic current variability and a synapse’s position on the dendritic tree?
How does synaptic noise influence dendritic signal integration?
Do central synapses change their 'noisiness' with development and aging?
Are there subgroups of synapses or cells with distinct magnitudes of conductance variability?
Are there synaptic circuits or brain regions with distinct magnitudes of noise?
Does the magnitude of synaptic noise correlate with a neural circuit’s ability to undergo plastic changes?
Does the magnitude of neural network noise correlate with the animal's ability to learn?
Alt-text: Outstanding Questions
Acknowledgements
This study was supported by the Wellcome Trust Principal Research Fellowship (212251_Z_18_Z), European Research Council (ERC) Advanced Grant (323113), and European Commission NEUROTWIN grant (857562), to D.A.R.; P.E.L. was supported by the Gatsby Charitable Foundation and the Wellcome Trust.
Glossary
- Bimodal
having two peaks in the distribution, so that typical values are either, say, high or low.
- Nonlinear system
a system in which the relationship between input and output is not linear.
- Release probability
average probability with which a synaptic vesicle is released upon the arrival of an action potential.
- Stochastic
probabilistic; containing an element of uncertainty.
- Stochastic resonance
a phenomenon in which adding noise increases the probability that a signal will reach threshold for detection.
- Subthreshold synaptic currents
synaptic receptor-generated membrane current that is too weak to a neuron to fire.
- Synaptic conductance
peak cell membrane conductance arising through the opening of postsynaptic receptor channels.
- Synaptic drive
integrated effect of all active synaptic input to a neuron.
Contributor Information
Dmitri A. Rusakov, Email: d.rusakov@ucl.ac.uk.
Leonid P. Savtchenko, Email: leonid.savtchenko@ucl.ac.uk.
Peter E. Latham, Email: pel@gatsby.ucl.ac.uk.
References
- 1.Tarr T.B. Are unreliable release mechanisms conserved from NMJ to CNS? Trends Neurosci. 2013;36:14–22. doi: 10.1016/j.tins.2012.09.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Stein R.B. Neuronal variability: noise or part of the signal? Nat. Rev. Neurosci. 2005;6:389–397. doi: 10.1038/nrn1668. [DOI] [PubMed] [Google Scholar]
- 3.McDonnell M.D., Ward L.M. The benefits of noise in neural systems: bridging theory and experiment. Nat. Rev. Neurosci. 2011;12:415–426. doi: 10.1038/nrn3061. [DOI] [PubMed] [Google Scholar]
- 4.Faisal A.A. Noise in the nervous system. Nat. Rev. Neurosci. 2008;9:292–303. doi: 10.1038/nrn2258. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Edwards P.J., Murray A.F. Analog synaptic noise – implications and learning improvements. Int. J. Neural Syst. 1993;4:427–433. doi: 10.1142/s0129065793000353. [DOI] [PubMed] [Google Scholar]
- 6.Murray A.F., Edwards P.J. Synaptic weight noise during multilayer perceptron training: fault tolerance and training improvements. IEEE Trans. Neural Netw. 1993;4:722–725. doi: 10.1109/72.238328. [DOI] [PubMed] [Google Scholar]
- 7.Varshney L.R. Optimal information storage in noisy synapses under resource constraints. Neuron. 2006;52:409–423. doi: 10.1016/j.neuron.2006.10.017. [DOI] [PubMed] [Google Scholar]
- 8.Kitajo K. Behavioral stochastic resonance within the human brain. Phys. Rev. Lett. 2003;90:218103. doi: 10.1103/PhysRevLett.90.218103. [DOI] [PubMed] [Google Scholar]
- 9.Lu L.L. Effects of noise and synaptic weight on propagation of subthreshold excitatory postsynaptic current signal in a feed-forward neural network. Nonlinear Dynam. 2019;95:1673–1686. [Google Scholar]
- 10.Stocks N.G. Suprathreshold stochastic resonance in multilevel threshold systems. Phys. Rev. Lett. 2000;84:2310–2313. doi: 10.1103/PhysRevLett.84.2310. [DOI] [PubMed] [Google Scholar]
- 11.Mino H., Durand D.M. Enhancement of information transmission of sub-threshold signals applied to distal positions of dendritic trees in hippocampal CA1 neuron models with stochastic resonance. Biol. Cybern. 2010;103:227–236. doi: 10.1007/s00422-010-0395-5. [DOI] [PubMed] [Google Scholar]
- 12.Stacey W.C., Durand D.M. Synaptic noise improves detection of subthreshold signals in hippocampal CA1 neurons. J. Neurophysiol. 2001;86:1104–1112. doi: 10.1152/jn.2001.86.3.1104. [DOI] [PubMed] [Google Scholar]
- 13.Redman S. Quantal analysis of synaptic potentials in neurons of the central nervous system. Physiol. Rev. 1990;70:165–198. doi: 10.1152/physrev.1990.70.1.165. [DOI] [PubMed] [Google Scholar]
- 14.Jack J.J.B. Quantal analysis of excitatory synaptic mechanisms in the mammalian central nervous system. Cold Spring Harb. Symp. Quant. Biol. 1990;55:57–67. doi: 10.1101/sqb.1990.055.01.008. [DOI] [PubMed] [Google Scholar]
- 15.Oertner T.G. Facilitation at single synapses probed with optical quantal analysis. Nat. Neurosci. 2002;5:657–664. doi: 10.1038/nn867. [DOI] [PubMed] [Google Scholar]
- 16.Emptage N.J. Optical quantal analysis reveals a presynaptic component of LTP at hippocampal Schaffer-associational synapses. Neuron. 2003;38:797–804. doi: 10.1016/s0896-6273(03)00325-8. [DOI] [PubMed] [Google Scholar]
- 17.Sylantyev S. Cannabinoid- and lysophosphatidylinositol-sensitive receptor GPR55 boosts neurotransmitter release at central synapses. Proc. Natl. Acad. Sci. U. S. A. 2013;110:5193–5198. doi: 10.1073/pnas.1211204110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Jensen T.P. Multiplex imaging relates quantal glutamate release to presynaptic Ca2+ homeostasis at multiple synapses in situ. Nat. Commun. 2019;10:1414. doi: 10.1038/s41467-019-09216-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Borst J.G.G. The low synaptic release probability in vivo. Trends Neurosci. 2010;33:259–266. doi: 10.1016/j.tins.2010.03.003. [DOI] [PubMed] [Google Scholar]
- 20.Hanse E., Gustafsson B. Paired-pulse plasticity at the single release site level: an experimental and computational study. J. Neurosci. 2001;21:8362–8369. doi: 10.1523/JNEUROSCI.21-21-08362.2001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Liu G. Variability of neurotransmitter concentration and nonsaturation of postsynaptic AMPA receptors at synapses in hippocampal cultures and slices. Neuron. 1999;22:395–409. doi: 10.1016/s0896-6273(00)81099-5. [DOI] [PubMed] [Google Scholar]
- 22.Franks K.M. Independent sources of quantal variability at single glutamatergic synapses. J. Neurosci. 2003;23:3186–3195. doi: 10.1523/JNEUROSCI.23-08-03186.2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Zheng K. Receptor actions of synaptically released glutamate: the role of transporters on the scale from nanometers to microns. Biophys. J. 2008;95:4584–4596. doi: 10.1529/biophysj.108.129874. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Groc L. Differential activity-dependent regulation of the lateral mobilities of AMPA and NMDA receptors. Nat. Neurosci. 2004;7:695–696. doi: 10.1038/nn1270. [DOI] [PubMed] [Google Scholar]
- 25.Triller A., Choquet D. New concepts in synaptic biology derived from single-molecule imaging. Neuron. 2008;59:359–374. doi: 10.1016/j.neuron.2008.06.022. [DOI] [PubMed] [Google Scholar]
- 26.Xie X.P. Novel expression mechanism for synaptic potentiation: alignment of presynaptic release site and postsynaptic receptor. Proc. Natl. Acad. Sci. U. S. A. 1997;94:6983–6988. doi: 10.1073/pnas.94.13.6983. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Savtchenko L.P., Rusakov D.A. Moderate AMPA receptor clustering on the nanoscale can efficiently potentiate synaptic current. Philos. Trans. R. Soc. Lond. Ser. B Biol. Sci. 2014;369:20130167. doi: 10.1098/rstb.2013.0167. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Tang A.H. A trans-synaptic nanocolumn aligns neurotransmitter release to receptors. Nature. 2016;536:210–214. doi: 10.1038/nature19058. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Nair D. Super-resolution imaging reveals that AMPA receptors inside synapses are dynamically organized in nanodomains regulated by PSD95. J. Neurosci. 2013;33:13204–13224. doi: 10.1523/JNEUROSCI.2381-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Savtchenko L.P. Central synapses release a resource-efficient amount of glutamate. Nat. Neurosci. 2013;16:10–12. doi: 10.1038/nn.3285. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Yarom Y., Hounsgaard J. Voltage fluctions in neurons: signal or noise? Physiol. Rev. 2011;91:917–929. doi: 10.1152/physrev.00019.2010. [DOI] [PubMed] [Google Scholar]
- 32.Bloss E.B. Single excitatory axons form clustered synapses onto CA1 pyramidal cell dendrites. Nat. Neurosci. 2018;21:353–363. doi: 10.1038/s41593-018-0084-6. [DOI] [PubMed] [Google Scholar]
- 33.Zador A. Impact of synaptic unreliability on the information transmitted by spiking neurons. J. Neurophysiol. 1998;79:1219–1229. doi: 10.1152/jn.1998.79.3.1219. [DOI] [PubMed] [Google Scholar]
- 34.Manwani A., Koch C. Detecting and estimating signals over noisy and unreliable synapses: information-theoretic analysis. Neural Comput. 2001;13:1–33. doi: 10.1162/089976601300014619. [DOI] [PubMed] [Google Scholar]
- 35.Ma Z. Cortical circuit dynamics are homeostatically tuned to criticality in vivo. Neuron. 2019;104:655–664. doi: 10.1016/j.neuron.2019.08.031. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Turrigiano G.G. Activity-dependent scaling of quantal amplitude in neocortical neurons. Nature. 1998;391:892–896. doi: 10.1038/36103. [DOI] [PubMed] [Google Scholar]
- 37.Levy W.B., Baxter R.A. Energy-efficient neuronal computation via quantal synaptic failures. J. Neurosci. 2002;22:4746–4755. doi: 10.1523/JNEUROSCI.22-11-04746.2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Salmasi M. Short-term synaptic depression can increase the rate of information transfer at a release site. PLoS Comput. Biol. 2019;15 doi: 10.1371/journal.pcbi.1006666. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Goldman M.S. Redundancy reduction and sustained firing with stochastic depressing synapses. J. Neurosci. 2002;22:584–591. doi: 10.1523/JNEUROSCI.22-02-00584.2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Goldman M.S. Enhancement of information transmission efficiency by synaptic failures. Neural Comput. 2004;16:1137–1162. doi: 10.1162/089976604773717568. [DOI] [PubMed] [Google Scholar]
- 41.Renart A. Mean-driven and fluctuation-driven persistent activity in recurrent networks. Neural Comput. 2007;19:1–46. doi: 10.1162/neco.2007.19.1.1. [DOI] [PubMed] [Google Scholar]
- 42.Roach J.P. Resonance with subthreshold oscillatory drive organizes activity and optimizes learning in neural networks. Proc. Natl. Acad. Sci. U. S. A. 2018;115:E3017–E3025. doi: 10.1073/pnas.1716933115. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.van Vreeswijk C., Sompolinsky H. Chaotic balanced state in a model of cortical circuits. Neural Comput. 1998;10:1321–1371. doi: 10.1162/089976698300017214. [DOI] [PubMed] [Google Scholar]
- 44.van Vreeswijk C., Sompolinsky H. Chaos in neuronal networks with balanced excitatory and inhibitory activity. Science. 1996;274:1724–1726. doi: 10.1126/science.274.5293.1724. [DOI] [PubMed] [Google Scholar]
- 45.Haefner R.M. Perceptual decision-making as probabilistic inference by neural sampling. Neuron. 2016;90:649–660. doi: 10.1016/j.neuron.2016.03.020. [DOI] [PubMed] [Google Scholar]
- 46.Aitchison L., Lengyel M. With or without you: predictive coding and Bayesian inference in the brain. Curr. Opin. Neurobiol. 2017;46:219–227. doi: 10.1016/j.conb.2017.08.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Knill D.C., Pouget A. The Bayesian brain: the role of uncertainty in neural coding and computation. Trends Neurosci. 2004;27:712–719. doi: 10.1016/j.tins.2004.10.007. [DOI] [PubMed] [Google Scholar]
- 48.Pouget A. Probabilistic brains: knowns and unknowns. Nat. Neurosci. 2013;16:1170–1178. doi: 10.1038/nn.3495. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Strange B.A. Information theory, novelty and hippocampal responses: unpredicted or unpredictable? Neural Netw. 2005;18:225–230. doi: 10.1016/j.neunet.2004.12.004. [DOI] [PubMed] [Google Scholar]
- 50.Grabska-Barwinska A. A probabilistic approach to demixing odors. Nat. Neurosci. 2017;20:98–106. doi: 10.1038/nn.4444. [DOI] [PubMed] [Google Scholar]
- 51.Navajas J. The idiosyncratic nature of confidence. Nat. Hum. Behav. 2017;1:810–818. doi: 10.1038/s41562-017-0215-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Aitchison L. Probabilistic synapses. ArXiv. 2017;1410:1029. [Google Scholar]
- 53.Masco C. The Song Overlap Null model Generator (SONG): a new tool for distinguishing between random and non-random song overlap. Bioacoustics. 2016;25:29–40. [Google Scholar]
- 54.London M. Sensitivity to perturbations in vivo implies high noise and suggests rate coding in cortex. Nature. 2010;466:123–127. doi: 10.1038/nature09086. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Zylberberg J., Strowbridge B. Mechanisms of persistent activity in cortical circuits: possible neural substrates for working memory. Annu. Rev. Neurosci. 2017;40:603–627. doi: 10.1146/annurev-neuro-070815-014006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Beck J.M. Not noisy, just wrong: the role of suboptimal inference in behavioral variability. Neuron. 2012;74:30–39. doi: 10.1016/j.neuron.2012.03.016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Migliore M. Role of an A-type K+ conductance in the back-propagation of action potentials in the dendrites of hippocampal pyramidal neurons. J. Comput. Neurosci. 1999;7:5–15. doi: 10.1023/a:1008906225285. [DOI] [PubMed] [Google Scholar]
- 58.Hines M.L., Carnevale N.T. NEURON: a tool for neuroscientists. Neuroscientist. 2001;7:123–135. doi: 10.1177/107385840100700207. [DOI] [PubMed] [Google Scholar]
- 59.Shannon C.E., Weaver W. The University of Illinois Press; 1949. The Mathematical Theory of Communication. [Google Scholar]