Abstract
Sequence learning, prediction and replay have been proposed to constitute the universal computations performed by the neocortex. The Hierarchical Temporal Memory (HTM) algorithm realizes these forms of computation. It learns sequences in an unsupervised and continuous manner using local learning rules, permits a context specific prediction of future sequence elements, and generates mismatch signals in case the predictions are not met. While the HTM algorithm accounts for a number of biological features such as topographic receptive fields, nonlinear dendritic processing, and sparse connectivity, it is based on abstract discrete-time neuron and synapse dynamics, as well as on plasticity mechanisms that can only partly be related to known biological mechanisms. Here, we devise a continuous-time implementation of the temporal-memory (TM) component of the HTM algorithm, which is based on a recurrent network of spiking neurons with biophysically interpretable variables and parameters. The model learns high-order sequences by means of a structural Hebbian synaptic plasticity mechanism supplemented with a rate-based homeostatic control. In combination with nonlinear dendritic input integration and local inhibitory feedback, this type of plasticity leads to the dynamic self-organization of narrow sequence-specific subnetworks. These subnetworks provide the substrate for a faithful propagation of sparse, synchronous activity, and, thereby, for a robust, context specific prediction of future sequence elements as well as for the autonomous replay of previously learned sequences. By strengthening the link to biology, our implementation facilitates the evaluation of the TM hypothesis based on experimentally accessible quantities. The continuous-time implementation of the TM algorithm permits, in particular, an investigation of the role of sequence timing for sequence learning, prediction and replay. We demonstrate this aspect by studying the effect of the sequence speed on the sequence learning performance and on the speed of autonomous sequence replay.
Author summary
Essentially all data processed by mammals and many other living organisms is sequential. This holds true for all types of sensory input data as well as motor output activity. Being able to form memories of such sequential data, to predict future sequence elements, and to replay learned sequences is a necessary prerequisite for survival. It has been hypothesized that sequence learning, prediction and replay constitute the fundamental computations performed by the neocortex. The Hierarchical Temporal Memory (HTM) constitutes an abstract powerful algorithm implementing this form of computation and has been proposed to serve as a model of neocortical processing. In this study, we are reformulating this algorithm in terms of known biological ingredients and mechanisms to foster the verifiability of the HTM hypothesis based on electrophysiological and behavioral data. The proposed model learns continuously in an unsupervised manner by biologically plausible, local plasticity mechanisms, and successfully predicts and replays complex sequences. Apart from establishing contact to biology, the study sheds light on the mechanisms determining at what speed we can process sequences and provides an explanation of fast sequence replay observed in the hippocampus and in the neocortex.
Introduction
Learning and processing sequences of events, objects, or percepts are fundamental computational building blocks of cognition [1–4]. Prediction of upcoming sequence elements, mismatch detection and sequence replay in response to a cue signal constitute central components of this form of processing. We are constantly making predictions about what we are going to hear, see, and feel next. We effortlessly detect surprising, non-anticipated events and adjust our behavior accordingly. Further, we manage to replay learned sequences, for example, when generating motor behavior, or recalling sequential memories. These forms of processing have been studied extensively in a number of experimental works on sensory processing [5, 6], motor production [7], and decision making [8].
The majority of existing biologically motivated models of sequence learning addresses sequence replay [9–12]. Sequence prediction and mismatch detection are rarely discussed. The Hierarchical Temporal Memory (HTM) [13] combines all three aspects: sequence prediction, mismatch detection and replay. Its Temporal Memory (TM) model [14] learns complex context dependent sequences in a continuous and unsupervised manner using local learning rules [15], and is robust against noise and failure in system components. Furthermore, it explains the functional role of dendritic action potentials (dAPs) and proposes a mechanism of how mismatch signals can be generated in cortical circuits [14]. Its capacity benefits from sparsity in the activity, and therefore provides a highly energy efficient sequence learning and prediction mechanism [16].
The original formulation of the TM model is based on abstract models of neurons and synapses with discrete-time dynamics. Moreover, the way the network forms synapses during learning is difficult to reconcile with biology. Here, we propose a continuous-time implementation of the TM model derived from known biological principles such as spiking neurons, dAPs, lateral inhibition, spike-timing-dependent structural plasticity, and homeostatic control of synapse growth. This model successfully learns, predicts and replays high-order sequences, where the prediction of the upcoming element is not only dependent on the current element, but also on the history. Bringing the model closer to biology allows for testing its hypotheses based on experimentally accessible quantities such as synaptic connectivity, synaptic currents, transmembrane potentials, or spike trains. Reformulating the model in terms of continuous-time dynamics moreover enables us to address timing-related questions, such as the role of the sequence speed for the prediction performance and the replay speed.
The study is organized as follows: the Methods describe the task, the network model, and the performance measures. The Results illustrate how the interaction of the model’s components gives rise to context dependent predictions and sequence replay, and evaluate the sequence processing speed and prediction performance. The Discussion finally compares the spiking TM model to other biologically motivated sequence learning models, summarizes limitations, and provides suggestions for future model extensions.
Methods
In the following, we provide an overview of the task and the training protocol, the network model, and the task performance analysis. A detailed description of the model and parameter values can be found in Tables 1 and 2.
Table 1. Description of the network model.
Summary | |
Populations | excitatory neurons (), inhibitory neurons (), external spike sources (); and composed of M disjoint subpopulations and (k = 1, …, M) |
Connectivity |
|
Neuron model |
|
Synapse model | exponential or alpha-shaped postsynaptic currents (PSCs) |
Plasticity | homeostatic spike-timing dependent structural plasticity in excitatory-to-excitatory connections |
Table 2. Model and simulation parameters.
Name | Value | Description |
---|---|---|
Network | ||
N E | 2100 | total number of excitatory neurons |
N I | 14 | total number of inhibitory neurons |
number of excitatory subpopulations (= number of external spike sources) | ||
number of excitatory neurons per subpopulation | ||
number of inhibitory neurons per subpopulation | ||
ρ | 20 | (target) number of active neurons per subpopulation after learning = minimal number of coincident excitatory inputs required to trigger a spike in postsynaptic inhibitory neurons |
(Potential) Connectivity | ||
K EE | 420 | number of excitatory inputs per excitatory neuron (EE in-degree) |
probability of potential (excitatory) connections | ||
number of inhibitory inputs per excitatory neuron (EI in-degree) | ||
number of excitatory inputs per inhibitory neuron (IE in-degree) | ||
K II | 0 | number of inhibitory inputs per inhibitory neuron (II in-degree) |
Excitatory neurons | ||
τ m,E | 10 ms | membrane time constant |
τ ref,E | 10 ms | absolute refractory period |
C m | 250 pF | membrane capacity |
V r | 0.0 mV | reset potential |
θ E | 20 mV (predictive mode), 5 mV (replay mode) | somatic spike threshold |
I dAP | 200 pA | dAP current plateau amplitude |
τ dAP | 60 ms | dAP duration |
θ dAP | 59 pA (predictive mode), 41.3 pA (replay mode) | dAP threshold |
Inhibitory neurons | ||
τ m,I | 5 ms | membrane time constant |
τ ref,I | 2 ms | absolute refractory period |
C m | 250 pF | membrane capacity |
V r | 0.0 mV | reset potential |
θ I | 15 mV | spike threshold |
Synapse | ||
γ | 5 | number co-active presynaptic neurons required to trigger a dAP in the postsynaptic neuron |
W | 12.98 pA | weight of mature EE connections (EPSC amplitude) |
0.9 mV (predictive mode), 0.12 mV (replay mode) | weight of IE connections (EPSP amplitude) | |
J IE | 581.19 pA (predictive mode), 77.49 pA (replay mode) | weight of IE connections (EPSC amplitude) |
−40 mV | weight of EI connections (IPSP amplitude) | |
weight of EI connections (IPSC amplitude) | ||
22 mV | weight of EX connections (EPSP amplitude) | |
weight of EX connections (EPSC amplitude) | ||
τ EE | 5 ms | synaptic time constant of EE connections |
τ IE | 0.5 ms | synaptic time constant of IE connections |
τ EI | 1 ms | synaptic time constant of EI connections |
τ EX | 2 ms | synaptic time constant of EX connection |
d EE | 2 ms | delay of EE connections (dendritic) |
d IE | 0.1 ms | delay of IE connections |
d EI | {0.1, 0.2} ms | delay of EI connections (non-default value used in Figs 10 and 11) |
d EX | 0.1 ms | delay of EX connections |
Plasticity | ||
λ+ | 0.08 (sequence set I), 0.28 (sequence set II) | potentiation rate |
λ− | 0.0015 (sequence set I), 0.0061 (sequence set II) | depression rate |
θ P | 20 | synapse maturity threshold |
minimum permanence | ||
P max | 20 | maximum permanence |
P 0,min | 0 | minimal initial permanence |
P 0,max | 8 | maximal initial permanence |
τ + | 20 ms | potentiation time constant |
z* | 1 | target dAP activity |
λh | 0.014 (sequence set I), 0.024 (sequence set II) | homeostasis rate |
τ h | 440 ms (sequence set I), 1560 ms (sequence set II) | homeostasis time constant |
y i | 1 | depression decrement |
Δtmin | 4 ms | minimum time lag between pairs of pre- and postsynaptic spikes at which synapses are potentiated |
Δtmax | 2ΔT | maximum time lag between pairs of pre- and postsynaptic spikes at which synapses are potentiated |
Input | ||
L | 1 | number of subpopulations per sequence element = number of target subpopulations per spike source |
S | 2 (sequence set I), 6 (sequence set II) | number of sequences per set |
C | 4 (sequence set I), 5 (sequence set II) | number of elements per sequence |
A | 14 | alphabet length (total number of distinct sequence elements) |
ΔT | {2,…,40,…,90} ms | inter-stimulus interval |
ΔTseq | max(2.5ΔT, τdAP) | inter-sequence interval |
ΔTcue | 80 ms | inter-cue interval |
Simulation | ||
Δt | 0.1 ms | time resolution |
K | {80, 100} | number of training episodes |
Task and training protocol
In this study, we develop a neuronal architecture that can learn and process an ensemble of S sequences of ordered discrete items ζi,j with , i ∈ [1, …, S]. The length of sequence si is denoted by Ci. Throughout this study, the sequence elements ζi,j ∈ {A, B, C, …} are represented by Latin characters, serving as placeholders for arbitrary discrete objects or percepts, such as images, numbers, words, musical notes, or movement primitives (Fig 1A). The order of the sequence elements within a given sequence represents the temporal order of item occurrence.
The tasks to be solved by the network consist of
-
i)
predicting subsequent sequence elements in response to the presentation of other elements,
-
ii)
detecting unanticipated stimuli and generating a mismatch signal if the prediction is not met, and
-
iii)
autonomously replaying sequences in response to a cue signal after learning.
The architecture learns sequences in a continuous manner: the network is exposed to repeated presentations of a given ensemble of sequences (e.g., {A, D, B, E} and {F, D, B, C} in Fig 1B). In the prediction mode (task i) and ii)), there is no distinction between a “training” and a “testing” phase. At the beginning of the learning process, all presented sequence elements are unanticipated and do not lead to a prediction (diffuse shades in Fig 1B, left). As a consequence, the network generates mismatch signals (flash symbols in Fig 1B, left). After successful learning, the presentation of some sequence element leads to a prediction of the subsequent stimulus (colored arrows in Fig 1B). In case this subsequent stimulus does not match the prediction, the network generates a mismatch signal (red arrow and flash symbol in Fig 1B, right). The learning process is entirely unsupervised, i.e., the prediction performance does not affect the learning. As described in Sequence replay, the network can be configured into a replay mode where the network autonomously replays learned sequences in response to a cue signal (task iii)).
In general, the sequences in this study are “high-order” sequences, similar to those generated by a high-order Markov chain; the prediction of an upcoming sequence element requires accounting for not just the previous element, but for (parts of) the entire sequence history, i.e., the context. Sequences within a given set of training data can be partially overlapping; they may share certain elements or subsequences (such as in {A, D, B, E} and {F, D, B, C}). Similarly, the same sequence element (but not the first one, see Limitations and outlook) may occur multiple times within the same sequence (such as in {A, D, B, D}). Throughout this work, we use two sequence sets:
Sequence set I
For an illustration of the learning process and the network dynamics in the prediction (section Sequence learning and prediction) and in the replay mode (section Sequence replay), as well as for the investigation of the sequence processing speed (section Dependence of prediction performance on the sequence speed), we start with a simple set of two partially overlapping sequences s1 = {A, D, B, E} and s2 = {F, D, B, C} (see Fig 1B).
Sequence set II
For a more rigorous evaluation of the sequence prediction performance (section Prediction performance), we consider a set of S = 6 high-order sequences: s1 = {E, N, D, I, J}, s2 = {L, N, D, I, K}, s3 = {G, J, M, C, N}, s4 = {F, J, M, C, I}, s5 = {B, C, K, H, I}, s6 = {A, C, K, H, F}, each consisting of C = 5 elements. The complexity of this sequence ensemble is comparable to the one used in [14], but is more demanding in terms of the high-order context dependence.
Results for two additional sequence sets are summarized in the Supporting information. The set used in S2 Fig is composed of sequences with recurring first elements. In S3 Fig, we show results for longer sequences with a larger number of overlapping elements.
Network model
Algorithmic requirements
To solve the tasks outlined in Task and training protocol, the network model needs to implement a number of algorithmic components. Here, we provide an overview of these components and their corresponding implementations:
Learning and storage of sequences: in both the original and our model, sequences are represented by specific subnetworks embedded into the recurrent network. During the learning process, these subnetworks are carved out in an unsupervised manner by a form of structural Hebbian plasticity.
Context specificity: in our model, learning of high-order sequences is enabled by a sparse, random potential connectivity, and by a homeostatic regulation of synaptic growth.
Generation of predictions: neurons are equipped with a predictive state, implemented by a nonlinear synaptic integration mimicking the generation of dendritic action potentials (dAPs).
Mismatch detection: only few neurons become active if a prediction matches the stimulus. In our model, this sparsity is realized by a winner-take-all (WTA) dynamics implemented in the form of inhibitory feedback. In case of non-anticipated stimuli, the WTA dynamics cannot step in, thereby leading to a non-sparse activation of larger neuron populations.
Sequence replay: autonomous replay of learned sequences in response to a cue signal is enabled by increasing neuronal excitability.
In the following paragraphs, the implementations of these components and the differences between the original and our model are explained in more detail.
Network structure
The network consists of a population of NE excitatory (“E”) and a population of NI inhibitory (“I”) neurons. The neurons in are randomly and recurrently connected, such that each neuron in receives KEE excitatory inputs from other randomly chosen neurons in . Note that these “EE” connections are potential connections in the sense that they can be either “mature” (“effective”) or “immature”. Immature connections have no effect on target neurons (see below). In the neocortex, the degree of potential connectivity depends on the distance between the neurons [17]. It can reach probabilities as high as 90% for neighboring neurons, and decays to 0% for neurons that are farther apart. In this work, the connection probability is chosen such that the connectivity is sufficiently dense, allowing for the formation of specific subnetworks, and sufficiently sparse for increasing the network capacity (see paragraph “Constraints on potential connectivity” below). The excitatory population is subdivided into M non-overlapping subpopulations , each of them containing neurons with identical stimulus preference (“receptive field”; see below). Each subpopulation thereby represents a specific element within a sequence (Fig 2A and 2B). In the original TM model [14], a single sequence element is represented by multiple (L) subpopulations (“minicolumns”). For simplicity, we identify the number M of subpopulations with the number of elements required for a specific set of sequences, such that each sequence element is encoded by just one subpopulation (L = 1). All neurons within a subpopulation are recurrently connected to a subpopulation-specific inhibitory neuron . The inhibitory neurons in are mutually unconnected. The subdivision of excitatory neurons into stimulus-specific subpopulations defines how external inputs are fed to the network (see next paragraph), but does not affect the potential excitatory connectivity, which is homogeneous and not subpopulation specific.
External inputs
During the prediction mode, the network is driven by an ensemble of M external inputs, representing inputs from other brain areas, such as thalamic sources or other cortical areas. Each of these external inputs xk represents a specific sequence element (“A”, “B”, …), and feeds all neurons in the subpopulation with the corresponding stimulus preference. The occurrence of a specific sequence element ζi,j at time ti,j is modeled by a single spike xk(t) = δ(t − ti,j) generated by the corresponding external source xk. Subsequent sequence elements ζi,j and ζi,j+1 within a sequence si are presented with an inter-stimulus interval ΔT = ti,j+1 − ti,j. Subsequent sequences si and si+1 are separated in time by an inter-sequence time interval . During the replay mode, we present only a cue signal encoding for first sequence elements ζi,1 at times ti,1. Subsequent cues are separated in time with an inter-cue time interval ΔTcue = ti+1,1 − ti,1. In the absence of any other (inhibitory) inputs, each external input spike is strong enough to evoke an immediate response spike in all target neurons . Sparse activation of the subpopulations in response to the external inputs is achieved by a winner-take-all mechanism implemented in the form of inhibitory feedback (see Sequence learning and prediction).
Neuron and synapse model
In the original TM model [14], excitatory (pyramidal) neurons are described as abstract three-state systems that can assume an active, a predictive, or a non-active state. State updates are performed in discrete time. The current state is fully determined by the external input in the current time step and the network state in the previous step. Each TM neuron is equipped with a number of dendrites (segments), modeled as coincidence detectors. The dendrites are grouped into distal and proximal dendrites. Distal dendrites receive inputs from other neurons in the local network, whereas proximal dendrites are activated by external sources. Inputs to proximal dendrites have a large effect on the soma and trigger the generation of action potentials. Individual synaptic inputs to a distal dendrite, in contrast, have no direct effect on the soma. If the total synaptic input to a distal dendritic branch at a given time step is sufficiently large, the neuron becomes predictive. This dynamic mimics the generation of dendritic action potentials (dAPs), NMDA spikes [18–20]), which result in a long-lasting depolarization (∼50–500 ms) of the somata of neocortical pyramidal neurons.
In contrast to the original study, the model proposed here employs neurons with continuous-time dynamics. For all types of neurons, the temporal evolution of the membrane potential is given by the leaky integrate-and-fire model Eq (10). The total synaptic input current of excitatory neurons is composed of currents in distal dendritic branches, inhibitory currents, and currents from external sources. Inhibitory neurons receive only inputs from excitatory neurons in the same subpopulation. Individual spikes arriving at dendritic branches evoke alpha-shaped postsynaptic currents, see Eq (12). The dendritic current includes an additional nonlinearity describing the generation of dAPs: if the dendritic current IED exceeds a threshold θdAP, it is instantly set to a the dAP plateau current IdAP, and clamped to this value for a period of duration τdAP, see Eq (16). This plateau current leads to a long lasting depolarization of the soma (see Fig 3B). The dAP threshold θdAP is chosen such that the co-activation of γ neurons with mature connections to the target neuron reliably triggers a dAP. In this work, we use a single dendritic branch per neuron. However, the model could easily be extended to include multiple dendritic branches. External and inhibitory inputs to excitatory neurons as well as excitatory inputs to inhibitory neurons trigger exponential postsynaptic currents, see Eqs (13)–(15). Similar to the original implementation, an external input strongly depolarizes the neurons and causes them to fire. To this end, the external weights JEX are chosen to be supra-threshold (see Fig 3A). Inhibitory interactions implement the WTA described in Sequence learning and prediction. The weights JIE of excitatory synapses on inhibitory neurons are chosen such that the collective firing of a subset of ρ excitatory neurons in the corresponding subpopulation causes the inhibitory neuron to fire. The weights JEI of inhibitory synapses on excitatory neurons are strong such that each inhibitory spike prevents all excitatory neurons in the same subpopulation that have not generated a spike yet from firing. All synaptic time constants, delays and weights are connection-type specific (see Table 1).
Plasticity dynamics
Both in the original [14] and in our model, the lateral excitatory connectivity between excitatory neurons (EE connectivity) is dynamic and shaped by a Hebbian structural plasticity mechanism mimicking principles known from the neuroscience literature [21–25]. All other connections are static. The dynamics of the EE connectivity is determined by the time evolution of the permanences Pij (), representing the synapse maturity, and the synaptic weights Jij. Unless the permanence Pij exceeds a threshold θP, the synapse {j → i} is immature, with zero synaptic weight Jij = 0. Upon threshold crossing, Pij ≥ θP, the synapse becomes mature, and its weight is assigned a fixed value Jij = W (∀i, j). Overall, the permanences evolve according to a Hebbian plasticity rule: the synapse {j → i} is potentiated, i.e., Pij is increased, if the activation of the postsynaptic cell i is immediately preceded by an activation of the presynaptic cell j. Otherwise, the synapse is depressed, i.e., Pij is decreased. At the beginning of the learning process or during relearning, the activity in the individual subpopulations is non-sparse. Hebbian learning alone would therefore lead to the strengthening of all existing synapses between two subsequently activated subpopulations, irrespective of the context these two subpopulations participate in. After learning, the subsets of neurons that are activated by a sequence element recurring in different sequences would therefore largely overlap. As a consequence, it becomes harder to distinguish between different contexts (histories) based on the activation patterns of these subsets. The original TM model [14] avoids this loss of context sensitivity by restricting synaptic potentiation to a small subset of synapses between a given pair of source and target subpopulations: if there are no predictive target neurons, the original algorithm selects a “matched” neuron from the set of active postsynaptic cells as the one being closest to becoming predictive, i.e., the neuron receiving the largest number of synaptic inputs on a given dendritic branch from the set of active presynaptic cells (provided this number is sufficiently large). Synapse potentiation is then restricted to this set of matched neurons. In case there are no immature synapses, the “least used” neuron or a randomly chosen neuron is selected as the “matched” cell, and connected to the winner cell of the previously active subpopulation. Restricting synaptic potentiation to synapses targeting such a subset of “matched” neurons is difficult to reconcile with biology. It is known that inhibitory inputs targeting the dendrites of pyramidal cells can locally suppress backpropagating action potentials and, hence, synaptic potentiation [26]. A selection mechanism based on such local inhibitory circuits would however involve extremely fast synapses and require fine-tuning of parameters. The model presented in this work circumvents the selection of “matched” neurons and replaces this with a homeostatic mechanism controlled by the postsynaptic dAP rate. In the following, the specifics of the plasticity dynamics used in this study are described in detail.
Within the interval [Pmin,ij, Pmax], the dimensionless permanences Pij(t) evolve according to a combination of an additive spike-timing-dependent plasticity (STDP) rule [27] and a homeostatic component [28, 29]:
(1) |
At the boundaries Pmin,ij and Pmax, Pij(t) is clipped. While the maximum permanences Pmax are identical for all EE connections, the minimal permanences Pmin,ij are uniformly distributed in the interval [P0,min, P0,max] to introduce a form of persistent heterogeneity. The first term on the right-hand side of Eq (1) corresponds to the spike-timing dependent synaptic potentiation triggered by the postsynaptic spikes at times . Here, denotes the set of all postsynaptic spike times for which the time lag exceeds Δtmin for all presynaptic spikes . The indicator function ensures that the potentiation (and the homeostasis; see below) is restricted to time lags in the interval (Δtmin, Δtmax) to avoid a growth of synapses between synchronously active neurons belonging to the same subpopulation, and between neurons encoding for the first elements in different sequences; see Eq 17. Note that the potentiation update times lag the somatic postsynaptic spike times by the delay dEE, which is here interpreted as a purely dendritic delay [27, 30]. The potentiation increment is determined by the dimensionless potentiation rate λ+, and the spike trace xj(t) of the presynaptic neuron j, which is updated according to
(2) |
The trace xj(t) is incremented by unity at each spike time , followed by an exponential decay with time constant τ+. The potentiation increment ΔPij at time therefore depends on the temporal distance between the postsynaptic spike time and all presynaptic spike times (STDP with all-to-all spike pairing; [27]). The second term in Eq (1) represents synaptic depression, and is triggered by each presynaptic spike at times . The depression decrement yi = 1 is treated as a constant, independently of the postsynaptic spike history. The depression magnitude is parameterized by the dimensionless depression rate λ−. The third term in Eq (1) corresponds to a homeostatic control triggered by postsynaptic spikes at times . Its overall impact is parameterized by the dimensionless homeostasis rate λh. The homeostatic control enhances or reduces the synapse growth depending on the dAP trace zi(t) of neuron i, the low-pass filtered dAP activity updated according to
(3) |
Here, τh represents the homeostasis time constant, and the onset time of the kth dAP in neuron i. According to Eq (1), synapse growth is boosted if the dAP activity zi(t) is below a target dAP activity z*. Conversely, high dAP activity exceeding z* reduces the synapse growth (Fig 4). This homeostatic regulation of the synaptic maturity controlled by the postsynaptic dAP activity constitutes a variation of previous models [28, 29] describing ‘synaptic scaling’ [31–33]. It counteracts excessive synapse formation during learning driven by Hebbian structural plasticity. In addition, the combination of Hebbian plasticity and synaptic scaling can introduce a competition between synapses [28, 29]. Here, we exploit this effect to ensure that synapses are generated in a context specific manner, and thereby reduce the overlap between neuronal subpopulations activated by the same sequence element occurring in different sequences. To this end, the homeostasis parameters z* = 1 and τh are chosen such that each neuron tends to become predictive, i.e., generate a dAP, at most once during the presentation of a single sequence ensemble of total duration ((C − 1)ΔT + ΔTseq)S (see Network model). The time constant τh is hence adapted to the parameters of the task. For sequence sets I and II and the default inter-stimulus interval ΔT = 40 ms, it is set to τh = 440 ms and τh = 1560 ms, respectively. In section Dependence of prediction performance on the sequence speed, we study the effect of the sequence speed (inter-stimulus interval ΔT) on the prediction performance for a given network parameterization. For these experiments, τh = 440 ms is therefore fixed even though the inter-stimulus interval ΔT is varied.
The prefactor in Eq (1) ensures that all learning rates λ+, λ− and λh are measured in units of the maximum permanence Pmax.
Constraints on potential connectivity
The sequence processing capabilities of the proposed network model rely on its ability to form sequence specific subnetworks based on the skeleton provided by the random potential connectivity. On the one hand, the potential connectivity must not be too diluted to ensure that a subset of neurons representing a given sequence element can establish sufficiently many mature connections to a second subset of neurons representing the subsequent element. On the other hand, a dense potential connectivity would promote overlap between subnetworks representing different sequences, and thereby slow down the formation of context specific subnetworks during learning (see Sequence learning and prediction). Here, we therefore identify the minimal potential connection probability p guaranteeing the existence of network motifs with a sufficient degree of divergent-convergent connectivity.
Consider the subset of ρ excitatory neurons representing the jth sequence element ζij in sequence si (see Task and training protocol and Network model). During the learning process, the plasticity dynamics needs to establish mature connections from to a second subset of neurons in another subpopulation representing the subsequent element ζi,j+1. Each neuron in must receive at least c = ⌈θdAP/W⌉ inputs from to ensure that synchronous firing of the neurons in can evoke a dAP after synapse maturing. For a random, homogeneous potential connectivity with connection probability p, the probability of finding these c potential connections for some arbitrary target neuron is given by
(4) |
For a successful formation of sequence specific subnetworks during learning, the sparse subset of presynaptic neurons needs to recruit at least ρ targets in the set of nE neurons representing the subsequent sequence element (Fig 5A). The probability of observing such a divergent-convergent connectivity motif is given by
(5) |
Note that the above described motif does not require the size of the postsynaptic subset to be exactly ρ. Eq (5) constrains the parameters p, c, nE and ρ to ensure such motifs exist in a random network. Fig 5B illustrates the dependence of the motif probability u on the connection probability p for our choice of parameters nE, c, and ρ. For p ≥ 0.2, the existence of the divergent-convergent connectivity motif is almost certain (u ≈ 1). For smaller connection probabilities p < 0.2, the motif probability quickly vanishes. Hence, p = 0.2 constitutes a reasonable choice for the potential connection probability.
Network realizations and initial conditions
For every network realization, the potential connectivity and the initial permanences are drawn randomly and independently. All other parameters are identical for different network realizations. The initial values of all state variables are given in Tables 1 and 2.
Simulation details
The network simulations are performed in the neural simulator NEST [34] under version 3.0 [35]. The differential equations and state transitions defining the excitatory neuron dynamics are expressed in the domain specific language NESTML [36, 37] which generates the required C++ code for the dynamic loading into NEST. Network states are synchronously updated using exact integration of the system dynamics on a discrete-time grid with step size Δt [38]. The full source code for the implementation with a list of other software requirements is available at Zenodo: https://doi.org/10.5281/zenodo.5578212.
Task performance measures
To assess the network performance, we monitor the dendritic currents reporting predictions (dAPs) as well as the somatic spike times of excitatory neurons. To quantify the prediction error, we identify for each last element in a sequence si all excitatory neurons that have generated a dAP in the time interval , where and ΔT denote the time of the external input corresponding to the last sequence element and the inter-stimulus interval, respectively (see Task and training protocol and Network model). All subpopulations with at least ρ/2 neurons generating a dAP are considered “predictive”. The prediction state of the network is encoded in an M dimensional binary vector o, where ok = 1 if the kth subpopulation is predictive, and ok = 0 else. The
(6) |
is defined as the Euclidean distance between o and the binary target vector v representing the pattern of external inputs for each last element , normalized by the number L of subpopulations per sequence element. Furthermore, we assess the
(7) |
and the
(8) |
where Θ(⋅) denotes the Heaviside function. In addition to these performance measures, we monitor for each last sequence element the level of sparsity by measuring the ratio between the number of active neurons and the total number LnE of neurons representing this element. During learning, we expose the network repetitively to the same set {s1, …, sS} of sequences for a number of training episodes K. To obtain the total prediction performance in each episode, we average the prediction error, the false negative and false positive rates, as well as the level of sparsity across the set of sequences.
Results
Sequence learning and prediction
According to the Temporal Memory (TM) model, sequences are stored in the form of specific paths through the network. Prediction and replay of sequences correspond to a sequential sparse activation of small groups of neurons along these paths. Non-anticipated stimuli are signaled in the form of non-sparse firing of these groups. This subsection describes how the model components introduced in Network model interact and give rise to the network structure and behavior postulated by TM. For illustration, we here consider a simple set of two partly overlapping sequences {A, D, B, E} and {F, D, B, C} corresponding to the sequence set I (see Fig 1B).
The initial sparse, random and immature network connectivity (Fig 2A and 2B) constitutes the skeleton on which the sequence-specific paths will be carved out during the learning process. To guarantee a successful learning, this initial skeleton must be neither too sparse nor too dense (see Methods). Before learning, the presentation of a particular sequence element causes all neurons with the corresponding stimulus preference to reliably and synchronously fire a somatic action potential due to the strong, suprathreshold external stimulus (Fig 3A). All other subpopulations remain silent (see Fig 6A and 6B). The lateral connectivity between excitatory neurons belonging to the different subpopulations is subject to a form of Hebbian structural plasticity. Repetitive and consistent sequential presentation of sequence elements turns immature connections between successively activated subpopulations into mature connections, and hence leads to the formation of sequence-specific subnetworks (see Fig 2C and 2D). Synaptic depression prunes connections not supporting the learned pattern, thereby reducing the chance of predicting wrong sequence items (false positives).
During the learning process, the number of mature connections grows to a point where the activation of a certain subpopulation by an external input generates dendritic action potentials (dAPs), a “prediction”, in a subset of neurons in the subsequent subpopulation (blue neurons in Fig 6C). The dAPs generate a long-lasting depolarization of the soma (Fig 3B). When receiving an external input, these depolarized neurons fire slightly earlier as compared to non-depolarized (non-predictive) neurons (Fig 3A, 3B and 3D). If the number of predictive neurons within a subpopulation is sufficiently large, their advanced spikes (Fig 3C) initiate a fast and strong inhibitory feedback to the entire subpopulation, and thereby suppress subsequent firing of non-predictive neurons in this population (Fig 6C and 6D). Owing to this winner-take-all dynamics, the network generates sparse spiking in response to predicted stimuli, i.e., if the external input coincides with a dAP-triggered somatic depolarization. In the presence of a non-anticipated, non-predicted stimulus, the neurons in the corresponding subpopulation fire collectively in a non-sparse manner, thereby signaling a “mismatch”.
In the model presented in this study, the initial synapse maturity levels, the permanences, are randomly chosen within certain bounds. During learning, connections with a higher initial permanence mature first. This heterogeneity in the initial permanences enables the generation of sequence specific sparse connectivity patterns between subsequently activated neuronal subpopulations (Fig 2D). For each pair of sequence elements in a given sequence ensemble, there is a unique set of postsynaptic neurons generating dAPs (Fig 6D). These different activation patterns capture the context specificity of predictions. When exposing a network that has learned the two sequences {A, D, B, E} and {F, D, B, C} to the elements “A” and “F”, different subsets of neurons are activated in “D” and “B”. By virtue of these sequence specific activation patterns, stimulation by {A, D, B} or {F, D, B} leads to correct predictions “E” or “C”, respectively (Fig 6C–6F).
Heterogeneity in the permanences alone, however, is not sufficient to guarantee context specificity. The subsets of neurons activated in different contexts may still exhibit a considerable overlap. This overlap is promoted by Hebbian plasticity in the face of the initial non-sparse activity, which leads to a strengthening of connections to neurons in the postsynaptic population in an unspecific manner (Fig 7A and 7B). Moreover, the reoccurrence of the same sequence elements in different co-learned sequences initially causes higher firing rates of the neurons in the respective populations (“D” and “B” in Fig 7). As a result, the formation of unspecific connections would even be accelerated if synapse formation was driven by Hebbian plasticity alone. The model in this study counteracts this loss of context specificity by supplementing the plasticity dynamics with a homeostatic component, which regulates synapse growth based on the rate of postsynaptic dAPs. This form of homeostasis prevents the same neuron from becoming predictive multiple times within the same set of sequences, and thereby reduces the overlap between subsets of neurons activated within different contexts (Fig 7C and S4 Fig). To further aid the formation of context specific paths, the density of the initial potential connectivity skeleton is set close to the minimum value ensuring the existence of the connectivity motifs required for a faithful prediction (see Methods).
Prediction performance
To quantify the sequence prediction performance, we repetitively stimulate the network with the sequences in sequence set I (see Task and training protocol), and continuously monitor the prediction error, the false-positive and false-negative rates, as well as the fraction of active stimulated neurons as a measure of encoding sparsity (Fig 8; Task performance measures). To ensure the performance results are not specific to a single network, the evaluation is repeated for a number of randomly instantiated network realizations with different initial potential connectivities. At the beginning of the learning process, all neurons of a stimulated subpopulation collectively fire in response to the external input. Non-stimulated neurons remain silent. As the connectivity is still immature at this point, no dAPs are triggered in postsynaptic neurons, and, hence, no predictions are generated. As a consequence, the prediction error, the false-negative rate and the number of active neurons (in stimulated populations) are at their maximum, and the false positive rate is zero (Fig 8). During the first training episodes, the consistent collective firing of subsequently activated populations leads to the formation of mature connections as a result of the Hebbian structural plasticity. Upon reaching of a critical number of mature synapse, first dAPs (predictions) are generated in postsynaptic cells (in Fig 8, this happens after about 10 learning episodes). As a consequence, the false negative rate decreases, and the stimulus responses become more sparse. At this early phase of the learning, the predictions of upcoming sequence elements are not yet context specific (for sequence set I, non-sparse activity in “B” triggers a prediction in both “E” and “C”, irrespective of the context). Hence, the false-positive rate transiently increases. As the context specific connectivity is not consolidated at this point, more and more presynaptic subpopulations fail at triggering dAPs in their postsynaptic targets when they switch to sparse firing. Therefore, the false-positive rate decreases again, and the false-negative rate increases. In other words, there exists a negative feedback loop in the interim learning dynamics where the generation of predictions leads to an increase in sparsity which, in turn, causes prediction failures (and, hence, non-sparse firing). With an increasing number of training episodes, synaptic depression and homeostatic regulation increase context selectivity and thereby break this loop. Eventually, sparse firing of presynaptic populations is sufficient to reliably trigger predictions in their postsynaptic targets. For sequence set I, the total prediction error becomes zero and the stimulus responses maximally sparse after about 30 training episodes (Fig 8). For a time resolved visualization of the learning dynamics, see S1 Video.
Up to this point, we illustrated the model’s sequence learning dynamics and performance for a simple set of two sequences (sequence set I). In the following, we assess the network’s sequence prediction performance for a more complex sequence set (II) composed of five high-order sequences (see Task and training protocol), each consisting of five elements. This sequence set is comparable to the one used in [14], but contains a larger amount of overlap between sequences. The overall pattern of the learning dynamics resembles the one reported for sequence set I (Fig 9). The prediction error, the false-positive and false-negative rates as well as the sparsity measure vary more smoothly, and eventually converge at minimal levels after about 40 training episodes. To compare the spiking TM model with the original, non-spiking TM model, we repeat the experiment based on the simulation code provided in [14], see S1 Table. With our parameterization, the learning rates λ+ and λ− of the spiking model are by a factor of about 10 smaller than in the original model. As a consequence, learning sequence set II with the original model converges faster than with the spiking model (compare black and gray curves in Fig 9). The ratio in learning speeds, however, is not larger than about 2. Increasing the learning rates, i.e., the permanence increments, would speed up the learning process in the spiking model, but bears the risk that a large fraction of connections mature simultaneously. This would effectively overwrite the permanence heterogeneity which is essential to form context specific connectivity patterns (see Sequence learning and prediction). As a result, the network performance would decrease. The original model avoids this problem by limiting the number of potentiated synapses in each update step (see “Plasticity dynamics” in Network model).
In sequence sets I and II, the maximum sequence order is 2 and 3, respectively. For the two sequences {E, N, D, I, J} and {L, N, D, I, K} in sequence set II, for example, predicting element “J” after activation of “I” requires remembering the element “E”, which occured three steps back into the past. The TM model can cope with sequences of much higher order. Each sequence element in a particular context activates a specific pattern, i.e., a specific subset of neurons. The number of such patterns that can be learned is determined by the size of each subpopulation and the sparsity [39]. In a sequence with repeating elements, such as {ABBBBBC}, the maximum order is limited by this number. Without repeating elements, the order could be arbitrarily high provided the number of subpopulations matches or exceeds the number of distinct characters. In S3 Fig, we demonstrate successful learning of two sequences {A, D, B, G, H, I, J, K, L, M, N, E}, {F, D, B, G, H, I, J, K, L, M, N, C} of order 10.
Dependence of prediction performance on the sequence speed
The reformulation of the original TM model in terms of continuous-time dynamics allows us to ask questions related to timing aspects. Here, we investigate the sequence processing speed by identifying the range of inter-stimulus intervals ΔT that permit a successful prediction performance (Fig 10). The timing of the external inputs affects the dynamics of the network in two respects. First, reliable predictions of sequence elements can only be made if the time interval ΔT between two consecutive stimulus presentations is such that the second input coincides with the somatic depolarization caused by the dAP triggered by the first stimulus. Second, the formation of sequence specific connections by means of the spike-timing-dependent structural plasticity dynamics depends on ΔT.
If the external input does not coincide with the somatic dAP depolarization, i.e., if ΔT is too small or to large, the respective target population responds in a non-sparse, non-selective manner (mismatch signal; Fig 10C), and in turn, generates false positives (Fig 10B). For small ΔT, the external stimulus arrives before the dAP onset, i.e., before it is predicted. In consequence, the false negative rate is high. For large ΔT, the false negative rate remains low as the network is still generating predictions (Fig 10B). The inter-stimulus interval ΔT in addition affects the formation of sequence specific connections due to the dependence of the plasticity dynamics on the timing of pre- and postsynaptic spikes, see Eqs (1) and (2). Larger ΔT results in smaller permanence increments, and thereby a slow-down of the learning process (red curve in Fig 10A).
Taken together, the model predicts a range of optimal inter-stimulus interval ΔT (Fig 10A). For our choice of network parameters, this range spans intervals between 10 ms and 75 ms. The lower bound depends primarily on the synaptic time constant τEE, the spike transmission delay dEE, and the membrane time constant τm. The upper bound is mainly determined by the dAP plateau duration τdAP.
Sequence replay
So far, we studied the network in the predictive mode, where the network is driven by external inputs and generates predictions of upcoming sequence elements. Another essential component of sequence processing is sequence replay, i.e., the autonomous generation of sequences in response to a cue signal (see Task and training protocol). After successful learning, the network model presented in this study is easily configured into the replay mode by increasing the neuronal excitability, such that the somatic depolarization caused by a dAP alone makes the neuron fire a somatic spike. Here, this is implemented by lowering the somatic spike threshold θE of the excitatory neurons. In the biological system, this increase in excitability could, for example, be caused by the effect of neuromodulators [40, 41], additional excitatory inputs from other brain regions implementing a top-down control, e.g, attention [42, 43], or propagating waves during sleep [44, 45].
The presentation of the first sequence element activates dAPs in the subpopulation corresponding to the expected next element in a previously learned sequence. Due to the reduced firing threshold in the replay mode, the somatic depolarization caused by these dAPs is sufficient to trigger somatic spikes during the rising phase of this depolarization. These spikes, in turn, activate the subsequent element. This process repeats, such that the network autonomously reactivates all sequence elements in the correct order, with the same context specificity and sparsity level as in the predictive mode (see Fig 11A and 11B). The latency between the activation of subsequent sequence elements is determined by the spike transmission delay dEE, the synaptic time constant τEE, the membrane time constant τm,E, the synaptic weights JEE,ij, the dAP current plateau amplitude IdAP, and the somatic firing threshold θE. For sequences that can be successfully learned (see previous section), the time required for replaying the entire sequence is independent of the inter-stimulus interval ΔT employed during learning (Fig 11C).
As shown in the previous section, sequences cannot be learned if the inter-stimulus interval ΔT is too small or too large. For small ΔT, connections between subpopulations corresponding to subsequent elements are strongly potentiated by the Hebbian plasticity due to the consistent firing of pre- and postsynaptic populations during the learning process. The network responses are, however, non-sparse, as the winner-take-all mechanism cannot be invoked during the learning (Fig 10C). In the replay mode, sequences are therefore replayed in a non-sparse and non-context specific manner (left gray region in Fig 11C). Similarly, connections between subsequent populations are slowly potentiated for very large ΔT. With sufficiently long learning, sequences can still be replayed in the right order, but the activity is non-sparse and therefore not context specific (right gray region in Fig 11C).
Discussion
Summary
In this work, we reformulate the Temporal Memory (TM) model [14] in terms of biophysical principles and parameters. We replace the original discrete-time neuronal and synaptic dynamics with continuous-time models with biologically interpretable parameters such as membrane and synaptic time constants and synaptic weights. We further substitute the original plasticity algorithm with a more biologically plausible mechanism, relying on a form of Hebbian structural plasticity, homeostatic control, and sparse random connectivity. Moreover, our model implements a winner-take-all dynamics based on lateral inhibition that is compatible with the continuous-time neuron and synapse models. We show that the revised TM model supports successful learning and processing of high-order sequences with a performance similar to the one of the original model [14].
A new aspect that we investigated in the context of our work is sequence replay. After learning, the model is able to replay sequences in response to a cue signal. The duration of sequence replay is independent of the sequence speed during training, and determined by the intrinsic parameters of the network. In general, sequence replay is faster than the sequence presentation during learning, consistent with sequence compression and fast replay observed in hippocampus [46–48] and neocortex [6, 49].
Finally, we identified the range of possible sequence speeds that guarantee a successful learning and prediction. Our model predicts an optimal range of processing speeds (inter-stimulus intervals) with lower and upper bounds constrained by neuronal and synaptic parameters (e.g., firing threshold, neuronal and synaptic time constants, coupling strengths, potentiation time constants). Within this range, the number of required training episodes is proportional to the inter-stimulus interval ΔT.
Relationship to other models
The model presented in this work constitutes a recurrent, randomly connected network of neurons with predefined stimulus preferences. The model learns sequences in an unsupervised manner using local learning rules. This is in essence similar to several other spiking neuronal network models for sequence learning [9–12, 50]. The new components employed in this work are dendritic action potentials (dAPs) and Hebbian structural plasticity. We use structural plasticity to be as close as possible to the original model, and Hebbian forms of this are also known from the literature [21, 22, 25]. However, preliminary results show that classical (non-structural) spike-timing-dependent plasticity (STDP) can yield similar performance (see S1 Fig). Dendritic action potentials are instrumental for our model for two reasons. First, they effectively lower the threshold for coincidence detection and thereby permit a reliable and robust propagation of sparse activity [51, 52]. In essence, our model bears similarities to the classical synfire chain [53], one difference being that our mature network is not a simple feedforward network but has an abundance of recurrent connections. As shown in [54], a stable propagation of synchronous activity requires a minimal number of neurons in each synfire group. Without active dendrites, this minimal number is in the range of ∼100 for plausible single-cell and synaptic parameters. In our (and in the original TM) model, coincidence detection happens in the dendrites. The number of presynaptic spikes needed to trigger a dAP is small, of the order of 10 [55–57]. This helps to reduce redundancy (only a small number of neurons needs to become active) and to increase the capacity of the network (the number of different patterns that can be learned is increased with pattern sparsity; [39]). Second, dAPs equip neurons with a third type of state (next to the quiescent and the firing state): the predictive state, i.e., a long lasting (∼ 50–200 ms) strong depolarization of the soma. Due to the prolonged depolarization of the soma, the inter-stimulus interval can be much larger than the synaptic time constants and delays. An additional benefit of dAPs, which is not exploited in the current version of our model, is that they equip individual neurons with more possible states if they comprise more than one dendritic branch. Each branch constitutes an independent pattern detector. The response of the soma may depend on the collective predictions in different dendritic branches. A single neuron could hence perform the types of computations that are usually assigned to multilayer perceptrons, i.e., small networks [58, 59].
Similar to a large class of other models in the literature, the TM network constitutes a recurrent network in the sense that the connectivity before and after learning forms loops at the subpopulation level. Recurrence in the immature connectivity permits the learning of arbitrary sequences without prior knowledge of the input data. In particular, recurrent connections enable the learning of sequences with repeating elements (such as in {A, B, B, C} or {A, B, C, B}). Further, bidirectional connections between subpopulations are needed to learn sequences where pairs of elements occur in different orders (such as in {A, B, C}, {D, C, B}). Apart from providing the capability to learn sequences with all possible combinations of sequence elements, recurrent connections play no further functional role in the current version of the TM model. They may, however, become more important for future versions of the model enabling the learning of sequence timing and duration (see below).
Most of the existing models have been developed to replay learned sequences in response to a cue signal. The TM model can perform this type of pattern completion, too. In addition, it can act as a quiet, sparsely active observer of the world that becomes highly active only in the case of unforeseen, non-anticipated events. In this work, we didn’t directly analyze the network’s mismatch detection performance. However, this could be easily achieved by equipping each population with a “mismatch” neuron that fires if a certain fraction of neurons in the population fires (threshold detectors). In our model, predicted stimuli result in sparse firing due to inhibitory feedback (WTA). For unpredicted stimuli, this feedback is not effective, resulting in non-sparse firing indicating a mismatch. In [60], a similar mechanism is employed to generate mismatch signals for novel stimuli. In this study, the strength of the inhibitory feedback needs to be learned by means of inhibitory synaptic plasticity. In our model, the WTA mechanism is controlled by the predictions (dAPs) and implemented by static inhibitory connections. Furthermore, the model in [60] can learn a set of elements, but not the order of these elements in the sequence.
In contrast to other sequence learning models [9, 11], our model is not able to learn an element specific timing and duration of sequence elements. The model in [9] relies on a clock network, which activates sequence elements in the correct order and with the correct timing. With this architecture, different sequences with different timings would require separate clock networks. Our model learns both sequence contents and order for a number of sequences without any auxiliary network. In an extension of our model, the timing of sequence element could be learned by additional plastic recurrent connections within each subpopulation. The model in [11] can learn and recall higher-order sequences with limited history by means of an additional reservoir network with sparse readout. The TM model presents a more efficient way of learning and encoding the context in high-order sequences, without prior assignment of context specificity to individual neuron populations [9], and without additional network components (such as reservoir networks in [11]).
An important sequence processing component that is not addressed in our work is the capability of identifying recurring sequences within a long stream of inputs. In the literature, this process is referred to as chunking, and constitutes a form of feature segmentation [3]. Sequence chunking has been illustrated, for example, in [61, 62]. Similar to our model, the network model in [61, 62] is composed of neurons with dendritic and somatic compartments, with the dendritic activity signaling a prediction of somatic spiking. Recurrent connections in their model improve the context specificity of neuronal responses, and thereby permit a context dependent feature segmentation. The model can learn high order sequences, but the history is limited. Although not explicitly tested here, our model is likely to be able to perform chunking if sequences are presented randomly across trials and without breaks. If the order of sequences is not systematic across trials, connections between neurons representing different sequences are not strengthened by spike-timing-dependent potentiation. Consecutive sequences are therefore not merged and remain distinct.
An earlier spiking neural network version of the HTM model has already been devised in [63]. It constitutes a proof-of-concept study demonstrating that the HTM model can be ported to an analog-digital neuromorphic hardware system. It is restricted to small simplistic sequences and does not address the biological plausibility of the TM model. In particular, it does not offer a solution to the question of how the model can perform online learning by known biological ingredients. Our study delivers a solution for this based on local plasticity rules and permits a direct implementation on a neuromorphic hardware system.
Limitations and outlook
The model developed in this study serves as a proof of concept demonstrating that the TM algorithm proposed in [14] can be implemented using biological ingredients. While it is still fairly simplistic, it may provide the basis for a number of future extensions.
Our results on the sequence processing speed revealed that the model presented here can process fast sequences with inter-stimulus intervals ΔT up to ∼75 ms. This range of processing speeds is relevant in many behavioral contexts such as motor generation, vision (saccades), music perception and generation, language, and many others [64]. However, slow sequences with inter-stimulus intervals beyond several hundreds of milliseconds cannot be learned by this model with biologically plausible parameters. This is problematic as behavioral time scales are often larger [64, 65]. By increasing the duration τdAP of the dAP plateau, the upper bound for ΔT could be extended to 500 ms, and maybe beyond [66]. However, for such long intervals, the synaptic potentiation would be very slow, unless the time constant τ+ of the structural STDP is increased and the depression rate λ− is adapted accordingly. Furthermore, while our model explains the fast replay observed in the hippocampus and cortex, it is not able to learn an element specific timing and duration of sequence elements [5, 67, 68]. This could be overcome by equipping the model with a working memory mechanism, which maintains the activity of the subpopulations for behaviorally relevant time scales [9, 69].
In the current version of the model, the number of subpopulations, the number of neurons within each subpopulation, the number of dendritic branches per neuron, as well as the number of synapses per neuron are far from realistic [14]. The number of sequences that can be successfully learned in this network is hence rather small. In addition, the current work is focusing on sequence processing at a single abstraction level, not accounting for a hierarchical network and task structure with both bottom-up and top-down projections. A further simplification in this work is that the lateral inhibition within a subpopulation is mediated by a single interneuron with unrealistically strong and fast connections to and from the pool of excitatory neurons. In future versions of this model, this interneuron could be replaced by a recurrently connected network of inhibitory neurons, thereby permitting more realistic weights, and simultaneously speeding up the interaction between inhibitory and excitatory cells by virtue of the fast-tracking property of such networks [70]. Similarly, the external inputs in our model are represented by single spikes, which are passed to the corresponding target population by a strong connection, and thereby lead to an immediate synchronous spike response. Replacing each external input by a population of synchronously firing neurons would be a more realistic scenario without affecting the model dynamics. The external neurons could even fire in a non-synchronous, rate modulated fashion, provided the spike responses of the target populations remain nearly synchronous and can coincide with the dAP-triggered somatic depolarization (see S6 Fig). The current version of the model relies on a nearly synchronous immediate response to ensure that a small set of (∼ 20) active neurons can reliably trigger postsynaptic dAPs, and that the predictive neurons (those depolarized by the dAPs) consistently fire earlier as compared to the non-predictive neurons, such that they can be selected by the WTA dynamics. Non-synchronous responses could possibly lead to a reliable generation of dAPs in postsynaptic neurons, but would require large active neuron populations (loss of sparsity) or unrealistically strong synaptic weights. The temporal separation between predictive and non-predictive neurons becomes harder for non-synchronous spiking. In future versions of the model, it could potentially be achieved by increasing the dAP plateau potential, and simultaneously equipping the excitatory neurons with a larger membrane time constant, such that non-depolarized neurons need substantially longer to reach the spike threshold. Increasing the dAP plateau potential, however, makes the model more sensitive to background noise (see below). Note that, in our model, only the immediate initial spike response needs to be synchronous. After successfully triggering the WTA circuit, the winning neurons could –in principle– continue firing in an asynchronous manner (for example, due the working-memory dynamics mentioned above). Similarly, long lasting or tonic external inputs could lead to repetitive firing of the neurons in the TM network. As long as these repetitive responses remain nearly synchronous, the network performance is likely to be preserved.
In the predictive mode, the statistics of the spiking activity generated by our model is primarily determined by the temporal structure of the external inputs. Upon presentation of a sequence element, a specific subset of excitatory neurons fires a single volley of synchronous spikes. If the stimulus is predicted, this subset is small. The spike response is therefore highly sparse both in time and space, in line with experimental findings [71]. For simplicity and illustration, the sequences in this study are presented in a serial manner with fixed order, and fixed inter-sequence and inter-element (inter-stimulus) intervals. As a consequence, the single-neuron spike responses are highly regular. The in-vivo spiking activity in cortical networks, in contrast, exhibits a high degree of irregularity [72]. A more natural presentation of sequences with irregular order and timing trivially leads to more irregular spike responses in our model. As long as the inter-stimulus intervals fall into the range depicted in Fig 10, the model can learn and predict irregular sequences. Spiking activity in the cortex is not only irregular, but also fairly asynchronous in the sense that the average level of synchrony for randomly chosen pairs of neurons is low [73, 74]. This, however, is not necessarily the case for any subset of neurons and at any point in time. It is well known that cortical neurons can systematically synchronize their firing with millisecond precision in relation to behaviorally relevant events (see, e.g., [75]). As demonstrated in [76], synchronous firing of small subsets of neurons may easily go unnoticed in the presence of subsampling. The model proposed in this study relies on (near) synchronous firing of small subsets of neurons. In cases where the model processes large sets of sequences in parallel, this synchrony will hardly be observable if only a small fraction of neurons is monitored (see S5 Fig). After learning, different sequences are represented by distinct subnetworks with little overlap. Hence, the network can process multiple sequences at the same time with little interference between subnetworks. The model could even learn multiple sequences in parallel, provided there is no systematic across-trial dependency between the sequences presented simultaneously. We dedicate the task of testing these ideas to future studies. While the synchrony predicted by the TM model may hardly be observable in experimental data suffering from strong subsampling, the predicted patterns of spikes could be identifiable using methods accounting for both spatial and temporal dependencies in the spike data [76–78]. There are other factors that may contribute to a more natural spiking activity in extended versions of the model. First, equipping the model with a working memory mechanism enabling the learning of slow sequences and sequence timing (see above) would likely lead to sustained asynchronous irregular firing. Second, replacing the inhibitory neurons by recurrent networks of inhibitory neurons (see above) would generate asynchronous irregular activity in the populations of inhibitory neurons and thereby contribute variability in the spike responses of the excitatory neurons. Third, the model proposed here may constitute a module embedded into a larger architecture and receive irregular inputs from other components. As shown in the supplementary S6 and S7 Figs, the spiking activity and the prediction performance of the TM model are robust with respect to low levels of synaptic background activity, and, hence, membrane potential fluctuations reminiscent of those observed in vivo [79]. For an increasing level of noise, the learning speed decreases. For high noise levels leading to additional, non-task related background spikes, the dAP triggered plateau depolarization is overwritten, such that the WTA dynamics fails at selecting predictive neurons, ultimately leading to a loss of context specificity in the responses. Hence, the prediction performance degrades for large noise amplitudes. A potential application of introducing background noise is to allow the network to perform probabilistic computations [80], such as replaying sequences in the presence of ambiguous cues.
Similar to the original TM model, the response of the population representing the first element in a sequence is non-sparse, indicating that the first sequence element is not anticipated and can therefore not be predicted. If a given first sequence element reoccurs within the same sequence (say, “A” in {A, B, A, C}) or in other sequences (e.g., in {D, E, A, F}), the non-sparse response of the respective population to a first sequence element leads to a simultaneous prediction of all possible subsequent elements, i.e., the generation of false positives. These false predictions would lead to a pruning of functional synapses as a response of the homeostatic regulation to the increased dAP activity. This could be overcome by three possible mechanisms: a) synaptic normalization avoiding excessive synapse growth [81, 82], b) removing breaks between sequences, or c) sparse, sequence specific firing of subpopulations representing first elements. Results of applying the last mechanism are shown in S2 Fig, where dAPs are externally activated in random subsets of neurons in the populations representing first elements. In a more realistic hierarchical network, a similar effect could be achieved by top-down projections from a higher level predicting sequences of sequences.
In the original model, synapses targeting silent postsynaptic cells are depressed, even if the presynaptic neuron is inactive. This pruning process, the freeing of unused synaptic resources, increases the network capacity while ensuring context sensitivity. According to the structural plasticity dynamics employed in our study, synapse depression is bound to presynaptic spiking, similar to other implementations of (non-structural) STDP [30]. As a consequence, strong connections originating from silent presynaptic neurons are not depressed (dark gray dots in Fig 2D). This may complicate or slow down the learning of new sequences, and could be overcome by synaptic normalization.
For the dAP-rate homeostasis used in this study, the target dAP rate is set to one to make sure that each neuron contributes at most one dAP during each training episode. As a consequence, the time constant of the dAP-rate homeostasis is adapted to the duration of a training episode, which is in the range of few seconds in this work. We are not aware of any biological mechanism that could account for such an adaptation. dAP-rate homeostasis is mediated by the intracellular calcium concentration, which, in turn, controls the synthesis of synaptic receptors, and hence, the synaptic strength. It is therefore known to be rather slow, acting on timescales of many minutes, hours or days [32, 33]. It is unclear to what extent the use of long homeostatic time constants and increased dAP target rates would alter the model performance. Alternatively, the dAP-rate homeostasis could be replaced by other mechanisms such as synaptic normalization.
Conclusion
Our work demonstrates that the principle mechanisms underlying sequence learning, prediction, and replay in the TM model can be implemented using biologically plausible ingredients. By strengthening the link to biology, our implementation permits a more direct evaluation of the TM model predictions based on electrophysiological and behavioral data. Furthermore, this implementation allows for a direct mapping of the TM model on neuromorphic hardware systems.
Supporting information
Acknowledgments
The authors thank Rainer Waser for valuable discussions on the project, Charl Linssen for help with the neuron model implementation using NESTML, Abigail Morrison for suggestions on the plasticity dynamics, Danylo Ulianych and Dennis Terhorst for code review, and Sebastian B.C. Lehmann for assistance with graphic design. All network simulations were carried out with NEST (http://www.nest-simulator.org).
Data Availability
The documented workflow and source code necessary to reproduce our findings are provided online at: https://doi.org/10.5281/zenodo.5578212.
Funding Statement
This project was funded by the Helmholtz Association Initiative and Networking Fund (project number SO-092, Advanced Computing Architectures) [YB, DJW, MD, TT], and the European Union’s Horizon 2020 Framework Programme for Research and Innovation under the Specific Grant Agreement No.~785907 (Human Brain Project SGA2) [YB, MD, TT] and No.~945539 (Human Brain Project SGA3) [YB, MD, TT]. Open access publication funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation, 491111487) [YB, MD, TT]. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1. Lashley KS. The problem of serial order in behavior. vol. 21. Bobbs-Merrill; 1951. [Google Scholar]
- 2. Hawkins J, Blakeslee S. On intelligence: How a new understanding of the brain will lead to the creation of truly intelligent machines. Macmillan; 2007. [Google Scholar]
- 3. Dehaene S, Meyniel F, Wacongne C, Wang L, Pallier C. The neural representation of sequences: from transition probabilities to algebraic patterns and linguistic trees. Neuron. 2015;88(1):2–19. doi: 10.1016/j.neuron.2015.09.019 [DOI] [PubMed] [Google Scholar]
- 4. Clegg BA, DiGirolamo GJ, Keele SW. Sequence learning. Trends Cogn Sci. 1998;2(8):275–281. doi: 10.1016/S1364-6613(98)01202-9 [DOI] [PubMed] [Google Scholar]
- 5. Gavornik JP, Bear MF. Learned spatiotemporal sequence recognition and prediction in primary visual cortex. Nat Neurosci. 2014;17(5):732. doi: 10.1038/nn.3683 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Xu S, Jiang W, Poo MM, Dan Y. Activity recall in a visual cortical ensemble. Nat Neurosci. 2012;15(3):449–455. doi: 10.1038/nn.3036 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Hahnloser RH, Kozhevnikov AA, Fee MS. An ultra-sparse code underlies the generation of neural sequences in a songbird. Nature. 2002;419(6902):65–70. doi: 10.1038/nature00974 [DOI] [PubMed] [Google Scholar]
- 8. Harvey CD, Coen P, Tank DW. Choice-specific sequences in parietal cortex during a virtual-navigation decision task. Nature. 2012;484(7392):62–68. doi: 10.1038/nature10918 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Maes A, Barahona M, Clopath C. Learning spatiotemporal signals using a recurrent spiking network that discretizes time. PLOS Comput Biol. 2020;16(1):e1007606. doi: 10.1371/journal.pcbi.1007606 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Klos C, Miner D, Triesch J. Bridging structure and function: A model of sequence learning and prediction in primary visual cortex. PLOS Comput Biol. 2018;14(6):e1006187. doi: 10.1371/journal.pcbi.1006187 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Cone I, Shouval HZ. Learning precise spatiotemporal sequences via biophysically realistic learning rules in a modular, spiking network. eLife. 2021;10:e63751. doi: 10.7554/eLife.63751 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Klampfl S, Maass W. Emergence of dynamic memory traces in cortical microcircuit models through STDP. J Neurosci. 2013;33(28):11515–11529. doi: 10.1523/JNEUROSCI.5044-12.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Hawkins J, Ahmad S, Dubinsky D, et al. Cortical learning algorithm and hierarchical temporal memory. Numenta Whitepaper. 2011;1(68):2. [Google Scholar]
- 14. Hawkins J, Ahmad S. Why neurons have thousands of synapses, a theory of sequence memory in neocortex. Front Neural Circuits. 2016;10:23. doi: 10.3389/fncir.2016.00023 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Cui Y, Ahmad S, Hawkins J. Continuous online sequence learning with an unsupervised neural network model. Neural Comput. 2016;28(11):2474–2504. doi: 10.1162/NECO_a_00893 [DOI] [PubMed] [Google Scholar]
- 16.Ahmad S, Hawkins J. Properties of sparse distributed representations and their application to hierarchical temporal memory. 10.48550/arXiv.1503.07469 [q-bio.NC]; 2015. [DOI]
- 17. Stepanyants A, Hirsch JA, Martinez LM, Kisvárday ZF, Ferecskó AS, Chklovskii DB. Local Potential Connectivity in Cat Primary Visual Cortex. Cereb Cortex. 2007;18(1):13–28. Available from: doi: 10.1093/cercor/bhm027 [DOI] [PubMed] [Google Scholar]
- 18. Antic SD, Zhou WL, Moore AR, Short SM, Ikonomu KD. The decade of the dendritic NMDA spike. J Neurosci Res. 2010;88(14):2991–3001. doi: 10.1002/jnr.22444 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Schiller J, Major G, Koester HJ, Schiller Y. NMDA spikes in basal dendrites of cortical pyramidal neurons. Nature. 2000;404(6775):285–289. Available from: doi: 10.1038/35005094 [DOI] [PubMed] [Google Scholar]
- 20. Larkum ME, Nevian T, Sandler M, Polsky A, Schiller J. Synaptic Integration in Tuft Dendrites of Layer 5 Pyramidal Neurons: A New Unifying Principle. Science. 2009;325(5941):756–760. Available from: doi: 10.1126/science.1171958 [DOI] [PubMed] [Google Scholar]
- 21. Liao D, Hessler NA, Malinow R. Activation of postsynaptically silent synapses during pairing-induced LTP in CA1 region of hippocampal slice. Nature. 1995;375(6530):400–404. doi: 10.1038/375400a0 [DOI] [PubMed] [Google Scholar]
- 22. Wu GY, Malinow R, Cline HT. Maturation of a Central Glutamatergic Synapse. Science. 1996;274(5289):972–976. Available from: doi: 10.1126/science.274.5289.972 [DOI] [PubMed] [Google Scholar]
- 23. Lüscher C, Noicoll RA, Malenka RC, Muller D. Synaptic plasticity and dynamic modulation of the postsynaptic membrane. Nat Neurosci. 2000;3(6):545–550. doi: 10.1038/75714 [DOI] [PubMed] [Google Scholar]
- 24. Nevian T, Sakmann B. Single Spine Ca2+ Signals Evoked by Coincident EPSPs and Backpropagating Action Potentials in Spiny Stellate Cells of Layer 4 in the Juvenile Rat Somatosensory Barrel Cortex. J Neurosci. 2004;24(7):1689–1699. doi: 10.1523/JNEUROSCI.3332-03.2004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Deger M, Helias M, Rotter S, Diesmann M. Spike-Timing Dependence of Structural Plasticity Explains Cooperative Synapse Formation in the Neocortex. PLOS Comput Biol. 2012;8(9):e1002689. doi: 10.1371/journal.pcbi.1002689 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Müllner FE, Wierenga CJ, Bonhoeffer T. Precision of inhibition: dendritic inhibition by individual GABAergic synapses on hippocampal pyramidal cells is confined in space and time. Neuron. 2015;87(3):576–589. doi: 10.1016/j.neuron.2015.07.003 [DOI] [PubMed] [Google Scholar]
- 27. Morrison A, Diesmann M, Gerstner W. Phenomenological models of synaptic plasticity based on spike-timing. Biol Cybern. 2008;98(6):459–478. Available from: doi: 10.1007/s00422-008-0233-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Abbott LF, Nelson SB. Synaptic plasticity: taming the beast. Nat Neurosci. 2000;3:1178–1183. Available from: doi: 10.1038/81453 [DOI] [PubMed] [Google Scholar]
- 29. Tetzlaff C, Kolodziejski C, Timme M, Wörgötter F. Synaptic scaling in combination with many generic plasticity mechanisms stabilizes circuit connectivity. Front Comput Neurosci. 2011;5:47. doi: 10.3389/fncom.2011.00047 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Morrison A, Aertsen A, Diesmann M. Spike-Timing Dependent Plasticity in Balanced Random Networks. Neural Comput. 2007;19:1437–1467. Available from: doi: 10.1162/neco.2007.19.6.1437 [DOI] [PubMed] [Google Scholar]
- 31. Turrigiano GG, Leslie KR, Desai NS, Rutherford LC, Nelson SB. Activity-dependent scaling of quantal amplitude in neocortical neurons. Nature. 1998;391:892–896. doi: 10.1038/36103 [DOI] [PubMed] [Google Scholar]
- 32. Turrigiano GG, Nelson SB. Homeostasic plasticity in the developing nervous system. Nat Rev Neurosci. 2004;5:97–107. doi: 10.1038/nrn1327 [DOI] [PubMed] [Google Scholar]
- 33. Turrigiano GG. The self-tuning neuron: synaptic scaling of excitatory synapses. Cell. 2008;135(3):422–435. doi: 10.1016/j.cell.2008.10.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Gewaltig MO, Diesmann M. NEST (NEural Simulation Tool). Scholarpedia J. 2007;2(4):1430. Available from: doi: 10.4249/scholarpedia.1430 [DOI] [Google Scholar]
- 35.Hahne J, Diaz S, Patronis A, Schenck W, Peyser A, Graber S, et al. NEST 3.0. Zenodo; 2021. Available from: 10.5281/zenodo.4739103. [DOI]
- 36.Plotnikov D, Blundell I, Ippen T, Eppler JM, Rumpe B, Morrison A. NESTML: a modeling language for spiking neurons. In: Oberweis A, Reussner R, editors. Modellierung 2016. vol. P-254 of Lecture Notes in Informatics (LNI). Modellierung 2016, Karlsruhe (Germany), 17 Mar 2016–19 Mar 2016. Gesellschaft für Informatik e.V. (GI); 2016. p. 93–108. Available from: http://juser.fz-juelich.de/record/826510.
- 37.Nagendra Babu P, Linssen C, Eppler JM, Schulte to Brinke T, Ziaeemehr A, Fardet T, et al. NESTML 4.0. Zenodo; 2021. Available from: 10.5281/zenodo.4740083. [DOI]
- 38. Rotter S, Diesmann M. Exact digital simulation of time-invariant linear systems with applications to neuronal modeling. Biol Cybern. 1999;81(5-6):381–402. Available from: doi: 10.1007/s004220050570 [DOI] [PubMed] [Google Scholar]
- 39.Ahmad S, Hawkins J. How do neurons operate on sparse distributed representations? A mathematical theory of sparsity, neurons and active dendrites. 10.48550/arXiv.1601.00720 [q-bio.NC]; 2016. [DOI]
- 40. Atherton LA, Dupret D, Mellor JR. Memory trace replay: the shaping of memory consolidation by neuromodulation. Trends Neurosci. 2015;38(9):560–570. Available from: doi: 10.1016/j.tins.2015.07.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Thomas SA. Neuromodulatory signaling in hippocampus-dependent memory retrieval. Hippocampus. 2015;25(4):415–431. doi: 10.1002/hipo.22394 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42. Baluch F, Itti L. Mechanisms of top-down attention. Trends Neurosci. 2011;34(4):210–224. Available from: doi: 10.1016/j.tins.2011.02.003 [DOI] [PubMed] [Google Scholar]
- 43. Noudoost B, Chang MH, Steinmetz NA, Moore T. Top-down control of visual attention. Curr Opin Neurobiol. 2010. April;20(2):183–190. Available from: doi: 10.1016/j.conb.2010.02.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44. Buzsáki G. Rhythms of the Brain. Oxford University Press; 2006. Available from: 10.1093/acprof:oso/9780195301069.001.0001. [DOI] [Google Scholar]
- 45. Grosmark AD, Mizuseki K, Pastalkova E, Diba K, Buzsáki G. REM Sleep Reorganizes Hippocampal Excitability. Neuron. 2012;75(6):1001–1007. Available from: doi: 10.1016/j.neuron.2012.08.015 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46. Nádasdy Z, Hirase H, Czurkó A, Csicsvari J, Buzsáki G. Replay and Time Compression of Recurring Spike Sequences in the Hippocampus. J Neurosci. 1999;19(21):9497–9507. Available from: doi: 10.1523/JNEUROSCI.19-21-09497.1999 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47. Lee AK, Wilson MA. Memory of Sequential Experience in the Hippocampus during Slow Wave Sleep. Neuron. 2002;36(6):1183–1194. Available from: doi: 10.1016/S0896-6273(02)01096-6 [DOI] [PubMed] [Google Scholar]
- 48. Davidson TJ, Kloosterman F, Wilson MA. Hippocampal replay of extended experience. Neuron. 2009;63(4):497–507. doi: 10.1016/j.neuron.2009.07.027 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49. Euston DR, Tatsuno M, McNaughton BL. Fast-Forward Playback of Recent Memory Sequences in Prefrontal Cortex During Sleep. Science. 2007;318(5853):1147–1150. Available from: doi: 10.1126/science.1148979 [DOI] [PubMed] [Google Scholar]
- 50. Lazar A. SORN: a Self-organizing Recurrent Neural Network. Front Comput Neurosci. 2009;3. Available from: doi: 10.3389/neuro.10.023.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51. Jahnke S, Timme M, Memmesheimer RM. Guiding Synchrony through Random Networks. Phys Rev X. 2012;2(4): doi: 10.1103/PhysRevX.2.041016 [DOI] [Google Scholar]
- 52. Breuer D, Timme M, Memmesheimer RM. Statistical physics of neural systems with nonadditive dendritic coupling. Phys Rev X. 2014;4(1): doi: 10.1103/PhysRevX.4.011053 [DOI] [Google Scholar]
- 53. Abeles M. Corticonics: Neural Circuits of the Cerebral Cortex. Cambridge: Cambridge University Press; 1991. 10.1017/CBO9780511574566 [DOI] [Google Scholar]
- 54. Diesmann M, Gewaltig MO, Aertsen A. Stable propagation of synchronous spiking in cortical neural networks. Nature. 1999;402(6761):529–533. doi: 10.1038/990101 [DOI] [PubMed] [Google Scholar]
- 55. Major G, Larkum M, Schiller J. Active Properties of Neocortical Pyramidal Neuron Dendrites. Annu Rev Neurosci. 2013;36:1–24. doi: 10.1146/annurev-neuro-062111-150343 [DOI] [PubMed] [Google Scholar]
- 56. Mengual UM, Wybo WA, Spierenburg LJ, Santello M, Senn W, Nevian T. Efficient low-pass dendro-somatic coupling in the apical dendrite of layer 5 pyramidal neurons in the anterior cingulate cortex. J Neurosci. 2020;40(46):8799–8815. doi: 10.1523/JNEUROSCI.3028-19.2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Diesmann M. Conditions for Stable Propagation of Synchronous Spiking in Cortical Neural Networks: Single Neuron Dynamics and Network Properties; 2002. Available from: https://hss-opus.ub.rub.de/opus4/frontdoor/index/index/docId/3292.
- 58. Gidon A, Zolnik TA, Fidzinski P, Bolduan F, Papoutsi A, Poirazi P, et al. Dendritic action potentials and computation in human layer 2/3 cortical neurons. Science. 2020;367(6473):83–87. Available from: doi: 10.1126/science.aax6239 [DOI] [PubMed] [Google Scholar]
- 59. Poirazi P, Brannon T, Mel BW. Pyramidal Neuron as Two-Layer Neural Network. Neuron. 2003;37(6):989–999. Available from: doi: 10.1016/S0896-6273(03)00149-1 [DOI] [PubMed] [Google Scholar]
- 60. Schulz A, Miehl C, Berry MJ, Gjorgjieva J. The generation of cortical novelty responses through inhibitory plasticity. eLife. 2021;10:e65309. Available from: doi: 10.7554/eLife.65309 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61. Asabuki T, Fukai T. Somatodendritic consistency check for temporal feature segmentation. Nat Commun. 2020;11(1):1–13. Available from: doi: 10.1038/s41467-020-15367-w [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62. Asabuki T, Fukai T. Neural mechanisms of context-dependent segmentation tested on large-scale recording data. BioRxiv. 2021; Available from: 10.1101/2021.04.25.441363. [DOI] [Google Scholar]
- 63.Billaudelle S, Ahmad S. Porting HTM models to the Heidelberg neuromorphic computing platform. 10.48550/arXiv.1505.02142 [q-bio.NC]; 2015. [DOI]
- 64. Mauk MD, Buonomano DV. The Neural Basis of Temporal Processing. Annu Rev Neurosci. 2004;27:307–340. doi: 10.1146/annurev.neuro.27.070203.144247 [DOI] [PubMed] [Google Scholar]
- 65. Paton JJ, Buonomano DV. The neural basis of timing: Distributed mechanisms for diverse functions. Neuron. 2018;98(4):687–705. doi: 10.1016/j.neuron.2018.03.045 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66. Milojkovic BA. A Strict Correlation between Dendritic and Somatic Plateau Depolarizations in the Rat Prefrontal Cortex Pyramidal Neurons. J Neurosci. 2005;25(15):3940–3951. Available from: doi: 10.1523/JNEUROSCI.5314-04.2005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67. Dave AS, Margoliash D. Song replay during sleep and computational rules for sensorimotor vocal learning. Science. 2000;290(5492):812–816. doi: 10.1126/science.290.5492.812 [DOI] [PubMed] [Google Scholar]
- 68. Louie K, Wilson MA. Temporally structured replay of awake hippocampal ensemble activity during rapid eye movement sleep. Neuron. 2001;29(1):145–156. doi: 10.1016/S0896-6273(01)00186-6 [DOI] [PubMed] [Google Scholar]
- 69. Tully PJ, Lindén H, Hennig MH, Lansner A. Spike-Based Bayesian-Hebbian Learning of Temporal Sequences. PLOS Comput Biol. 2016;12(5):e1004954. Available from: doi: 10.1371/journal.pcbi.1004954 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70. van Vreeswijk C, Sompolinsky H. Chaotic Balanced State in a Model of Cortical Circuits. Neural Comput. 1998;10(6):1321–1371. Available from: doi: 10.1162/089976698300017214 [DOI] [PubMed] [Google Scholar]
- 71. Barth AL, Poulet JF. Experimental evidence for sparse firing in the neocortex. Trends Neurosci. 2012;35(6):345–355. doi: 10.1016/j.tins.2012.03.008 [DOI] [PubMed] [Google Scholar]
- 72. Shadlen MN, Newsome WT. The Variable Discharge of Cortical Neurons: Implications for Connectivity, Computation, and Information Coding. J Neurosci. 1998;18(10):3870–3896. Available from: doi: 10.1523/JNEUROSCI.18-10-03870.1998 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73. Ecker AS, Berens P, Keliris GA, Bethge M, Logothetis NK. Decorrelated Neuronal Firing in Cortical Microcircuits. Science. 2010;327(5965):584–587. Available from: doi: 10.1126/science.1179867 [DOI] [PubMed] [Google Scholar]
- 74. Renart A, De La Rocha J, Bartho P, Hollender L, Parga N, Reyes A, et al. The asynchronous State in Cortical Circuits. Science. 2010;327:587–590. Available from: doi: 10.1126/science.1179850 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75. Riehle A, Grün S, Diesmann M, Aertsen A. Spike synchronization and rate modulation differentially involved in motor cortical function. Science. 1997;278(5345):1950–1953. doi: 10.1126/science.278.5345.1950 [DOI] [PubMed] [Google Scholar]
- 76. Schrader S, Grün S, Diesmann M, Gerstein GL. Detecting synfire chain activity using massively parallel spike train recording. J Neurophysiol. 2008;100:2165–2176. Available from: doi: 10.1152/jn.01245.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77. Quaglio P, Yegenoglu A, Torre E, Endres DM, Grün S. Detection and evaluation of spatio-temporal spike patterns in massively parallel spike train data with spade. Front Comput Neurosci. 2017;11:41. doi: 10.3389/fncom.2017.00041 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78. Quaglio P, Rostami V, Torre E, Grün S. Methods for identification of spike patterns in massively parallel spike trains. Biol Cybern. 2018;p. 1–24. doi: 10.1007/s00422-018-0755-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79. DeWeese MR, Zador AM. Non-Gaussian Membrane Potential Dynamics Imply Sparse, Synchronous Activity in Auditory Cortex. J Neurosci. 2006;26(47):12206–12218. doi: 10.1523/JNEUROSCI.2813-06.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80. Jordan J, Petrovici MA, Breitwieser O, Schemmel J, Meier K, Diesmann M, et al. Deterministic networks for probabilistic computing. Sci Rep. 2019;9:18303. Available from: https://www.nature.com/articles/s41598-019-54137-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81. Turrigiano GG, Leslie KR, Desai NS, Rutherford LC, Nelson SB. Activity-dependent scaling of quantal amplitude in neocortical neurons. Nature. 1998;391(6670):892–896. Available from: doi: 10.1038/36103 [DOI] [PubMed] [Google Scholar]
- 82. Elliott T. An Analysis of Synaptic Normalization in a General Class of Hebbian Models. Neural Comput. 2003;15(4):937–963. Available from: doi: 10.1162/08997660360581967 [DOI] [PubMed] [Google Scholar]