Skip to main content

This is a preprint.

It has not yet been peer reviewed by a journal.

The National Library of Medicine is running a pilot to include preprints that result from research funded by NIH in PMC and PubMed.

bioRxiv logoLink to bioRxiv
[Preprint]. 2025 May 27:2025.05.22.655599. [Version 1] doi: 10.1101/2025.05.22.655599

Cellular and subcellular specialization enables biology-constrained deep learning

Alessandro R Galloni 1, Ajay Peddada 1, Yash Chennawar 1,2, Aaron D Milstein 1,*
PMCID: PMC12154792  PMID: 40501912

Summary

Learning and memory in the brain depend on changes in the strengths of synaptic connections between neurons. While the molecular and cellular mechanisms of synaptic plasticity have been extensively studied experimentally, much of our understanding of how plasticity is organized across populations of neurons during task learning comes from training artificial neural networks (ANNs) using computational methods. However, the architectures of modern ANNs and the algorithms used to train them are not compatible with fundamental principles of neuroscience, leaving a gap in understanding how the brain coordinates learning across multiple layers of neural circuitry. Here we leverage recent experimental evidence to test an emergent theory that biological learning depends on specialization of distinct neuronal cell types and compartmentalized signaling within neuronal dendrites. We demonstrate that multilayer ANNs comprised of separate recurrently connected excitatory and inhibitory cell types, and neuronal units with separate soma and dendrite compartments, can be trained to accurately classify images using a fully biology-compatible deep learning algorithm called dendritic target propagation. By adhering to strict biological constraints, this model is able to provide unique insight into the biological mechanisms of learning and to make experimentally testable predictions regarding the roles of specific neuronal cell types in coordinating learning across different brain regions.

Main

When an animal learns to perform a task, patterns of neuronal activity in the brain change over time as behavioral performance improves1. Mathematically, this learning can be described as an optimization process than minimizes task performance error by modifying the parameters in the brain that control behavior. Thus, by definition, when learning is successful, neuronal parameters must have changed in a direction that minimizes performance error2. However, it is not known how the brain utilizes information about behavioral outcomes to adjust the right parameters in the right neurons during learning – the credit assignment problem3 – or how it transmits that information to coordinate learning across brain areas or multiple layers of neural circuitry4.

In standard models of learning in neural circuits, the primary parameters that can be adjusted during training are the strengths of synaptic connections between neuronal units, often referred to as “synaptic weights.” In the machine learning field, synaptic weights in artificial neural networks (ANNs) are typically optimized using gradient descent. This requires calculating the impact of each individual synapse on the overall performance error of the network (the error gradient with respect to the network weights), and precisely adjusting each synaptic weight proportionally. In multilayer ANNs, an algorithm called “backpropagation” is typically used to transmit gradient information across multiple neural circuit layers5,6. However, it has not been demonstrated how backpropagation could be implemented using entirely biology-compatible mechanisms7. In contrast, in the neuroscience field, synaptic weights in neural circuit models are typically modified using Hebbian synaptic plasticity rules that have been observed experimentally, like spike-timing-dependent plasticity (STDP)810, but those rules do not necessarily adjust synaptic weights in the direction of the error gradient or guarantee improvement in task performance4,11. In this study, we aim to reconcile the discrepancies between these two approaches and to demonstrate a mechanism for performance error to guide learning in multilayer neural circuits in a way that is compatible with known principles of neurobiology.

In particular, we impose the following experimentally-supported biological constraints to address multiple biologically-implausible features of standard ANNs and the backpropagation algorithm: 1) the cellular diversity constraint: while neuronal units in standard ANNs are uniform, biological neurons are specialized into diverse cell types with different functions12 (Fig. 1); 2) the signed synapse constraint: while gradient descent can cause a synaptic weight to change sign (from positive to negative, or vice versa), biological synapses cannot change sign, and neurons are typically pre-specified during development as either excitatory (all outgoing synapses are positive) or inhibitory (all outgoing synapses are negative) (this constraint is often referred to as Dale’s law13); 3) the continuous signaling constraint: while backpropagation requires that sensory inference and error propagation occur in two separate phases (sensory responses flow bottom-up exclusively during the sensory phase, and error signals flow top-down exclusively during the learning phase), in biological circuits, neurons continuously send and receive both bottom-up and top-down signals; and 4) the non-symmetric synapse constraint: while backpropagation requires that the synaptic weights used to propagate sensory information bottom-up are exactly symmetrical to those used to propagate gradient signals top-down, in biological circuits separate synapses with distinct weights are responsible for processing either bottom-up or top-down signals (this is often referred to as the “weight transport problem”1416).

Figure 1. Incorporating biological cellular and subcellular specialization into artificial neural networks.

Figure 1.

Top left: Architecture of multi-layer neural networks trained to classify images of handwritten digits. The number of excitatory units in each layer is indicated in parentheses. Top right: Each excitatory-inhibitory artificial neural network (EIANN) has two hidden layers containing 500 excitatory (E) units and 50 inhibitory (I) units that are recurrently connected, consistent with experimentally observed proportions. Networks trained with the proposed biology-constrained algorithm dendritic target propagation have two separate populations of 50 inhibitory neurons each that target either the soma (Is) or the dendrite (Id). Networks trained with backpropagation receive top-down error signals, Hebbian networks receive no top-down inputs, and networks trained with dendritic target propagation receive top-down target signals and compute error locally. Bottom: Table summarizes proposed computational roles of different biological cell types.

Finally, we propose a biologically-constrained mechanism for information about performance error to be computed and propagated within a multilayer neural circuit. Our approach depends on the morphology of biological neurons, which contain a tree-like structure of branched subcellular compartments called dendrites. We leverage insight from recent experimental and theoretical work supporting specialized roles for electrically compartmentalized dendrites in processing top-down signals and controlling synaptic plasticity1734. The algorithm, which we refer to as dendritic target propagation, locally computes an error signal in the distal dendrites of excitatory neurons as a difference between top-down excitation and lateral feedback inhibition18,19,35 (Fig. 1). In each neuron, this dendritic error signal has two effects: it determines the sign and magnitude of synaptic weight updates, and it “nudges” the activity level of the neuron either up or down. These changes in activity are then relayed to lower layers of the network via top-down projections. This, in turn, signals to neurons in lower layers that they must also adjust their synaptic weights to reduce network performance error. We demonstrate that dendritic target propagation can be used to effectively train multilayer neural networks with biologically-constrained recurrent architectures to perform stimulus classification tasks. We also show that this algorithm is compatible with multiple experimentally and theoretically supported synaptic plasticity rules, and remains effective even when the weights of bottom-up and top-down synapses are independently initialized and learned. In summary, we demonstrate that neurobiological principles like cell-type diversity, subcellular computation in dendrites, and synaptic plasticity can support target-directed deep learning in the brain.

Training recurrent networks with separate excitatory and inhibitory neurons

First, we imposed the biological cellular diversity and signed synapse constraints on the architecture of a 3-layer ANN (Fig. 1) by dividing neurons within each layer into separate excitatory (E) and inhibitory (I) populations. Within each layer, both excitatory and inhibitory cells receive bottom-up inputs from excitatory neurons in the previous layer, and lateral inputs from inhibitory neurons in the same layer. This establishes a feedforward inhibitory circuit motif common in mammalian hippocampal and cortical circuits, which has been associated with specific computations such as input normalization and stimulus feature subtraction36,37 (Fig. 1). Inhibitory neurons also receive a lateral input from excitatory neurons in the same layer. This establishes a recurrent feedback inhibitory circuit motif, which can impose competition between ensembles of excitatory neurons, limit the number of neurons that can be simultaneously active (i.e. enforce sparsity), shape their stimulus selectivity, and decorrelate their activity3840 (Fig. 1).

Despite potential advantages of this functional division-of-labor across neuronal cell types (Fig. 1), this biological microcircuit architecture imposes some additional challenges. Depending on the strength of connections and timescales of input integration, recurrent excitatory-inhibitory (EI) networks can be prone to unstable or oscillatory dynamics4143, which can interfere with learning44. Furthermore, biological neuronal circuits are typically comprised of a majority of excitatory neurons and a minority of inhibitory neurons45,46, potentially imposing a bottleneck for memory storage47 and computations that require subtraction48,49. Thus, before testing biological learning mechanisms, we first verified that the backpropagation algorithm can be used to train a multilayer excitatory-inhibitory ANN (EIANN) to perform an image classification task (handwritten digit characters) (Fig. 1; Methods). We tuned model hyperparameters to ensure that dynamic network responses to static image stimuli converged to a non-oscillatory equilibrium (Supplementary Fig. S1; Methods). The resulting trained EIANN exhibited classification accuracy comparable to a standard feedforward network (Fig. 2a,b,g). We used an activity maximization method to visualize the stimulus preferences of individual neurons after training50, and found that excitatory units in the EIANN had more spatially structured receptive fields than the non-sign-constrained units in the standard ANN (Fig. 2d,e,i; Methods)51. Importantly, these receptive fields contained some regions of the stimulus space that were excitatory (Fig. 2d,e; black), and some that were inhibitory (Fig. 2d,e; blue), indicating that separating excitation and inhibition into distinct neuronal cell types did not interfere with the capability of the network to perform stimulus feature subtraction – an important elementary computation. Importantly, training a network without any inhibitory neurons (all positive weights) resulted in strongly reduced performance (Supplementary Table S1), indicating that the task requires a source of negative weights to perform stimulus feature subtraction52.

Figure 2. Coordinating learning across network layers requires top-down signaling.

Figure 2.

a-c, Average neuronal population activity in response to each class of handwritten digits. a, Shown are 500 excitatory (E) units in the second hidden layer (H2) of a feedforward ANN trained with backpropagation. b, Same as a for a recurrent EIANN trained with backpropagation. c, Same as a for a recurrent EIANN trained with a normalized Hebbian rule. d-f, Example receptive fields selected from H2E units from the networks shown in a-c. Numbers in the top left corners indicate the label of the class that produces the largest activity (averaged across samples) for each unit. g, Classification performance accuracy. Shading indicates standard deviation across five instances of each network. h, Selectivity of E units (in both hidden layers) over stimulus classes. i, Spatial structure of the receptive fields of E units (in both hidden layers), as measured by spatial autocorrelation (Moran’s I).

Following training, excitatory neurons exhibited a greater degree of stimulus selectivity than inhibitory neurons (Fig. 2a,b,h and Supplementary Fig. S2), consistent with experimental recordings from the mammalian hippocampus and cortex 5357. Interestingly, we found a similar degree of stimulus selectivity in inhibitory cells when all incoming and outgoing synapses to and from inhibitory cells were fixed at their random initial weights and not learned (Supplementary Fig. S2). Fixed inhibition also resulted in comparable classification performance, suggesting that plasticity of inhibition is not required for the task (Supplementary Fig. S2).

Next, we sought to impose the biological continuous signaling constraint, which disallows neurons from switching between separate signaling modes in which their activities represent either bottom-up responses to the sensory environment or top-down error gradients. The simplest implementation of this constraint is to completely remove top-down connections between circuit layers, and to use unsupervised learning rules to promote self-organization of each layer such that they produce discriminable patterns of activity in response to different classes of stimuli5861. Then, an error-dependent learning rule can be used in the top layer to perform image classification without propagating error signals to earlier network layers. A recent study62 used a combination of experimentally-supported Hebbian and homeostatic learning rules6365 at all synapses within a single layer of an EIANN to achieve highly selective unit responses and highly discriminating population responses that uniformly tiled a stimulus space, without any supervision. In that network, feedback inhibition imposed competition between units (Fig. 1), encouraging different units to represent different stimuli. At each synapse between a presynaptic neuron j and a postsynaptic neuron i, the connection weight wij was updated proportional to the correlation cij of the activities of the two units:

cij=aiaj (1)
Δwij=ηcij (2)

where η is a learning rate. After each weight update, the total pool of synaptic resources available in each postsynaptic neuron for a specific projection type (e.g. bottom-up excitatory inputs) was maintained constant by applying the following weight normalization inspired by experimentally-observed heterosynaptic plasticity62,63,65:

wijwsumwij+Δwijkwik+Δwik (3)

where wsum is a scale factor. This normalization prevents unbounded growth of synaptic strengths, and enhances stimulus selectivity by imposing competition between synapses within a neuron62. We implemented this unsupervised Hebbian learning rule with weight normalization in our multilayer EIANN (Fig. 1 and Fig. 2c; Methods) and found that, while hidden units became highly stimulus-selective after learning (Fig. 2c,h), their receptive fields showed little diversity, resembling the average image within a class (Fig. 2f and Supplementary Fig. S3). Some neurons responded similarly to more than one class of stimuli (Fig. 2c and Supplementary Fig. S3), resulting in poor image classification performance (Fig. 2g). These results suggest that, while unsupervised Hebbian learning rules are effective at self-organizing neuronal EI networks to produce sparse and selective sensory responses, without top-down signaling between layers in a multilayer network, a purely unsupervised algorithm is unable to coordinate learning across layers or to guarantee that the learned stimulus representations will be useful for accurate image classification (Supplementary Fig. S3).

Biology-constrained multilayer credit assignment via dendritic target propagation

The above results suggested that top-down propagation of error information between circuit layers is essential for target-directed learning. However, the biological continuous signaling constraint dictates that neurons simultaneously propagate their activity both bottom-up and top-down, which means that activity generated in higher layers could interfere with sensory responses in lower layers, and likewise sensory information could be aliased with information about learning targets or performance error. Previous models have proposed various solutions to this problem, including i) limiting the strengths of top-down connections66, ii) applying different temporal filters to bottom-up and top-down pathways24,26,27, iii) processing bottom-up and top-down signals in separate neurons6769, iv) processing them in separate subcellular compartments (soma and dendrite)2124,70, and/or v) limiting the influence of top-down inputs by predicting and cancelling them with a separate lateral or top-down input pathway21,23,27. Only recently has experimental data from awake behaving animals become available that tests some of the predictions of these models and provides insight into the roles of top-down excitation and lateral inhibition to the distal apical dendrites of pyramidal neurons in modulating neuronal activity and plasticity during target-directed learning1719,31,35,7176.

We sought to incorporate into our EIANN model the following experimental findings from the mammalian hippocampus and cortex: 1) sufficient input to the apical dendrites of excitatory pyramidal cells can evoke a long (~50 ms to > 1 s) regenerative event called a dendritic calcium spike29,7780; 2) dendritic calcium spikes can propagate to the soma, resulting in an increase in output firing rate; 3) dendritic calcium spikes can rapidly induce either increases or decreases in synaptic strength through a potent form of synaptic plasticity called behavioral timescale synaptic plasticity (BTSP)1719,29,31, 4) top-down excitatory inputs onto distal apical dendrites carry information about behaviorally-relevant context, cues, and goals, and strongly promote dendritic calcium spikes and associated plasticity19,31,71,76, and 5) dendritic inhibition from local interneurons strongly inhibits dendritic calcium spikes and associated plasticity35,57,74,75,81. Accordingly, we modified the excitatory neurons in our EIANN to comprise separate soma and distal apical dendrite compartments (Fig. 1 and Fig. 3a, Methods). Bottom-up excitatory inputs and lateral inhibitory inputs from soma-targeting interneurons arrive at the soma compartment (representing the cell body as well as perisomatic dendrites) (Fig. 1). We also added an additional specialized population of dendrite-targeting interneurons, which receive lateral excitatory inputs and recurrent inhibitory inputs from each other (Fig. 1 and Fig. 3a). The apical dendrite compartment of excitatory neurons receives top-down excitatory inputs, and lateral inhibitory inputs from the dendrite-targeting interneurons (Fig. 1 and Fig. 3a). To train this dendritic EIANN, we designed a learning algorithm termed dendritic target propagation, in which the activation state of the dendritic compartments of excitatory neurons, D, acts both as a gate for plasticity and as a source of input to the soma compartments that affects their output firing rates.

Figure 3. Adaptation of dendrite-targeting inhibition improves deep credit assignment.

Figure 3.

a, Diagram of the dendritic target propagation algorithm. Left: Network responses to sensory stimuli are propagated both forward and backward through the network layers. During this sensory phase, lateral dendritic inhibition aims to accurately cancel top-down dendritic excitation, resulting in dendritic states close to zero. Right: During a supervisory phase, output layer excitatory units are nudged up (red) or down (blue) to their target activity. These changes in activity are propagated to lower layer dendrites, where they are unpredicted and remain uncancelled by dendritic inhibition, resulting in dendritic states that approximate local error gradients in each neuron. b, Example temporal dynamics of the dendritic states (top) and resulting activities (bottom) of the two example hidden layer neurons illustrated in a. The caret symbol (^) represents values immediately after the nudge is applied, while the tilde symbol (~) represents values after additional time steps of equilibration. c, Average neuronal population activity in response to each class of handwritten digits are shown for dendrite-targeting inhibitory neurons (DendI) in the second hidden layer (H2) for networks using three different learning schemes for their incoming weights. Left: Weights onto DendI neurons are frozen at their random initialization. Middle: Weights onto DendI neurons are learned by using backpropagation locally within each hidden layer to minimize the magnitude of excitatory neuron dendritic states. Right: Weights onto DendI neurons are learned using a fully unsupervised normalized Hebbian learning rule. d, Selectivity of DendI units (in both hidden layers) over stimulus classes. e, Average absolute value of the sensory-phase dendritic states of hidden layer E cells during training. f, Classification performance accuracy. g, For each model, the bottom-up excitatory (𝑊) weight updates prescribed by the model were compared to those prescribed by backpropagation using cosine similarity. An alignment angle of 90° indicates an update that is orthogonal to the gradient, while 0° reflects identical weight updates. e-g, Shading indicates standard deviation across five instances of each network.

In a network with L layers, the dendritic states in the output layer D(L) are initially set to zero during stimulus presentation, in the absence of any supervisory input. However, any activity A(E,L) in output layer excitatory neurons excites the dendrites of neurons in the previous hidden layer l=L1 via top-down projections with weights B(l):

di(l)=jbij(l)aj(E,l+1)+kyik(DendI,l)ak(DendI,l) (4)

where di(l) is the dendritic state of a postsynaptic excitatory neuron i in layer l,aj(E,l+1) is the activity of a presynaptic excitatory neuron j in layer l+1 connected to di(l) via weight bij(l), and ak(DendI,l) is the activity of a presynaptic dendrite-targeting inhibitory neuron k in layer l connected to di(l) via weight yik(DendI,l) (note that here we use capitalized symbols to indicate population activity vectors and weight matrices, and lowercase symbols to indicate single scalar values for one cell, compartment, or synapse). We initially set the top-down weights B(l) (from layer l+1 to layer l) as symmetric to the bottom-up weights W(l+1) (from layer l to layer l+1), as prescribed by the backpropagation algorithm. Later we will consider methods to separately learn the independent weights B(l).

In order for the firing rates of neurons to cleanly reflect their sensory responses without interference from top-down inputs, ideally the dendritic states D(l) in each hidden layer l would also be close to zero during stimulus presentation (Fig. 3a,b). Within the dendritic EIANN architecture, this can be achieved if the sum of lateral inhibitory inputs to each dendrite exactly predicts and cancels the sum of its top-down excitatory inputs (Equation (4)). Therefore, we applied a learning rule to the lateral inhibitory synapses onto excitatory neuron dendrites (with weights Y(l)) that aims to minimize the dendritic states D(l):

Δyik(DendI,l)=ηY(DendI,l)di(l)ak(DendI,l) (5)

for postsynaptic excitatory neuron i and presynaptic dendrite-targeting inhibitory neuron k in layer l, where ηY(DendI,l) is a learning rate.

Following equilibration of neuronal activities during the sensory phase, we then applied supervisory inputs N to the dendrite compartments D(L) of excitatory neurons in the output layer to "nudge" their activity A(E,L) towards target values T that would lead to correct stimulus categorization (Fig. 3a,b):

N=TA(E,L) (6)
D^(L)=N (7)

where the caret symbol (^) indicates variables after the supervisory inputs have been applied. In each layer l, the "nudged" dendritic states D^(l) in turn "nudge" the excitatory neuron firing rates Aˆ(E,l) (Fig. 3a):

Aˆ(E,l)=fS(E,l)+D^(l)G(E,l) (8)

where S(E,l) is the state of the soma compartments of the excitatory neurons, f is the rectified linear unit activation function (ReLU) (see Methods), and G(E,l)=fS(E,l) is the derivative of the ReLU activation function, which is one when S(E,l)>0 and zero otherwise. This change in output layer activity then propagates to lower layers via top-down projections (see Methods). Since lateral dendritic inhibition is tuned to cancel the hidden layer dendritic states D(l) during the sensory phase, these changes in top-down excitation during the supervisory phase are not predicted by the inhibitory neurons and remain uncancelled (Fig. 3a,b). This results in the "nudged" dendritic states D^(l) in each layer reflecting a weighted sum of the supervisory nudges D^(l+1) applied to the layer above. Finally, the weights W(l) of the bottom-up excitatory inputs onto layer l excitatory neurons are updated with the following dendritic state-dependent learning rule:

Δwij(l)=ηW(l)dˆi(l)gi(E,l)aj(E,l1) (9)

for postsynaptic excitatory unit i in layer l and presynaptic excitatory unit j in layer l1. To the extent that D^(l) approaches the value of the error gradient A(E,l) used in backpropagation (see Methods), this weight update will also move the network in a direction that minimizes error.

The above scheme locally computes error-related learning signals in the dendrites of excitatory neurons (Fig. 3a). However, it does not yet address whether or how synaptic weights should be updated at the inputs to dendrite-targeting interneurons to ensure that their activity provides an appropriate basis for cancelling the top-down excitatory inputs to the excitatory neuron dendrites. These interneurons receive lateral excitatory inputs with weights Q(DendI,l) and recurrent inhibitory inputs with weights R(DendI,l) (Fig. 3a). First, we asked whether the dendritic target propagation algorithm could effectively train the dendritic EIANN if the weights of these projections were simply held fixed at their random initial values, and not learned, as proved effective for soma-targeting inhibitory neurons (Supplementary Fig. S2). Fixed dendritic inhibition resulted in low stimulus selectivity in dendrite-targeting interneurons (Fig. 3c,d), ineffective cancelling of top-down sensory input to excitatory neuron dendrites (Fig. 3e), and poor image classification performance by the network (Fig. 3f). Updates to the bottom-up excitatory weights W(l) were poorly aligned with the direction of the error gradient (Fig. 3g). This implies that, unlike in soma-targeting interneurons, plasticity in dendrite-targeting interneurons may be required to improve local error computation and task performance.

Before testing biologically-plausible learning mechanisms in these dendrite-targeting interneurons, we first took a normative approach and tested whether task performance could be improved if the weights of inputs onto dendrite-targeting inhibitory neurons were directly optimized by gradient descent to minimize the dendritic states of the excitatory neurons. To do this, we defined a separate loss function (l) in each hidden layer as the mean squared dendritic state of the excitatory neurons in the layer:

(l)=1nindi(l)2 (10)

Then, in each hidden layer, we used backpropagation to adjust weights in proportion to the gradient of this local loss function with respect to the synaptic weights Q(DendI,l) and R(DendI,l) (see Methods). Learning dendritic inhibition with backpropagation resulted in dendrite-targeting interneurons that were highly stimulus selective (Fig. 3c,d), excitatory dendritic states that were effectively minimized (Fig. 3e), and increased task performance (Fig. 3f).

With this configuration of dendritic inhibition learned by gradient descent, the dendritic target propagation algorithm prescribed updates to the bottom-up excitatory weights W(l) that were well-aligned with those prescribed by the backpropagation algorithm (Fig. 3g). Given this gradient-based baseline to compare against, we next explored how the weights of inputs to the dendrite-targeting interneurons could be learned with biologically-plausible learning rules. To address this, we revisited the normalized Hebbian plasticity rule previously used to self-organize individual circuit layers without supervisory inputs62 (Fig. 1 and Fig. 2c; see Methods). This approach has been shown to be directly related to principal component analysis (PCA) and independent component analysis (ICA), and can enable a population of neurons to learn a compressed (reduced dimensionality) representation of their inputs 62,82. We found that applying the normalized Hebbian rule to the incoming projections onto dendrite-targeting interneurons resulted in high stimulus selectivity (Fig. 3c,d), effective minimization of excitatory dendritic states (Fig. 3e), increased classification performance (Fig. 3f and Fig. 4b), and bottom-up weight updates that better aligned with backpropagation (Fig. 3g). This strategy also resulted in highly selective hidden layer excitatory neurons (Fig. 4a,d) with highly structured receptive fields (Fig. 4c,d and Supplementary Fig. S3). These results demonstrate that unsupervised Hebbian learning in inhibitory neurons can be combined with dendritic-state-dependent learning in excitatory neurons to achieve biology-compatible training of multilayer networks. They also highlight that while unsupervised Hebbian rules are limited in the complete absence of top-down supervision (Fig. 1 and Fig. 2c), they can be used to effectively self-organize some components of a neural circuit, as long as they are supplemented by top-down modulation of learning in other components of the circuit, which can provide a scaffold to organize around.

Figure 4. Dendritic target propagation enables biology-constrained deep learning.

Figure 4.

a, Average neuronal population activity in response to each class of handwritten digits. Left: Shown are 500 excitatory (E) units in the second hidden layer (H2) of a recurrent EIANN trained with backpropagation. Right: Same as left for a dendritic EIANN trained with dendritic target propagation. b, Classification performance accuracy. c, Example receptive fields selected from H2E units from the networks shown in a. Numbers in the top left corners indicate the label of the class that produces the largest activity (averaged across samples) for each unit. d, Left: Selectivity of E units (in both hidden layers) over stimulus classes. Right: Spatial structure of the receptive fields of E units (in both hidden layers), as measured by spatial autocorrelation (Moran’s I). e, Left: Following training on the 2D spirals classification task, nonlinear decision boundaries are shown for an EIANN trained with backpropagation. Data points with colors that do not match the background region were misclassified. Right: Same as left for a dendritic EIANN trained with dendritic target propagation. f, Classification performance accuracy on the 2D spirals task. The networks shown in e are compared to standard feedforward ANNs with and without learned neuronal bias parameters. In b and f, shading indicates standard deviation across five instances of each network.

For simplicity, the models described above were all trained by adjusting synaptic connection weights, but did not include learnable neuronal bias parameters. Biases are typically implemented as modifiable levels of positive or negative background input to each neuronal unit that affect how much additional input a cell needs to cross its activation threshold:

aiE,l=fsiE,l+βil (11)

Some nonlinear classification tasks cannot be solved without adjustable biases. Thus, we tested the effectiveness of the dendritic target propagation algorithm on a 2D spiral classification task that requires learned biases (Supplementary Table S2). We added neuronal bias parameters βi(l) to excitatory neurons only, and updated them with a learning rule that depends linearly on the dendritic state:

Δβi(l)=ηβ(l)dˆi(l)gi(E,l) (12)

The resulting nonlinear classification boundaries are visualized in Fig. 4e. This method resulted in classification accuracy comparable to backpropagation (Fig. 4f and Supplementary Table S2), demonstrating that the dendritic target propagation algorithm is also applicable to highly nonlinear classification tasks that have a strong dependence on learned biases.

Dendritic target propagation is compatible with multiple biological synaptic plasticity rules

In the above section, we described a biology-constrained method that performs deep synaptic credit assignment by computing local error signals in the activation states of the distal dendrites of excitatory neurons. These dendritic error signals were used both 1) to “nudge” the activity of excitatory neurons towards target values that would reduce task performance error, and 2) to update bottom-up synaptic weights with a learning rule that depended linearly on the dendritic state (linear dendritic state rule; LDS; Equation (9)). This scheme was inspired by the recently experimentally-observed BTSP rule, which depends on dendritic calcium spikes17,18. However, while the LDS rule was designed to directly approximate gradient descent, the BTSP rule has additional unique features not captured by the simple LDS rule, including saturable weights and the ability to modify inputs that are active at long time delays (up to ~5 s) from a dendritic calcium spike17,18. Thus, we investigated whether the more complex BTSP rule could also be used to update synaptic weights once the dendritic states of excitatory neurons have been set by dendritic target propagation. We also considered an alternative possibility that, once dendritic error signals have been used to “nudge” somatic activities, classical Hebbian plasticity rules that depend on somatic firing rates could also be used to update synaptic weights60,66,8386. For this purpose, we tested two variants of Hebbian learning that have been well-supported by experimental and theoretical evidence: a “temporally contrastive Hebbian” (TCH) rule66,84, which is related to experimentally-observed STDP810, and the Bienenstock-Cooper-Munro (BCM) rule85, which features an adaptive plasticity threshold and is commonly used to model learning with firing rate homeostasis, such as during visual system development60,86.

The TCH rule weakens connections with correlated activity during the sensory phase, and strengthens connections that are correlated during the supervised phase:

cij(E,l)=ai(E,l)aj(E,l1) (13)
Δwij(l)=ηW(l)c˜ij(E,l)cij(E,l) (14)

where the tilde symbol (~) reflects variables after supervisory inputs have been applied and the network has re-equilibrated (Fig. 3a and Fig. 5b; see Methods). We expected this rule to perform similarly to the LDS rule, as TCH indirectly recovers the dendritic error signal by detecting how much somatic activities have changed after receiving the “nudge” from their dendrites:

Δwij(l)a˜i(E,l)a˜j(E,l1)ai(E,l)aj(E,l1)ai(E,l)+dˆi(l)aj(E,l1)+dˆj(l1)ai(E,l)aj(E,l1)ai(E,l)+dˆi(l)aj(E,l1)ai(E,l)aj(E,l1)+dˆi(l)aj(E,l1)+ (15)

Figure 5. Dendritic target propagation is compatible with multiple biological synaptic plasticity rules.

Figure 5.

a-b, Diagrams summarize biological synaptic plasticity rules. a, Two plasticity rules that depend on dendritic state are represented. Left: The linear dendritic state (LDS) rule updates weights proportionally to dendritic state, and approximates backpropagation to the extent that dendritic state approximates the error gradient. Right: The behavioral timescale synaptic plasticity (BTSP) rule considers interactions between dendritic calcium spikes and presynaptic activity at long time delays, enabling interactions between temporally adjacent stimuli. BTSP also features a dependence on synaptic weight such that weak weights increase and strong weights decrease. b, Two plasticity rules that depend on somatic activity are represented. Left: The temporally contrastive Hebbian (TCH) rule updates weights proportionally to differences in activity correlations between a sensory phase and a supervised phase. Right: The BCM rule is a model of learning with firing rate homeostasis and an adaptive plasticity threshold that depends on the recent history of somatic activity. c, Classification performance accuracy. d, Average absolute value of the sensory-phase dendritic states of hidden layer E cells during training. e, Alignment angle between bottom-up excitatory (𝑊) weight updates prescribed by each model and those prescribed by backpropagation. Shading indicates standard deviation across five instances of each network.

Indeed, we found that the TCH rule results in low amplitude dendritic states during the sensory phase (Fig. 5d) and high task performance comparable to the LDS rule (Fig. 5c), though it prescribed weight updates that were less aligned with backpropagation (Fig. 5e).

The BCM rule strengthens active connections when the postsynaptic activity is above an adaptive threshold θ(l), and depresses them otherwise:

Δwijl=ηWla˜iE,lθila˜iE,la˜jE,l1 (16)

where the threshold θi(l) is a running average of activity a˜i(E,l) over recent stimulus samples (Fig. 5b; see Methods). While this rule is calibrated to produce stable and selective receptive fields, it refers only to target-aligned activities but does not incorporate any form of error signals, even indirectly. Using this rule to train a dendritic EIANN resulted in very low dendritic states during the sensory phase (Fig. 5d), reduced image classification accuracy (Fig. 5c), and weight updates that were nearly orthogonal to gradient descent (Fig. 5e).

The BTSP rule is distinct in that it updates the weights of inputs that are active within a broad window spanning seconds before, during, and after a dendritic calcium spike17,18. It achieves this by referring to slowly decaying traces of recent presynaptic activity (eligibility traces ET(l1) and recent postsynaptic dendritic calcium spikes (instructive signals IS(l))18:

Δwij(l)=ηWdˆiE,lgiE,lwmax(l)wij(l)ETj(l1)ISi(l)wij(l)σETj(l1)ISi(l) (17)

where wmax(l) is a maximum weight and 𝜎 is a sigmoidal function (see Methods). This plasticity rule promotes selectivity by favoring potentiation of inputs activated by the current stimulus, and depressing inputs activated by temporally adjacent stimuli (Fig. 5a). When the BTSP rule was used to update the bottom-up weights of a dendritic EIANN trained with dendritic target propagation, it resulted in effective dendritic error computation (Fig. 5d) and task performance comparable to the LDS rule (Fig. 5c). This is particularly interesting given that the BTSP rule sometimes prescribes weight updates with opposite sign compared to the LDS rule and backpropagation, resulting in intermediate alignment between the two weight updates (Fig. 5e).

In summary, when the biology-constrained dendritic target propagation algorithm is used to locally compute error signals in the distal dendrites of excitatory neurons in a multilayer EIANN, a variety of experimentally-supported biological plasticity rules can flexibly be used to update the synaptic weights of bottom-up excitatory inputs. We found that synaptic plasticity rules like BTSP or TCH that refer either directly or indirectly to an error signal outperform rules like BCM that refer only to somatic activity.

Learning the weights of top-down projections

So far, we have shown that supervised learning can be achieved in a dendritic EIANN that conforms to multiple biological constraints, including the cellular diversity constraint, the signed synapse constraint, and the continuous signaling constraint. However, in order to perform synaptic credit assignment comparable to backpropagation, the above results were obtained with top-down weights B(l) set as symmetric with the bottom-up weights W(l+1):

bji(l)wij(l+1) (18)

This violates the biological non-symmetric synapse constraint, which acknowledges that two different synapses in two different neurons do not share identical histories and may not undergo identical weight updates throughout training. One well-known solution proposed in previous work is to randomly initialize the top-down weights B(l), and to simply fix them at their initial values throughout training15. This approach, called “feedback alignment,” has been shown to work surprisingly well in some supervised learning tasks, and works by encouraging the bottom-up weights W(l+1) to gradually increase their symmetry with the top-down weights B(l) over the course of learning. Related work has shown that the most important effect of this alignment is driven by symmetry in the signs of the W(l+1) and B(l) weights, even if their magnitudes are non-symmetric87. Interestingly, this symmetrical sign constraint is implicitly guaranteed by our cellular diversity and signed synapse constraints, which enforce that both the bottom-up and top-down connection weights are all excitatory (positive sign), resulting in partial alignment (45° angle) at initialization. However, unlike in this prior work, we found that fixing B(l) caused alignment with W(l+1) to progressively worsen over the course of training, rather than improve (Fig. 6b).

Figure 6. Independent learning of bottom-up and top-down synaptic weights.

Figure 6.

a, Diagram highlights that the bottom-up weights (W) and top-down weights (B) are distinct and may be learned independently. b, Performance metrics for dendritic EIANN networks trained with dendritic target propagation with bottom-up weights (W) learned with the LDS rule and different strategies for learning the top-down weights (B): symmetric (red) sets B exactly symmetric to W, fixed random B (grey) fixes B at its random initialization, and learned B uses a temporally contrastive Hebbian (TCH) learning rule with weight normalization to update B. Left: Classification performance accuracy. Middle: Alignment angle between bottom-up excitatory (W) weight updates prescribed by each model and those prescribed by backpropagation. Right: Alignment angle between W and B. c, Same as b but for networks with bottom-up weights (W) learned with BTSP (symmetric model in yellow). b-c, Shading indicates standard deviation across five instances of each network.

The question of how to learn the top-down weights has also been considered by prior work, and most proposed solutions fall into two broad categories: 1) methods that learn B(l) in a separate phase from W(l+1), in which each neuron is driven by a statistically independent source of noise rather than sensory stimuli16,8890, and 2) methods that simultaneously update both B(l) and W(l+1) by the same amount in each train step (sometimes referred to as the “Kolen-Pollack algorithm” )14,16,26,91. Neither of these approaches are fully biologically plausible, as real neurons share non-independent sources of noise (even during sleep periods), and transporting weight updates is a non-local operation. We therefore formulated an alternative solution that approximates the second approach while adhering more closely to biological constraints.

Since the top-down synapse bji(l) resides in the lower-layer neuron j (in layer l), it has local access to the activity ai(l+1) of the presynaptic upper-layer neuron i (in layer l+1 ), but not to its dendritic state dˆi(l+1), which is required to explicitly match the bottom-up update Δwij(l+1) (Equation (9)). However, as we showed with the TCH rule (Equation 1415), a neuron's response to a dendritic "nudge" can be used to locally infer the magnitude of that nudge through a temporal comparison of its activity. Thus, we formulated a learning rule for the top-down weights B(l) that leverages the same temporal comparison to align their updates with the dendritic signals that drive updates to the bottom-up weights W(l+1):

Δbji(l)=ηB(l)a˜i(E,l+1)ai(E,l+1)aj(E,l) (19)

This is followed by a rescaling of the weights onto each postsynaptic neuron, as done in the Hebbian rule with weight normalization62 (Fig. 2c) (Equation (3)):

bji(l)bsum(l)bji(l)+Δbji(l)k(bjk(l)+Δbjk(l)) (20)

We found that learning top-down weights using this TCH rule with weight normalization resulted in improved symmetry of bottom-up weights W(l+1) and top-down weights B(l), compared to using fixed top-down weights (Fig. 6b). This improvement held across models using either the LDS or BTSP rule for learning the bottom-up weights (Fig. 6b,c). These results demonstrate that dendritic target propagation can be used to train multilayer deep neural networks while satisfying multiple biological constraints that have proved challenging to incorporate into ANN models, including the non-symmetric synapse constraint.

Discussion

In this study, we imposed strict biological constraints on the architecture, information processing modes, and synaptic learning rules of multilayer artificial neural networks to better understand the biological mechanisms that coordinate changes in synaptic weights across multiple circuit layers during task learning. The design of our network and learning algorithm was informed by recent neuroscience experiments, which showed that top-down feedback signals from cortex provide target signals to instruct learning in the hippocampus19, and that this plasticity is mediated by dendritic calcium spikes initiated in the distal apical dendrites of excitatory pyramidal neurons17,18. Our dendritic target propagation algorithm implements separate soma and dendrite compartments in excitatory neurons, and computes plasticity-driving error signals locally in dendrites as the difference between top-down excitation and lateral inhibition from local dendrite-targeting interneurons. Our results demonstrate that using dendritic signaling to modulate somatic activity and synaptic plasticity is an effective biology-constrained mechanism to perform deep credit assignment, adjusting the right synapses in the right neurons to improve task performance, and relaying that information across multiple layers of neuronal circuitry.

Other recent theoretical work has proposed that biological neurons might perform credit assignment by multiplexing bottom-up sensory signals to neuronal soma compartments with top-down error signals to distal dendrites26,27. These models rely heavily on low-pass filtering bottom-up inputs to selectively pass “regular” spikes, and high-pass filtering top-down inputs to selectively pass bursts. They also consider dendritic activation to influence only the rate of bursts at the soma, but not the rate of regular spikes. It has not yet been demonstrated if those algorithms would be robust to the incorporation of the biological signed synapse constraint, or to relaxing the assumptions on synaptic filtering to match the more moderate and heterogenous dynamics observed experimentally92. However, that approach to deep credit assignment is complementary and not mutually exclusive to the one proposed in the current study, and it may be beneficial in the future to combine them, using both privileged processing of multiplexed spikes and bursts and gating of plasticity by local dendritic inhibition. Another line of research has focused on the advantages of nonlinear computations in basal and proximal apical dendrites for processing bottom-up inputs93, with theoretical work showing that single biological neurons with complex dendritic morphologies can classify input patterns comparably to entire deep networks94,95. That work is also complementary to ours, which focuses instead on the separate problem of processing top-down signals containing target information for synaptic credit assignment. Future dendritic EIANNs could be designed that both use dendritic target propagation for error computation, and contain more complex perisomatic dendrites for nonlinear processing of bottom-up sensory inputs.

An advantage of building computational models using components that directly correspond to biological counterparts is that the models can make precise predictions about specific circuit elements that can be tested experimentally. A key observation from our models is that using a learning rule to adjust both the synapses onto dendrite-targeting interneurons and their inhibitory connections onto the dendrites plays an important role in enabling the activation state of distal dendrites to approximate the local error gradient. While experiments have indicated that incoming and outgoing synapses to and from inhibitory cells can be plastic9699, it has not yet been tested directly whether calcium spikes or dendritic voltage in pyramidal neurons modify dendritic inhibitory strengths, or whether the stimulus tuning of dendrite-targeting inhibitory cells adapts during learning. Interestingly, recent experiments have demonstrated that different dendritic compartments can express different synaptic learning rules73,100102, which is consistent with the idea that learning in dendritic subcircuits might be specially tuned for optimal local error computation.

We note that, while the components of our model were heavily inspired by experimental observations and the microcircuit architectures of the hippocampus and neocortex, the layered network architecture of our model is generic, and is not meant to exactly mimic the architecture of any particular brain region. A natural extension of our work would be to more directly model the architectures of specific brain regions, for example by including recurrent lateral connections between excitatory neurons, as seen in cortical layer 2/3 or the CA3 region of the hippocampus. This could enable additional functionality, including selectivity for temporal sequences of inputs103,104, and the ability to retrieve associative memories from partial or degraded inputs105,106. Another possibility would be to more accurately model the top-down connectivity of the hippocampus, in which multiple circuit layers receive direct feedback projections from entorhinal cortex, rather than each layer receiving feedback only from adjacent layers107,108. Such models could help to inform the design of brain-region- and cell-type-specific circuit interventions for neurological disorders.

In this work, we demonstrate a biologically-plausible mechanism for the brain to approximate learning by gradient descent. However, to the extent that an algorithm precisely aligns with backpropagation, it will also suffer from the same drawbacks as backpropagation, including the well-known problem of “catastrophic forgetting,” where sequential training on multiple tasks causes degraded performance on previously learned tasks109. Interestingly, some previously proposed methods for reducing memory interference and improving task generalization involved imposing constraints on growth of synaptic weights, sparsity of population activity, or sparsity of gradients110113. These regularization methods have direct analogies to many of the biological mechanisms implemented in our dendritic EIANNs, including weight normalization and regulation of activity and plasticity by feedback inhibition. In the future, it will be important to identify which aspects of biological learning allow the brain to mitigate memory interference and to learn efficiently from few training examples114.

In conclusion, our work leverages the computational power of biological neuronal microcircuits and neurons with compartmentalized signaling in dendrites to implement a deep learning algorithm that can solve nonlinear classification tasks in a manner fully compatible with principles of neuroscience. The recent discovery of behavioral timescale synaptic plasticity (BTSP) expanded the repertoire of known biological plasticity rules, and its unique features appear to offer advantages over previously characterized unsupervised learning rules, like spike timing-dependent plasticity (STDP), for target-directed learning11. However, the theoretical implications of BTSP for understanding and modeling biological learning have only begun to be explored18,32,33. Our results provide new insight into the role BTSP plays in enabling behavioral outcomes to drive synaptic changes distributed across multiple brain areas.

Methods

Network architecture and dynamics

Specification and training of ANN and EIANN models were performed in Python using a custom extension to PyTorch115,116. To train networks on the MNIST handwritten digit character dataset117, pixel intensities of image stimuli (28 × 28 pixels) were normalized between zero and one and encoded in the activities of 784 input neurons. Hidden layers (l=1 and l=2) each contained 500 excitatory neurons, 50 soma-targeting inhibitory neurons, and, when included, 50 dendrite-targeting inhibitory neurons. The output layer (L=3) contained 10 excitatory neurons (one for each stimulus class) and 10 soma-targeting inhibitory neurons. For networks trained on the 2D spirals dataset, stimuli were encoded in the activities of 2 input neurons. Hidden layer l=1 contained 128 excitatory neurons, 32 soma-targeting inhibitory neurons and, when included, 54 dendrite-targeting inhibitory neurons. Hidden layer l=2 contained 32 excitatory neurons, 8 soma-targeting inhibitory neurons, and when included, 13 dendrite-targeting inhibitory neurons. The output layer (L=3) contained 4 excitatory neurons (one for each stimulus class) and 4 soma-targeting inhibitory neurons.

During training, stimuli were presented to a network one at a time (batch size of 1) to mimic serial sampling of the sensory environment by biological organisms. For networks with recurrent (lateral) connections between separate excitatory and inhibitory cells (EIANNs), stimulus inputs were held constant for 15 time steps during a sensory phase while the somatic states and output firing rates of the neurons in all layers were equilibrated (Supplementary Fig. S1). Excitatory neuron somatic states (S(E,l)) were updated in each time step as:

τΔsi(E,l)[t]=si(E,l)[t1]+jwij(l)aj(E,l1)[t]+kyik(SomaI,l)ak(SomaI,l)[t1] (21)

where τ=3 is a time constant. Soma-targeting inhibitory neuron somatic states were updated as:

τΔsi(SomaI,l)[t]=si(SomaI,l)[t1]+jwij(SomaI,l)aj(E,l1)[t]+mqim(SomaI,l)am(E,l)[t1]+krik(SomaI,l)ak(SomaI,l)[t1] (22)

Dendrite-targeting inhibitory neuron somatic states were updated as:

τΔsiDendI,lt=siDendI,lt1+jqijiajE,lt1+krikDendI,laikDendI,lt1 (23)

The weights of self-connections (i=k) between recurrently connected interneurons (rik(SomaI,l) or rik(DendI,l)) were set to zero. Finally, the ReLU activation function was used to determine the activities of neurons from their somatic states:

A=f(Z)=0,Z0Z,Z>0 (24)

At the end of the sensory phase, output layer dendritic states D^(L) and activities Aˆ(L) were updated by a supervisory "nudge" according to Equations (6)(8). In hidden layers, the "nudged" dendritic states were then computed as follows:

dˆi(l)=jbij(l)aˆj(E,l+1)+kyik(DendI,l)ak(DendI,l) (25)

Dendritic states D^(l) were then bounded between −1 and 1 and held constant during the supervised phase for an additional 15 time steps while the somatic states S˜(l) and firing rates A˜(l) of all neurons re-equilibrated according to Equations (21)(24) and (8).

Learning rules

When backpropagation was used to learn the bottom-up excitatory weights W(E,l) (Fig. 2a,b and Fig. 4), a global loss function was defined as the mean squared error of the output neurons relative to a one-hot target vector:

=1nintiai(E,L)2 (26)

Then, each synaptic weight was updated proportionally to the gradient of loss with respect to the weight:

Δwij(E,l)=ηW(E,l)ai(E,l)gi(E,l)aj(E,l1) (27)

In networks with EIANN recurrent architectures, the loss gradient was backpropagated through 3 equilibration time steps. When the incoming and outgoing projections of soma-targeting interneurons were learned by backpropagation (Fig. 2b, and Supplementary Fig. S1S3), the bottom-up and lateral weights (W(SomaI,l),Q(SomaI,l), and R(SomaI,l)) were updated using the same method.

In Fig. 2c, the Hebbian learning rule with weight normalization described in Equations (1)(3) was used to update the bottom-up (W(E,l)) and lateral (Y(SomaI,l)) weights onto excitatory neurons and the bottom-up (W(SomaI,l)) and lateral (Q(SomaI,l) and R(SomaI,l)) weights onto soma-targeting inhibitory neurons in the hidden layers. For that network, the bottom-up excitatory weights W(E,L) onto output layer excitatory neurons were updated using the error-dependent rule described in Equations (6)(9). The bottom-up (W(SomaI,l)) and lateral (Q(SomaI,l) and R(SomaI,l)) weights onto soma-targeting inhibitory neurons in the output layer were updated according to Equations (1)(3,) after activities A˜(E,L) and A˜(SomaI,l) had equilibrated during the supervised phase for an additional 15 time steps after the "nudge" was applied to the excitatory neurons.

When the BCM rule was used to update the bottom-up excitatory weights W(E,l) using Equation (16) (Fig. 5b), sliding plasticity thresholds θ(l) were updated after each stimulus presentation as:

τθ(E,l)Δθi(l)=θi(l)+(a˜i(E,l))2k(l) (28)

where τθ(E,l) is a time constant, k(l) is a constant scale factor, and a˜i(E,l) is the activity of postsynaptic neuron i after supervisory inputs have been applied and the network has reequilibrated.

The BTSP rule defined in Equation (17) refers to presynaptic eligibility traces ET(l1) and postsynaptic instructive signals IS(l). When both presynaptic activity a˜j(E,l1)[x] and postsynaptic dendritic state dˆi(l)[x] were greater than zero in response to the current stimulus x, then ETj(l1)[x]=a˜j(E,l1)[x] (bounded to be <=1), ISi(l)[x]=1, and the weight wij(E,l) was updated according to Equation (17). The sigmoidal function σ(z) referenced in Equation (17) is scaled and offset to meet the edge constraints σ(0)=0 and σ(1)=1:

σˆ(z)=11+e(z0.01)/0.005σ(z)=σˆ(z)σˆ(0)σˆ(1)σˆ(0) (29)

The BTSP rule also prescribes weight updates for inputs active at long time delays from a dendritic calcium spike18. Here we modeled this aspect in the following way: when a presynaptic unit j was active in response to the previous stimulus x1, then the eligibility trace ETj(l1)[x1] was decayed by a factor of λ(l) by the time of stimulus x, resulting in ETj(l1)[x1]λ(l)ETj(l1)[x1] and an additional weight update according to:

Δwij(l)=ηWdˆi(E,l)[x]gi(E,l)(wmax(l)wij(l)ETj(l1)[x1]ISi(l)[x]wij(l)σETj(l1)[x1]ISi(l)[x]) (30)

When a postsynaptic uniťs dendritic state dˆi(l)[x1] in response to the previous stimulus x1 was greater than zero, then the instructive signal ISi(l)[x1] was decayed by a factor of λ(l) by the time of stimulus x, resulting in ISi(l)[x1]=λ(l) and an additional weight update according to:

Δwij(l)=ηWdˆiE,lx1giE,l(wmax(l)wij(l)ETj(l1)[x]ISi(l)[x1]wij(l)σETj(l1)[x]ISi(l)[x1]) (31)

Finally, while the experimentally-observed BTSP learning rule is defined when calcium spikes are triggered by dendritic depolarization above rest (positive dendritic state), it is not known how weights may be updated when dendrites are hyperpolarized below rest (negative dendritic state). In this case when dˆi(l)[x]<0, weights were updated using the LDS rule described in Equation (9).

Network training and hyperparameter optimization

The weights of each synaptic projection between cell types and layers were initialized from uniform distributions, respecting the signed synapse constraint, with projection-specific scales selected during hyperparameter optimization (discussed below; see Supplementary Table S3S6). During training, five instances of each network were separately initialized using different random seeds, and separately trained on the same stimulus samples, but presented in a different shuffled order. For the MNIST handwritten digit classification task, each network instance was trained by serially presenting either 20,000 or 50,000 images (Supplementary Table S1) without repeat (less than a single epoch of the dataset), with one weight update applied after each image. Final task accuracy was evaluated by presenting a separate test set of 10,000 images not used during training, and quantified as the percent of test samples where the most active (excitatory) output neuron corresponded to the correct stimulus class. During hyperparameter optimization, task accuracy was evaluated by presenting a separate validation set of 10,000 images not used during either training or final testing, and averaging the validation accuracy across the five network instances. For the 2D spirals task, networks were trained for either 1 or 10 epochs on a train set of 5,600 samples (Supplementary Table S2). Accuracy was evaluated with distinct train and validation sets of 1,200 samples each.

Each network was specified by a set of hyperparameters that included projection-specific scales for weight initialization or weight normalization, projection-specific learning rates, and parameters internal to each synaptic learning rule (Supplementary Table S3S6). For each network, these hyperparameters were searched within specified bounds using custom Python code implementing an iterative population-based search based on the basin-hopping algorithm118,119 that evaluated 30,000 variants of each network. Optimization aimed to minimize the following objectives, each computed as an average across the five instances of each network variant: 1) MSE loss (Equation (26)); 2) validation accuracy error (100% - validation accuracy); 3) sensory-phase dendritic state, averaged across cells; and 4) residual oscillations in activity after sensory-phase equilibration, averaged across cells (Supplementary Figure S1). The resulting hyperparameters for each network are shown in Supplementary Table S3S6. For an excitatory projection with n presynaptic neurons and weights W(l), a hyperparameter Winit(l) was used to initialize weights from a uniform distribution between zero and Winit(l)/n. When top-down weights B(l) were set as symmetric to bottom-up weights W(l+1), a hyperparameter Bscale(l) was used to rescale the top-down weights as B(l)=Bscale(l)W(l+1).

Analysis of neuronal stimulus selectivity

The stimulus selectivity of a neuron (Fig. 24 and Supplementary Fig. S2 and S4) was quantified as:

S=1kak/n2kak2/n11/n (32)

where ak is the activity in response to stimulus k, and n is the number of stimuli in a dataset120. A value of zero corresponds to uniform responses to all stimuli, and a value of one indicates that a unit responds maximally to a single stimulus and minimally to all other stimuli.

For networks trained to classify handwritten digits, we also analyzed the stimulus preferences of individual neurons by visualizing their receptive fields in image space using an algorithm adapted from previous work50 (Fig. 2 and 4 and Supplementary Fig. S3S5). We selected 1000 random images from the MNIST test dataset and presented them to trained networks with frozen parameters. For each unit in the network, we defined a loss function equal to the negative of the unit’s activity, and computed the gradient of this loss function with respect to each pixel in each presented image. These gradients were then averaged over the 1000 images and normalized, resulting in a visualization representing the average positive or negative influence of each pixel on the unit’s activity. We quantified the spatial structure of these two-dimensional receptive fields (Fig. 2 and 4 and Supplementary Fig. S4) using Moran’s I measure of spatial autocorrelation:

I=NWi=1Nj=1Nwijxixxjxi=1Nxix2 (33)

where N is the number of pixels, xi and xj are the intensities of a pair of pixels i and j,x is the mean pixel intensity, wij is a spatial weight between a pair of pixels, and W is the sum of all weights wij. The weights were set to one for neighboring pixels and zero otherwise, resulting in I values close to zero for random noise images and approaching one for images with strong correlations across pixels within spatial domains.

Supplementary Material

Supplement 1

Acknowledgements

This work was supported by funding from Rutgers Biomedical and Health Sciences (A.P., A.D.M.), NIMH grant R01MH121979 (A.P., A.R.G., A.D.M.), NIMH grant R01MH135576 (A.R.G., A.D.M.), EMBO postdoctoral fellowship ALTF 473–2022 (A.R.G.), and Rutgers Aresty Research Center (Y.C.). Large-scale compute resources were provided by Rutgers OARC (Amarel) and NSF-supported allotments from TACC (Frontera) and ACCESS (Bridges-2). We appreciated the opportunity to discuss and work out early concepts and code with students Tyler Rivera, Sandhya Senthilkumar, Barbara Gruszka, Maya Salameh, Michael Finch, Anshul Voleti, and Michael Cheng.

References

  • 1.Sweatt J. D. Neural plasticity and behavior - sixty years of conceptual advances. J Neurochem 13G Suppl 2, 179–199, (2016). [DOI] [PubMed] [Google Scholar]
  • 2.Richards B. A. & Kording K. P. The study of plasticity has always been about gradients. J Physiol 601, 3141–3149, (2023). [DOI] [PubMed] [Google Scholar]
  • 3.Minsky M. Steps toward Artificial Intelligence. P Ire 4G, 8–30, (1961). [Google Scholar]
  • 4.Richards B. A., Lillicrap T. P. et al. A deep learning framework for neuroscience. Nat Neurosci 22, 1761–1770, (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Rumelhart D. E., Hinton G. E. et al. Learning Representations by Back-Propagating Errors. Nature 323, 533–536, (1986). [Google Scholar]
  • 6.Schmidhuber J. Deep learning in neural networks: An overview. Neural Networks 61, 85–117, (2015). [DOI] [PubMed] [Google Scholar]
  • 7.Lillicrap T. P., Santoro A. et al. Backpropagation and the brain. Nature Reviews Neuroscience 21, 335–346, (2020). [DOI] [PubMed] [Google Scholar]
  • 8.Scellier B. & Bengio Y. Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation. Front Comput Neurosci 11, 24, (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Bengio Y., Mesnard T. et al. STDP-compatible approximation of backpropagation in an energy-based model. Neural computation 2G, 555–577, (2017). [DOI] [PubMed] [Google Scholar]
  • 10.Markram H., Gerstner W. et al. Spike-timing-dependent plasticity: a comprehensive overview. Front Synaptic Neurosci 4, 2, (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Magee J. C. & Grienberger C. Synaptic Plasticity Forms and Functions. Annu Rev Neurosci 43, 95–117, (2020). [DOI] [PubMed] [Google Scholar]
  • 12.Luo L. Architectures of neuronal circuits. Science 373, eabg7285, (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Strata P. & Harvey R. Dale's principle. Brain Res Bull 50, 349–350, (1999). [DOI] [PubMed] [Google Scholar]
  • 14.Kolen J. F. & Pollack J. B. in Proceedings of 1SS4 IEEE International Conference on Neural Networks (ICNN'S4). 1375–1380 (IEEE; ). [Google Scholar]
  • 15.Lillicrap T. P., Cownden D. et al. Random synaptic feedback weights support error backpropagation for deep learning. Nat Commun 7, 13276, (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Akrout M., Wilson C. et al. Deep learning without weight transport. Advances in neural information processing systems 32, (2019). [Google Scholar]
  • 17.Bittner K. C., Milstein A. D. et al. Behavioral time scale synaptic plasticity underlies CA1 place fields. Science 357, 1033–1036, (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Milstein A. D., Li Y. et al. Bidirectional synaptic plasticity rapidly modifies hippocampal representations. Elife 10, (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Grienberger C. & Magee J. C. Entorhinal cortex directs learning-related changes in CA1 representations. Nature 611, 554–562, (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Pagkalos M., Makarov R. et al. Leveraging dendritic properties to advance machine learning and neuro-inspired computing. Curr Opin Neurobiol 85, 102853, (2024). [DOI] [PubMed] [Google Scholar]
  • 21.Sacramento J., Costa R. P. et al. in Advances in Neural Information Processing Systems. 8721–8732. [Google Scholar]
  • 22.Guerguiev J., Lillicrap T. P. et al. Towards deep learning with segregated dendrites. Elife 6, (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Körding K. P. & König P. Learning with two sites of synaptic integration. Network: Computation in Neural Systems 11, 25–39, (2000). [PubMed] [Google Scholar]
  • 24.Kording K. P. & Konig P. Supervised and unsupervised learning with two sites of synaptic integration. J Comput Neurosci 11, 207–215, (2001). [DOI] [PubMed] [Google Scholar]
  • 25.Richards B. A. & Lillicrap T. P. Dendritic solutions to the credit assignment problem. Curr Opin Neurobiol 54, 28–36, (2019). [DOI] [PubMed] [Google Scholar]
  • 26.Payeur A., Guerguiev J. et al. Burst-dependent synaptic plasticity can coordinate learning in hierarchical circuits. Nat Neurosci 24, 1010–1019, (2021). [DOI] [PubMed] [Google Scholar]
  • 27.Greedy W., Zhu H. W. et al. Single-phase deep learning in cortico-cortical networks. Advances in neural information processing systems 35, 24213–24225, (2022). [Google Scholar]
  • 28.Francioni V., Tang V. D. et al. Vectorized instructive signals in cortical dendrites during a brain-computer interface task. bioRxiv, (2025). [Google Scholar]
  • 29.Takahashi H. & Magee J. C. Pathway interactions and synaptic plasticity in the dendritic tuft regions of CA1 pyramidal neurons. Neuron 62, 102–111, (2009). [DOI] [PubMed] [Google Scholar]
  • 30.Dudman J. T., Tsay D. et al. A role for synaptic inputs at distal dendrites: instructive signals for hippocampal long-term plasticity. Neuron 56, 866–879, (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Bittner K. C., Grienberger C. et al. Conjunctive input processing drives feature selectivity in hippocampal CA1 neurons. Nat Neurosci 18, 1133–1142, (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Li P. Y. & Roxin A. Rapid memory encoding in a recurrent network model with behavioral time scale synaptic plasticity. PLoS Comput Biol 1G, e1011139, (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Wu Y. & Maass W. A simple model for Behavioral Time Scale Synaptic Plasticity (BTSP) provides content addressable memory with binary synapses and one-shot learning. Nat Commun 16, 342, (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Cone I. & Shouval H. Z. Behavioral Time Scale Plasticity of Place Fields: Mathematical Analysis. Front Comput Neurosci 15, 640235, (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Rolotti S. V., Ahmed M. S. et al. Local feedback inhibition tightly controls rapid formation of hippocampal place fields. Neuron 110, 783–794 e786, (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Pouille F., Marin-Burgin A. et al. Input normalization by global feedforward inhibition expands cortical dynamic range. Nature Neuroscience 12, 1577–U1131, (2009). [DOI] [PubMed] [Google Scholar]
  • 37.Wilson N. R., Runyan C. A. et al. Division and subtraction by distinct cortical inhibitory networks. Nature 488, 343–348, (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Pérez-Ortega J., Akrouh A. et al. Stimulus encoding by specific inactivation of cortical neurons. Nature Communications 15, (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.McKenzie S. Inhibition shapes the organization of hippocampal representations. Hippocampus 28, 659–671, (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Bernacchia A. & Wang X. J. Decorrelation by Recurrent Inhibition in Heterogeneous Neural Circuits. Neural Computation 25, 1732–1767, (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Wilson H. R. & Cowan J. D. Excitatory and Inhibitory Interactions in Localized Populations of Model Neurons. Biophysical Journal 12, 1–24, (1972). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Hansel D. & Mato G. Asynchronous states and the emergence of synchrony in large networks of interacting excitatory and inhibitory neurons. Neural Computation 15, 1–56, (2003). [DOI] [PubMed] [Google Scholar]
  • 43.Brunel N. & Wang X. J. What determines the frequency of fast network oscillations with irregular neural discharges? I. Synaptic dynamics and excitation-inhibition balance. Journal of Neurophysiology G0, 415–430, (2003). [DOI] [PubMed] [Google Scholar]
  • 44.Li P. S., Cornford J. et al. Learning better with Dale's Law: A Spectral Perspective. Adv Neur In, (2023). [Google Scholar]
  • 45.DeFelipe J., Alonso-Nanclares L. et al. Microstructure of the neocortex: comparative aspects. J Neurocytol 31, 299–316, (2002). [DOI] [PubMed] [Google Scholar]
  • 46.Freund T. F. & Buzsaki G. Interneurons of the hippocampus. Hippocampus 6, 347–470, (1996). [DOI] [PubMed] [Google Scholar]
  • 47.Chapeton J., Fares T. et al. Efficient associative memory storage in cortical circuits of inhibitory and excitatory neurons. P Natl Acad Sci USA 10G, E3614–E3622, (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.King P. D., Zylberberg J. et al. Inhibitory Interneurons Decorrelate Excitatory Cells to Drive Sparse Code Formation in a Spiking Model of V1. Journal of Neuroscience 33, 5475–5485, (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Rozell C. J., Johnson D. H. et al. Sparse coding via thresholding and local competition in neural circuits. Neural Computation 20, 2526–2563, (2008). [DOI] [PubMed] [Google Scholar]
  • 50.Erhan D., Bengio Y. et al. Visualizing higher-layer features of a deep network. University of Montreal 1341, 1, (2009). [Google Scholar]
  • 51.Moran P. A. Notes on continuous stochastic phenomena. Biometrika 37, 17–23, (1950). [PubMed] [Google Scholar]
  • 52.Wang Ǫ., Powell M. A. et al. in Proceedings of the IEEE/CVF International Conference on Computer Vision. 22551–22559. [Google Scholar]
  • 53.Scholl B., Pattadkal J. J. et al. Local Integration Accounts for Weak Selectivity of Mouse Neocortical Parvalbumin Interneurons. Neuron 87, 424–436, (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Kerlin A. M., Andermann M. L. et al. Broadly tuned response properties of diverse inhibitory neuron subtypes in mouse visual cortex. Neuron 67, 858–871, (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Ego-Stengel V. & Wilson M. A. Spatial selectivity and theta phase precession in CA1 interneurons. Hippocampus 17, 161–174, (2007). [DOI] [PubMed] [Google Scholar]
  • 56.Geiller T., Vancura B. et al. Large-Scale 3D Two-Photon Imaging of Molecularly Identified CA1 Interneuron Dynamics in Behaving Mice. Neuron 108, 968–983 e969, (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Grienberger C., Milstein A. D. et al. Inhibitory suppression of heterogeneously tuned excitation enhances spatial coding in CA1 place cells. Nat Neurosci 20, 417–426, (2017). [DOI] [PubMed] [Google Scholar]
  • 58.Journé A., Rodriguez H. G. et al. Hebbian deep learning without feedback. arXiv preprint arXiv:220S.11883, (2022). [Google Scholar]
  • 59.Diehl P. U. & Cook M. Unsupervised learning of digit recognition using spike-timing-dependent plasticity. Frontiers in Computational Neuroscience G, (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Krotov D. & Hopfield J. J. Unsupervised learning by competing hidden units. P Natl Acad Sci USA 116, 7723–7731, (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Illing B., Gerstner W. et al. Biologically plausible deep learning — But how far can we go with shallow networks? Neural Networks 118, 90–101, (2019). [DOI] [PubMed] [Google Scholar]
  • 62.Eckmann S., Young E. J. et al. Synapse-type-specific competitive Hebbian learning forms functional recurrent networks. Proc Natl Acad Sci U S A 121, e2305326121, (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Turrigiano G. G. The dialectic of Hebb and homeostasis. Philos Trans R Soc Lond B Biol Sci 372, (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Keck T., Toyoizumi T. et al. Integrating Hebbian and homeostatic plasticity: the current state of the field and future research directions. Philos Trans R Soc Lond B Biol Sci 372, (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Royer S. & Paré D. Conservation of total synaptic weight through balanced synaptic depression and potentiation. Nature 422, 518–522, (2003). [DOI] [PubMed] [Google Scholar]
  • 66.Xie X. & Seung H. S. Equivalence of backpropagation and contrastive Hebbian learning in a layered network. Neural Comput 15, 441–454, (2003). [DOI] [PubMed] [Google Scholar]
  • 67.Bengio Y. How auto-encoders could provide credit assignment in deep networks via target propagation. arXiv preprint arXiv:1407.7S0C, (2014). [Google Scholar]
  • 68.Lee D.-H., Zhang S. et al. 498–515 (Springer International Publishing; ). [Google Scholar]
  • 69.Whittington J. C. R. & Bogacz R. An Approximation of the Error Backpropagation Algorithm in a Predictive Coding Network with Local Hebbian Synaptic Plasticity. Neural Comput 2G, 1229–1262, (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Meulemans A., Farinha M. T. et al. Credit Assignment in Neural Networks through Deep Feedback Control. Advances in Neural Information Processing Systems 34 (Neurips 2021) 34, (2021). [Google Scholar]
  • 71.Dorian C. C., Taxidis J. et al. Behavioral timescale synaptic plasticity in the hippocampus creates non-spatial representations during learning and is modulated by entorhinal inputs. bioRxiv, (2025). [Google Scholar]
  • 72.Fan L. Z., Kim D. K. et al. All-optical physiology resolves a synaptic basis for behavioral timescale plasticity. Cell 186, 543–559 e519, (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Gonzalez K. C., Negrean A. et al. Synaptic basis of feature selectivity in hippocampal neurons. Nature 637, 1152–1160, (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Murayama M., Perez-Garci E. et al. Dendritic encoding of sensory stimuli controlled by deep cortical interneurons. Nature 457, 1137–1141, (2009). [DOI] [PubMed] [Google Scholar]
  • 75.Lovett-Barron M., Kaifosh P. et al. Dendritic inhibition in the hippocampus supports fear learning. Science 343, 857–863, (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Bowler J. C. & Losonczy A. Direct cortical inputs to hippocampal area CA1 transmit complementary signals for goal-directed navigation. Neuron 111, 4071–4085 e4076, (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Murayama M. & Larkum M. E. Enhanced dendritic activity in awake rats. Proc Natl Acad Sci U S A 106, 20482–20486, (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Xu N. L., Harnett M. T. et al. Nonlinear dendritic integration of sensory and motor input during an active sensing task. Nature 4G2, 247–251, (2012). [DOI] [PubMed] [Google Scholar]
  • 79.Ranganathan G. N., Apostolides P. F. et al. Active dendritic integration and mixed neocortical network representations during an adaptive sensing behavior. Nat Neurosci 21, 1583–1590, (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Major G., Larkum M. E. et al. Active properties of neocortical pyramidal neuron dendrites. Annu Rev Neurosci 36, 1–24, (2013). [DOI] [PubMed] [Google Scholar]
  • 81.Milstein A. D., Bloss E. B. et al. Inhibitory Gating of Input Comparison in the CA1 Microcircuit. Neuron 87, 1274–1289, (2015). [DOI] [PubMed] [Google Scholar]
  • 82.Oja E. Learning in nonlinear constrained Hebbian networks. Artificial neural networks, (1991). [Google Scholar]
  • 83.Song Y., Millidge B. et al. Inferring neural activity before plasticity as a foundation for learning beyond backpropagation. Nature Neuroscience 27, 348–358, (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.O'Reilly R. C. Biologically Plausible Error-Driven Learning Using Local Activation Differences: The Generalized Recirculation Algorithm. Neural Computation 8, 895–938, (1996). [Google Scholar]
  • 85.Bienenstock E. L., Cooper L. N. et al. Theory for the development of neuron selectivity: orientation specificity and binocular interaction in visual cortex. J Neurosci 2, 32–48, (1982). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86.Abraham W. C. Metaplasticity: tuning synapses and networks for plasticity. Nat Rev Neurosci G, 387, (2008). [DOI] [PubMed] [Google Scholar]
  • 87.Liao Ǫ., Leibo J. et al. in Proceedings of the AAAI Conference on Artificial Intelligence. [Google Scholar]
  • 88.Meulemans A., Carzaniga F. et al. A theoretical framework for target propagation. Advances in Neural Information Processing Systems 33, 20024–20036, (2020). [Google Scholar]
  • 89.Ernoult M. M., Normandin F. et al. in Proceedings of the 3Sth International Conference on Machine Learning Vol. 162 (eds Kamalika Chaudhuri et al. ) 5968--5987 (PMLR, Proceedings of Machine Learning Research, 2022). [Google Scholar]
  • 90.Kunin D., Nayebi A. et al. Two Routes to Scalable Credit Assignment without Weight Symmetry. arXiv preprint arXiv:2003.01513, (2020). [Google Scholar]
  • 91.Hinton G. E. Deterministic Boltzmann Learning Performs Steepest Descent in Weight-Space. Neural Computation 1, 143–150, (1989). [Google Scholar]
  • 92.Campagnola L., Seeman S. C. et al. Local connectivity and synaptic dynamics in mouse and human neocortex. Science 375, 1144–+, (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93.Chavlis S. & Poirazi P. Dendrites endow artificial neural networks with accurate, robust and parameter-efficient learning. Nat Commun 16, 943, (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94.Jones I. S. & Kording K. P. Might a Single Neuron Solve Interesting Machine Learning Problems Through Successive Computations on Its Dendritic Tree? Neural Comput 33, 1554–1571, (2021). [DOI] [PubMed] [Google Scholar]
  • 95.Beniaguev D., Segev I. et al. Single cortical neurons as deep artificial neural networks. Neuron 10G, 2727–2739 e2723, (2021). [DOI] [PubMed] [Google Scholar]
  • 96.Wang J. H. & Stelzer A. Shared calcium signaling pathways in the induction of long-term potentiation and synaptic disinhibition in CA1 pyramidal cell dendrites. Journal of Neurophysiology 75, 1687–1702, (1996). [DOI] [PubMed] [Google Scholar]
  • 97.Honoré E., Khlaifia A. et al. Hippocampal Somatostatin Interneurons, Long-Term Synaptic Plasticity and Memory. Front Neural Circuit 15, (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98.Tzilivaki A., Tukker J. J. et al. Hippocampal GABAergic interneurons and memory. Neuron 111, 3154–3175, (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99.McFarlan A. R., Chou C. Y. C. et al. The plasticitome of cortical interneurons. Nature Reviews Neuroscience 24, 80–97, (2023). [DOI] [PubMed] [Google Scholar]
  • 100.Wright W. J., Hedrick N. G. et al. Distinct synaptic plasticity rules operate across dendritic compartments in vivo during learning. Science 388, 322–328, (2025). [DOI] [PubMed] [Google Scholar]
  • 101.Caya-Bissonnette L., Naud R. et al. Cellular Substrate of Eligibility Traces. bioRxiv, 2023.2006.2029.547097, (2023). [Google Scholar]
  • 102.O'Hare J. K., Gonzalez K. C. et al. Compartment-specific tuning of dendritic feature selectivity by intracellular Ca(2+) release. Science 375, eabm1670, (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 103.Farovik A., Dupont L. M. et al. Distinct roles for dorsal CA3 and CA1 in memory for sequential nonspatial events. Learn Memory 17, 801–806, (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 104.Deveau C. E., Zhou Z. et al. Recurrent cortical networks encode natural sensory statistics via sequence filtering. bioRxiv, (2024). [Google Scholar]
  • 105.Rolls E. T. The mechanisms for pattern completion and pattern separation in the hippocampus. Front Syst Neurosci 7, 74, (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 106.Shin H., Ogando M. B. et al. Recurrent pattern completion drives the neocortical representation of sensory inference. bioRxiv, (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 107.Meulemans A., Farinha M. T. et al. in International Conference on Machine Learning. 15458–15483 (PMLR; ). [Google Scholar]
  • 108.Sucevic J. & Schapiro A. C. A neural network model of hippocampal contributions to category learning. Elife 12, (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109.French R. M. Catastrophic forgetting in connectionist networks. Trends Cogn Sci 3, 128–135, (1999). [DOI] [PubMed] [Google Scholar]
  • 110.Wang Z. F., Zhan Z. et al. SparCL: Sparse Continual Learning on the Edge. Advances in Neural Information Processing Systems 35 (Neurips 2022), (2022). [Google Scholar]
  • 111.Hurtado J., Lobel H. et al. Overcoming Catastrophic Forgetting Using Sparse Coding and Meta Learning. Ieee Access G, 88279–88290, (2021). [Google Scholar]
  • 112.Moraitis T., Toichkin D. et al. SoftHebb: Bayesian inference in unsupervised Hebbian soft winner-take-all networks. Neuromorph Comput En 2, (2022). [Google Scholar]
  • 113.Kirkpatrick J., Pascanu R. et al. Overcoming catastrophic forgetting in neural networks. P Natl Acad Sci USA 114, 3521–3526, (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 114.Galloni A. R., Yuan Y. et al. Neuromorphic one-shot learning utilizing a phase-transition material. Proceedings of the National Academy of Sciences 121, e2318362121, (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 115.Ansel J., Yang E. et al. in Proceedings of the 2Sth ACM International Conference on Architectural Support for Programming Languages and Operating Systems, Volume 2. 929–947. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 116.Galloni A. R., Chennawar Y. et al. Code repository for training biology-constrained dendritic EIANNs, <https://github.com/Milstein-Lab/EIANN> (2025).
  • 117.LeCun Y. The MNIST database of handwritten digits, <http://yann.lecun.com/exdb/mnist/> (1998).
  • 118.Wales D. J. & Doye J. P. Global optimization by basin-hopping and the lowest energy structures of Lennard-Jones clusters containing up to 110 atoms. The Journal of Physical Chemistry A 101, 5111–5116, (1997). [Google Scholar]
  • 119.Milstein A. D. Code repository for nested: parallel multi-objective optimization software, <https://github.com/neurosutras/nested> (2021).
  • 120.Vinje W. E. & Gallant J. L. Sparse coding and decorrelation in primary visual cortex during natural vision. Science 287, 1273–1276, (2000). [DOI] [PubMed] [Google Scholar]
  • 121.Whittington J. C. R. & Bogacz R. Theories of Error Back-Propagation in the Brain. Trends Cogn Sci 23, 235–250, (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 122.Bartunov S., Santoro A. et al. Assessing the scalability of biologically-motivated deep learning algorithms and architectures. arXiv preprint arXiv:1807.04587, (2018). [Google Scholar]
  • 123.Krotov D. & Hopfield J. J. Unsupervised learning by competing hidden units. Proc Natl Acad Sci U S A 116, 7723–7731, (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 124.Song Y., Lukasiewicz T. et al. Can the brain do backpropagation?---exact implementation of backpropagation in predictive coding networks. Advances in neural information processing systems 33, 22566–22579, (2020). [PMC free article] [PubMed] [Google Scholar]
  • 125.Lansdell B. J., Prakash P. R. et al. Learning to solve the credit assignment problem. arXiv preprint arXiv:1S0C.0088S, (2019). [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplement 1

Articles from bioRxiv are provided here courtesy of Cold Spring Harbor Laboratory Preprints

RESOURCES