Significance
The brain represents information with neural activity patterns. At the periphery, these patterns contain correlations, which are detrimental to stimulus discrimination. We study the peripheral olfactory circuit of the Drosophila larva, which preprocesses neural representations before relaying them downstream. A comprehensive understanding of this preprocessing is, however, lacking. We formulate a principle-driven framework based on similarity-matching and, using neural input activity, derive a circuit model that largely explains the biological circuit’s synaptic organization. It also predicts that inhibitory neurons cluster odors and facilitate decorrelation and normalization of neural representations. If equipped with Hebbian synaptic plasticity, the circuit model autonomously adapts to different environments. Our work provides a comprehensive approach to deciphering the relationship between structure and function in neural circuits.
Keywords: olfaction, connectome, encoding, clustering, normative approach
Abstract
One major question in neuroscience is how to relate connectomes to neural activity, circuit function, and learning. We offer an answer in the peripheral olfactory circuit of the Drosophila larva, composed of olfactory receptor neurons (ORNs) connected through feedback loops with interconnected inhibitory local neurons (LNs). We combine structural and activity data and, using a holistic normative framework based on similarity-matching, we formulate biologically plausible mechanistic models of the circuit. In particular, we consider a linear circuit model, for which we derive an exact theoretical solution, and a nonnegative circuit model, which we examine through simulations. The latter largely predicts the ORN LN synaptic weights found in the connectome and demonstrates that they reflect correlations in ORN activity patterns. Furthermore, this model accounts for the relationship between ORN LN and LN–LN synaptic counts and the emergence of different LN types. Functionally, we propose that LNs encode soft cluster memberships of ORN activity, and partially whiten and normalize the stimulus representations in ORNs through inhibitory feedback. Such a synaptic organization could, in principle, autonomously arise through Hebbian plasticity and would allow the circuit to adapt to different environments in an unsupervised manner. We thus uncover a general and potent circuit motif that can learn and extract significant input features and render stimulus representations more efficient. Finally, our study provides a unified framework for relating structure, activity, function, and learning in neural circuits and supports the conjecture that similarity-matching shapes the transformation of neural representations.
Technological advances in connectomics (1, 2) and neural population activity imaging (3) enable the anatomical and physiological characterization of neural circuits at unprecedented scales and detail. However, it remains unclear how to combine these datasets to advance our understanding of brain computation. To address this, we focus on the peripheral olfactory system of the first instar Drosophila larva—a small and genetically tractable circuit with available connectivity and activity imaging datasets (4, 5).
This circuit is an analogous but simpler version of the well-studied olfactory circuit in adult flies and vertebrates (6). It contains 21 olfactory receptor neurons (ORNs), each expressing a different receptor type (Fig. 1A). ORN axons are reciprocally connected to a web of multiple interconnected inhibitory local neurons (LNs) through feedforward excitation and feedback inhibition. The connectome dataset contains not only the presence or absence of a connection between two neurons, but also the number of synaptic contacts in parallel (4), which is an estimate of the connection strength (2, 7–9) (nonetheless, other factors like release probability and active zone properties also affect synaptic strength (10, 11)).
Previous studies examined the role of LNs in transforming the neural representation of odors from ORN somas to downstream projection neurons (PNs). In adult Drosophila, this circuit was suggested to perform gain control and divisive normalization (12, 13), which equalizes different odor concentrations and decorrelates input channels. In the zebrafish larva, an analogous circuit was suggested to whiten the input, leading to pattern decorrelation, which helps odor discrimination downstream (14, 15).
However, the underlying mechanistic principles of computation remain elusive. For example, while different types of LNs have different connectivity patterns with ORNs in the Drosophila larva (4), the role of different LN types, their multiplicity, and their specific connectivity is not yet understood. Furthermore, the peripheral olfactory circuit of adult Drosophila exhibits synaptic plasticity in response to changes in the olfactory environment (16–19), but the functional role of this plasticity is unclear.
To address these shortcomings, we use a combination of data analysis and modeling and develop a holistic theoretical framework that links circuit structure, function, activity data, and learning. Our contribution is fivefold: 1) We find that the vectors of the number of synapses between ORNs and LNs reflect features of the independently acquired ORN activity pattern dataset (Figs. 2 and 3). 2) Building upon the normative similarity-matching framework (20, 21), we develop an optimization problem solvable by a biologically realistic circuit model with the same architecture as the ORN-LN circuit. 3) The model, driven by the ORN activity dataset, largely predicts the following observations in the structural dataset (Figs. 3 and 4): the ORNs LN synaptic weights, the emergence of LN groups, and the relationship between feedforward ORN LN and lateral LN–LN connections. 4) Using our model, we characterize the circuit computation (Figs. 5 and 6), and propose that LNs play a dual role in rendering the neural representation of odors in ORNs more efficient and extracting useful features that are transmitted downstream. 5) We show that the synaptic weights that enable this computation can, in principle, be learned in an unsupervised manner via Hebbian plasticity. Note that, given the connectome (4) originates from a 6-h-old first instar Drosophila larva, new synaptic contact formation can take longer than 6 h (11), and no study has yet demonstrated activity-dependent plasticity in the larval ORN-LN circuit, it is unknown whether the observed synaptic counts in this connectome could result from activity-dependent synaptic plasticity.
In this study, we further our understanding of LNs and their computations. We highlight the importance of minutely organized ORN–LN and LN–LN connection weights, which allow LNs to encode different significant features of input activity and dampen them in ORN axons. The transformation from the representation in ORN somas to that in ORN axons consists of a partial equalization of PCA variances, which enables a more efficient stimulus encoding (22). In fact, this results in a decorrelation and equalization of ORNs and odor representations, which correspond to two fundamental computations in the brain: partial ZCA (zero-phase) whitening (23, 24) and divisive normalization (25). In essence, we uncover an elegant neural circuit motif that can extract features and perform two critical computations. If endowed with Hebbian plasticity, the circuit can also adapt and perform its functions in different stimulus environments. Thus, we present a framework that allows us to quantitatively link synaptic weights in the structural data with the circuit’s function and with the circuit adaptation to input correlations, thus making a crucial step toward a more integrated understanding of neural circuits.
The results are organized as follows. First, we show that the connectome is adapted to ORN activity patterns. Second, we propose a normative approach leading to two circuit models: a linear circuit (LC) model, and a nonnegative circuit (NNC) model. Third, we show that the NNC reproduces key structural observations. Finally, we describe the computations performed by the LC and NNC in general and on the ORN activity dataset in particular.
Results
ORN-LN Circuit.
ORNs in the Drosophila larva carry odor information from the antennas to the antennal lobe, where they synapse onto LNs and PNs. There, olfactory information is reformatted and transferred through ORN axons and LNs to PNs. LNs, which synapse bidirectionally with ORN axons and PN dendrites, strongly contribute to the reformatting in ORNs and PNs through presynaptic and postsynaptic inhibition, as shown mainly in the adult fly (12, 13, 26–30). LNs project to several uni- and multiglomerular PNs, and PNs project to higher brain areas such as the mushroom body and the lateral horn (4).
We study the circuit and computation presynaptic to PNs, i.e., occurring from ORN somas to ORN axons and LNs. Specifically, we examine the subcircuit formed by all ORNs and those 4 LN types (on each side of the brain) that reciprocally connect with ORNs (4) (Fig. 1A, SI Appendix, Fig. S1). The 4 LN types include 3 Broad Trio (BT) neurons, 2 Broad Duet (BD) neurons, 1 Keystone (KS, bilateral connections) neuron, and 1 Picky 0 (P0) neuron (SI Appendix, Figs. S1 and S2A). This amounts to 8 ORNs–LN connections per side (3 BTs, 2 BDs, 2 KSs, and 1 P0s) and 16 on both sides. See SI Appendix, Tables S1 and S2 for a list of all acronyms and mathematical variables used in the paper.
We use the number of synaptic contacts in parallel between two neurons as a proxy for the synaptic weight (2, 7–9) (but see refs. 10 and 11). In the linear approximation, the change in the postsynaptic neuron activity due to a change in the presynaptic neuron activity is proportional to the synaptic weight connecting them.
We focus our analysis on the synaptic counts of the feedforward ORNs LN connections. We call the dimensional vector containing the synaptic counts of the connections from the 21 ORNs to one LN. Because all the entries of this synaptic count vector share the same postsynaptic neuron, this vector is likely proportional to the corresponding synaptic weight vector. Conversely, the synaptic count vector from one LN to all 21 ORNs may not be proportional to the corresponding synaptic weight vector, because each connection affects a different postsynaptic ORN, which potentially has different electrical properties. This makes the entries of a feedback synaptic count vector not directly comparable. Yet, the feedforward and feedback synaptic count vectors are somewhat correlated (SI Appendix, Fig. S2).
While the study (4) divided LNs into the above types based on their neuronal lineage, morphology, and qualitative connectivity, we also find that these types are innervated differently by ORNs (Fig. 1B). Indeed, the average correlation of s within each LN type is higher than between LN types (Fig. 1C). Thus, for a part of our study (Figs. 2 and 3 A and B) we use the 4 average , where n is the number of connection vectors for that LN type.
ORNs LN Synaptic Count Vectors Are Adapted to Odor Representations in ORNs.
Several studies proposed that LNs could facilitate the decorrelation of the neural representation of odors (14, 15, 32–35). To perform such decorrelation, the circuit must be adapted to or “know about” the correlations in the activity patterns (36). We investigate whether this is the case in this olfactory circuit by testing whether the s contain signatures of ORN activity patterns.
An ensemble of ORN activity patterns () was obtained using Ca2+ fluorescence imaging of ORN somas in response to a set of 34 odorants at 5 dilutions (5) (Fig. 2A and SI Appendix). These odorants were chosen from the components of fruits and plant leaves from the larva’s natural environment to stimulate ORNs as broadly and evenly as possible, with many odorants activating just a single ORN at the lowest concentration (i.e., the highest dilution).
We examine the Pearson correlation coefficients between the activity patterns and the ORNs LN synaptic count vectors (Fig. 2 C and D for and two odors; Fig. 2B for all four s and all activity patterns ). After controlling for multiple comparisons (31), we find that the s for the Broad Trio and Picky 0 maintain significant correlations () with a selection of ORN activity patterns, BT being highly correlated with the largest set of s. This suggests that the synaptic count vectors of at least these two LN types are more adapted to these activity patterns than would be expected by chance (see SI Appendix, Fig. S4 and SI Appendix for additional evidence). This supports the hypothesis that the circuit is at least partially adapted to ORN activity patterns and that it could perform a computation like decorrelation of input stimuli.
Each exhibits a different “connectivity tuning curve” shape (Fig. 2G), being correlated with the largest set of s, and the most highly correlated to a few s, and the and the most weakly correlated. Biologically, this could signify that the BT type is activated by the largest set of odors and P0 only by a few odors. One possibility is that a different set of odors activates each LN class.
A Normative and Mechanistic Model of the ORN-LN Circuit.
We aim to understand the circuit’s computation and organization using a top-down, normative (also called principle-driven) approach, which involves formulating an optimization problem. Such an approach provides us with a theoretical understanding of the computation and organizational principles of the circuit. Although a bottom-up modeling approach requires unavailable physiological circuit parameters, we verified our predictions with a connectome-constrained model (Fig. 6).
Previous studies suggest that analogous circuits perform stimulus whitening or decorrelation (14, 15, 32–35), and our analysis above supports the possibility of such a computation. A class of optimization problems based on the similarity-matching principle and solvable by circuits similar to the ORN-LN one has been shown to be capable of implementing whitening, principal subspace extraction, and clustering (20, 21, 37). Note that the circuit’s synaptic weights are adapted (optimized) to the ensemble of inputs to perform such computation.
To understand the circuit, we first postulate an optimization problem (Eq. 4) based on the similarity-matching principle and solvable by a circuit with the ORN-LN architecture (see Methods and SI Appendix). To match this architecture, similarity-matching takes place between ORN axon and LN activities, which seeks to maintain that distances (similarities) between neural representations at the level of ORN axons and LNs. Specifically, if the representations of two odors are similar (dissimilar) in ORN axons, their representations will also tend to be similar (dissimilar) in LNs. Second, we derive the circuit models (Eqs. 5–7) that solve this optimization problem with the recorded ORN soma activity described above (5) as input. Third, we compare the synaptic weight organization of the circuit model with the connectome (4) (Figs. 2, 3, and 4) and find that the circuit model accounts for multiple experimental observations. We thus conclude that the similarity-matching principle and the optimization problem widely explain the biological circuit’s organization. Lastly, we describe in detail the computation performed by the circuit model (Figs. 5 and 6).
Mathematically, given a set of T activity patterns in D ORN somas as input, , the optimization provides us as output the activity patterns in the D ORN axons and K LNs . The circuit model performing the computation of the optimization has the following parameters: and , which are proportional to the connection weights between ORNs and LNs, and between LNs, respectively. In addition to K, the number of LNs, the model contains only one effective parameter , corresponding to the ratio between feedback inhibition and feedforward excitation strengths.
We consider two optimization problems leading to two circuit models, differing in their domain of optimization: 1) a linear circuit, LC-K with K LNs, Eq. 6, with no constraint on the optimization domain; 2) a nonnegative circuit, NNC-K, Eq. 7, with nonnegative constrains on ORN axon and LN activity (). This constraint renders the NNC more biologically plausible than the LC, and the NNC indeed predicts the structural data better than the LC (below). However, only for the LC we can derive the analytical expressions for , , , and , whereas for the NNC we must rely on numerical simulations (SI Appendix). Because both models are closely related, we examine the analytical solution of the LC to quantitatively understand the relationship between input and output variables, describe the circuit’s function in a mathematically tractable manner, and substantiate the numerical results for the NNC.
Given an input , the optimal synaptic weights can be found by solving the optimization problem offline (Eqs. 4 and 5), or online with Hebbian plasticity (Eq. 8). The latter implies that the circuit model’s synaptic weights can adapt to solve the optimization problem on any ORN activity patterns ensemble, in an unsupervised manner. This would correspond to activity-dependent synaptic plasticity in the biological circuit, which was, so far, only observed in the adult Drosophila (16–19). Given the specific wiring of some LNs such as Keystone and Picky 0 in the biological circuit (4), it is very likely that the synaptic weights of these (and potentially other) LNs are largely genetically predetermined and were set over evolutionary time scales (similar to an offline setting). It is unknown which mechanisms determine the synaptic weights in the biological circuit, and it is beyond the scope of this study to elucidate them.
Next, we characterize the computation performed by the LC and the NNC as well as the connectivity (in terms of and ) that supports the computation. In short, in the LC, LNs extract and encode the top K PCA subspace of the input in ORN somas and the ORNs LN synaptic weight vectors span that subspace. In the NNC, LNs encode soft cluster/feature memberships of the odor representations in ORN somas and are related to cluster locations. In both models, the ORN axons encode a partially whitened and normalized version of the ORN soma activity due to LN feedback inhibition.
Predictions of the ORN–LN Connection Weight Vectors.
We start by analyzing our models’ predictions in terms of circuit connectivity. In the LC-K, the (proportional to the ORNs LN connection weight vectors) are linearly independent and span the same K dimensional subspace as the top K PCA directions of the uncentered input (SI Appendix):
[1] |
This ensures that LNs extract the top K PCA subspace of the input (below). The are coefficients with a degree of freedom, arising from the nonuniqueness of the optimization solution. Thus, the s do not necessarily correspond to specific PCA directions of the input and are not orthogonal. Because the model predictions rely on “uncentered PCA,” i.e., PCA without prior centering of the data, we use such PCA throughout the paper.
We probe this structural prediction by testing the alignment between the four ORNs LN synaptic count vectors, and the first 5 PCA directions of the ORN activity data, (Fig. 2E, F, and H). We find that only is significantly correlated with the first PCA direction. Because this is uncentered PCA, this direction closely resembles the mean activity direction. We compare with the top 5 (instead of 4, as the number of s) PCA directions to account for the potential discrepancy between this ORN activity dataset and the true ORN activity.
Next, to test Eq. 1 directly, we examine the alignment of the subspaces spanned by the four s and the top five PCA directions of (SI Appendix, Fig. S7). While more dimension is significantly aligned than is randomly expected, supporting the results of Fig. 2H, there is no complete alignment. In summary, although aligns with the top PCA direction of , and the connectivity and activity subspaces are more aligned than expected by chance, the LC does not account for the connectivity of most LN types.
Next, we study the predicted by the NNC-4 ( as the number of LN types) optimized on (Fig. 2A), for . For , three of the four s align significantly with a (BT, BD, and P0, Fig. 3 A and B). In a perfect fit between model and data, each is aligned one . is not significantly correlated with any of the s, but NNC-5 has one significantly aligned with (SI Appendix, Fig. S6H). The significant alignment of with both and could arise due to partial correlation between s (Fig. 1C). Furthermore, we find a similarity between the model and the data in terms of alignment of the ORNs LN connection weight vectors with the ORN activity vectors (SI Appendix, Fig. S8).
In summary, the ORN LN connection weights predicted by the NNC model strongly resemble the synaptic counts in , but do not provide an exact one-to-one correspondence. This analysis confirms that all the s are adapted to ORN activity patterns. It also corroborates the hypothesis that the similarity-matching principle and the optimization problem have explanatory power for the organization of the biological circuit. Later we discuss the potential reasons for the nonexact alignment between the model and the data.
Emergence of LN Groups in the NNC.
In the connectome, LNs are grouped by type and several s are similar (Figs. 1 B and C and 3C). Do LN groups naturally emerge in our models? In the LC, spans the top K-dimensional principal subspace of the input , resulting in distinct s and thus no LN group emerges.
In the NNC, however, we observe the formation of LN groups. For example, in NNC-8 (8 LNs as on each side of the larva) trained on , several s are similar, especially for smaller (Fig. 3D). Given that the s point toward the cluster locations in the ORN axon activity space, the grouping of s is influenced by 1) ORN activity pattern statistics (closer clusters elicit more aligned s), 2) the number of LNs (having more LNs than clusters lead to several similar s), and 3) the value of (higher leads to more separated clusters in ORN axons and thus dissimilar s) (SI Appendix, Figs. S9 and S10).
For the biological circuit, we lack exact measures of the factors (e.g., and ) that influence grouping. Nevertheless, we inquire whether NNC-8 can, in principle, generate a grouping similar to the biological circuit for different values of . At , the mean rectified correlation coefficient () between all s of the NNC-8 matched that of the connectome (Fig. 3E). While this value of , which corresponds to a relatively low feedback inhibition in the model, should not be interpreted as the “true” value in the actual biological circuit, it falls within the range found above ().
In summary, within a reasonable parameter range, the NNC reproduces another property of the biological circuit: the emergence of LN groups.
Relation between LN–LN and Feedforward ORNs LN Connection Weights.
The ORN-LN circuit contains reciprocal inhibitory LN–LN connections (Fig. 4A) whose connectivity patterns and roles are not fully understood. In our models, these connections are symmetric: the synaptic weights and are equal. This is largely verified in the connectome, except for the P0, which inhibits the KSs, but is not strongly inhibited by them. Theoretical predictions of the LC-K model (with K LNs) state that the strength of LN–LN connections () and ORN–LN connections () are related (SI Appendix):
[2] |
where is the matrix transpose. This relationship is exact for the LC and approximate for the NNC. The ith column of , , is the LNs LNi (and LNi LNs) synaptic weight vector. The ith column of , , is proportional to the ORNs LNi (and LNi ORNs) synaptic weight vector. From Eq. 2 follows that: 1) , i.e., the ratio between the magnitude of the ORNs LN and LNs LN synaptic weight vectors is the same at each LN. The magnitude is a proxy for the total synaptic strength of a synaptic weight vector. 2) , where is the angle between two vectors and . Thus 2 LNs with a similar (different) connectivity pattern with the ORNs have a similar (different) connectivity pattern with LNs.
We test whether Eq. 2 holds in the connectome (Fig. 4), and find a significant correlation (, ) between the off-diagonal entries of matrices and , suggesting a meticulous co-organization of the ORN–LN and LN–LN connections. We lack the values of the LN neural leaks, which correspond to the diagonal entries of (Eqs. 6 and 7).
In summary, the synaptic weight organization in the NNC model resembles that the connectome in several key ways: the synaptic counts , the emergence of LN groups, and the relationship between ORNs LN and LN–LN. The LC model, on the other hand, fails at explaining several of these structural features.
Circuit Model Computation and Coding Efficiency.
We next explore the computations of the LC and NNC. In both models, upon ORN soma activation, the computation is implemented dynamically through the ORN–LN loop and converges exponentially to a steady state (Eqs. 6 and 7). Given inputs , the circuit’s outputs are the converged representations in ORN axons, , and LNs, .
Efficient encoding of odor representations in ORN is crucial for downstream processing. Odor representations can be visualized as points in a neural space, where each axis is the activity of an ORN. We consider a circuit with just ORNs and LNs, and an artificial input dataset of two odors A and B (Fig. 5 A and D). Given and the representations of the two odors: the larger the angle , the easier the two odors can be discriminated, and the more efficiently the space is utilized. We quantify the efficiency of the encoding by the coefficient of variation of the PCA variances, , of the representation: . If all the variances are equal (), the representation is white, and the encoding space is efficiently used (38). A larger indicates a less optimal space utilization. We study the PCA variances and “whiteness” of uncentered data because we assume downstream neurons experience uncentered activity. We further describe the computation in terms of the modification of the stimulus representations.
LC: Extraction of the Principal Subspace by LNs and Partial Equalization of PCA Variances in ORN Axons.
We first describe the computation in the LC. Given activity patterns in the D ORN somas, we call and () the PCA directions and variances of the uncentered (Fig. 5D). The activity of the K LNs, , encodes the top K PCA subspace of , i.e., spanned by (Fig. 5B). How exactly LNs encode the subspace is a degree of freedom of the optimization, and thus the activity of individual LNs does not necessarily align with the PCA directions of the input. When , LNs perform a dimensionality reduction of the ORN soma activity.
LNs inhibit ORN axons, altering their odor representation (Fig. 5D). However, the PCA directions of ORN axon activity remain the same as in ORN somas, i.e., . Thus, this transformation from soma to axons only stretches and does not rotate the cloud of representations in the neural space. This absence of rotation (called “zero-phase”) makes the axonal and somatic activity maximally similar (23). This is advantageous for downstream processing because the evolving representation in ORN axons, computed dynamically via LN activation, is thus maximally close to the converged representation, allowing meaningful downstream processing before the complete representation convergence.
The PCA variances and of and are (Fig. 5 D and F):
Hence, the variances of the last D–K PCA directions in ORN somas () remain unaltered in ORN axons (). The variances of top K PCA directions in ORN somas are diminished according to Eq. 3a (Fig. 5 G and H): relatively large PCA variances in ORN somas () are shrunken with a cubic root in ORN axons (), relatively small PCA variances () remain virtually unchanged (). The PCA variances in LN activity () are proportional to those in ORN axon activity () (Eq. 3c). (Note the indices i of the PCA directions and variances in ORN axons have been set to match those in ORN somas, and do not follow the usual decreasing order).
This transformation generally results in a smaller coefficient of variation of PCA variances, , in the output than in the input (SI Appendix, see below, Fig. 6D). The PCA variances are then less spread and the odor representations are encoded more efficiently. Because the PCA variances are partially equated and no rotation occurs, this transformation is a partial (Zero-phase) ZCA-whitening.
NNC: Clustering by LNs and Partial Equalization of PCA Variances in ORN Axons.
We next explore the computation of the NNC, where LN () and ORN axon () activities are nonnegative. LNs implement symmetric nonnegative matrix factorization (SNMF) on ORN axon activity, which consists of clustering and feature discovery (SI Appendix) (37). SNMF is essentially “soft” K-means clustering, allowing inputs to belong to multiple clusters. Clustering satisfies the optimization’s objective of nonnegative LN activity and maximally conserved distances between stimulus representations in ORN axons and LNs. Thus LN activity, , encodes the cluster membership of odor representations in ORN axons (), and the ORN LN synaptic weight vectors, , point toward clusters (Fig. 5 C and E). Unlike the LC, there is no degree of freedom in LN activity.
The activity in ORN axons in NNC resembles that in LC, only without negative values, and the PCA variances are also similar (Figs. 5D–F).
Circuit Model Computation on the ORN Activity Dataset.
Next, to better comprehend the potential computation of the ORN-LN circuit, we study the computation of the NNC on the dataset of odor representation in ORNs, (Fig. 2A). We also show the LC. We set the parameter that regulates the inhibition strength to clearly represent the effect of the odor representation transformation in ORNs. K, the number of LNs, is set to 1, 4 (as the number of LN types) or 8 (as the number of LNs on one side of the larva). We also examine the computation of a nonnegative circuit model (NNC-conn) with connectivity weights proportional to the synaptic counts of the connectome (SI Appendix). Because for NNC-conn multiple unknown model parameters need to be guessed, and this circuit might not be adapted to the specific statistics of , its computation might not accurately reflect that of the true circuit, and the discrepancies with the normative models might be a consequence of this. Nevertheless, we find many similarities between NNC-conn and NNC-8, further supporting our predictions regarding circuit computation. Fig. 6 exhibits the main results, SI Appendix, Figs. S13, S14, and S15 display additional analysis of the LC, NNC, and NNC-conn, respectively.
As above, LNs in the LC encode the top K-dimensional PCA subspace of ORN soma activity (SI Appendix, Fig. S11B). LNs in the NNC softly cluster odors, as observed by their sparser activity and their correspondence with ORN activity patterns (Fig. 6A). LN activity in NNC-conn is also rather sparse.
In all models, ORN axon activity () is weaker than in somas (Fig. 6B). While it is also sparser and nonnegative in the NNC models, in the LC, it contains negative values, which may not be biologically plausible.
Next, we compare the PCA variances of the odor representations in ORN somas () and axons () (Fig. 6C). In the NNC models, variances decrease for all PCA directions. In the LC, however, only the variances of the top K PCA directions decrease. This difference results from the nonnegativity constraint in the NNC models, which affects all stimulus directions. The spread of PCA variances decreases in all models (smaller , Fig. 6D) indicating a whiter representation in the ORN axons. This effect is the weakest in the NNC-conn. Changing the number of LNs impacts the NNC less than the LC. In the LC, only the order of the PCA directions of and changes, because K of them are shrunken (SI Appendix, Fig. S12 A and B). For the NNC, the PCA directions are slightly altered, but their order mostly remains (SI Appendix, Fig. S12 C and D). In the NNC-conn, the PCA directions are modified more strongly (SI Appendix, Fig. S12E).
Considering the decreased spread of PCA variances, we inquire whether activity becomes more evenly distributed among ORNs, an important property of efficient coding. Both the LC and NNC decrease the (uncentered) activity variance of “high-variance ORNs” and leave “low-variance ORNs” virtually unaffected, reducing the CV of ORN variance (Fig. 6 E and F). The NNC-conn, however, exhibits an increase in CV due to several “high-variance ORNs” being not strongly dampened.
Subsequently, we investigate changes in the magnitude of ORN soma and axon activity patterns. The magnitude is the length of an activity pattern vector in the dimensional ORN space and is a proxy for the total activity of all ORNs in response to an odor. Similarly to ORN variances, the magnitude of large-magnitude patterns decreases, whereas small-magnitude patterns remain unchanged, decreasing the spread of pattern magnitudes (Fig. 6 G and H). These effects resemble a divisive normalization-type computation, also reported in Drosophila (13, 25).
In line with the less dispersed PCA variances in ORN axons, in all models ORNs and odor representations are more decorrelated in the axons than in the somas (Fig. 6I–L), consistent with partial whitening.
Additionally, we investigate the effect of adjusting the model parameter , which regulates feedback inhibition strength. A higher (, SI Appendix, Fig. S16) leads to decreased activity in ORN axons and smaller PCA variances, reduced spread of PCA variances, channels and patterns norms, stronger decorrelation of ORNs and patterns. When inhibition is eliminated (), the axonal and somatic ORN activity become identical. Although it is unknown if inhibition is modulated in the real circuit, altering this parameter allows us to understand this circuit’s potential.
In summary, NNC analysis predicts that the ORN-LN circuit clusters odors with LNs and performs partial ZCA-whitening and normalization of odor representations in ORN axons. This results in a more efficiently encoded output with more decorrelated and equalized ORNs and odor representations, ultimately enhancing odor discrimination downstream.
Computation without LN–LN Connections.
Lastly, we investigate the role of LN–LN connections by considering two alternative circuit models. First, we consider an LC or NNC circuit adapted to an input ensemble (i.e., Fig. 6) and remove the LN–LN connections, which corresponds to setting the off-diagonal elements in to 0 (SI Appendix, Fig. S17). This manipulation leads to less sparse LN activity in the NNC, altered PCA directions in the axonal activity relatively to the soma, increased inhibition, and more dissimilar odor representations in ORN axons compared to somas. Thus, in an already “adapted” circuit, LN–LN connections improve clustering in LNs for the NNC, regulate inhibition, and maintain similar representation in ORN axons and somas.
Second, we consider the slightly different optimization problem that leads to an ORN-LN circuit without LN–LN connections (SI Appendix) (39). In the linear case, the whitening is complete (i.e., the first PCA variances that are larger than become equal) and the K LNs still encode the top K dimensional subspace of the input. However, with nonnegativity constraints on ORN axon and LN activity, all LNs display the same activity, lacking differentiation (SI Appendix, Fig. S18). Thus, in this case, LN–LN connections are imperative for clustering.
Discussion
Combining the Drosophila larva olfactory circuit connectome, ORN activity data, and a normative model, we advance the understanding of sensory computation and adaptation, quantitatively link ORN activity statistics, functional data, and connectome, and make testable predictions. We reveal a canonical circuit model capable of autonomously adapting to different environments, while maintaining the critical computations of partial whitening, normalization, and feature extraction. Such a circuit architecture may arise in other brain areas and may be applicable in machine learning and signal processing. Using ORN activity patterns as input, our normative framework accounts for the biological circuit structural organization and identifies in the connectome signatures of circuit function and adaptation to ORN activity. Such an approach offers a general framework to understand circuit computation (40, 41) and could provide valuable insights into more neural circuits, whose structural and activity data become available (1, 2).
Model and Biological Circuit: Similarities and Differences.
In this paper, we compare the structural predictions of our normative approach to the connectome. The NNC model, when adapted to the ORN activity dataset (5), accounts for key structural characteristics (Figs. 3 and 4), for example, the ORNs LN connection weight vectors. We ask two questions: 1) Why does the strong resemblance between model and data arise, when the available odor dataset most probably imperfectly matches the true larva odor environment? 2) Why isn’t the resemblance even greater, and could the imperfect fit suggest that the model inadequately explains the biological circuit?
For 1), a possibility is that generic correlations between ORNs arise in large enough ORN activity datasets, causing robust features in the model connectivity. These correlations could result from the intrinsic chemical properties of ORN receptors. Odor statistics would also influence the connection weights, but to a lesser degree. Thus, a more naturalist activity dataset could further improve model predictions.
For 2), first, due to intrinsic noise and variability, no model could be 100% accurate in predicting connectivity. In fact, variability in synaptic count and innervation arises for Drosophilas raised in similar environments (27, 42), indicating potential “imprecision” of development and/or learning. We also observe variability in the left vs. right side connectivity (Fig. 1B). Second, incomplete ORN activity statistics may decrease prediction accuracy. Third, synaptic count might not exactly reflect synaptic strength (11). Finally, our model being a simplification of reality misses additional factors shaping circuit connectivity.
Our analysis indicates that the matches between model and data likely do not result from chance only, suggesting that the similarity-matching principle influences circuit organization. However, our unsupervised approach assumes that no odor is “special” for the animal, and thus LNs in the circuit model cluster odors solely based on their representations in the ORN activity space. This contrasts with the biological ORN-LN circuit, where LNs such as Keystone and Picky 0 have specific downstream connections likely related to survival needs and different hardwired animal behaviors (4, 43), requiring them to detect particular odors. Consequently, the connectivity of such LNs might contribute to the imperfect one-to-one correspondence between the model and the connectome (e.g., KS in NNC-4, Fig. 3A).
The circuit model can learn the optimal connection weights autonomously via Hebbian learning, offering the capacity to adapt to different environments. Studies in adult Drosophila reveal that glomeruli sizes (and thus ORN–LN or ORN–PN synaptic weights) or activity depend on the environment in which the Drosophila grew up (16–19). It is, however, unknown if activity-dependent plasticity also occurs in the larval ORN-LN circuit and whether the observed synaptic counts are a result of such plasticity. If present, it is unclear whether the short 6-h life of the larva from which the connectome was reconstructed allows substantial learning to occur and whether changes in synaptic weights would translate to different synaptic counts (11).
Resolving connectomes of larvae raised in different odor environments and at different times of their life, probing synaptic plasticity, and recording ORN responses to the full odor ensemble present in its environment would help clarify the influence of noise, plasticity, and genetics in circuit shaping.
Roles of LNs.
LNs form a significant part of the neural populations in the brain, perform diverse computational functions, and exhibit extremely varied morphologies and excitabilities (27, 44). We propose a dual role for LNs in this olfactory circuit: altering the odor representation in ORNs and extracting ORN activity features, available for downstream use (4). In the olfactory system of Drosophila and zebrafish, LNs perform multiple computations, such as gain control, normalization of odor representations, and pattern and channel decorrelation (12–15, 32, 45), which is consistent with our results. Also, in Drosophila the LN population expands the temporal bandwidth of synaptic transmission and temporally tunes PN responses (28, 29, 46), which was not addressed here.
In topographically organized circuits, such as in the visual periphery or in the auditory cortex, distinct LN types uniformly tile the topographic space, and each LN type extracts a specific feature of the input, e.g., in the retina (47). In nontopographically organized networks, however, the organization and role of LNs remains a matter of research and controversy (27, 48). We study a subcircuit with four LN types, and most types contain several similarly connected LNs (Fig. 1). What is the function of multiple similar LNs in the ORN-LN circuit, as also observed in the NNC (Fig. 3C–E)? First, LNs might differentiate further as the larva grows. Second, several LNs might help expand the dynamic range of a single LN. What are the features extracted by LNs in the Drosophila larva? Our NNC model and the distinct connectivity patterns of LN types in the connectome (4), suggest that different LN types are activated in response to different sets of odors. The extracted features might relate to clusters in ORN activity and to prewired, animal-relevant odors. Since several ORNs LN connection weight vectors in the NNC model resemble those in the biological circuit, the odor clusters identified by the model likely correspond to the set of odors that activate LNs in the biological circuit. The feedforward synaptic count vector from ORNs to the Broad Trio , which aligns with the first PCA direction of ORN activity and with an ORNs LN connection weight vector in the NNC model (Figs. 2H, 3 A and B) could potentially encode the mean ORN activity and thus be related to the global odor concentration (26). Other LNs might encode features of odors, such as aromatic vs. long-chain alcohols (5), or specific information influencing larva behavior (4, 43), but more experiments are required to definitely resolve the features. While our conclusions differ from a study that found that LN activation is invariant to odor identity (48), that study imaged several LNs simultaneously and might thus have missed the selectivity of individual LNs.
The connectome reveals LN–LN connections, which we propose play a key role in clustering and shaping the odor representation, and are co-organized with the ORN–LN connections (Fig. 4). To the best of our knowledge, the role of LN–LN connections and their relationship to ORN–LN connections is relatively unexplored.
In summary, our study emphasizes the importance of the different ORN–LN and LN–LN connection strengths and argues that LNs are minutely selective and organized to extract features and render the representation of odors more efficient.
Circuit Computation, Partial ZCA-Whitening, and Divisive Normalization.
We propose that the circuit’s effect on the neural representation of odors in ORNs corresponds to partial ZCA-whitening and divisive normalization (Figs. 5 and 6). Such computations, which reduce correlations originating from the sensory system and the environment, have appeared in efficient coding and redundancy reduction theories (22, 25, 36, 38, 49, 50). Partial whitening is in fact a solution to mutual information maximization in the presence of input noise (38). In this circuit too, complete whitening might also not be desirable due to potential noise amplification. Thus, keeping low-variance signal directions of the input unchanged and dampening larger ones is consistent with mutual information maximization. Our conclusions are in line with reports of pattern decorrelation and/or whitening in the olfactory system in zebrafish (14, 15, 32, 33) and mice (34, 35).
The computation in our model also resembles divisive normalization, an ubiquitous computation in the brain (25), proposed for the analogous circuit in the adult Drosophila (12, 13). In its simplest form, divisive normalization is defined as , where is the response of neuron j, is the driving input of neuron i, is the maximum response of the output neuron and and n determine the offset and slope of the neuronal sigmoidal response curve, respectively (25). Divisive normalization captures two effects of neuronal and circuit computation: 1) neural response saturation with increasing input up to a maximum spiking rate , arising from the neuron’s biophysical properties; 2) dampening of the response of a given neuron when other neurons also receive input, often due to lateral inhibition (but see ref. 51). Aspect (1) is absent in our model but could be implemented with a saturating nonlinearity. Depending on the biological value of the maximum output, our model might not accurately capture responses for high-magnitude inputs. However, signatures of (2) are evident in the saturation of the activity pattern magnitudes in ORN axons for increasing ORN soma activity pattern magnitudes (Fig. 6G). Activity patterns of large magnitude correspond to activity at higher odor concentrations and with a high number of active ORNs. Because such input directions are more statistically significant in our dataset, these stimuli are more strongly dampened by LNs (which encode such directions) than those with few ORNs active. Thus, our model presents a possible linear implementation of a crucial aspect of divisive normalization, which in itself is a nonlinear operation.
Although the basic form of divisive normalization performs channel decorrelation, and not activity pattern decorrelation (13, 14, 32), our models perform both channel and pattern decorrelation. Nevertheless, a modified version of divisive normalization, which includes different coefficients for the driving inputs in the denominator (52), performs pattern decorrelation too, as our circuit model. The proposed neural implementations of divisive normalization usually require multiplication by the feedback (52, 53), which might not be as biologically realistic as our circuit implementation.
Several neural architectures similar to ours have been proposed to learn to decorrelate channels, perform normalization, or learn sparse representations in an unsupervised manner (21, 37, 52, 54–59). However, these studies either lack a normative/optimization approach or have a different circuit architecture or synaptic learning rules. Using a normative approach has the advantage of directly investigating the underlying principles of neural functioning and also potentially providing a mathematically tractable understanding of the circuit structure and function.
Our study complements machine learning approaches to understand neural circuit organization (60, 61). These approaches use supervised learning and backpropagation to train an artificial neural network to perform tasks such as odor or visual classification. In the olfactory system, circuit configurations arising from this optimization, which could mimic the evolutionary process, display many connectivity features found in biology (61). Unlike these approaches, we propose a general principle governing the transformation of neural representations, similarity-matching, and also a mechanism to learn autonomously during the animal’s lifetime.
Materials and Methods
Optimization Problems Describing the ORN-LN Circuit.
We use a normative approach to study the ORN-LN circuit. We formulate two optimization problems that can be solved by a circuit model with the ORN-LN architecture. Studying the circuit model computation is then equivalent to studying the solution of an optimization problem. We derive analytical expressions describing different aspects of the computation and the circuit synaptic organization (SI Appendix).
We define the following variables: an input matrix of T samples, and outputs , . and are D-dimensional vectors, while are K-dimensional. , , and represent the activity patterns of D ORN somas (i.e., the inputs), D ORN axons and K LNs, respectively. We call an optimal value (solution) of a variable b. In the results section, we drop the . We postulate the following similarity-matching-inspired optimization problem (e.g., ref. 20), which seeks the optimal output activities and given an input :
[4] |
where is the square of the matrix Frobenius (Euclidean) norm. The term drives the activity of the ORN axons toward the activity of ORN somas and ensures that when there is no activity in the LNs. The terms and align the similarities between the activities of ORN axons and LNs and puts a 4th order penalty on the norm of ; they correspond to the bidirectional all-to-all connectivity between ORN axons and LNs, as well as between LNs, but no direct connectivity between ORN axons; such similarity-matching terms permit a significant change of neural representation and a change of dimensionality, which takes place between ORN axons and LNs. is a parameter related to the strength of the dampening in and affects both the optima and .
We consider this optimization in two search domains for and . One without any constraints on and , representing the linear circuit (LC) model, and one with nonnegativity constraints (, ), representing the nonnegative circuit (NNC) model. Nonnegativity constraints account for the fact that neural activity is usually nonnegative, or at least not symmetric in the negative and positive directions. The optimal and can be found analytically for the LC, and through numerical simulations for the NNC. Note that one cannot always guarantee converging to a global optimum for the NNC (62).
We prove that a neural circuit with ORN-LN architecture can solve this optimization problem (SI Appendix, Online algorithm). In brief, we introduce into the optimization problem two auxiliary matrices and , which naturally map onto ORNs–LNs and LNs–LNs synaptic weights, respectively. By construction, is symmetric, i.e., = . The new objective function is then optimized over the variables , , , and . Writing the gradient descent/ascent over and provides the neural dynamics equations, with and related to the synaptic weights (Eqs. 6 and 7). The optimal and :
[5] |
can be found “offline” by obtaining the optimal and in Eq. 4, or in the “online setting,” through unsupervised, Hebbian learning, where and are updated after each stimulus presentation (Eq. 8, see below).
Circuit Neural Dynamics.
A solution to the optimization problem Eq. 4 without the nonnegativity constraints can be implemented by the following differential equations describing the LC, whose steady-state solutions correspond to the optima for and for given and (SI Appendix, Online algorithm). These equations naturally map onto the ORN-LN neural circuit dynamics (dropping the sample index for simplicity of notation):
[6] |
where , , and are D, D, and K-dimensional vectors, and represent the activity (e.g., spiking rate) of the ORN somas, ORN axons, and LNs, respectively. and are neural time constants, is the local time evolution (not to be confused with the (t) sample index). The elements of the matrices and contain the synaptic weights of the feedforward ORNs LN and feedback LN ORNs connections, respectively. Thus, the feedforward connection vectors are proportional to the feedback vectors, and the parameter sets the ratio. The assumption of proportionality is reasonable considering the connectivity data (SI Appendix, Fig. S2 A, B, and D). Off-diagonal elements of the matrix contain the weights of the LN - LN inhibitory connections, whereas the diagonal entries encode the LN leaks. In the absence of LN activity and at steady state, the equations satisfy , i.e., somatic and axonal activities of ORNs are identical. In the absence of input () both and decay exponentially to , because of the terms and , respectively. In summary, these equations effectively model the ORN-LN circuit dynamics by implementing that 1) the ORN axonal activity is driven by the input in ORN somas and inhibited by the feedback from the LNs through the term and 2) LN activity is driven by the activity in ORN axonal terminals by and inhibited by LNs through the term . Note that changing in the objective function leads to different optimal and .
When optimized online, the optimization problem Eq. 4 with the nonnegativity constraints gives rise to the following equations describing the NNC:
[7] |
where is the step size parameter and is a component-wise rectification. Here, is a discrete-time variable. These equations are analog to Eq. 6, but also satisfying constraints on the activity: . Such constraints are implemented by formulating circuit dynamics in discrete time and using a projected gradient descent.
We call LC-K the linear circuit model implemented by Eq. 6 and NNC-K the nonnegative circuit model implemented by Eq. 7, with K LNs.
Note that there is a manifold of implementations of the same computation by a circuit model. First, one can introduce a parameter (SI Appendix), that scales the feedforward and feedback connections as well as the magnitude of LN activity, in such a way that the ORN axon activity remains the same. Second, multiplying the whole equation in Eq. 6 or Eq. 7 would not alter the converged output, but would scale the circuit time constants and synaptic weights.
Synaptic Plasticity.
The circuit model is capable of reaching the optimal synaptic weights and , which solve the optimization problem Eq. 4, in an unsupervised manner, with Hebbian plasticity. In practice, as the circuit receives a stimulus (ORN soma activation), it performs a computation that yields a steady state output activity in ORN axons and LNs (with Eq. 6 or Eq. 7); the synaptic weights are then updated using Hebbian rules:
[8] |
where are learning rates. These equations arise when optimizing Eq. 4 online. We assume that the ORN soma activation is present long enough so that and reach steady state values. During this iterative process of synaptic updating, where the circuit model “learns”/“adapts” to the stimulus ensemble , the synaptic weights converge toward “optimum” steady state Eq. 5 (which might require multiple learning epochs over the ). Note that the neural leaks of LNs (diagonal values of ) are set (Eq. 5) and updated (Eq. 8) similarly to the synaptic weights ( and off-diagonal of ).
Supplementary Material
Acknowledgments
We thank Aravinthan D.T. Samuel, Jacob Baron, Guangwei Si, Thomas Frank, Victor Minden, Anirvan Sengupta, Eftychios A. Pnevmatikakis, Siavash Golkar, David Lipshutz, and Shiva GhaaniFarashahi for discussions and/or comments on the manuscript. C.P. was supported by an NSF Award DMS-2134157 and the Intel Corporation through the Intel Neuromorphic Research Community.
Author contributions
N.M.C., C.P., and D.B.C. designed research; C.P. and D.B.C. formulated the optimization problem; N.M.C. and C.P. performed theoretical derivations; N.M.C. wrote the code, analyzed the data, and performed numerical simulations; and N.M.C. wrote the paper with input from C.P. and D.B.C.
Competing interests
The authors declare no competing interest.
Footnotes
This article is a PNAS Direct Submission.
Data, Materials, and Software Availability
The connectome and activity datasets are available in refs. (4) and (5). Code for generating the analysis and all the figures is available in GitHub (https://github.com/chapochn/ORN-LN_circuit) (63).
Supporting Information
References
- 1.Eichler K., et al. , The complete connectome of a learning and memory centre in an insect brain. Nature 548, 175–182 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Scheffer L. K., et al. , A connectome and analysis of the adult Drosophila central brain. eLife 9, e57443 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Aimon S., et al. , Fast near-whole-brain imaging in adult Drosophila during responses to stimuli and behavior. PLOS Biol. 17, e2006732 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Berck M. E., et al. , The wiring diagram of a glomerular olfactory system. eLife 5, e14859 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Si G., et al. , Structured odorant response patterns across a complete olfactory receptor neuron population. Neuron 101, 950–962.e7 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Wilson R. I., Early olfactory processing in drosophila: Mechanisms and principles. Annu. Rev. Neurosci. 36, 217–241 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.C. L. Barnes, D. Bonnéry, A. Cardona, “Synaptic counts approximate synaptic contact area in Drosophila” (Tech. Rep., bioRxiv, December 2020). [DOI] [PMC free article] [PubMed]
- 8.Takemura S., et al. , A visual motion detection circuit suggested by Drosophila connectomics. Nature 500, 175–181 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Holderith N., et al. , Release probability of hippocampal glutamatergic terminals scales with the size of the active zone. Nat. Neurosci. 15, 988–997 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Akbergenova Y., Cunningham K. L., Zhang Y. V., Weiss S., Littleton J. T., Characterization of developmental and molecular factors underlying release heterogeneity at Drosophila synapses. eLife 7, e38268 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Bailey C. H., Kandel E. R., Harris K. M., Structural components of synaptic plasticity and memory consolidation. Cold Spring Harbor Perspect. Biol. 7, a021758 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Olsen S. R., Wilson R. I., Lateral presynaptic inhibition mediates gain control in an olfactory circuit. Nature 452, 956–960 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Olsen S. R., Bhandawat V., Wilson R. I., Divisive normalization in olfactory population codes. Neuron 66, 287–299 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Wanner A. A., Friedrich R. W., Whitening of odor representations by the wiring diagram of the olfactory bulb. Nat. Neurosci. 23, 433–442 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Friedrich R. W., Neuronal computations in the olfactory system of zebrafish. Annu. Rev. Neurosci. 36, 383–402 (2013). [DOI] [PubMed] [Google Scholar]
- 16.Devaud J.-M., Acebes A., Ferrús A., Odor exposure causes central adaptation and morphological changes in selected olfactory glomeruli in Drosophila. J. Neurosci. 21, 6274–6282 (2001). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Sudhakaran I. P., et al. , Plasticity of recurrent inhibition in the Drosophila antennal lobe. J. Neurosci. 32, 7225–7231 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Sachse S., et al. , Activity-dependent plasticity in an olfactory circuit. Neuron 56, 838–850 (2007). [DOI] [PubMed] [Google Scholar]
- 19.Das S., et al. , Plasticity of local GABAergic interneurons drives olfactory habituation. Proc. Natl. Acad. Sci. U.S.A. 108, E646–E654 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Pehlevan C., Sengupta A., Chklovskii D. B., Why do similarity matching objectives lead to Hebbian/anti-Hebbian networks? Neural Comput. 30, 84–124 (2018). [DOI] [PubMed] [Google Scholar]
- 21.C. Pehlevan, D. B. Chklovskii, “Optimization theory of Hebbian/anti-Hebbian networks for PCA and whitening” in 2015 53rd Annual Allerton Conference on Communication, Control, and Computing, Allerton 2015 (2016), pp. 1458–1465.
- 22.H. B. Barlow, “Possible principles underlying the transformations of sensory messages” in Sensory Communication, W. A. Rosenblith, Ed. (The MIT Press, 1961), pp. 217–234.
- 23.Kessy A., Lewin A., Strimmer K., Optimal whitening and decorrelation. Am. Stat. 72, 309–314 (2018). [Google Scholar]
- 24.Bell A. J., Sejnowski T. J., The “independent components’’ of natural scenes are edge filters. Vis. Res. 37, 3327–3338 (1997). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Carandini M., Heeger D. J., Normalization as a canonical neural computation. Nat. Rev. Neurosci. 13, 51–62 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Asahina K., Louis M., Piccinotti S., Vosshall L. B., A circuit supporting concentration-invariant odor perception in Drosophila. J. Biol. 8, 9 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Chou Y.-H., et al. , Diversity and wiring variability of olfactory local interneurons in the Drosophila antennal lobe. Nat. Neurosci. 13, 439–449 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Kim A. J., Lazar A. A., Slutskiy Y. B., Projection neurons in Drosophila antennal lobes signal the acceleration of odor concentrations. eLife 4, e06651 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Nagel K. I., Hong E. J., Wilson R. I., Synaptic and circuit mechanisms promoting broadband transmission of olfactory stimulus dynamics. Nat. Neurosci. 18, 56–65 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Laurent G., Olfactory network dynamics and the coding of multidimensional signals. Nat. Rev. Neurosci. 3, 884–895 (2002). [DOI] [PubMed] [Google Scholar]
- 31.Benjamini Y., Hochberg Y., Controlling the false discovery rate: A practical and powerful approach to multiple testing. J. R. Stat. Soc., B: Stat. Methodol. 57, 289–300 (1995). [Google Scholar]
- 32.Friedrich R. W., Wiechert M. T., Neuronal circuits and computations: Pattern decorrelation in the olfactory bulb. FEBS Lett. 588, 2504–2513 (2014). [DOI] [PubMed] [Google Scholar]
- 33.Friedrich R. W., Laurent G., Dynamic optimization of odor representations by slow temporal patterning of mitral cell activity. Science 291, 889–894 (2001). [DOI] [PubMed] [Google Scholar]
- 34.Gschwend O., et al. , Neuronal pattern separation in the olfactory bulb improves odor discrimination learning. Nat. Neurosci. 18, 1474–1482 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Giridhar S., Doiron B., Urban N. N., Timescale-dependent shaping of correlation by olfactory bulb lateral inhibition. Proc. Natl. Acad. Sci. U.S.A. 108, 5843–5848 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Simoncelli E. P., Olshausen B. A., Natural image statistics and neural representation. Annu. Rev. Neurosci. 24, 1193–1216 (2001). [DOI] [PubMed] [Google Scholar]
- 37.C. Pehlevan, D. B. Chklovskii, “A Hebbian/Anti-Hebbian network derived from online non-negative matrix factorization can cluster and discover sparse features” in Conference Record - Asilomar Conference on Signals, Systems and Computers (2015), pp. 769–775.
- 38.Atick J. J., Redlich A. N., What does the retina know about natural scenes? Neural Comput. 4, 196–210 (1992). [Google Scholar]
- 39.Lipshutz D., Pehlevan C., Chklovskii D.B, Interneurons accelerate learning dynamics in recurrent neural networks for statistical adaptation. arXiv [Preprint] (2022). http://arxiv.org/abs/2209.10634 (Accessed 3 October 2022).
- 40.Bahroun Y., Chklovskii D., Sengupta A., A similarity-preserving network trained on transformed images recapitulates salient features of the fly motion detection circuit. Adv. Neural Inf. Process. Syst. 32 (2019). [Google Scholar]
- 41.Golkar S., Lipshutz D., Bahroun Y., Sengupta A., Chklovskii D., A simple normative network approximates local non-Hebbian learning in the cortex. Adv. Neural Inf. Process. Syst. 33, 7283–7295 (2020). [Google Scholar]
- 42.Tobin W. F., Wilson R. I., Lee W.-C. A., Wiring variations that enable and constrain neural computation in a sensory microcircuit. eLife 6, e24838 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Vogt K., et al. , Internal state configures olfactory behavior and early sensory processing in Drosophila larvae. Sci. Adv. 7, eabd6900 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Hattori R., Kuchibhotla K. V., Froemke R. C., Komiyama T., Functions and dysfunctions of neocortical inhibitory neuron subtypes. Nat. Neurosci. 20, 1199–1208 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Zhu P., Frank T., Friedrich R. W., Equalization of odor representations by a network of electrically coupled inhibitory interneurons. Nat. Neurosci. 16, 1678–1686 (2013). [DOI] [PubMed] [Google Scholar]
- 46.Nagel K. I., Wilson R. I., Mechanisms underlying population response dynamics in inhibitory interneurons of the Drosophila antennal lobe. J. Neurosci. 36, 4325–4338 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Masland R. H., The neuronal organization of the retina. Neuron 76, 266–280 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Hong E. J., Wilson R. I., Simultaneous encoding of odors by channels with diverse sensitivity to inhibition. Neuron 85, 573–589 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.M. D. Plumbley, “A Hebbian/anti-Hebbian network which optimizes information capacity by orthonormalizing the principal subspace” in Proceedings of IEE Conference on Artificial Neural Networks (1993), pp. 86–90.
- 50.Linsker R., Self-organization in a perceptual network. Computer 21, 105–117 (1988). [Google Scholar]
- 51.Sato T. K., Haider B., Häusser M., Carandini M., An excitatory basis for divisive normalization in visual cortex. Nat. Neurosci. 19, 568–570 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Westrick Z. M., Heeger D. J., Landy M. S., Pattern adaptation and normalization reweighting. J. Neurosci. 36, 9805–9816 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Heeger D. J., Normalization of cell responses in cat striate cortex. Vis. Neurosci. 9, 181–197 (1992). [DOI] [PubMed] [Google Scholar]
- 54.King P. D., Zylberberg J., DeWeese M. R., Inhibitory interneurons decorrelate excitatory cells to drive sparse code formation in a spiking model of V1. J. Neurosci. 33, 5475–5485 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Zhu M., Rozell C. J., Modeling inhibitory interneurons in efficient sensory coding models. PLoS Comput. Biol. 11, e1004353 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Olshausen B. A., Field D. J., Sparse coding with an overcomplete basis set: A strategy employed by V1? Vis. Res. 37, 3311–3325 (1997). [DOI] [PubMed] [Google Scholar]
- 57.Koulakov A. A., Rinberg D., Sparse incomplete representations: A potential role of olfactory granule cells. Neuron 72, 124–136 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Wick S. D., Wiechert M. T., Friedrich R. W., Riecke H., Pattern orthogonalization via channel decorrelation by adaptive networks. J. Comput. Neurosci. 28, 29–45 (2010). [DOI] [PubMed] [Google Scholar]
- 59.Atick J. J., Redlich A. N., Convergent algorithm for sensory receptive field development. Neural Comput. 5, 45–60 (1993). [Google Scholar]
- 60.Yamins D. L. K., DiCarlo J. J., Using goal-driven deep learning models to understand sensory cortex. Nat. Neurosci. 19, 356–365 (2016). [DOI] [PubMed] [Google Scholar]
- 61.Wang P. Y., Sun Y., Axel R., Abbott L. F., Yang G. R., Evolving the olfactory system with machine learning. Neuron 109, 3879–3892.e5 (2021). [DOI] [PubMed] [Google Scholar]
- 62.Sengupta A. M., Tepper M., Pehlevan C., Genkin A., Chklovskii D. B., “Manifold-tiling localized receptive fields are optimal in similarity-preserving neural networks” in Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS 2018 (Curran Associates Inc., Red Hook, NY, 2018), pp. 7080–7090.
- 63.Chapochnikov N. M., Pehlevan C., Chklovskii D. B., Python code for paper “Normative and mechanistic model of an adaptive circuit for efficient encoding and feature extraction”. GitHub. https://github.com/chapochn/ORN-LN_circuit/. Deposited 28 June 2023. [DOI] [PMC free article] [PubMed]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The connectome and activity datasets are available in refs. (4) and (5). Code for generating the analysis and all the figures is available in GitHub (https://github.com/chapochn/ORN-LN_circuit) (63).