Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2016 Apr 11;12(4):e1004850. doi: 10.1371/journal.pcbi.1004850

A Robust Feedforward Model of the Olfactory System

Yilun Zhang 1,2, Tatyana O Sharpee 1,2,*
Editor: Lyle J Graham3
PMCID: PMC4827830  PMID: 27065441

Abstract

Most natural odors have sparse molecular composition. This makes the principles of compressed sensing potentially relevant to the structure of the olfactory code. Yet, the largely feedforward organization of the olfactory system precludes reconstruction using standard compressed sensing algorithms. To resolve this problem, recent theoretical work has shown that signal reconstruction could take place as a result of a low dimensional dynamical system converging to one of its attractor states. However, the dynamical aspects of optimization slowed down odor recognition and were also found to be susceptible to noise. Here we describe a feedforward model of the olfactory system that achieves both strong compression and fast reconstruction that is also robust to noise. A key feature of the proposed model is a specific relationship between how odors are represented at the glomeruli stage, which corresponds to a compression, and the connections from glomeruli to third-order neurons (neurons in the olfactory cortex of vertebrates or Kenyon cells in the mushroom body of insects), which in the model corresponds to reconstruction. We show that should this specific relationship hold true, the reconstruction will be both fast and robust to noise, and in particular to the false activation of glomeruli. The predicted connectivity rate from glomeruli to third-order neurons can be tested experimentally.

Author Summary

Many olfactory systems are capable of accurately sensing a minimum of thousands of different odorants using as few as hundreds of different receptors. This compression raises the possibility that the mathematical properties of compressed sensing might be relevant to olfaction, similar to how these properties were found relevant to other sensory systems. In olfaction, previous applications of compressed sensing algorithms relied on the dynamics of neural circuits to reconstruct high dimensional signals. Such approaches are relatively temporally inefficient and sensitive to noise. To overcome these problems, we propose a purely feedforward compressed sensing model of the olfactory system where high dimensional signals can be recovered with a single feedforward layer of neural processing. The reconstructions are shown to be robust to noise, account for a number of experimental observations, and because of the feedforward structure are temporally efficient. Using the model, we make predictions that can be tested in future experiments with respect to optimal connectivity within the olfactory system. Our results indicate that feedforward neural architectures can provide an efficient way to implement compressed sensing in neural systems.

Introduction

Although it is still debated how many different odorants humans can perceive, the most commonly cited number is on the order of 104 [13], much greater than the 500 olfactory receptor neuron (ORNs) types. Many other species, including both vertebrates and insects, have the same order of magnitude of ORN types or even fewer (around 1000 in mice, 50 in Drosophila). The order of magnitude difference between the number of odorants and ORN types implies that humans as well as other species rely on compressed representations, potentially following the principles of compressed sensing [47].

In the compressed sensing framework [4], sparse high dimensional signals can be accurately reconstructed using a small number of measurements provided that the input signals are sparse. Natural odors are sparse in the sense that they are dominated by a few molecular components [810]. The relevance of compressed sensing algorithms to olfactory coding is reinforced by the anatomical organization of the olfactory system. High dimensional odor signals are compressed into a low-dimensional representation in terms of the activity of a relatively small number of glomeruli in the olfactory bulb, in the case of vertebrates, or the antennal lobe in the case of invertebrates. The standard compressed sensing algorithm performs signal reconstruction as a constrained 1 minimization [4]. Such optimization can be solved through neural dynamics [5, 6], but the resulting reconstructions were considerably less fault tolerant than observed experimentally. For example, mice olfactory discrimination remains essentially intact when half of glomeruli are disabled [11] whereas theoretical reconstructions fail at this level of signal interference [5]. Furthermore, signal reconstruction based on dynamical optimization by construction requires more time for signal recognition compared to feedforward reconstruction schemes. Here we describe a feedforward reconstruction scheme based on compressed sensing ideas that is both fault tolerant and matches the main features of the organization of the olfactory system. The results demonstrate that a purely feedforward network is capable of robustly compressing/decompressing binary signal without dynamical optimization.

Models and Methods

A compressed sensing model of the olfactory system

We begin by reviewing the main results from compressed sensing literature as they pertain to olfactory coding. The odor signal s0 can be described as a binary vector of length N where each element is either 1 or 0 depending upon whether a given molecular component is present or not in the odor. We refer to the number K of nonzero components in the odor as the odor sparsity. The main premise of compressed sensing is that a sparse signal s0 can be compressed into a vector x = As0 of length M < N and then recovered with high reconstruction quality provided KN. The encoding matrix A has dimensions M × N; its matrix elements can be chosen randomly. With this setup, the original signal s0 can be recovered exactly from the convex 1 optimization problem [4]

s^=argmin||s||1subjecttox=As0. (1)

Although the 1 minimization problem can be solved in polynomial time, it is not straightforward to implement such optimization algorithms in a neural circuit. One solution involves a two-layer neural network that perform similar 1 minimization through neural dynamics [6]. However, this imposes certain requirements on the structure of recurrent connections in the second layer together with a static nonlinear activation function. Another alternative implementation relies on 2 minimization instead of 1. In this case, the reconstruction is obtained simply as s^=(ATA)-1ATx where the −1 represents a pseudo-inverse relation. However, such an approach does not produce exact signal reconstruction [7] and would predict much larger errors than observed in olfactory experiments.

Robust feedforward reconstruction of sparse odors

We now propose a model for the olfactory system, which can compress and robustly recover sparse binary signal with high probability, without using any dynamical optimization. The solution is based on a nonlinear binary encoding model instead of the linear encoding model used in the conventional compressed sensing approach. Specifically, the compressed vector x has the form of a threshold function xi=H(xil-θc) where xl = As0 and H is the Heaviside step function with H(0)=1. We assume that the measurement matrix (affinity matrix) A is a M×N random binary matrix where each element is chosen independently to be either 1 or 0 with equal probability p and 1 − p, respectively. It is worth mentioning that while we use a random connectivity matrix in our model, we do not assume that this matrix differs across individuals. Rather, the randomness is meant to characterize how well the system works in the absence of specificity between odorants and glomeruli identity. By extending the definition of H to vectors, the measurement vector x can be compactly written as

x=H(As0-θc), (2)

where θc = 1, reflects that all measurements larger than 1 are set to 1 so that x is binary. This corresponds to a binary model of glomeruli activity described by the binary vector x. The threshold value of θc = 1 corresponds to a logical OR operation, so that glomerulus k will be activated if any of the odor components that are associated with inputs to this glomerulus are activated.

To reconstruct the original signal, the glomeruli activity x are projected to another layer of neurons (neurons in the olfactory cortex of vertebrates or Kenyon cells in the mushroom body of insects) which has the same dimension as the original signal s0. The activity of neurons in this layer is denoted by vector s^ which has the same dimensionality N as the original signal s0. The reconstructed signal can be computed as

s^=H(WTx-θr), (3)

where θr is the activation threshold for neurons in the reconstruction layer. The reconstruction matrix W equals the measurement matrix A normalized to 1 by column, i.e. Wki = Aki/∑k Aki. With this normalization, the reconstruction threshold θr = 1 corresponds to logical AND operation. That is, odor component i will be detected as present if all glomeruli that feed signals to node i in the reconstruction layer are activated. Below we will present most of the results for θr = 1 and then analyze how the reconstruction quality and recovery robustness depend on this threshold. We will also determine the optimal connectivity ratio from the compression to the reconstruction layer that maximizes the fidelity of reconstructions.

Maximal information transmission

Our feedforward model can be thought of as an information transmission channel that compresses, transmits, and decompresses a sparse binary signal. To find the optimal network configuration, we seek to maximize mutual information between the input and output of the channel as has been done to characterize performance in the visual and other sensory systems. The mutual information between s0 and s^ is given by

I(s0,s^)=s0s^P(s^|s0)P(s0)log2P(s^|s0)P(s^). (4)

For a given signal sparsity K, the conditional probability P(s^|s0) of the reconstructed signal s^ given the original signal s0 can be computed as:

P(s^|s0)=pfalseNerr(1-pfalse)(N-K-Nerr), (5)

where pfalseP(s^i=1|si0=0) is the probability of false detection for an odor component and Nerr=||s^||0-K is the number of false detection events for the odor s0. We note that for θr = 1, the probability to miss an odor component is zero provided this odor component activates at least one of the glomeruli. In this regime, the information is fully determined by the false detection rate pfalse, and as we show below decreases proportionally with pfalse.

Assuming a uniform prior over individual odor components P(s0)=1/(NK), one can also compute the probability distribution of reconstructed signals:

P(s^)=s0P(s^|s0)P(s0)=(K+NerrK)(NK)pfalseNerr(1pfalse)(NKNerr) (6)

Putting together Eqs (4)–(6), the mutual information can be written as

I(s0,s^) =log2(NK)Nerr=0NK(NKNerr)pfalseNerr(1pfalse)(NKNerr)log2(K+NerrK).

When (NK)pfalse ≪ 1, the summation above can be well approximated by its leading nonzero term

Nerr=0NK(NKNerr)pfalseNerr(1pfalse)(NKNerr)log2(K+NerrK)(NK)pfalselog2(K+1), (7)

so that the expression for the mutual information becomes:

I(s0,s^)log2(NK)(NK)pfalselog2(K+1). (8)

Thus, for given N and K, maximizing I(s0,s^) can be approximated by minimizing the probability of false detection pfalse.

Results

Optimal connectivity rate

The false detection rate that appears in Eq 8 can be computed as

pfalseP(s^i=1|si0=0)=k=1MP(s^i=1|||Ti||0=k)P(||Ti||0=k|si0=0)=k=1M1-(1-p)Kk(Mk)pk(1-p)M-k1-(1-p)M=11-(1-p)Mk=0M1-(1-p)Kk(Mk)pk(1-p)M-k-(1-p)M1-(1-p)M=1-p(1-p)KM-(1-p)M1-(1-p)M, (9)

where Ti ≡ {xkx|Aki = 1}, and p is the average connectivity rate from the compression to the reconstruction layer. In the last line above we use the binomial expansion. Because we are interested in the regime where M is large, we have (1 − p)M ≪ [1 − p(1 − p)K]M ≪ 1 as long as p is not too small. Thus, Eq 9 can be approximated with great accuracy by the following simple equation:

pfalse=1-p(1-p)KM. (10)

As shown in the inset of Fig 1B, Eq 10 provides an accurate approximation when the connectivity p is not too sparse. Since our main interest is near the optimal connectivity rate (see below) where Eq 10 is very accurate, we will use Eq 10 unless specified.

Fig 1.

Fig 1

(A) Illustration of the model structure. An odor is represented by a sparse binary vector s0 of its mono-molecular components. This signal is compressed into the activities of M glomeruli represented by a binary vector x through a binary measurement matrix A. The signal is then recovered as the activities of N neurons in the mushroom body or olfactory cortex represented by a binary vector s^ through another matrix WT. (B) The signal-to-noise ratio (SNR) as a function of signal sparsity K where N = 10000 and M = 500. For a given K, there is a optimal connectivity rate p = pm that maximizes SNR. At the same time, even for a system optimized to a given K, decreasing K still increases SNR. Inset: false detection rate pfalse as a function of average connectivity p; M = 500 and K = 15 are chosen for this illustration. Solid line is exact formula, while dashed line is the approximation using Eq 10. We can see that Eq 10 is a very good approximation to the exact formula when p is not too small.

As expected, the false detection rate pfalse decreases as the number of glomeruli M increases and as the signal sparseness K decreases. Importantly, for a given M and K, there is an optimal p, which we refer to as pm, that minimizes pfalse, as shown in Fig 1B. Taking ∂pfalse/∂p = 0 leads to

pm=1K+1. (11)

It is worth noticing that the optimal connectivity pm is independent of the number of glomeruli M, and depends only on the signal sparseness K. Thus, optimal connectivity depends exclusively on the level of sparseness of signals in the environment and can be determined prior to any measurements on neural circuits.

For an optimal connectivity p = pm, the probability of fault activation decreases exponentially as M increases and thus can be very small. This indicates that the proposed feedforward compression-reconstruction scheme from Fig 1A can achieve exact recovery with high probability.

To test the reconstruction quality, we compute the signal-to-noise-ratio (SNR) of the recovered signal. Since all nonzero components in the original will be recovered, the only source of errors in the reconstructed signal are due to false detection rates. Therefore, we can define the SNR of recovered signal as

SNR=||s0||0<||s^||0>-||s0||0=K(N-K)pfalse, (12)

as shown in Fig 2A–2C, where < ⋅ > denotes the expectation value. We can see from Fig 2B that the SNR increases exponentially with M. For our case where KN, we can achieve a high SNR for a number of glomeruli M much smaller than the number of odor components N or, equivalently, the number of third-order neurons.

Fig 2. Signal-To-Noise-Ratio (SNR) of the recovered signal in our model.

Fig 2

N = 10000 is used. (A) SNR as a function of K and M. Black is shown for SNR > 106. The blue line shows SNR = 1, and the red line shows SNR = K, i.e. one error occurs on average. (B) Optimal SNR as a function of M. (C) Optimal SNR as a function of K. (D) Number of glomeruli required to reach threshold SNR when optimal connectivity rate is used.

Compression rate and sparsity

A key characteristic of a compression algorithm is the compression ratio αM/N. In previous compressed sensing frameworks, the critical compression ratio αc above which the signal can be perfectly recovered was shown to only depend on the relative signal sparsity fK/N. As f → 0, αc(f)∼ −f log f [12]. To compute the critical compression ratio for our reconstruction algorithm, we note that from Eq 12, log pfalse = log f − log(1 − f) − log SNR. In the strong compression limit where fK/N is small, this yields

logpfalselogf-logSNR. (13)

On the other hand, for the optimal connectivity rate pm and large K, log pfalse can also be simplified using Eq 10 as follows:

logpfalse=Mlog1-1K+11+1K-KMlog1-1K+1e-1-MeK=-αSNRef. (14)

where αSNR is defined as the compression rate to achieve a certain SNR. Combining Eqs 13 and 14, in the limit of strong compression where f → 0, the critical compression ratio behaves as αSNR ∼ −f log f. We note that care should be taken when the SNR becomes comparable to or larger than N because 1/f = N/KN, so that log SNR cannot be neglected when f → 0.

The obtained critical compression rate can be compared to its theoretical limit. The latter corresponds to the minimal number of bits required to encode a sparse signal:

Mmin=log2(NK), (15)

where ⌈x⌉ is the smallest integer not less than x. When N and K are large but fK/N is small, using Stirling’s approximation, we obtain that

Mmin×log2NlogN-KlogK-(N-K)log(N-K)KlogN-KlogK+K=K-Klogf, (16)

This yields that the theoretically possible compression ratio αmin in the strong compression limit of f → 0 as

αminflog2e/f, (17)

which also yields αmin ∼ −f log f as f → 0.

Notice that although both αSNR and αmin behave as −f log f for f → 0, they have different proportionality coefficients. To be more specific, αSNRef log 1/f while αmin ∼ (log 2)−1 f log 1/f. As a result, αSNR/αmine log 2 ≈ 1.88 as f → 0. Thus, the number of glomeruli needed in our model is about twice the theoretical limit but is achieved here with an extremely simple feedforward encoding model.

As shown in Fig 2D, the number of required glomeruli increases sub-linearly with K, and logarithmically with SNR. In practice, with only a few times more glomeruli than the theoretical limit, a very high SNR can be achieved.

Robustness and fault tolerance

Advances in experimental techniques provide opportunities to test our theory under the circumstances of extreme genetic manipulations. For example, following a genetic manipulation that caused most olfactory receptor neurons to express a single odorant receptor M71, the M71 ligand acetophenone activates half of the glomeruli. Despite this drastic manipulation, mice can still readily detect other odors in the presence of acetophenone, while their discrimination performance is only moderately compromised [11]. This result is consistent with our model. Assume there are M glomeruli in our model and half of them are always turned on (corrupted). Such a system is equivalent to a model with only M/2 glomeruli, since the anomalously activated glomeruli will not affect signal recovery. Thus, the odor signal can still be recovered, but the SNR is decreased, which is in agreement with the experimental result. As a comparison, in previous compressed sensing framework, one can only allow a small percentage of corrupted glomeruli even when M > N [4].

In another set of experimental studies, part of the glomeruli in mice are removed or disabled [1315]. It is shown that the ability to discriminate odors and simple odor mixtures is not impaired even when most of the glomeruli are removed or disabled. This seemingly surprising finding is also consistent with our model. From previous results, one can see that decreasing M will only lead to larger noise in the recovered odor signal but not to a failure of the system if the activation threshold for neurons in the reconstruction layer can be properly adapted to the new M. Assume the mice need SNR > ν to discriminate odors. When K is small, the minimal M needed for discrimination is

Mlow=logKNνlog[1-p(1-p)K]. (18)

From experiment data, p ≈ 0.05 (although this is a very rough estimation, see [11, 1618]). One can check that the equation above is insensitive to variations in K and over a broad range. If we assume K < 10 (as in the experiments) and is within the range of 104 ∼ 105, then Mlow is roughly between 200 and 300, or around 20% of the glomeruli, which is in good agreement with the data in those experiments.

On the other hand, our model can tolerate negative gloleruli noise (false negative) by changing its recovery threshold θr. Although we use θr = 1 in our results for analytical solution, it is very likely that real biological systems would use a lower threshold θr. With θr < 1, the SNR is somewhat lower, as shown in Fig 3, yet the system is more robust to noise in the reconstruction stage since the activation of a third-order neuron doesn’t require all of its connected gloleruli to be active and it also leaves room for odor generalization and pattern completion [19]. Indeed, when the threshold at the reconstruction stage is less than 1, the reconstruction can tolerate some incompleteness in the glomeruli activation patterns. Real biological systems likely have the ability to adaptively change the activation threshold in order to balance the needs of high quality reconstruction and pattern completion.

Fig 3. Demonstration of the accuracy-robustness trade-off.

Fig 3

N = 10000, K = 15, M = 1000 and the optimal connectivity rate are used. (A) pfalse and SNR for different activation thresholds at the reconstruction stage. With lower recovery thresholds, the robustness of the system to recovery noise increases, while the false detection rate increases, and the SNR of recovered signal decreases. (B) An example of the recovered signal with different recovery thresholds. True signal is shown in big colored dots, while the reconstruction error is represented by small colored dots. As we lower the threshold, the recovered signal becomes noisier.

Our model is shown to be very robust and fault tolerant, and this robustness is achieved with accuracy. As one can see, each glomerulus in the model only contains part of the information about the original signal. Because the measurement matrix A is random, no single glomerulus or cluster contains more or unique information, so any subset of the glomeruli could recover the original signal. The more glomeruli there are, the better recovery quality (SNR) can be achieved. Thus, removing or disabling part of the glomeruli will not change the system qualitatively, but will make the recovered signal more noisy, up to a point where noise becomes comparable to the true signal at which point the reconstruction fails. For a real biological system, it is reasonable to assume that the recovered signal has very high SNR, which also means high redundancy, as is observed experimentally.

Discussion

Predicted optimal connectivity rate compared with experimental data

From our analysis we observed that for a given level of signal sparseness K, there is an optimal connectivity rate pm that maximizes SNR as well as the mutual information. Assuming that the biological system is adapted to a given value of odor sparseness in its environmental niche, one can essentially make predictions on the connectivity rate of matrix A. This is followed by another prediction that the percentage of glomeruli activated by a single odorant should be close to the percentage of glomeruli that could activate a neuron in olfactory cortex or a Kenyon cell, and this number should be similar among species which operate in similar olfactory environments. The latter prediction should be easier to test, since the number of coexisting odorants in the environment is hard to measure.

Fortunately, previous experiments have gathered sufficient data to test our prediction indirectly. It has been shown that in Drosophila, 9% of the glomeruli have a strong response to an odorant [20], while the connectivity rate between glomeruli and Kenyon Cells is 6.5% [21] to 12.5% [22]. (The latter number is obtained based on the average number of claws per Kenyon cell measured in [22]) These estimates are consistent with model predictions. Furthermore, in the locust, a typical projection neuron responds to about half of the odorants [23], while the connectivity rate between projection neurons and Kenyon Cell is also around 50% [24], which is also consistent with our prediction.

We can see that the connectivity rate is very different between species. Such differences can be unified in our model as the adaptation to different environmental niches. The locust has an anomalously high connectivity rate (50%), which in our model implies that its olfactory system is adapted to extreme odor sparseness tuned to odors with primarily a single component (pm = 0.5 when K = 1). Similarly, Drosophila is adapted to sense odors composed of a mixture of about 10 odor components, while mice are tuned to detect a mixture of about 20 mono-molecular odors. In general, our model predicts that species with sparse connectivity will behave better in environments with complex odor mixtures, while species with dense connectivity have better performance in detecting simple odor mixtures.

Structural and functional evidence

In addition to the predictions above, further experimental evidence supports the structure of our model, in particular the approximate logical OR/AND operations associated with the compression/reconstruction stages, respectively. For example, it has been observed experimentally that Kenyon Cells in Drosophila receive convergent input from different glomeruli and require several inputs to be co-active to spike [25]. This is consistent with our threshold activation function which at the reconstruction stage uses a logical AND operation.

Functionally, experiments have shown that locust Kenyon cells are individually much better than projection neurons from glomeruli at detecting a single odorant; Kenyon cells that respond to an odorant also often respond to odor mixtures containing it [26]. This observation agrees with our assumption that each Kenyon cell only responds to one odorant and it will respond when an odor mixture contains that odorant.

Stereotyped versus non-stereotyped connectivity

Since the affinity matrix A is determined genetically, all the connections in our model are predetermined before birth. There is some debate about such stereotypy versus random connectivity, and a compressed sensing model of olfaction based on random connections from glomeruli to mushroom body has been proposed [27]. Yet, our model supports both stereotyped and non-stereotyped projection from glomeruli to the mushroom body/olfactory cortex because the model is invariant under the exchange of neurons within the same layer. In order to verify such predetermination, one needs to obtain a detailed connectivity map from glomeruli to the mushroom body/olfactory cortex for different individuals, which is experimentally very challenging. An indirect approach to verify the predetermined connectivity hypothesis could be through an examination of innate behaviors that should depend primarily on predetermined connections. If one could relate innate behaviors to projections between glomeruli and the mushroom body/olfactory cortex, it would then provide additional supporting evidence for the genetically predetermined structural connectivity of the feedforward model.

Effective feedforward model for non-feedforward structure

The feedforward structure of our model is an effective approximation to the more complicated structure of biological olfactory system where recurrent and feedforward-feedback connections exist. For example, it has been observed that inhibitory interneurons modulate neuronal responses in the olfactory bulb [28, 29]. In linear dynamic systems, such feedforward-feedback structure could be mathematically modeled as a pure feedforward system with different effective feedforward connectivity. Suppose that we add a layer of interneurons z in Fig 1 that is connected to the glomeruli layer x by feedforward-feedback connectivity B. Then the linear dynamics of the system are x˙=-x+As0-BTz and z˙=-z+Bx, where we assume B is feedforward excitatory and feedback inhibitory. The steady state solution is x = (I + BT B)−1 As0, which is the same for a pure feedforward system, except that connectivity A is replaced by (I + BT B)−1 A. This analysis is not exact if the activation function is nonlinear. In general, the feedforward-feedback system in steady state with a nonlinear activation function does not have an equivalent feedforward system, but one can still write the linear perturbation when neurons receive only weak inputs, which allows a feedforward approximation. Such a feedforward approximation is supported by experimental observations that the representations of odor mixtures in mouse glomeruli can be explained well by the summation of the glomeruli responses to their components [30].

One advantange of the effective feedforward model is that it enables an adaptive affinity matrix even with pre-determined connectivity. In the feedforward-feedback architecture mentioned above, the effective affinity matrix is (I + BT B)−1 A, where A is the pre-determined affinity matrix encoded in the genes, while B could be a learned matrix adapted to the environment. From this perspective, the existence of interneurons in both insects and vertebrates [31, 32], as well as adult neurogenesis in the olfactory bulb of mammals [33], could play the role of adjusting the effective affinity matrix for the purpose of adaptation.

Comparison with 1 minimization algorithm

We compare the performance of our feedforward architecture with the often-used LASSO 1 minimization algorithm [34] provided by the Python scikit-learn library

mins^12M||As^-x||22+β||s^||1, (19)

where N = 1000, M = 500, β = 0.001 are used. Linear measurement x = As0 is used for LASSO. For each K, we conduct 100 experiments with different random measurement matrices and signals, and compute the average of the reconstruction errors ||s^-s0||1 as well as the number of iterations used in LASSO. We also compute the mean reconstruction error when only 5 iterations are used in LASSO as a comparison. The results are shown in Fig 4. As shown in the figure, the feedforward architecture has a lower reconstruction error when the signal is very sparse, while LASSO has a lower reconstruction error than the feedforward architecture when K becomes larger. However, the number of iterations also increases as the signal becomes denser. If we restrict the number of iterations to 5 in the LASSO (equivalent to setting a maximum response time), LASSO performs much worse when the signal is very sparse. But as K increases, it still has a lower reconstruction error than the feedforward architecture.

Fig 4. Comparison of the performance of feedforward architecture with that of LASSO.

Fig 4

For this example, we chose N = 1000 and M = 500. Linear measurement is used for LASSO. Feedforward architecture performs well when the signal is very sparse, while LASSO has lower reconstruction error as K increases, at the price of increasingly more iterations. On the other hand, if we constrain the number of iterations, LASSO still performs better when K is large, but significantly worse with very sparse signals.

Performance with non-sparse signal

One drawback of this feedforward architecture is that it may not be able to achieve both compression and high-quality reconstruction simultaneously when the signal is not sparse. Unlike the 1 minimization method where the number of measurements required to reconstruct the signal will never exceed signal length N (N/2 for binary signal)[35, 36], the feedforward architecture may need more measurements than the signal length to accurately reconstruct the signal. This can be seen by restoring the term in Eq 13 that we have previously neglected assuming that f is small

logpfalse=logf-log(1-f)-logSNR. (20)

Combining this with Eq 14 that remains the same when f is not small, we obtain:

αSNR=eflogSNR+eflog(f-1-1), (21)

which could be larger than 1 when f is not small. Thus, the feedforward computation may require number of measurements that are larger than the input dimensionality to achieve reliable reconstruction.

From another perspective, we can compute the upper bound on the reconstruction SNR that can be achieved for a given compression level. From Eq 21 and αSNR < 1 we get

logSNR<1ef-log(f-1-1), (22)

which only depends on signal sparsity. For example, if f = 0.1, then SNR < 4.4, and the reconstructed signal will not be accurate.

Extension to continuous variables and other activation functions

Although our analysis above is based on a binary signal / measurement matrix / glomeruli activity and threshold activation function, our results can be extended to positive real-valued signal / measurement matrix / glomeruli activity and any monotonically increasing activation function. Consider the case where the signal s0 and the element of measurement matrix Aij could take any positive value rather than just 0 and 1. Denoting xl = As0, and letting the activation function g be any monotonically increasing function, the output at the glomerulus stage can be written as xi=g(xil). Now, signal reconstruction can proceed based on the evaluation of a minimum function (rather than the logical AND function that was used in the case of binary inputs and binary measurement matrices). Indeed, when the ith component of the reconstructed signal s^i is computed as the smallest value {g−1(xj)/Aji} across the set of its inputs (i.e. where Aji ≠ 0), then our analysis remains valid. The only modification is that now the distribution of the signal and the measurement matrix elements are both required to compute the noise magnitude. This procedure ensures that the recovered components are still recovered exactly, while corrupted components are still corrupted. As a practical aside, we note that the minimum function can be implemented by short-term synaptic plasticity, see S1 Text and S1 and S2 Figs.

Supporting Information

S1 Text. Neural implementation of min function using short-term plasticity.

We show by simulation that the output firing rate of a Leaky Integrate and Fire neuron could be well approximated by its minimal input firing rate when synaptic weight is controlled by Short-Term Plasticity.

(PDF)

S1 Fig. Simulation results of S1 Text.

(EPS)

S2 Fig. Simulation results of S1 Text.

(EPS)

Acknowledgments

We thank Fred. H. Gage, Charles Stevens, Ryan Rowekamp and other members of the CNL-T group for helpful discussions.

Data Availability

All relevant data are within the paper.

Funding Statement

This research was supported by McDonnel Foundation, NSF CAREER Award IIS-1254123 and NSF grant IOS-1556388, as well as NEI grants P30 EY019005 and T32 EY020503. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Gilbert AN. What the nose knows: the science of scent in everyday life. Crown Publishers New York; 2008. [Google Scholar]
  • 2. Hummel T, Welge-Lüssen A. Taste and smell: An update. vol. 63. Karger Medical and Scientific Publishers; 2006. [Google Scholar]
  • 3.Kandel ER, Schwartz JH, Jessell TM, Siegelbaum SA, Hudspeth A. Principles of neural science; 2014.
  • 4. Candes EJ, Tao T. Decoding by linear programming. Information Theory, IEEE Transactions on. 2005;51(12):4203–4215. 10.1109/TIT.2005.858979 [DOI] [Google Scholar]
  • 5. Tootoonian S, Lengyel M. A Dual Algorithm for Olfactory Computation in the Locust Brain In: Advances in Neural Information Processing Systems; 2014. p. 2276–2284. [Google Scholar]
  • 6. Rozell CJ, Johnson DH, Baraniuk RG, Olshausen BA. Sparse coding via thresholding and local competition in neural circuits. Neural computation. 2008;20(10):2526–2563. 10.1162/neco.2008.03-07-486 [DOI] [PubMed] [Google Scholar]
  • 7. Ganguli S, Sompolinsky H. Compressed sensing, sparsity, and dimensionality in neuronal information processing and data analysis. Annual review of neuroscience. 2012;35:485–508. 10.1146/annurev-neuro-062111-150410 [DOI] [PubMed] [Google Scholar]
  • 8. Jouquand C, Chandler C, Plotto A, Goodner K. A sensory and chemical analysis of fresh strawberries over harvest dates and seasons reveals factors that affect eating quality. Journal of the American Society for Horticultural Science. 2008;133(6):859–867. [Google Scholar]
  • 9. Bengtsson M, Bäckman AC, Liblikas I, Ramirez MI, Borg-Karlson AK, Ansebo L, et al. Plant odor analysis of apple: antennal response of codling moth females to apple volatiles during phenological development. Journal of agricultural and food chemistry. 2001;49(8):3736–3741. 10.1021/jf0100548 [DOI] [PubMed] [Google Scholar]
  • 10. Karlsson MF, Birgersson G, Cotes Prado AM, Bosa F, Bengtsson M, Witzgall P. Plant odor analysis of potato: response of Guatemalan moth to above-and belowground potato volatiles. Journal of agricultural and food chemistry. 2009;57(13):5903–5909. 10.1021/jf803730h [DOI] [PubMed] [Google Scholar]
  • 11. Fleischmann A, Shykind BM, Sosulski DL, Franks KM, Glinka ME, Mei DF, et al. Mice with a “monoclonal nose”: perturbations in an olfactory map impair odor discrimination. Neuron. 2008;60(6):1068–1081. 10.1016/j.neuron.2008.10.046 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Ganguli S, Sompolinsky H. Statistical mechanics of compressed sensing. Physical review letters. 2010;104(18):188701 10.1103/PhysRevLett.104.188701 [DOI] [PubMed] [Google Scholar]
  • 13. Guthrie KM, Holmes J, Leon M. Focal 2-DG uptake persists following olfactory bulb lesions. Brain research bulletin. 1995;38(2):129–134. 10.1016/0361-9230(95)00078-S [DOI] [PubMed] [Google Scholar]
  • 14. Lu XC, Slotnick B. Olfaction in rats with extensive lesions of the olfactory bulbs: implications for odor coding. Neuroscience. 1998;84(3):849–866. 10.1016/S0306-4522(97)00520-4 [DOI] [PubMed] [Google Scholar]
  • 15. McBride K, Slotnick B, Margolis FL. Does intranasal application of zinc sulfate produce anosmia in the mouse? An olfactometric and anatomical study. Chemical senses. 2003;28(8):659–670. 10.1093/chemse/bjg053 [DOI] [PubMed] [Google Scholar]
  • 16. Fantana AL, Soucy ER, Meister M. Rat olfactory bulb mitral cells receive sparse glomerular inputs. Neuron. 2008;59(5):802–814. 10.1016/j.neuron.2008.07.039 [DOI] [PubMed] [Google Scholar]
  • 17. Uchida N, Takahashi YK, Tanifuji M, Mori K. Odor maps in the mammalian olfactory bulb: domain organization and odorant structural features. Nature neuroscience. 2000;3(10):1035–1043. 10.1038/79857 [DOI] [PubMed] [Google Scholar]
  • 18. Wachowiak M, Cohen LB. Representation of odorants by receptor neuron input to the mouse olfactory bulb. Neuron. 2001;32(4):723–735. 10.1016/S0896-6273(01)00506-2 [DOI] [PubMed] [Google Scholar]
  • 19. Khan AG, Parthasarathy K, Bhalla US. Odor representations in the mammalian olfactory bulb. Wiley Interdisciplinary Reviews: Systems Biology and Medicine. 2010;2(5):603–611. [DOI] [PubMed] [Google Scholar]
  • 20. Hallem EA, Carlson JR. Coding of odors by a receptor repertoire. Cell. 2006;125(1):143–160. 10.1016/j.cell.2006.01.050 [DOI] [PubMed] [Google Scholar]
  • 21. Caron SJ, Ruta V, Abbott L, Axel R. Random convergence of olfactory inputs in the Drosophila mushroom body. Nature. 2013;497(7447):113–117. 10.1038/nature12063 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Butcher NJ, Friedrich AB, Lu Z, Tanimoto H, Meinertzhagen IA. Different classes of input and output neurons reveal new features in microglomeruli of the adult Drosophila mushroom body calyx. Journal of Comparative Neurology. 2012;520(10):2185–2201. 10.1002/cne.23037 [DOI] [PubMed] [Google Scholar]
  • 23. Perez-Orive J, Mazor O, Turner GC, Cassenaer S, Wilson RI, Laurent G. Oscillations and sparsening of odor representations in the mushroom body. Science. 2002;297(5580):359–365. 10.1126/science.1070502 [DOI] [PubMed] [Google Scholar]
  • 24. Jortner RA, Farivar SS, Laurent G. A simple connectivity scheme for sparse coding in an olfactory system. The Journal of neuroscience. 2007;27(7):1659–1669. 10.1523/JNEUROSCI.4171-06.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Gruntman E, Turner GC. Integration of the olfactory code across dendritic claws of single mushroom body neurons. Nature neuroscience. 2013;16(12):1821–1829. 10.1038/nn.3547 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Shen K, Tootoonian S, Laurent G. Encoding of mixtures in a simple olfactory system. Neuron. 2013;80(5):1246–1262. 10.1016/j.neuron.2013.08.026 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Stevens CF. What the fly’s nose tells the fly’s brain. Proceedings of the National Academy of Sciences. 2015;112(30):9460–9465. 10.1073/pnas.1510103112 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Olsen SR, Wilson RI. Lateral presynaptic inhibition mediates gain control in an olfactory circuit. Nature. 2008;452(7190):956–960. 10.1038/nature06864 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Lledo PM, Saghatelyan A, Lemasson M. Inhibitory interneurons in the olfactory bulb: from development to function. The Neuroscientist. 2004;10(4):292–303. 10.1177/1073858404263460 [DOI] [PubMed] [Google Scholar]
  • 30. Shea SD, Katz LC, et al. Representation of natural stimuli in the rodent main olfactory bulb. Neuron. 2006;50(6):937–949. 10.1016/j.neuron.2006.03.021 [DOI] [PubMed] [Google Scholar]
  • 31. Chou YH, Spletter ML, Yaksi E, Leong JC, Wilson RI, Luo L. Diversity and wiring variability of olfactory local interneurons in the Drosophila antennal lobe. Nature neuroscience. 2010;13(4):439–449. 10.1038/nn.2489 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Shepherd GM, Chen WR, Willhite D, Migliore M, Greer CA. The olfactory granule cell: from classical enigma to central role in olfactory processing. Brain research reviews. 2007;55(2):373–382. 10.1016/j.brainresrev.2007.03.005 [DOI] [PubMed] [Google Scholar]
  • 33. Lazarini F, Lledo PM. Is adult neurogenesis essential for olfaction? Trends in neurosciences. 2011;34(1):20–30. [DOI] [PubMed] [Google Scholar]
  • 34. Candè EJ, Wakin MB. An introduction to compressive sampling. Signal Processing Magazine, IEEE. 2008;25(2):21–30. 10.1109/MSP.2007.914731 [DOI] [Google Scholar]
  • 35. Donoho D, Reeves G. The sensitivity of compressed sensing performance to relaxation of sparsity In: Information Theory Proceedings (ISIT), 2012 IEEE International Symposium on. IEEE; 2012. p. 2211–2215. [Google Scholar]
  • 36. Stojnic M. Recovery thresholds for L1 optimization in binary compressed sensing In: Information Theory Proceedings (ISIT), 2010 IEEE International Symposium on. IEEE; 2010. p. 1593–1597. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

S1 Text. Neural implementation of min function using short-term plasticity.

We show by simulation that the output firing rate of a Leaky Integrate and Fire neuron could be well approximated by its minimal input firing rate when synaptic weight is controlled by Short-Term Plasticity.

(PDF)

S1 Fig. Simulation results of S1 Text.

(EPS)

S2 Fig. Simulation results of S1 Text.

(EPS)

Data Availability Statement

All relevant data are within the paper.


Articles from PLoS Computational Biology are provided here courtesy of PLOS

RESOURCES