Abstract
The study of complex information processing systems requires appropriate theoretical tools to help unravel their underlying design principles. Information theory is one such tool, and has been utilized extensively in the study of the neural code. Although much progress has been made in information theoretic methodology, there is still no satisfying answer to the question: “What is the information that a given property of the neural population activity (e.g., the responses of single cells within the population) carries about a set of stimuli?” Here, we answer such questions via the minimum mutual information (MinMI) principle. We quantify the information in any statistical property of the neural response by considering all hypothetical neuronal populations that have the given property and finding the one that contains the minimum information about the stimuli. All systems with higher information values necessarily contain additional information processing mechanisms and, thus, the minimum captures the information related to the given property alone. MinMI may be used to measure information in properties of the neural response, such as that conveyed by responses of small subsets of cells (e.g., singles or pairs) in a large population and cooperative effects between subunits in networks. We show how the framework can be used to study neural coding in large populations and to reveal properties that are not discovered by other information theoretic methods.
Keywords: neural coding, information theory, maximum entropy, population coding
Some of the greatest challenges to science today involve complex systems, such as the brain and gene regulatory networks. Such systems are characterized by a very large number of interacting units that potentially cooperate in complex ways to produce ordered behavior. Some of the more interesting systems may be viewed, to a certain degree, as input–output systems. The brain, for example, receives multiple inputs from the environment and processes them to generate behavior. In order to obtain insight into such systems, this transmission of information needs to be quantified. An attractive mathematical tool in this context is information theory (IT), introduced by Claude Shannon in his mathematical theory of communication (1). IT has been used in neuroscience since the 1950s (2, 3), yielding insights into design principles in neural coding (4, 5) and offering new methods for analyzing data obtained in neurophysiological experiments (6). The main information theoretic measure used in the literature is the mutual information (MI), which quantifies the level of dependence between 2 variables. Experimental works typically employ IT by studying the MI between aspects of the external world [e.g., motor activity (7) or visual stimuli (8)] and aspects of the neural response [e.g., spike counts (9) or precise spike times (10) among others].
Empirical studies of complex systems in general and information theoretic analyses in particular are fundamentally limited by the fact that the space of possible system states is extremely large. Thus any measurement of the system is bound to be limited and reveal only a subset of its possible states. For example, it is not practical to fully characterize the statistics of a 100-ms spike train of even a single neuron, because of its high dimensionality (2100 for 1-ms precision) and the limited number of experimental trials. The problem of limited measurements is more acute for multiple neurons for 2 reasons. First, the dimension of the response space grows exponentially with the number of neurons. Second, neurons are often not recorded simultaneously but rather over several recording sessions, so their joint statistics are not accessible. However, it is possible to reliably estimate partial measurements, or statistics, of the system, such as firing rates of single neurons or correlations between pairs of neurons. Consider for example a population of a hundred neurons where we experimentally characterize the response of each neuron to a given set of stimuli. These measurements provide a partial characterization of the population response, and as such should also help in estimating the MI between this response and the stimulus.
Here, we present a framework for evaluating MI in such settings. At the basis of our approach is the assumption that the partial measurements hold for the true underlying system, whose complete characterization cannot be measured. We next consider all hypothetical systems that are consistent with the observed partial measurements (e.g., all populations of 100 neurons whose single neuron responses are identical to the ones we measured). Clearly, there is a large set of such systems, each with its own value of MI between input and output. Our goal is to find the value of information that can be attributed only to the given measurements. Intuitively, the systems with relatively high MI in the hypothetical set have some additional structure that cannot be inferred based on the given partial measurements alone. However, the system with minimum information in this set cannot be simplified further (in the MI sense) and its information can thus be taken to reflect the information available in the given measurements alone. Our minimum information (MinMI) principle thus states that given a set of measurements of a system, the MI available in these measurements is the minimum mutual information between input and output in any system consistent with the given measurements. An immediate implication of the above construction is that this minimum information is a lower bound on the information in the true underlying system, because the true system is also in the set we minimized over.
A conceptual tool which has previously been used to tackle partial measurements is the maximum entropy (MaxEnt) principle (11–13). MaxEnt posits that the underlying distribution is the one with “least structure” among all those that could have generated the given measurements. The MaxEnt principle is oblivious to the stimulus–response structure of a system and thus may arrive at incorrect conclusions regarding the information content of the neural activity. In contrast, MinMI directly considers information transfer, and yields a bound on the information in the true system, whereas MaxEnt does not. The MinMI and MaxEnt distributions are also mathematically very different, as we show below.
Minimum Information Bound
Consider a system with a discrete input stimulus S ∈ {1,…,ns} and an output response R ∈{1,…,nr} (our formalism applies to continuous variables as well; we focus on discrete ones for presentation purposes). The mutual information between S and R is a measure of the dependence between these 2 variables. Denote by p(s,r) the joint distribution of S and R. This distribution fully characterizes the input–output relations in the system. The mutual information between S and R is defined as
where p(s),p(r) are the marginal distributions of S,R (14). The MI is zero if and only if the variables are independent. High MI indicates that the response R encodes properties of the stimulus S.
We focus on the case where p(s,r) is not known. Rather, we have access to partial measurements of it, given by the expected value of some function of R given S. Formally, we consider a set of d functions → ℝd and assume we know their expected values given S. We denote these expected values by (s), so that
where the expectation operator 〈〉 is defined by 〈f(x)〉p(x) = ∑xf(x)p(x). Such expected values may be the firing rates of individual neurons in a population, the number of coincident spikes for pairs of neurons, or any other measurable statistic of the spatiotemporal activity. The expected values are typically estimated from experimental data. Denote the experimental data by the set of pairs (s1,r1),…,(sn,rn). Then (s) is estimated by , where ms is the number of data pairs where si = s. Because of finite sample effects, this empirical estimate will typically not equal the true expected value. In what follows, we shall assume that these values are identical. However, our approach may be extended to the cases where the expected values are known up to some confidence interval. We further assume that the prior probabilities of the stimulus variable are known, and denote these by p(s). This assumption is reasonable because these probabilities are usually determined by the experimentalist. The above formalism encompasses a wide range of response characteristics, from the response of single neurons, through that of neurons over time, to joint statistics of any order.
To bound the information in our system from below, we consider all hypothetical joint distributions that could yield the given partial measurements. These distributions are those that yield expected values of (r) that are equal to the measured ones
The true underlying distribution p(s,r) is clearly in P((s),p(s)). Thus, the true underlying information Ip(R;S) is lower bounded by the minimum information attainable in P((s),p(s))
where denotes the minimum information value. The distribution MI(r|s) that achieves the above minimum may be characterized by introducing Lagrange multipliers (s) ∈ ℝd (1 vector per s value), yielding the following characterization (which can be viewed as a generalization of the result in ref. 15)
where γ(s) is a normalization factor. Note that MI(r) depends on MI(r|s) through marginalization: MI(r) = ∑sp(s)MI(r|s). Thus, Eq. 5 is not a closed form solution but, rather, a set of equations involving the variables MI(r|s) and (s). The variables (s) should be chosen to satisfy the constraints P((s),p(s)) and may be found by using an iterative algorithm, as we show in Methods. The value of the minimum information turns out to be a simple function of the optimal (s):
The function (r) may be any function of the response space. In analyzing neural codes, we shall be specifically interested in the case where the expected values are kth order marginals of the true distribution p(r|s). We denote the minimum information given the set of all kth order marginals by I(k) (see Minimum Information in kth-order Statistics in SI Appendix). In what follows, we shall specifically demonstrate how I(1) and I(2) may be used to study aspects of neural coding.
It is interesting to contrast the MinMI solution with that of MaxEnt (see Relation to Maximum Entropy Modeling in SI Appendix). The MaxEnt approach will seek the distribution ∈ P((s),p(s)) with maximum entropy. This distribution (and its MI value) is very different from the MinMI one, as shown in the following simple example. Consider a set of N binary neurons with the same first-order responses p(ri|s) to 2 stimuli: p(ri = 1|s = 1) = α and p(ri = 1|s = 2) = β, and assume p(s) = 0.5. Clearly the information minimizing distribution is one where neurons are completely correlated (i.e., all fire or don't fire simultaneously). Thus the MinMI information will equal the information in a single neuron. On the other hand, the MaxEnt distribution in this case will correspond to neurons being conditionally independent given the stimulus. As N → ∞, the information in the MaxEnt distribution will approach one (as long as α ≠ β), because an observer of the response R will be able to perfectly predict the identity of the stimulus S by averaging over the N neurons to obtain (with probability 1) the values α,β for s = 1,2. Thus, the MaxEnt approach becomes inadequate for measuring information in large populations, whereas the MinMI approach does not have this limitation. This difference will be illustrated in the experiments reported below.
Synergy and Redundancy Measures
A key issue in neural coding is the importance of high-order statistics and their contribution with respect to lower-order statistics. One approach to quantifying this contribution is to compare the MI in a model based on higher order statistics with one based on lower-order statistics. A positive difference indicates synergy: information in higher-order interactions, whereas a negative difference indicates redundancy. Several such measures have been suggested in the literature (9, 16–19). The SynSum measure (9) is defined as:
It measures the difference between the full information and the sum of individual (first-order) informations. One shortcoming of the above measure is that the second term becomes dominant as N grows [the first is always bounded by H(S)]. Thus, large populations will always appear redundant. Another possible measure compares the full information to the information in the case where neurons are conditionally independent (CI) given the stimulus
where (SynCI was denoted by ΔInoise in ref. 17 and by ΔI(1,2) in ref. 19). Note that this measure does not grow with N and will equal zero when the neurons are CI. Another related measure based on the CI case, but not directly using information, was introduced in ref. 16.
Both SynSum and SynCI compare the full information to that in first order statistics. Moreover, the typical implementation of these measures is for the 2-neuron case, where the only statistics less than full order are first order. The generalization of synergy/redundancy measures to higher-order statistics, and to N > 2 populations, poses an important challenge. The SynSum measure has been generalized to this scenario in ref. 20, where it was decomposed into elements measuring synergy in kth order correlations. MinMI offers an elegant approach for generalizing the SynCI measure to higher orders. At first sight, it seems like a reasonable approach is to take the difference between the informations of the MaxEnt distributions for orders k and k − 1. However, these 2 numbers will saturate as N → ∞ (see Relation to Maximum Entropy Modeling in SI Appendix), and thus this measure will be zero at the limit. MinMI offers a way around this problem, as we now illustrate for second-order statistics. The I(2) measure quantifies the information available in a population given only its (first- and) second-order statistics. To turn it into a synergy/redundancy measure, we need to subtract the second-order information in the CI model. If the neurons are CI, the pairwise statistics are expected to be p(ri,rj|s) = p(ri|s)p(rj|s). We denote the minimum information in these pairwise statistics by ICI(2). A natural measure of synergy is then the difference
When the true population is CI, we have SynI(2) = 0, as expected. Furthermore, when N = 2, we have that SynI(2) = SynCI. Thus MinMI generalizes SynCI to the study of pairwise interactions in large populations. Furthermore, the MinMI information does not saturate as MaxEnt does, and thus this measure is meaningful even as N → ∞. The SynI(2) measure may be extended to the kth order case by replacing ICI(2) with the minimum information subject to kth order statistics given by a MaxEnt model of order k − 1.
Results
Neural population codes may be studied at several levels, corresponding to different coding strategies. The basic level is the single neuron code. Next is the relation between the codes of different single neurons. Higher-order interactions between neurons constitute yet another level, along with the relation between multiple higher-order interactions. Finally, temporal structure may also be used to enhance coding efficiency. In the applications below, we show how the MinMI principle may be used to study various neural coding schemes and to quantify the level to which different populations use these schemes.
Two Binary Neurons and a Binary Stimulus.
We begin with an illustration of MinMI calculation for the case of 2 toy binary neurons R1, R2 where each neuron has 2 possible responses: 0 or 1. The stimulus S is also taken to be binary. We assume that only first-order statistics p(r1|s),p(r2|s),p(s) are known and that p(r1,r2|s) is unknown. We are interested in the minimum information I(1), i.e., the information available in a distribution (r1,r2,s) satisfying the first-order constraints (ri|s) = p(ri|s), i = 1,2. Note that any such distribution is completely defined by 2 numbers (r1 = 1,r2 = 1|s) (for s = 1,2), because for each S value (r1,r2|s) has 4 free parameters and has to satisfy 3 constraints (2 first-order constraints and one normalization constraint). The space of possible distributions (s,r) can thus be visualized in 2 dimensions, as in Fig. 1. The figure shows the value of the MI for each possible distribution in (s,r) satisfying the constraints above. This is done for two different pairs of neurons, with different first-order responses. The figure shows (in yellow circles) the location of the MinMI distribution MI(r1,r2|s). Also shown (in white squares) is the distribution under which the neurons are CI given the stimulus: (r1,r2|s) = (r1|s)(r2|s). By definition, this distribution has higher MI than I(1). In the first example (Fig. 1A–C) the 2 neurons have the same response distributions p(r1|s) = p(r2|s). The MinMI distribution shown in the figure is then the one in which the neurons are completely correlated and thus lies on the boundary of the space of possible distributions. It is intuitively clear why this is the minimum: the 2 neurons are, in the worst case, equivalent to a single neuron. In this case the CI information is higher because when the 2 neurons are CI, one can average over the noise to obtain more information about the stimulus.
In contrast, when the 2 neurons differ in their response distributions (Fig. 1D–F), they cannot be completely correlated. Thus, the information minimizing distribution will not lie on the boundary as in the previous example (compare Fig. 1C with 1F) but will still be lower than the CI information (compare circle with square in Fig. 1F).
Coding Redundancy in Single Neurons.
We next illustrate the use of MinMI in the study of single neuron codes and their combination in a population. As an example, consider a population of neurons where each neuron is tuned to some preferred direction of movement (PD) in the stimulus (e.g., the direction of hand movement in motor neurons, or stimulus motion in visual neurons). A population of neurons may have different distributions of such PDs. In one extreme, all neurons have the same PD, whereas in the other extreme PDs are uniformly distributed among neurons. It is intuitively clear that the second scenario is advantageous in terms of coding. However, it is not clear how to quantify this intuition in terms of information, especially when the joint distribution of the population cannot be estimated.
The MinMI principle provides a natural framework for tackling the above problem. Ideally, in studying information in populations, we are interested in the quantity I(R1,…,RN;S). More specifically, we are interested in the contribution of single neuron codes to this information. Our I(1) measure provides precisely that. To illustrate how I(1) differentiates between different single neurons coding schemes, we simulate data from 3 hypothetical neuronal populations, with different degrees of overlap between single neuron codes. Fig. 2 shows the code structure for these populations and the respective I(1) values. The results correspond to the intuition mentioned above: low I(1) values correspond to populations with high overlap between single neuron codes, and high values correspond to low overlap. Note that the MinMI calculation is model-free, and thus does not use the concept of directional tuning or preferred direction. It can thus detect differences in population coding in considerably more complex scenarios, which may be harder to visualize. Fig. 2 also compares I(1) to the information in a distribution where neurons are CI given the stimulus (i.e., the MaxEnt distribution subject to first-order statistics). We denote this information by ICI(1) (see Calculating the CI Information in SI Appendix for ICI(1) calculation method). The shortcoming of the CI assumption is that for large populations, the information will asymptotically reach its maximum value, because one can average over the responses of independent neurons and recover the exact stimulus as the population size grows. This behavior is apparent in the results in Fig. 2, where the ICI(1) measure approaches the maximum value of 3 bits for randomly distributed PDs and a similarly high value for the uniformly spaced PDs. Thus, unlike I(1), the measure ICI(1) does not differentiate between the different neuronal populations.
Pairwise Coding in Populations.
Second-order statistics between neurons have been shown to play a part in neural coding in the sense that their joint activity provides more information about a stimulus than their individual responses (7, 21). Most research has been devoted to studying pairs of neurons, mostly due to sample size limitations (but see refs. 13 and 22). It is intuitively clear, however, that if the second-order statistics between all pairs in a population provide information, but about the same property of the stimulus, this should result in less information than if different pairs encoded different stimulus properties. This situation is the pairwise equivalent of the single neuron coding issue discussed in the previous section.
The information available from grouped pairwise responses in a population can be quantified using the SynI(2) measure. Fig. 3 shows 2 toy populations, 4 neurons each, with identical pairwise synergy values: the set of SynSum(Ri,Rj;S) (i,j ∈{1,…,4}) values is identical in both populations. Furthermore, in this case SynSum = SynCI because I(Ri;S) = 0 for all neurons. In one population (Fig. 3B), all synergistic coding provides information about the same property of the stimulus, whereas in the other (Fig. 3C), the pairwise codes are designed to provide disparate information. The difference between these 2 populations is clearly seen in their SynI(2) values. Thus the MinMI principle can be used to differentiate between populations with different pairwise code designs. Although MaxEnt models (13) can also be applied to this case, they suffer from the same asymptotic behavior that we encountered for the ICI(1) case, and will not be able to discriminate between different populations for large N.
Temporal Coding.
Temporal response profiles of single neurons may transmit information about behaviorally relevant variables (23–25). Intuitively, one could argue that if different behavioral parameters induce different response profiles, as measured by a peristimulus time histogram (PSTH), then the temporal response carries information about the behavior. Our MinMI formalism allows us to make this statement explicit and to calculate the resulting information.
The temporal response function of a neuron can be given by its response in a series of time bins p(rt|s), t = 1…T. A PSTH is an example of such a profile where rt is a binary variable, and one plots the rate function p(rt = 1|s). The responses p(rt|s) are merely a set of first order statistics and we can calculate I(1) for these statistics, so as to obtain a measure of information in a PSTH.
Fig. 4 illustrates the application of MinMI to temporal coding in recordings from the primary motor cortex of behaving monkeys (see Experimental Procedures and Data Analysis in SI Appendix). We consider the response to a binary laterality signal (a visual stimulus), which instructs the monkey which hand to move. Fig. 4 shows a PSTH of a neuron, where the total spike count over a period of 600 ms after stimulus is similar for both conditions. However, the temporal profiles differ between the 2 conditions. To analyze this coding using I(1), we partitioned the 600 -ms period into time windows of 600,300,200,150, and 100 ms, and calculated p(rt|s) and the corresponding I(1) for each partition. We then shuffled the trials between laterality signals and compared the shuffled values with the raw I(1) in order to test whether the raw information was significantly different from zero. For the neuron in Fig. 4A, we found that it was not significant for window sizes of 300 ms and above but was significant for all lower-sized windows. This indicates that MinMI may be used to detect information related to temporal structure. We repeated the above procedure for the entire population of 827 neurons and counted the number of significant neurons for each window size. Fig. 4B shows this number as a function of window size. A large increase can be seen when moving from 600 to 200 ms, indicating relevant temporal structure at these time constants. The number then flattens for lower window sizes, suggesting that no information about the stimulus is added at these time scales.
Discussion
We have presented a framework for estimating MI in systems given partial measurements. Our MinMI principle has 2 attractive properties. The first is the ability to obtain a bound on information from a wide array of statistical observations and experimental scenarios. The second is the extension of standard information measures (such as information in single neurons or pairs of neurons) to large populations, allowing the detection of complex coding schemes not evident when analyzing neurons individually. Also, unlike previous decomposition methods (e.g., ref. 26) MinMI does not require knowledge of the complete joint distribution. These advantages improve on current IT-based methods in neuroscience and provide a comprehensive framework for tackling fundamental issues in neural coding.
The MinMI principle seeks a distribution p(s,r) that minimizes an information theoretic measure. This in fact is the mathematical structure of the central coding theorems in information theory (14), where information is either minimized (as in the rate distortion theorem) or maximized (as in the channel capacity theorem) under some constraints on the joint distribution p(s,r). Our approach is most closely related to rate distortion theory (RDT), which sets the achievable limits on “lossy” compression, i.e. compression which results in some distortion of the original signal. It turns out (15) that the optimal compression bound is obtained by minimizing MI between the input and output alphabets subject to a fixed prior on the inputs, and a constraint on the allowed input–output distortion. In MinMI, we also fix the input prior p(s) but introduce additional constraints on p(r|s) via its expected values. This can be understood as searching for a distribution p(r|s) as in RDT but with the single distortion constraint replaced by multiple constraints on expected values of several functions (interestingly, such a setting was mentioned in ref. 27 as a possible extension of RDT).
Another application of information minimization is in cases where one seeks to transform a signal X1,…,Xn into a set of independent variables Y1,…,Ym [e.g., the ICA method of ref. 28 or the method of spatially incoherent features (29)]. In this case information minimization originates from an assumption that the data was generated via a transformation on independent variables, and the reverse transformation is sought. It is thus very different from the MinMI scenario where no such assumptions are made, and the focus is on measurement of information rather than transformation on variables.
MI can be interpreted as a measure of the predictive power between 2 variables. A different measure is the optimal Bayes error e*, defined as the minimum probability of error incurred in predicting S from R. Interestingly, the MI can be used to obtain an upper bound on e* given by e* ≤ 0.5(Hp(S) − Ip(R;S)), where Hp(S) is the entropy of the stimulus (30). This implies that in the MinMI case, although the distribution p(s,r) is not known, we are guaranteed that e* ≤ 0.5(Hp(S) − Imin). MinMI thus yields an upper bound on e* of the true underlying distribution. The MinMI distribution MI(s|r) may also be used directly to predict S from R, and in fact it can be shown that the resulting error is bounded from above by Hp(S) − Imin (See Section 6 in ref. 31).
A common approach to calculating information in complex responses is to apply some quantization to R via a function f(r) (e.g., the spike count in a spike train r), such that the quantized variable f(R) has relatively few values and p(s,f(r)) may be estimated from small samples. The data processing inequality (32) then states that Ip(S;f(R)) ≤ Ip(S;R), and thus the quantized information always provides a lower bound on the true information. It can be shown (see Relation to the Data Processing Inequality in SI Appendix) that Ip(S;f(R)) is in fact the outcome of the following MinMI problem: what is the minimum information in a distribution p(s,r) whose quantized version is p(s,f(r)). Thus, MinMI may be viewed as generalizing the data processing inequality.
Because the MinMI bound represents a worst case scenario it may considerably underestimate the true information value in some cases. Also, the brain has inherent constraints such as energy consumption and response latency, that might make it impossible for it to employ the coding strategy obtained by MinMI. However, the MinMI bound can be increased as more measurements are added, and additional internal constraints are considered. Generally, as additional reliable measurements are made available, the MinMI bound may be gradually refined to better approximate the true information. Additionally, even in cases where the bound considerably underestimates the true information, it still serves as a quantifier of the information in the given measurement, and as such can be used to compare coding schemes (see Results).
In presenting the method, we made the assumption that partial measurements are exact. Because these measurements are commonly estimated from finite samples, their exact values are usually not known, but rather lie in some range of values (with high probability) which can be determined from the size of the sample (33). The expectation constraints (Eq. 3) can then be constrained to be in this range. The solution in this case still has the general form of Eq. 5, and corresponding algorithms may be derived.
Although the results presented here were applied to neural coding, the MinMI principle is general and may be used for studying a wide array of complex systems. For instance, it may be used to estimate the information in a set of gene expression profiles about external conditions (34) and thus help in analyzing their functional role and in comparing different gene regulatory networks.
Methods
This section describes algorithms for calculating the MinMI bound. The constrained optimization problem in Eq. 4 is convex, because its objective is convex (14), and the constraints are linear in (s,r). It thus has no local minima and can be solved by using convex optimization machinery (35). Here, we present 2 specialized iterative algorithms to solve it. The first algorithm is exact, but has complexity O(nr). Because nr may sometimes be large, we also present a second, approximate, algorithm to handle such cases.
The basic building block of our first, exact, algorithm is the I-projection (36). The I-projection of a distribution q(r) on a set of distributions F is defined as the distribution p* ∈ F, which minimizes the Kullback-Leibler (KL) divergence to the distribution q(r): p* = arg minp ∈ FDKL[p|q], where DKL[p|q] is defined as . The I-projection has a particularly simple form when F is determined by expectation constraints
The I-projection is then given by
where * are a set of Lagrange multipliers, chosen to fit the desired expected values, and γ* is a normalization factor. The values of * can be found by using several optimization techniques. All involve the computation of the expected value of (r) under distributions of the form q(r)e(r)·. Here, we use an L-BFGS-based algorithm (37).
The structural similarity between the form of Eq. 11 and the characterization of MI(r|s) in Eq. 5 suggests that MI(r|s) is an I-projection of MI(r) on the set F((r),(s)). The fact that MI(r) depends on MI(r|s) through marginalization suggests that the minimization problem may be solved using an iterative algorithm where marginalization and projection are performed at each step. Each iteration thus consists of the following steps (31):
For all s, set t+1(r|s) to be the I-projection of t(r) on F((r),(s)).
Set t+1(r) = ∑st+1(r|s)p(s).
The above procedure can be shown to converge to the minimum information (see Algorithm Convergence Proof in SI Appendix).
The exact algorithm presented above is feasible when the size of the input space nr is small enough to allow O(nr) memory and computational resources. For systems containing many elements, this is often not the case. For instance, when R is a response of 100 binary neurons, nr = 2100. To derive an approximate algorithm for the large system case, we first note that after t iterations of the exact iterative algorithm, the distribution pt(r) is a mixture of the form
where every iteration increases the number of components by a factor of ns. For the approximate algorithm, we limit the number of elements in this mixture to some constant K by clustering its components after each iteration using a K-means algorithm with K centroids (see chapter 10 in ref. 38). The resulting mixture is represented using its mixing probabilities ck and parameters k (resulting in O(K) parameters). We denote the resulting approximate distribution by ′t(r). The algorithm then proceeds as in the exact method, only with ′t(r) instead of the exact t(r).
For the I(1) case, the kth element in the mixture has the form . Recall that in order to perform the I-projection one needs to calculate expected values for distributions of the form
Because of the factorized form of each element in the sum, the first-order marginals are straightforward to calculate and so is the I-projection of ′t(r) on the relevant constraints. For the higher-order cases (e.g., I(2)) the mixture marginals do not have a closed-form solution, and require approximation methods such as Gibbs sampling. For the applications presented here, we used the approximate algorithm only for the I(1) case. We have found empirically that the above approximation scheme works well for cases where we could compare it with the exact algorithm (up to nr = 50,000). In the applications reported here, we used the exact algorithm for nr ≤ 50,000.
Supplementary Material
Acknowledgments.
We thank Rony Paz for sharing data, and Robert Gütig and Olivia White for reading the manuscript. This work was supported in part by the US–Israel Binational Science Foundation, by the Israel Science Foundation Centers of Excellence Grant, and by special contributions from the Rosetrees Trust and the Ida Baruch Fund. E.V. is the Jack H. Skirball Chair of Brain Research. A.G. was supported by the Eshkol foundation and by a fellowship from the Rothschild Foundation–Yad Hanadiv.
Footnotes
The authors declare no conflict of interest.
This article is a PNAS Direct Submission.
This article contains supporting information online at www.pnas.org/cgi/content/full/0806782106/DCSupplemental.
References
- 1.Shannon C. A mathematical theory of communication. Bell Syst Tech J. 1948;27:379–423. 632–656. [Google Scholar]
- 2.Attneave F. Some informational aspects of visual perception. Psych Rev. 1954;61:183–193. doi: 10.1037/h0054663. [DOI] [PubMed] [Google Scholar]
- 3.Miller G. The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psych Rev. 1956;63:81–97. [PubMed] [Google Scholar]
- 4.Barlow H. The coding of sensory messages. In: Thorpe W, Zangwill OL, editors. Current Problems in Animal Behaviour. Cambridge, UK: Cambridge Univ Press; 1960. pp. 331–360. [Google Scholar]
- 5.Linsker R. Self-organization in a perceptual network. IEEE Comput. 1988;21:105–117. [Google Scholar]
- 6.Rieke F, Warland D, de Ruyter van Steveninck R, Bialek W. Spikes. Cambridge, MA: MIT Press; 1997. [Google Scholar]
- 7.Hatsopoulos N, Ojakangas C, Paninski L, Donoghue J. Information about movement direction obtained from synchronous activity of motor cortical neurons. Proc Natl Acad Sci USA. 1998;95:15706–15711. doi: 10.1073/pnas.95.26.15706. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Bialek W, Rieke F, de Ruyter van Steveninck R, Warland D. Reading a neural code. Science. 1991;252:1854–1857. doi: 10.1126/science.2063199. [DOI] [PubMed] [Google Scholar]
- 9.Gawne T, Richmond B. How independent are the messages carried by adjacent inferior temporal cortical neurons? J Neurosci. 1993;13:2758–2771. doi: 10.1523/JNEUROSCI.13-07-02758.1993. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Dan Y, Alonso J, Usrey W, Reid R. Coding of visual information by precisely correlated spikes in the lateral geniculate nucleus. Nat Neurosci. 1998;1:501–507. doi: 10.1038/2217. [DOI] [PubMed] [Google Scholar]
- 11.Jaynes E. Information theory and statistical mechanics. Phys Rev. 1957;106:620–630. [Google Scholar]
- 12.Martignon L, et al. Neural coding: Higher-order temporal patterns in the neurostatistics of cell assemblies. Neural Comput. 2000;12:2621–2653. doi: 10.1162/089976600300014872. [DOI] [PubMed] [Google Scholar]
- 13.Schneidman E, Berry M, Segev R, Bialek W. Weak pairwise correlations imply strongly correlated network states in a neural population. Nature. 2006;440:1007–1012. doi: 10.1038/nature04701. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Cover T, Thomas J. Elements of Information Theory. New York: Wiley Interscience; 1991. [Google Scholar]
- 15.Shannon C. Coding theorems for a discrete source with a fidelity criterion. IRE Natl Conv Rec. 1959;4:142–163. [Google Scholar]
- 16.Nirenberg S, Latham P. Decoding neuronal spike trains: how important are correlations? Proc Natl Acad Sci USA. 2003;100:7348–7353. doi: 10.1073/pnas.1131895100. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Schneidman E, Bialek W, Berry M. Synergy, redundancy, and independence in population codes. J Neurosci. 2003;23:11539–11553. doi: 10.1523/JNEUROSCI.23-37-11539.2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Tononi G, Sporns O, Edelman G. Measures of degeneracy and redundancy in biological networks. Proc Natl Acad Sci USA. 1999;96:3257–3262. doi: 10.1073/pnas.96.6.3257. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Stark E, Globerson A, Asher I, Abeles M. Correlations between groups of premotor neurons carry information about prehension. J Neurosci. 2008;28:10618–10630. doi: 10.1523/JNEUROSCI.3418-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Schneidman E, Still S, Berry M, Bialek W. Network information and connected correlations. Phys Rev Lett. 2003;91:238701. doi: 10.1103/PhysRevLett.91.238701. [DOI] [PubMed] [Google Scholar]
- 21.Vaadia E, et al. Dynamics of neuronal interactions in monkey cortex in relation to behavioral events. Nature. 1995;373:515–518. doi: 10.1038/373515a0. [DOI] [PubMed] [Google Scholar]
- 22.Narayanan N, Kimchi E, Laubach M. Redundancy and synergy of neuronal ensembles in motor cortex. J Neurosci. 2005;25:4207–4216. doi: 10.1523/JNEUROSCI.4697-04.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Optican L, Richmond B. Temporal encoding of two-dimensional patterns by single units in primate inferior temporal cortex III. Information theoretic analysis. J Neurophysiol. 1987;57:162–177. doi: 10.1152/jn.1987.57.1.162. [DOI] [PubMed] [Google Scholar]
- 24.Osborne L, Bialek W, Lisberger S. Time course of information about motion direction in visual area MT of macaque monkeys. J Neurosci. 2004;24:3210–3222. doi: 10.1523/JNEUROSCI.5305-03.2004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Victor J, Purpura K. Nature and precision of temporal coding in visual cortex: A metric-space analysis. J Neurophysiol. 1996;76:1310–1326. doi: 10.1152/jn.1996.76.2.1310. [DOI] [PubMed] [Google Scholar]
- 26.Pola G, Thiele A, Hoffmann K, Panzeri S. An exact method to quantify the information transmitted by different mechanisms of correlational coding. Netw Comput Neural Syst. 2003;14:35–60. doi: 10.1088/0954-898x/14/1/303. [DOI] [PubMed] [Google Scholar]
- 27.Goblick T. Coding for a Discrete Information Source with a Distortion Measure. Cambridge, MA: Massachusetts Institute of Technology; 1962. PhD thesis. [Google Scholar]
- 28.Bell A, Sejnowski T. An information maximisation approach to blind separation and blind deconvolution. Neural Comput. 1995;7:1129–1159. doi: 10.1162/neco.1995.7.6.1129. [DOI] [PubMed] [Google Scholar]
- 29.Haykin S. Neural Networks: A Comprehensive Foundation. Upper Saddle River, NJ: Prentice–Hall; 2007. [Google Scholar]
- 30.Hellman M, Raviv J. Probability of error, equivocation, and the Chernoff bound. IEEE Trans Inf Theory. 1970;16:368–372. [Google Scholar]
- 31.Globerson A, Tishby N. The minimum information principle in discriminative learning. In: Chickering M, Halpern J, editors. Proceedings of the UAI. Assoc for Uncertainty in Artificial Intelligence; 2004. pp. 193–200. [Google Scholar]
- 32.Borst A, Theunissen F. Information theory and neural coding. Nat Neurosci. 1999;2:947–957. doi: 10.1038/14731. [DOI] [PubMed] [Google Scholar]
- 33.Dudík M, Phillips S, Schapire RE. Performance guarantees for regularized maximum entropy density estimation. In: Shawe-Taylor J, Singer Y, editors. Proceedings of COLT. New York: Springer; 2004. pp. 472–486. [Google Scholar]
- 34.Gasch A, et al. Genomic expression programs in the response of yeast cells to environmental changes. Mol Biol Cell. 2000;11:4241–4257. doi: 10.1091/mbc.11.12.4241. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Boyd S, Vandenberghe L. Convex Optimization. Cambridge, UK: Cambridge Univ Press; 2004. [Google Scholar]
- 36.Csiszar I. I-divergence geometry of probability distributions and minimization problems. Ann Prob. 1975;3:146–158. [Google Scholar]
- 37.Sha F, Pereira F. North American Chapter of the Association for Computational Linguistics—Human Language Technologies (NAACL HLT) Boulder, CO: Assoc for Comput Linguistics; 2003. Shallow parsing with conditional random fields; pp. 134–141. [Google Scholar]
- 38.Duda RO, Hart PE, Stork DG. Pattern Classification. New York: Wiley Interscience; 2000. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.