Skip to main content
Frontiers in Computational Neuroscience logoLink to Frontiers in Computational Neuroscience
. 2015 Aug 28;9:105. doi: 10.3389/fncom.2015.00105

Time resolution dependence of information measures for spiking neurons: scaling and universality

Sarah E Marzen 1, Michael R DeWeese 1,2, James P Crutchfield 3,*
PMCID: PMC4551861  PMID: 26379538

Abstract

The mutual information between stimulus and spike-train response is commonly used to monitor neural coding efficiency, but neuronal computation broadly conceived requires more refined and targeted information measures of input-output joint processes. A first step toward that larger goal is to develop information measures for individual output processes, including information generation (entropy rate), stored information (statistical complexity), predictable information (excess entropy), and active information accumulation (bound information rate). We calculate these for spike trains generated by a variety of noise-driven integrate-and-fire neurons as a function of time resolution and for alternating renewal processes. We show that their time-resolution dependence reveals coarse-grained structural properties of interspike interval statistics; e.g., τ-entropy rates that diverge less quickly than the firing rate indicated by interspike interval correlations. We also find evidence that the excess entropy and regularized statistical complexity of different types of integrate-and-fire neurons are universal in the continuous-time limit in the sense that they do not depend on mechanism details. This suggests a surprising simplicity in the spike trains generated by these model neurons. Interestingly, neurons with gamma-distributed ISIs and neurons whose spike trains are alternating renewal processes do not fall into the same universality class. These results lead to two conclusions. First, the dependence of information measures on time resolution reveals mechanistic details about spike train generation. Second, information measures can be used as model selection tools for analyzing spike train processes.

Keywords: statistical complexity, excess entropy, entropy rate, renewal process, alternating renewal process, integrate and fire neuron, leaky integrate and fire neuron, quadratic integrate and fire neuron

PACS numbers: 05.45.Tp 02.50.Ey 87.10.Vg 87.19.ll 87.19.lo 87.19.ls

1. Introduction

Despite a half century of concerted effort (Mackay and McCulloch, 1952), neuroscientists continue to debate the relevant timescales of neuronal communication as well as the basic coding schemes at work in the cortex, even in early sensory processing regions of the brain thought to be dominated by feedforward pathways (Softky and Koch, 1993; Bell et al., 1995; Shadlen and Newsome, 1995; Stevens and Zador, 1998; Destexhe et al., 2003; DeWeese and Zador, 2006; Jacobs et al., 2009; Koepsell et al., 2010; London et al., 2010). For example, the apparent variability of neural responses to repeated presentations of sensory stimuli has led many to conclude that the brain must average across tens or hundreds of milliseconds or across large populations of neurons to extract a meaningful signal (Shadlen and Newsome, 1998). Whereas, reports of reliable responses suggest shorter relevant timescales and more nuanced coding schemes (Berry et al., 1997; Reinagel and Reid, 2000; DeWeese et al., 2003). In fact, there is evidence for different characteristic timescales for neural coding in different primary sensory regions of the cortex (Yang and Zador, 2012). In addition to questions about the relevant timescales of neural communication, there has been an ongoing debate regarding the magnitude and importance of correlations among the spiking responses of neural populations (Meister et al., 1995; Nirenberg et al., 2001; Averbeck et al., 2006; Schneidman et al., 2003, 2006).

Most studies of neural coding focus on the relationship between a sensory stimulus and the neural response. Others consider the relationship between the neural response and the animal's behavioral response (Britten et al., 1996), the relationship between pairs or groups of neurons at different stages of processing (Linsker, 1989; Dan et al., 1996), or the variability of neural responses themselves without regard to other variables (Schneidman et al., 2006). Complementing the latter studies, we are interested in quantifying the randomness and predictability of neural responses without reference to stimulus. We consider the variability of a given neuron's activity at one time and how this is related to the same neuron's activity at other times in the future and the past.

Along these lines, information theory (Shannon, 1948; Cover and Thomas, 2006) provides an insightful and rich toolset for interpreting neural data and for formulating theories of communication and computation in the nervous system (Rieke et al., 1999). In particular, Shannon's mutual information has developed into a powerful probe that quantifies the amount of information about a sensory stimulus encoded by neural activity (Mackay and McCulloch, 1952; Barlow, 1961; Stein, 1967; Laughlin, 1981; Sakitt and Barlow, 1982; Srinivasan et al., 1982; Linsker, 1989; Bialek et al., 1991; Theunissen and Miller, 1991; Atick, 1992; Rieke et al., 1999). Similarly, the Shannon entropy has been used to quantify the variability of the resulting spike-train response. In contrast to these standard stimulus- and response-averaged quantities, a host of other information-theoretic measures have been applied in neuroscience, such as the Fisher information (Cover and Thomas, 2006) and various measures of the information gained per observation (DeWeese and Meister, 1999; Butts and Goldman, 2006).

We take an approach that complements more familiar informational analyses. First, we consider “output-only” processes, since their analysis is a theoretical prerequisite to understanding information in the stimulus-response paradigm. Second, we analyze rates of informational divergence, not only nondivergent components. Indeed, we show that divergences, rather than being a kind of mathematical failure, are important and revealing features of information processing in spike trains.

We are particularly interested in the information content of neural spiking on fine timescales. How is information encoded in spike timing and, more specifically, in interspike intervals? In this regime, the critical questions turn on determining the kind of information encoded and the required “accuracy” of individual spike timing to support it. At present, unfortunately, characterizing communication at submillisecond time scales and below remains computationally and theoretically challenging.

Practically, a spike train is converted into a binary sequence for analysis by choosing a time bin size and counting the number of spikes in successive time bins. Notwithstanding Strong et al. (1998) and Nemenman et al. (2008), there are few studies of how estimates of communication properties change as a function of time bin size, though there are examples of both short (Panzeri et al., 1999) and long (DeWeese, 1996; Strong et al., 1998) time expansions. Said most plainly, it is difficult to directly calculate the most basic quantities—e.g., communication rates between stimulus and spike-train response—in the submillisecond regime, despite progress on undersampling (Treves and Panzeri, 1995; Nemenman et al., 2004; Archer et al., 2012). Beyond the practical, the challenges are also conceptual. For example, given that a stochastic process' entropy rate diverges in a process-characteristic fashion for small time discretizations (Gaspard and Wang, 1993), measures of communication efficacy require careful interpretation in this limit.

Compounding the need for better theoretical tools, measurement techniques will soon amass enough data to allow serious study of neuronal communication at fine time resolutions and across large populations (Alivisatos et al., 2012). In this happy circumstance, we will need guideposts for how information measures of neuronal communication vary with time resolution so that we can properly interpret the empirical findings and refine the design of nanoscale probes.

Many single-neuron models generate neural spike trains that are renewal processes (Gerstner and Kistler, 2002). Starting from this observation, we use recent results (Marzen and Crutchfield, 2015) to determine how information measures scale in the small time-resolution limit. This is exactly the regime where numerical methods are most likely to fail due to undersampling and, thus, where analytic formulae are most useful. We also extend the previous analyses to structurally more complex, alternating renewal processes and analyze the time-resolution scaling of their information measures. This yields important clues as to which scaling results apply more generally. We then show that, across several standard neuronal models, the information measures are universal in the sense that their scaling does not depend on the details of spike-generation mechanisms.

Several information measures we consider are already common fixtures in theoretical neuroscience, such as Shannon's source entropy rate (Strong et al., 1998; Nemenman et al., 2008). Others have appeared at least once, such as the finite-time excess entropy (or predictable information) (Bialek et al., 2001; Crutchfield and Feldman, 2003) and statistical complexity (Haslinger et al., 2010). And others have not yet been applied, such as the bound information (Abdallah and Plumbley, 2009, 2012; James et al., 2011, 2014).

The development proceeds as follows. Section 2 reviews notation and definitions. To investigate the dependence of causal information measures on time resolution, Section 3 studies a class of renewal processes motivated by their wide use in describing neuronal behavior. Section 4 then explores the time-resolution scaling of information measures of alternating renewal processes, identifying those scalings likely to hold generally. Section 5 evaluates continuous-time limits of these information measures for common single-neuron models. This reveals a new kind of universality in which the information measures' scaling is independent of detailed spiking mechanisms. Taken altogether, the analyses provide intuition and motivation for several of the rarely-used, but key informational quantities. For example, the informational signatures of integrate-and-fire model neurons differ from both simpler, gamma-distributed processes and more complex, compound renewal processes. Finally, Section 6 summarizes the results, giving a view to future directions and mathematical and empirical challenges.

2. Background

We can only briefly review the relevant physics of information. Much of the phrasing is taken directly from background presented in Marzen and Crutchfield (2014, 2015).

Let us first recall the causal state definitions (Shalizi and Crutchfield, 2001) and information measures of discrete-time, discrete-state processes introduced in Crutchfield et al. (2009), James et al. (2011). The main object of study is a process P: the list of all of a system's behaviors or realizations {…x−2, x−1, x0, x1, …} and their probabilities, specified by the joint distribution Pr(…X−2, X−1, X0, X1, …). We denote a contiguous chain of random variables as X0:L = X0X1XL−1. We assume the process is ergodic and stationary—Pr(X0:L) = Pr(Xt:L+t) for all t ∈ ℤ —and the measurement symbols range over a finite alphabet: xA. In this setting, the present X0 is the random variable measured at t = 0, the past is the chain X:0 = …X−2X−1 leading up the present, and the future is the chain following the present X1: = X1X2⋯ (We suppress the infinite index in these).

As the Introduction noted, many information-theoretic studies of neural spike trains concern input-output information measures that characterize stimulus-response properties; e.g., the mutual information between stimulus and resulting spike train. In the absence of stimulus or even with a non-trivial stimulus, we can still study neural activity from an information-theoretic point of view using “output-only” information measures that quantify intrinsic properties of neural activity alone:

  • How random is it? The entropy rate hμ = H[X0|X:0], which is the entropy in the present observation conditioned on all past observations (Cover and Thomas, 2006).

  • What must be remembered about the past to optimally predict the future? The causal states S+, which are groupings of pasts that lead to the same probability distribution over future trajectories (Crutchfield and Young, 1989; Shalizi and Crutchfield, 2001).

  • How much memory is required to store the causal states? The statistical complexity Cμ=H[S+], or the entropy of the causal states (Crutchfield and Young, 1989).

  • How much of the future is predictable from the past? The excess entropy E = I[X:0; X0:], which is the mutual information between the past and the future (Crutchfield and Feldman, 2003).

  • How much of the generated information (hμ) is relevant to predicting the future? The bound information bμ = I[X0; X1:|X:0], which is the mutual information between the present and future observations conditioned on all past observations (Abdallah and Plumbley, 2009; James et al., 2011).

  • How much of the generated information is useless—neither affects future behavior nor contains information about the past? The ephemeral information rμ = H[X0|X:0, X1:], which is the entropy in the present observation conditioned on all past and future observations (Verdú and Weissman, 2006; James et al., 2011).

The information diagram of Figure 1 illustrates the relationship between hμ, rμ, bμ, and E. When we change the time discretization Δt, our interpretation and definitions change somewhat, as we describe in Section 3.

Figure 1.

Figure 1

Information diagram illustrating the anatomy of the information H[X0] in a process' single observation X0 in the context of its past X:0 and its future X1:. Although the past entropy H[X:0] and the future entropy H[X1:] typically are infinite, space precludes depicting them as such. They do scale in a controlled way, however: H[X − ℓ:0] ∝ hμℓ and H[X1:ℓ] ∝ hμℓ. The two atoms labeled bμ are the same, since we consider only stationary processes. (After James et al., 2011, with permission.)

Shannon's various information quantities—entropy, conditional entropy, mutual information, and the like—when applied to time series are functions of the joint distributions Pr(X0:L). Importantly, for a given set of random variables they define an algebra of atoms out of which information measures are composed (Yeung, 2008). James et al. (2011) used this to show that the past and future partition the single-measurement entropy H(X0) into the measure-theoretic atoms of Figure 1. These include those—rμ and bμ—already mentioned and the enigmatic information:

qμ=I[X0;X:0;X1:] ,

which is the co-information between past, present, and future. One can also consider the amount of predictable information not captured by the present:

σμ=I[X:0;X1:|X0].

which is the elusive information (Ara et al., 2015). It measures the amount of past-future correlation not contained in the present. It is nonzero if the process has “hidden states” and is therefore quite sensitive to how the state space is “observed” or coarse-grained.

The total information in the future predictable from the past (or vice versa)—the excess entropy—decomposes into particular atoms:

E=bμ+ σμ+ qμ.

The process's Shannon entropy rate hμ is also a sum of atoms:

hμ=rμ+bμ .

This tells us that a portion of the information (hμ) a process spontaneously generates is thrown away (rμ) and a portion is actively stored (bμ). Putting these observations together gives the information anatomy of a single measurement X0:

H[X0]=qμ+2bμ+rμ . (1)

Although these measures were originally defined for stationary processes, they easily carry over to a nonstationary process of finite Markov order.

Calculating these information measures in closed-form given a model requires finding the ϵ-machine, which is constructed from causal states. Forward-time causal states S+ are minimal sufficient statistics for predicting a process's future (Crutchfield and Young, 1989; Shalizi and Crutchfield, 2001). This follows from their definition—a causal state σ+S+ is a sets of pasts grouped by the equivalence relation ~+:

x:0~+x:0           Pr(X0:|X:0=x:0)=Pr(X0:|X:0=x:0) . (2)

So, S+ is a set of classes—a coarse-graining of the uncountably infinite set of all pasts. At time t, we have the random variable St+ that takes values σ+S+ and describes the causal-state process ,S-1+,S0+,S1+,. St+ is a partition of pasts X:t that, according to the indexing convention, does not include the present observation Xt. In addition to the set of pasts leading to it, a causal state σt+ has an associated future morph—the conditional measure Pr(Xt:|σt+) of futures that can be generated from it. Moreover, each state σt+ inherits a probability π(σt+) from the process's measure over pasts Pr(X:t). The forward-time statistical complexity is then the Shannon entropy of the state distribution π(σt+) (Crutchfield and Young, 1989): Cμ+=H[S0+]. A generative model is constructed out of the causal states by endowing the causal-state process with transitions:

Tσσ(x)=Pr(St+1+=σ,Xt=x|St+=σ) ,

that give the probability of generating the next symbol x and ending in the next state σ′, if starting in state σ (Residing in a state and generating a symbol do not occur simultaneously. Since symbols are generated during transitions there is, in effect, a half time-step difference in the indexes of the random variables Xt and St+. We suppress notating this.) To summarize, a process's forward-time is the tuple {A, S+, {T(x):xA}}.

For a discrete-time, discrete-alphabet process, the ϵ-machine is its minimal unifilar hidden Markov model (HMM) (Crutchfield and Young, 1989; Shalizi and Crutchfield, 2001) (For general background on HMMs see Paz, 1971; Rabiner and Juang, 1986; Rabiner, 1989). Note that the causal state set can be finite, countable, or uncountable; the latter two cases can occur even for processes generated by finite-state HMMs. Minimality can be defined by either the smallest number of states or the smallest entropy H[S0+] over states (Shalizi and Crutchfield, 2001). Unifilarity is a constraint on the transition matrices T(x) such that the next state σ′ is determined by knowing the current state σ and the next symbol x. That is, if the transition exists, then Pr(St+1+|Xt=x,St+=σ) has support on a single causal state.

3. Infinitesimal time resolution

One often treats a continuous-time renewal process, such as a spike train from a noisy integrate-and-fire neuron, in a discrete-time setting (Rieke et al., 1999). With results of Marzen and Crutchfield (2015) in hand, we can investigate how artificial time binning affects estimates of a model neuron's spike train's randomness, predictability, and information storage in the limit of infinitesimal time resolution. This is exactly the limit in which analytic formulae for information measures are most useful, since increasing the time resolution artificially increases the apparent range of temporal correlations as shown in Figure 3.

Time-binned neural spike trains of noisy integrate-and-fire neurons have been studied for quite some time (Mackay and McCulloch, 1952) and, despite that history, this is still an active endeavor (Rieke et al., 1999; Cessac and Cofre, 2013). Our emphasis and approach differ, though. We do not estimate statistics or reconstruct models from simulated spike train data using nonparametric inference algorithms—e.g., as done in Haslinger et al. (2010). Rather, we ask how ϵ-machine extracted from a spike train process and information measures calculated from them vary as a function of time coarse-graining. Our analytic approach highlights an important lesson about such studies in general: A process' ϵ-machine and information anatomy are sensitive to time resolution. A secondary and compensating lesson is that the manner in which the ϵ-machine and information anatomy scale with time resolution conveys much about the process' structure.

Suppose we are given a neural spike train with interspike intervals independently drawn from the same interspike interval (ISI) distribution ϕ(t) with mean ISI 1/μ. To convert the continuous-time point process into a sequence of binary spike-quiescence symbols, we track the number of spikes emitted in successive time bins of size Δt. Our goal, however, is to understand how the choice of Δt affects reported estimates for Cμ, hμ, E, bμ, and σμ. The way in which each of these vary with Δt reveals information about the intrinsic time scales on which a process behaves; cf., the descriptions of entropy rates in Costa et al. (2002, 2005) and Gaspard and Wang (1993). We concern ourselves with the infinitesimal Δt limit, even though the behavior of these information atoms is potentially most interesting when Δt is on the order of the process' intrinsic time scales.

In the infinitesimal time-resolution limit, when Δt is smaller than any intrinsic timescale, the neural spike train is a renewal process with interevent count distribution:

F(n)ϕ(nΔt) Δt (3)

and survival function:

w(n)nΔtϕ(t)dt . (4)

The interevent distribution F(n) is the probability distribution that the silence separating successive events (bins with spikes) is n counts long. While the survival function w(n) is the probability that the silence separating successive events is at least n counts long. The ϵ-machine transition probabilities therefore change with Δt. The mean interevent count 〈T〉 + 1 is not the mean interspike interval 1/μ since one must convert between counts and spikes1:

T+1=1μΔt . (5)

In this limit, the ϵ-machine of spike-train renewal processes can take one of the topologies described in Marzen and Crutchfield (2015).

Here, we focus only on two of these ϵ-machine topologies. The first topology corresponds to that of an eventually Poisson process, in which the ISI distribution takes the form ϕ(t) = ϕ(T)e−λ(tT) for some finite T and λ > 0. A Poisson neuron with firing rate λ and refractory period of time T, for instance, eventually (t > T) generates a Poisson process. Hence, we refer to them as eventually Poisson processes; see Figure 2B. A Poisson process is a special type of eventually Poisson process with T = 0; see Figure 2A. However, the generic renewal process has topology shown in Figure 2C. Technically, only non-eventually-Δ Poisson processes have this ϵ-machine topology, but for our purposes, this is the ϵ-machine topology for any renewal process not generated by a Poisson neuron; see Marzen and Crutchfield (2015).

Figure 2.

Figure 2

ϵ-Machines of processes generated by Poisson neurons and by integrate-and-fire neurons (left to right): (A) The ϵ-machine for a Poisson process. (B) The ϵ-machine for an eventually Poisson process; i.e., a Poisson neuron with a refractory period of length ñΔt. (C) The ϵ-machine for a generic renewal process—the not eventually Δ-Poisson process of Marzen and Crutchfield (2015); i.e., the process generated by noise-driven integrate-and-fire neurons. Edge labels p|x denote emitting symbol x (“1” is “spike”) with probability p. (Reprinted with permission from Marzen and Crutchfield, 2015.)

At present, inference algorithms can only infer finite ϵ-machines. So, such algorithms applied to renewal processes will yield an eventually Poisson topology. (Compare Figure 2C to the inferred approximate of an integrate-and-fire neuron in Figure 2 in Haslinger et al., 2010.) The generic renewal process has an infinite ϵ-machines, though, for which the inferred ϵ-machines are only approximations.

We calculated E and Cμ using the expressions given in Marzen and Crutchfield (2015). Substituting in Equations (3), (4), and (5), we find that the excess entropy E tends to:

limΔt0E(Δt)=0μtϕ(t)log2(μϕ(t))dt                20μΦ(t)log2(μΦ(t))dt , (6)

where Φ(t)=tϕ(t)dt is the probability that an ISI is longer than t. It is easy to see that Et) limits to a positive and (usually) finite value as the time resolution vanishes, with some exceptions described below. Similarly, using the expression in Marzen and Crutchfield (2015)'s Appendix II, one can show that the finite-time excess entropy2 E(T) takes the form:

limΔt0E(T)=(0TμΦ(t)dt)log21μ                20TμΦ(t)log2Φ(t)dt                μTΦ(t)dtlog2(μTΦ(t)dt)           +0TμtF(t)log2F(t)dt           +TTμF(t)log2F(t)dt (7)

As T → ∞, E(T) → E. Note that these formulae apply only when mean firing rate μ is nonzero.

Even if E limits to a finite value, the statistical complexity typically diverges due to its dependence on time discretization Δt. Suppose that we observe an eventually Poisson process, such that ϕ(t) = ϕ(T)e−λ(tT) for t > T. Then, from formulae in Marzen and Crutchfield (2015), statistical complexity in the infinitesimal time-resolution limit becomes:

Cμ(Δt)~(μ0TΦ(t)dt)log21Δt                  0T(μΦ(t))log2(μΦ(t))dt                  (μTΦ(t)dt)log2(μTΦ(t)dt), (8)

ignoring terms of Ot) or higher. The first term diverges, and its rate of divergence is the probability of observing a time since last spike less than T. This measures the spike train's deviation from being Δ-Poisson and so reveals the effective dimension of the underlying causal state space. Cμ's remaining nondivergent component is equally interesting. In fact, it is the differential entropy of the time since last spike distribution.

An immediate consequence of the analysis is that this generic infinitesimal renewal process is highly cryptic (Crutchfield et al., 2009). It hides an arbitrarily large amount of its internal state information: Cμ diverges as Δt → 0 but E (usually) asymptotes to a finite value. We have very structured processes that have disproportionately little in the future to predict. Periodic processes constitute an important exception to this general rule of thumb for continuous-time processes. A neuron that fires every T seconds without jitter has E = Cμ, and both E and Cμ diverge logarithmically with 1/Δt.

It is straightforward to show that any information measure contained within the present—H[X0], hμ, bμ, rμ, and qμ (recall Figure 1)—all vanish as Δt tends to 0. Therefore, limΔt0σμ=limΔt0 E and the entropy rate becomes:

hμ~μ(log2(Δt)+0ϕ(t)log2ϕ(t)dt)Δt . (9)

With Δt → 0, hμ nominally tends to 0: As we shorten the observation time scale, spike events become increasingly rare. There are at least two known ways to address hμ apparently not being very revealing when so defined. On the one hand, rather than focusing on the uncertainty per symbol, as hμ does, we opt to look at the uncertainty per unit time: hμt. This is the so-called Δt-entropy rate (Gaspard and Wang, 1993) and it diverges as −μ log Δt. Such divergences are to be expected: The large literature on dimension theory characterizes a continuous set's randomness by its divergence scaling rates (Farmer et al., 1983; Mayer-Kress, 1986). Here, we are characterizing sets of similar cardinality—infinite sequences. On the other hand, paralleling sequence block-entropy definition of entropy rate (hμ =ℓ → ∞H[X0:ℓ]/ℓ) (Crutchfield and Feldman, 2003), continuous-time entropy rates are often approached within a continuous-time framework using:

hμ=limTH(T)/T ,

where H(T) is path entropy, the continuous-time analog of the block entropy H(ℓ) (Girardin, 2005). In these analyses, any log Δt terms are regularized away using Shannon's differential entropy (Cover and Thomas, 2006), leaving the nondivergent component -μ0ϕ(t)logϕ(t)dt. Using the Δt-entropy rate but keeping both the divergent and nondivergent components, as in Equations (8) and (9), is an approach that respects both viewpoints and gives a detailed picture of time-resolution scaling.

A major challenge in analyzing spike trains concerns locating the timescales on which information relevant to the stimulus is carried. Or, more precisely, we are often interested in estimating what percentage of the raw entropy of a neural spike train is used to communicate information about a stimulus; cf. the framing in Strong et al. (1998). For such analyses, the entropy rate is often taken to be Ht, T)/T, where T is the total path time and Ht, T) is the entropy of neural spike trains over time T resolved at time bin size Δt. In terms of previously derived quantities and paralleling the well known block-entropy linear asymptote H(ℓ) = E + hμℓ (Crutchfield and Feldman, 2003), this is:

H(Δt,T)T=hμ(Δt)Δt+E(T,Δt)T .

From the scaling analyses above, the extensive component of Ht, T)/T diverges logarithmically in the small Δt limit due to the logarithmic divergence (Equation 9) in hμt)/Δt. If we are interested in accurately estimating the entropy rate, then the above is one finite-time T estimate of it. However, there are other estimators, including:

H(Δt,T)H(Δt,TΔt)Δthμ(Δt)Δt+E(T,Δt)T .

This estimator converges more quickly to the true entropy rate hμt)/Δt than does Ht, T)/T.

No such log Δt divergences occur with bμ. Straightforward calculation, not shown here, reveals that:

limΔt0bμΔt=μ(0ϕ(t)0ϕ(t)log2ϕ(t+t)dtdt             +1log20ϕ(t)log2ϕ(t)dt) . (10)

Since limΔt0bμ(Δt)/Δt< and limΔt0hμ(Δt)/Δt diverges, the ephemeral information rate rμt)/Δt also diverges as Δt → 0. The bulk of the information generated by such renewal processes is dissipated and, having no impact on future behavior, is not useful for prediction.

Were we allowed to observe relatively microscopic membrane voltage fluctuations rather than being restricted to the relatively macroscopic spike sequence, the Δt-scaling analysis would be entirely different. Following Marzen and Crutchfield (2014) or natural extensions thereof, the statistical complexity diverges as −log ϵ, where ϵ is the resolution level for the membrane voltage, the excess entropy diverges as log1/Δt, the time-normalized entropy rate diverges as log2πeDΔtΔt, and the time-normalized bound information diverges as 1/2Δt. In other words, observing membrane voltage rather than spikes makes the process far more predictable. The relatively more macroscopic modeling at the level of spikes throws away much detail of the underlying biochemical dynamics.

To illustrate the previous points, we turn to numerics and a particular neural model. Consider an (unleaky) integrate-and-fire neuron driven by white noise whose membrane voltage (after suitable change of parameters) evolves according to:

dVdt=b+Dη(t), (11)

where η(t) is white noise such that 〈η(t)〉 = 0 and 〈η(t)η(t′)〉 = δ(tt′). When V = 1, the neuron spikes and the voltage is reset to V = 0; it stays at V = 0 for a time τ, which enforces a hard refractory period. Since the membrane voltage resets to a predetermined value, the interspike intervals produced by this model are independently drawn from the same interspike interval distribution:

ϕ(t)={0t<τλ2π(tτ)3eλ(μ(tτ)1)2/2(tτ)tτ . (12)

Here, 1/μ = 1/b is the mean interspike interval and λ = 1/D is a shape parameter that controls ISI variance. This neural model is not as realistic as that of a linear leaky integrate-and-fire neural model (Gerstner and Kistler, 2002), but is complex enough to illustrate the points made earlier about the scaling of information measures and time resolution.

For illustration purposes, we assume that the time-binned neural spike train is well approximated by a renewal process, even when Δt is as large as one millisecond. This assumption will generally not hold, as past interevent counts could provide more detailed historical information that more precisely places the last spike within its time bin. Even so, the reported information measure estimates are still useful. The estimated hμ is an upper bound on the true entropy rate; the reported E is a lower bound on the true excess entropy using the Data Processing Inequality (Cover and Thomas, 2006); and the reported Cμ will usually be a lower bound on the true process' statistical complexity.

Employing the renewal process assumption, numerical analysis corroborates the infinitesimal analysis above. Figure 3 plots F(n)—the proxy for the full, continuous-time, ISI distribution—for a given set of neuronal parameter values as a function of time resolution. Figure 4 then shows that hμ and Cμ exhibit logarithmic scaling at millisecond time discretizations, but that E does not converge to its continuous-time value until we reach time discretizations on the order of hundreds of microseconds. Even when Δt = 100 μs, bμt)/Δt still has not converged to its continuous-time values.

Figure 3.

Figure 3

An unleaky integrate-and-fire neuron driven by white noise has varying interevent count distributions F(n) that depend on time bin size Δt. Based on the ISI distribution ϕ(t) given in Equation (12) with τ = 2 ms, 1/μ = 1 ms, and λ = 1 ms. Data points represent exact values of F(n) calculated for integer values of N. Dashed lines are interpolations based on straight line segments connecting nearest neighbor points.

Figure 4.

Figure 4

How spike-train information measures (or rates) depend on time discretization Δt for an unleaky integrate-and-fire neuron driven by white noise. Top left: Statistical complexity Cμ as a function of both the ISI distribution shape parameters and the time bin size Δt. The horizontal axis is Δt in milliseconds on a log-scale and the vertical axis is Cμ in bits on a linear scale for three different ISI distributions following Equation (12) with τ = 2 ms. Top right: Entropy rate hμ also as a function of both shape parameters and Δt. Axes labeled as in the previous panel and the same three ISI distributions are used. Bottom left: Excess entropy E as a function of both the shape parameters and Δt. For the blue line limΔt0E(Δt)=0.75 bits; purple line, limΔt0E(Δt)=0.86 bits; and yellow line, limΔt0E(Δt)=0.41 bits. All computed from Equation (6). Bottom right: Bound information rate bμt)/Δt parameterized as in the previous panels. For the blue line limΔt0bμ(Δt)Δt=0.73 bits per second; purple line, limΔt0bμ(Δt)Δt=1.04 bits per second; and yellow line, limΔt0bμ(Δt)Δt=0.30 bits per second. All computed from Equation (10).

The statistical complexity Cμ increases without bound, as Δt → 0; see the top left panel of Figure 4. As suggested in the infinitesimal renewal analysis, hμ vanishes, whereas hμt diverges at a rate of μlog21/Δt, as shown in the top right plots of Figure 4. As anticipated, E tends to a finite, ISI distribution-dependent value when Δt tends to 0, as shown in the bottom left panel in Figure 4. Finally, the lower right panel plots bμt)/Δt.

One conclusion from this simple numerical analysis is that one should consider going to submillisecond time resolutions to obtain accurate estimates of limΔt0E(Δt) and limΔt0bμ(Δt)/Δt, even though the calculated informational values are a few bits or even less than one bit per second in magnitude.

4. Alternating renewal processes

The form of the Δt-scalings discussed in Section 3 occur much more generally than indicated there. Often, our aim is to calculate the nondivergent component of these information measures as Δt → 0, but the rates of these scalings are process-dependent. Therefore, these divergences can be viewed as a feature rather than a bug; they contain additional information about the process' structure (Gaspard and Wang, 1993).

To illustrate this point, we now investigate Δt-scalings for information measures of alternating renewal processes (ARPs), which are structurally more complex than the standard renewal processes considered above. For instance, these calculations suggest that rates of divergence of the τ-entropy rate smaller than the firing rate, such as those seen in Nemenman et al. (2008), are indicative of strong ISI correlations. Calculational details are sequestered in Appendix A.

In an ARP, an ISI is drawn from one distribution ϕ(1)(t), then another distribution ϕ(2)(t), then the first ϕ(1)(t) again, and so on. We refer to the new piece of additional information—the ISI distribution currently being drawn from—as the modality. Under weak technical conditions, the causal states are the modality and time since last spike. The corresponding, generic ϵ-machine is shown in Figure 5. We define the modality-dependent survival functions as Φi(t)=tϕ(i)(t)dt, the modality-dependent mean firing rates as:

Figure 5.

Figure 5

ϵ-Machine for an alternating renewal process in which neither interevent count distribution is Δ-Poisson and they are not equal almost everywhere. State label nm denotes n counts since the last event and present modality m.

μ(i)=1/0ϕ(i)(t)dt, (13)

the modality-dependent differential entropy rates:

hμ(i)=μ(i)0ϕ(i)log2ϕ(i)(t)dt ,

the modality-dependent continuous-time statistical complexity:

Cμ(i)=0μ(i)Φ(i)(t)log2(μ(i)Φ(i)(t))dt ,

and the modality-dependent excess entropy:

E(i)=0μ(i)tϕ(i)(t)log2(μ(i)ϕ(i)(t))dt        20μ(i)Φ(i)(t)log2(μ(i)Φ(i)(t))dt . (14)

It is straightforward to show, as done in Appendix A, that the time-normalized entropy rate still scales with log21/Δt:

hμ(Δt)Δt~12μlog2(1Δt)+μ(2)hμ(1)+μ(1)hμ(2)μ(1)+μ(2), (15)

where μ=2μ(1)μ(2)μ(1)+μ(2). As expected, the statistical complexity still diverges:

Cμ(Δt)~2log2(1Δt)+μ(2)Cμ(1)+μ(1)Cμ(2)μ(1)+μ(2)                 +Hb(μ1μ1+μ2) , (16)

where Hb(p) = −p log2 p − (1 − p) log2 (1 − p) is the entropy in bits of a Bernoulli random variable with bias p. Finally, the excess entropy still limits to a positive constant:

limΔt0E(Δt)=Hb(μ1μ1+μ2)+μ(2)E(1)+μ(1)E(2)μ(1)+μ(2) . (17)

The additional terms Hb(·) come from the information stored in the time course of modalities.

As a point of comparison, we ask what these information measures would be for the original (noncomposite) renewal process with the same ISI distribution as the ARP. As described in Appendix B, the former entropy rate is always less than the true hμ; its statistical complexity is always less than the true Cμ; and its excess entropy is always smaller than the true E. In particular, the ARP's hμ divergence rate is always less than or equal to the mean firing rate μ. Interestingly, this coincides with what was found empirically in the time series of a single neuron; see Figure 5C in Nemenman et al. (2008).

The ARPs here are a first example of how one can calculate information measures of the much broader and more structurally complex class of processes generated by unifilar hidden semi-Markov models, a subclass of hidden semi-Markov models (Tokdar et al., 2010).

5. Information universality

Another aim of ours is to interpret the information measures. In particular, we wished to relate infinitesimal time-resolution excess entropies, statistical complexities, entropy rates, and bound information rates to more familiar characterizations of neural spike trains—firing rates μ and ISI coefficient of variations CV. To address this, we now analyze a suite of familiar single-neuron models. We introduce the models first, describe the parameters behind our numerical estimates, and then compare the information measures.

Many single-neuron models, when driven by temporally uncorrelated and stationary input, produce neural spike trains that are renewal processes. We just analyzed one model class, the noisy integrate-and-fire (NIF) neurons in Section 3, focusing on time-resolution dependence. Other common neural models include the linear leaky integrate-and-fire (LIF) neuron, whose dimensionless membrane voltage, after a suitable change of parameters, fluctuates as:

dVdt=bV+aη(t) , (18)

and when V = 1, a spike is emitted and V is instantaneously reset to 0. We computed ISI survival functions from empirical histograms of 105 ISIs; we varied b ∈ [1.5, 5.75] in steps of 0.25 and a ∈ [0.1, 3.0] in steps of 0.1 to a = 1.0 and in steps of 0.25 thereafter.

The quadratic integrate-and-fire (QIF) neuron has membrane voltage fluctuations that, after a suitable change of variables, are described by:

dVdt=b+V2+aη(t) , (19)

and when V = 100, a spike is emitted and V is instantaneously reset to −100. We computed ISI survival functions from empirical histograms of trajectories with 105 ISIs; we varied b ∈ [0.25, 4.75] in steps of 0.25 and a ∈ [0.25, 2.75] in steps of 0.25. The QIF neuron has a very different dynamical behavior from the LIF neuron, exhibiting a Hopf bifurcation at b = 0. Simulation details are given in Appendix B.

Finally, ISI distributions are often fit to gamma distributions, and so we also calculated the information measures of spike trains with gamma-distributed ISIs (GISI).

Each neural model—NIF, LIF, QIF, and GISI—has its own set of parameters that governs its ISI distribution shape. Taken at face value, this would make it difficult to compare information measures across models. Fortunately, for each of these neural models, the firing rate μ and coefficient of variation CV uniquely determine the underlying model parameters (Vilela and Lindner, 2009). As Appendix B shows, the quantities limΔt0E(Δt), limΔt0Cμ+log2(μΔt), limΔt0hμ(Δt)/μΔt+log2(μΔt), and limΔt0bμ(Δt)/μΔt depend only on the ISI coefficient of variation CV and not the mean firing rate μ.

We estimated information measures from the simulated spike train data using plug-in estimators based on the formulae in Section 3. Enough data was generated that even naive plug-in estimators were adequate except for estimating bμ when CV was larger than 1. See Appendix B for estimation details. That said, binned estimators are likely inferior to binless entropy estimators (Victor, 2002), and naive estimators tend to have large biases. This will be an interesting direction for future research, since a detailed analysis goes beyond the present scope.

Figure 6 compares the statistical complexity, excess entropy, entropy rate, and bound information rate for all four neuron types as a function of their CV. Surprisingly, the NIF, LIF, and QIF neuron's information measures have essentially identical dependence on CV. That is, the differences in mechanism do not strongly affect these informational properties of the spike trains they generate. Naturally, this leads one to ask if the informational indifference to mechanism generalizes to other spike train model classes and stimulus-response settings.

Figure 6.

Figure 6

Information universality across distinct neuron dynamics. We find that several information measures depend only on the ISI coefficient of variation CV and not the ISI mean firing rate μ for the following neural spike train models: (i) neurons with Gamma distributed ISIs (GISI, blue), (ii) noisy integrate-and-fire neurons governed by Equation (11) (NIF, green), (iii) noisy linear leaky integrate-and-fire neurons governed by Equation (18) (LIF, dotted red), and (iv) noisy quadratic integrate-and-fire neurons governed by Equation (19) (QIF, dotted blue). Top left: limΔt0Cμ(Δt)+log2(μΔt). Top right: limΔt0hμ(Δt)μΔt+log2(μΔt). Bottom left: limΔt0E(Δt). Bottom right: limΔt0bμ(Δt)μΔt). In the latter, ISI distributions with smaller CV were excluded due to the difficulty of accurately estimating 00ϕ(t)ϕ(t)log2ϕ(t+t)dtdt from simulated spike trains. See text for discussion.

Figure 6's top left panel shows that the continuous-time statistical complexity grows monotonically with increasing CV. In particular, the statistical complexity increases logarithmically with ISI mean and approximately linearly with the ISI coefficient of variation CV. That is, the number of bits that must be stored to predict these processes increases in response to additional process stochasticity and longer temporal correlations. In fact, it is straightforward to show that the statistical complexity is minimized and excess entropy maximized at fixed μ when the neural spike train is periodic. This is unsurprising since, in the space of processes, periodic processes are least cryptic (CμE = 0) and so knowledge of oscillation phase is enough to completely predict the future. (See Appendix B.)

The bottom left panel in Figure 6 shows that increasing CV tends to decrease the excess entropy E—the number of bits that one can predict about the future. E diverges for small CV, dips at the CV where the ISI distribution is closest to exponential, and limits to a small number of bits at large CV. At small CV, the neural spike train is close to noise-free periodic behavior. When analyzed at small but nonzero Δt, E encounters an “ultraviolet divergence” (Tchernookov and Nemenman, 2013). Thus, E diverges as CV → 0, and a simple argument in Appendix B suggests that the rate of divergence is log2(1/CV). At an intermediate CV ~ 1, the ISI distribution is as close as possible to that of a memoryless Poisson process and so E is close to vanishing. At larger CV, the neural spike train is noise-driven. Surprisingly, completely noise-driven processes still have a fraction of a bit of predictability: knowing the time since last spike allows for some power in predicting the time to next spike.

The top right panel shows that an appropriately rescaled differential entropy rate varies differently for neural spike trains from noisy integrate-and-fire neurons and neural spike trains with gamma-distributed ISIs. As expected, the entropy rate is maximized at CV near 1, consistent with the Poisson process being the maximum entropy distribution for fixed mean ISI. Gamma-distributed ISIs are far less random than ISIs from noisy integrate-and-fire neurons, holding μ and CV constant.

Finally, the continuous-time bound information (bμ) rate varies in a similar way to E with CV. (Note that since the plotted quantity is limΔt0bμ, one could interpret the normalization by 1/μ as a statement about how the mean firing rate μ sets the natural timescale.) At low CV, the bμ rate diverges as 1CV2, as described in Appendix B. Interestingly, this limit is singular, similar to the results in Marzen and Crutchfield (2014): at CV = 0, the spike train is noise-free periodic and so the bμ rate is 0. For CV ≈ 1, it dips for the same reason that E decreases. For larger CV, bμ's behavior depends rather strongly on the ISI distribution shape. The longer-ranged gamma-distribution results in ever-increasing bμ rate for larger CV, while the bμ rate of neural spike trains produced by NIF neurons tends to a small positive constant at large CV. The variation of bμ deviates from that of E qualitatively at larger CV in that the GISI spike trains yield smaller total predictability E than that of NIF neurons, but arbitrarily higher predictability rate.

These calculations suggest a new kind of universality for neuronal information measures within a particular generative model class. All of these distinct integrate-and-fire neuron models generate ISI distributions from different families, yet their informational properties exhibit the same dependencies on Δt, μ, and CV in the limit of small Δt. Neural spike trains with gamma-distributed ISIs did not show similar informational properties. And, we would not expect neural spike trains that are alternating renewal processes to show similar informational properties either. (See Section 4.) These coarse information quantities might therefore be effective model selection tools for real neural spike train data, though more groundwork must be explored to ascertain their utility.

6. Conclusions

We explored the scaling properties of a variety of information-theoretic quantities associated with two classes of spiking neural models: renewal processes and alternating renewal processes. We found that information generation (entropy rate) and stored information (statistical complexity) both diverge logarithmically with decreasing time resolution for both types of spiking models, whereas the predictable information (excess entropy) and active information accumulation (bound information rate) limit to a constant. Our results suggest that the excess entropy and regularized statistical complexity of different types of integrate-and-fire neurons are universal in the sense that they do not depend on mechanism details, indicating a surprising simplicity in complex neural spike trains. Our findings highlight the importance of analyzing the scaling behavior of information quantities, rather than assessing these only at a fixed temporal resolution.

By restricting ourselves to relatively simple spiking models we have been able to establish several key properties of their behavior. There are, of course, other important spiking models that cannot be expressed as renewal processes or alternating renewal processes, but we are encouraged by the robust scaling behavior of the entropy rate, statistical complexity, excess entropy, and bound information rate over the range of models we considered.

There was a certain emphasis here on the entropy rate and hidden Markov models of neural spike trains, both familiar tools in computational neuroscience. On this score, our contributions are straightforward. We determined how the entropy rate varies with the time discretization and identified the possibly infinite-state, unifilar HMMs required for optimal prediction of spike-train renewal processes. Entropy rate diverges logarithmically for stochastic processes (Gaspard and Wang, 1993), and this has been observed empirically for neural spike trains for time discretizations in the submillisecond regime (Nemenman et al., 2008). We argued that the hμ divergence rate is an important characteristic. For renewal processes, it is the mean firing rate; for alternating renewal processes, the “reduced mass” of the mean firing rates. Our analysis of the latter, more structured processes showed that a divergence rate less than the mean firing rate—also seen experimentally (Nemenman et al., 2008)—indicates that there are strong correlations between ISIs. Generally, the nondivergent component of the time discretization-normalized entropy rate is the differential entropy rate; e.g., as given in Stevens and Zador (1996).

Empirically studying information measures as a function of time resolution can lead to a refined understanding of the time scales over which neuronal communication occurs. Regardless of the information measure chosen, the results and analysis here suggest that much can be learned by studying scaling behavior rather than focusing only on neural information as a single quantity estimated at a fixed temporal resolution. While we focused on the regime in which the time discretization was smaller than any intrinsic timescale of the process, future and more revealing analyses would study scaling behavior at even smaller time resolutions to directly determine intrinsic time scales (Crutchfield, 1994).

Going beyond information generation (entropy rate), we analyzed information measures—namely, statistical complexity and excess entropy—that have only recently been used to understand neural coding and communication. Their introduction is motivated by the hypothesis that neurons benefit from learning to predict their inputs (Palmer et al., 2013), which can consist of the neural spike trains of upstream neurons. The statistical complexity is the minimal amount of historical information required for exact prediction. To our knowledge, the statistical complexity has appeared only once previously in computational neuroscience (Haslinger et al., 2010). The excess entropy, a closely related companion, is the maximum amount of information that can be predicted about the future. When it diverges, then its divergence rate is quite revealing of the underlying process (Crutchfield, 1994; Bialek et al., 2001), but none of the model neural spike trains studied here had divergent excess entropy. Finally, the bound information rate has yet to be deployed in the context of neural coding, though related quantities have drawn attention elsewhere, such as in nonlinear dynamics (James et al., 2014), music (Abdallah and Plumbley, 2009), spin systems (Abdallah and Plumbley, 2012), and information-based reinforcement learning (Martius et al., 2013). Though its potential uses have yet to be exploited, it is an interesting quantity in that it captures the rate at which spontaneously generated information is actively stored by neurons. That is, it quantifies how neurons harness randomness.

Our contributions to this endeavor are more substantial than the preceding points. We provided exact formulae for the above quantities for renewal processes and alternating renewal processes. The new expressions can be developed further as lower bounds and empirical estimators for a process' statistical complexity, excess entropy, and bound information rate. This parallels how the renewal-process entropy-rate formula is a surprisingly accurate entropy-rate estimator (Gao et al., 2008). By deriving explicit expressions, we were able to analyze time-resolution scaling, showing that the statistical complexity diverges logarithmically for all but Poisson processes. So, just like the entropy rate, any calculations of the statistical complexity—e.g., as in Haslinger et al. (2010)—should be accompanied by the time discretization dependence. Notably, the excess entropy and the bound information rate have no such divergences.

To appreciate more directly what neural information processing behavior these information measures capture in the continuous-time limit, we studied them as functions of the ISI coefficient of variation. With an appropriate renormalization, simulations revealed surprising simplicity: a universal dependence on the coefficient of variation across several familiar neural models. The simplicity is worth investigating further since the dynamics and biophysical mechanisms implicit in the alternative noisy integrate-and-fire neural models are quite different. If other generative models of neural spike trains also show similar information universality, then these information measures might prove useful as model selection tools.

Finally, we close with a discussion of a practical issue related to the scaling analyses—one that is especially important given the increasingly sophisticated neuronal measurement technologies coming online at a rapid pace (Alivisatos et al., 2012). How small should Δt be to obtain correct estimates of neuronal communication? First, as we emphasized, there is no single “correct” estimate for an information quantity, rather its resolution scaling is key. Second, results presented here and in a previous study by others (Nemenman et al., 2008) suggest that extracting information scaling rates and nondivergent components can require submillisecond time resolution. Third, and to highlight, the regime of infinitesimal time resolution is exactly the limit in which computational efforts without analytic foundation will fail or, at a minimum, be rather inefficient. As such, we hope that the results and methods developed here will be useful to these future endeavors and guide how new technologies facilitate scaling analysis.

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

The authors thank C. Hillar and R. G. James for helpful comments. They also thank the Santa Fe Institute for its hospitality during visits. JC is an SFI External Faculty member. This material is based upon work supported by, or in part by, the U.S. Army Research Laboratory and the U.S. Army Research Office under contracts W911NF-13-1-0390 and W911NF-13-1-0340. SM was funded by a National Science Foundation Graduate Student Research Fellowship and the U.C. Berkeley Chancellor's Fellowship.

Appendix A

Alternating renewal process information measures

A discrete-time alternating renewal process draws counts from F1(n), then F2(n), then F1(n), and so on. We now show that the modality and counts since last event are causal states when F1F2 almost everywhere and when neither F1 nor F2 is eventually Δ-Poisson. We present only a proof sketch.

Two pasts x:0 and x:0 belong to the same causal state when Pr(X0:|X:0=x:0)=Pr(X0:|X:0=x:0). We can describe the future uniquely by a sequence of interevent counts Ni, i ≥ 1, and the counts till next event N0. Likewise, we could describe the past as a sequence of interevent counts Ni, i < 0, and the counts since last event N0-N0. Let Mi be the modality at time step i. So, for instance, M0 is the present modality.

First, we claim that one can infer the present modality from a semi-infinite past almost surely. The probability that the present modality is 1 having observed the last 2M events is:

Pr(M0=1|N2M:1=n2M:1)           =i=1,odd2MF2(ni)F1(ni1) .

Similarly, the probability that the present modality is 2 having observed the last 2M events is:

Pr(M0=2|N2M:1=n2M:1)           =i=1,odd2MF1(ni)F2(ni1) .

We are better served by thinking about the normalized difference of the corresponding log likelihoods:

Q:=12MlogP(M0=1|N2M:1=n2M:1)P(M0=2|N2M:1=n2M:1) .

Some manipulation leads to:

Q=12(1Mi=1,odd2MlogF2(ni)F1(ni)+1Mi=1,even2MlogF1(ni)F2(ni)),

and, almost surely in the limit of M → ∞:

1Mi=1,odd2MlogF1(ni)F2(ni){D[F2||F1]M0=1D[F1||F2]M0=2 , (A1)

where D[P||Q] is the information gain between P and Q (Cover and Thomas, 2006). And, we also have:

1Mi=1,even2MlogF2(ni)F1(ni){D[F1||F2]M0=1D[F2||F1]M0=2 .

This implies that:

limMQ=D[F2||F1]D[F1||F2]2{1M0=11M0=2 .

We only fail to identify the present modality almost surely from the semi-infinite past if limMQ=0. Otherwise, the unnormalized difference of the log likelihoods:

logPr(M0=1|N:1=n:1)Pr(M0=2|N:1=n:1)

tends to ±∞, implying that one of the two probabilities has vanished. From the expression, limMQ=0 only happens when D[F2||F1] = D[F1||F2]. However, equality requires that F1(n) = F2(n) almost everywhere.

Given the present modality, we also need to know the counts since the last event in order to predict the future as well as possible. The proof of this is very similar to those given in Marzen and Crutchfield (2015). The conditional probability distribution of future given past is:

Pr(X0:|X:0=x:0)=Pr(N1:|N0,X:0=x:0)Pr(N0|X:0=x:0) .

Since the present modality is identifiable from the past x:0, and since interevent counts are independent given modality:

Pr(N1:|N0,X:0=x:0)=Pr(N1:|M0=m0(n:1)) .

So, it is necessary to know the modality in order to predict the future as well as possible. By virtue of how the alternating renewal process is generated, the second term is:

Pr(N0|X:0=x:0)=Pr(N0|N0=n0,M0=m0(n:1)) .

A very similar term was analyzed in Marzen and Crutchfield (2015), and that analysis revealed that it was necessary to store the counts since last spike when neither F1 nor F2 is eventually Δ-Poisson.

Identifying causal states S+ as the present modality M0 and the counts since last event N0 immediately allows us to calculate the statistical complexity and entropy rate. The entropy rate can be calculated via:

hμ=H[X0|M0,N0]    =π(M0=1)H[X0|M0=1,N0]         +π(M0=2)H[X0|M0=2,N0].

The statistical complexity is:

Cμ=H[S+]     =H[M0,N0]     =H[M0]+π(M0=1)H[N0|M0=1]              +π(M0=2)H[N0|M0=2] . (A2)

Finally, it is straightforward to show that the modality M1 at time step 1 and the counts to next event are the reverse-time causal states under the same conditions on F1 and F2. Therefore:

E=I[S+;S]   =I[M0,N0;M1,N0N0]   =I[M0;M1,N0N0]             +I[N0;M1,N0N0|M0] .

One can continue in this way to find formulae for other information measures of a discrete-time alternating renewal process.

These formulae can be rewritten terms of the modality-dependent information measures of Equations (13) and (14) if we recognize two things. First, the probability of a particular modality is proportional to the average amount of time spent in that modality. Second, for reasons similar to those outlined in Marzen and Crutchfield (2015), the probability of counts since last event given a particular present modality i is proportional to wi(n). Hence, in the infinitesimal time discretization limit, the probability of modality 1 is:

π(M0=1)=μ(1)μ(1)+μ(2)

and similarly for modality 2. Then, the entropy rate out of modality i is:

H[X1|M0=i,N0]~Δt(μ(i)log21Δt+hμ(i)(Δt)) ,

and the modality-dependent statistical complexity diverges as:

H[N0|M0=i]~log21/Δt+Cμ(Δt) .

Finally, in continuous-time M0 and M1 limit to the same random variable, such that:

limΔt0E(Δt)=H[M0]+limΔt0I[N0;N0N0|M0] .

Note that E(i)=limΔt0I[N0;N0-N0|M0=i].

Bringing these results together, we substitute the above components into Equation (A2)'s expression for Cμ and, after details not shown here, find the expression quoted in the main text as Equation (16). Similarly, for hμ and E, yielding the the formulae presented in the main text in Equations (15) and (17), respectively.

As a last task, as our hypothetical null model, we wish to find the information measures for the corresponding renewal process approximation. The ISI distribution of the alternating renewal process is:

ϕ(t)=μ(2)ϕ(1)(t)+μ(1)ϕ(2)(t)μ(1)+μ(2) (A3)

and its survival function is:

Φ(t)=μ(2)Φ(1)(t)+μ(1)Φ(2)(t)μ(1)+μ(2) . (A4)

Hence, its mean firing rate is:

μ=11/μ(1)+1/μ(2) . (A5)

From Section 3, the entropy rate of the corresponding renewal process is:

hμren(Δt)Δt~μlog21Δt+μH[ϕ(t)] ;

compare Equation (15). And, the statistical complexity of the corresponding renewal process is:

Cμren(Δt)~log21Δt+H[μΦ(t)] .

The rate of divergence of Cμren(Δt) is half the rate of divergence of the true Cμt), as given in Equation (16). Trivial manipulations, starting from 0(1μ(1)-1μ(2))2, imply that the rate of entropy-rate divergence is always less than or equal to the mean firing rate for an alternating renewal process. Jensen's inequality implies that each of the nondivergent components of these information measures for the renewal process is less than or equal to that of the alternating renewal process. The Data Processing Inequality (Cover and Thomas, 2006) also implies that the excess entropy calculated by assuming a renewal process is a lower bound on the true process' excess entropy.

Appendix B

Simplicity in complex neurons

Recall that our white noise-driven linear leaky integrate-and-fire (LIF) neuron has governing equation:

V˙=bV+aη(t) , (A6)

and, when V = 1, a spike is emitted and V is instantaneously reset to 0. We computed ISI survival functions from empirical histograms of 105 ISIs. These ISIs were obtained by simulating Equation (A6) in Python/NumPy using an Euler integrator with time discretization of 1/1000 of log b/(b−1), which is the ISI in the noiseless limit.

The white noise-driven quadratic integrate-and-fire (QIF) neuron has governing equation:

V˙=b+V2+aη(t) , (A7)

and, when V = 100, a spike is emitted and V is instantaneously reset to −100. We computed ISI survival functions also from empirical histograms of trajectories with 105 ISIs. These ISIs were obtained by simulating Equation (A7) in Python/NumPy using an Euler stochastic integrator with time discretization of 1/1000 of πb, which is the ISI in the noiseless limit when threshold and reset voltages are +∞ and −∞, respectively.

Figure 6 shows estimates of the following continuous-time information measures from this simulated data as they vary with mean firing rate μ and ISI coefficient of variation CV. This required us to estimate μ, CV, and:

CμCT:=limΔt0(Cμ(Δt)+log2Δt),
ECT:=limΔt0E(Δt) ,
hμCT:=limΔt0(hμ(Δt)Δt+μlog2Δt),  and
bμCT:=limΔt0bμ(Δt)Δt ,

where the superscript CT is a reminder that these are appropriately regularized information measures in the continuous-time limit.

We estimated μ and CV using the sample mean and sample coefficient of variation with sufficient samples so that error bars (based on studying errors as a function of data size) were negligible. The information measures required new estimators, however. From the formulae in Section 3, we see that:

CμCT=log21μμ0Φ(t)log2Φ(t)dt , (A8)
ECT=0μtϕ(t)log2(μϕ(t))dt          20μΦ(t)log2Φ(t)dt, (A9)
hμCT=μ0ϕ(t)log2ϕ(t) ,  and (A10)
bμCT=μ(0ϕ(t)0ϕ(t)log2ϕ(t+t)dtdt           +1log20ϕ(t)log2ϕ(t)dt) . (A11)

It is well known that the sample mean is a consistent estimator of the true mean, that the empirical cumulative density function is a consistent estimator of the true cumulative density function almost everywhere, and thus that the empirical ISI distribution is a consistent estimator of the true cumulative density function almost everywhere. In estimating the empirical cumulative density function, we introduced a cubic spline interpolator. This is still a consistent estimator as long as Φ(t) is three-times differentiable, which is the case for ISI distributions from integrate-and-fire neurons. We then have estimators of CμCT, ECT, hμCT, and bμCT that are based on consistent estimators of μ, Φ(t), and ϕ(t) and that are likewise consistent.

We now discuss the finding evident in Figure 6, that the quantities limΔt0E(Δt) and limΔt0Cμ+log2(μΔt) depend only on the ISI coefficient of variation CV and not the mean firing rate μ. Presented in a different way, this is not so surprising. First, we use Marzen and Crutchfield (2015)'s expression for Cμ to rewrite:

Q1=limΔt0(Cμ(Δt)+log2(μΔt))    =μ0Φ(t)log2Φ(t)dt

and Equation (6) to rewrite:

Q2=limΔt0E(Δt)    =2Q1+0μtϕ(t)log2(μϕ(t))dt .

So, we only need to show that -μ0Φ(t)log2Φ(t)dt and 0μtϕ(t)log2(μϕ(t))dt are independent of μ for two-parameter families of ISI distributions.

Consider a change of variables from t to t′ = μt; then:

Q1=0Φ(t/μ)log2(Φ(t/μ))dt (A12)

and

Q2=2Q1+0tϕ(t/μ)log2(ϕ(t/μ))dt . (A13)

For all of the ISI distributions considered here, ϕ(tμ) is still part of the same two-parameter family as ϕ(t), except that its mean firing rate is 1 rather than μ. Its CV is unchanged. Hence, Q1 and Q2 are the same for a renewal process with mean firing rate 1 and μ, as long as the CV is held constant. It follows that limΔt0E(Δt) and limΔt0Cμ+log2(μΔt) are independent of μ and only depend on CV for the two-parameter families of ISI distributions considered in Section 5. Similar arguments apply to understanding the universal CV-dependence of limΔt0bμ(Δt)μΔt and limΔt0hμ(Δt)μΔt+log2(μΔt).

In Figure 6, we also see that E seems to diverge as CV → 0. Consider the following plausibility argument that suggests it diverges as log21/CV as CV → 0. These two-parameter ISI distributions with finite mean firing rate μ and small CV ≪ 1 can be approximated as Gaussians with mean 1/μ and standard deviation CV/μ. Recall from Equation (6) that we have:

E=20μΦ(t)log2(μΦ(t))dt          +0μtϕ(t)log2(μϕ(t))dt   =log2μ2μ0Φ(t)log2Φ(t)dt         +μ0tϕ(t)log2ϕ(t)dt .

Note that as CV → 0:

Φ(t){1t<1μ12t=1μ0t>1μ (A14)

and so:

limCV00Φ(t)log2Φ(t)dt=0 .

We assumed that for small CV, we can approximate:

ϕ(t)12πCV2/μ2exp((μt1)22CV2) ,

which then implies that:

μ0tϕ(t)log2ϕ(t)dtlog2μ2πCV12. (A15)

So, for any ISI distribution tightly distributed about its mean ISI, we expect:

Elog21CV ,

so that E diverges in this way. A similar asymptotic analysis also shows that as CV → 0,

limΔt0bμ(Δt)Δt1log2(12CV212) , (A16)

thereby explaining the divergence of limΔt0bμ(Δt)/Δt evident in Figure 6.

Finally, a straightforward argument shows that Cμ is minimized at fixed μ when the neural spike train is periodic. We can rewrite Cμ in the infinitesimal time resolution limit as:

Cμ(Δt)~log2(1μΔt)+μ0Φ(t)log21Φ(t)dt .

Note that 0 ≤ Φ (t) ≤ 1, and so 0Φ(t)log21Φ(t)dt0. We set it equal to zero by using the step function given in Equation (A14), which corresponds to a noiseless periodic process. So, the lower bound on Cμt) is log21/μΔt, and this bound is achieved by a periodic process.

Footnotes

1As the subscript context makes clear, the mean count μ is not related to that μ in Cμ and related quantities. In the latter it refers to the measure μ(s) over bi-infinite sequences s generated by a process.

2In the theoretical neuroscience literature, E(T) is sometime called the predictive information Ipred(T) and is a useful indicator of process complexity when E diverges (Bialek et al., 2001).

References

  1. Abdallah S. A., Plumbley M. D. (2009). Information dynamics: patterns of expectation and surprise in the perception of music. Connect. Sci. 21, 89–117. 10.1080/09540090902733756 [DOI] [Google Scholar]
  2. Abdallah S. A., Plumbley M. D. (2012). A measure of statistical complexity based on predictive information with application to finite spin systems. Phys. Lett. A. 376, 275–281. 10.1016/j.physleta.2011.10.066 [DOI] [Google Scholar]
  3. Alivisatos A. P., Chun M., Church G. M., Greenspan R. J., Roukes M. L., Yuste R. (2012). The brain activity map project and the challenge of functional connectomics. Neuron 74, 970–974. 10.1016/j.neuron.2012.06.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Ara P. M., James R. G., Crutchfield J. P. (2015). The elusive present: hidden past and future dependence and why we build models. Available online at: http://arXiv.org:1507.00672 [cond-mat.stat-mech]. [DOI] [PubMed]
  5. Archer E., Park I. M., Pillow J. W. (2012). Bayesian estimation of discrete entropy with mixtures of stick-breaking priors. Adv. Neural Info. Proc. Sys. 25, 2015–2023. [Google Scholar]
  6. Atick J. J. (1992). Could information theory provide an ecological theory of sensory processing?, in Princeton Lectures on Biophysics, ed Bialek W. (Singapore: World Scientific; ), 223–289. [Google Scholar]
  7. Averbeck B. B., Latham P. E., Pouget A. (2006). Neural correlations, population coding and computation. Nat. Rev. Neurosci. 7, 358–366. 10.1038/nrn1888 [DOI] [PubMed] [Google Scholar]
  8. Barlow H. B. (1961). Possible principles underlying the transformation of sensory messages, in Sensory Communication, ed Rosenblith W. (Cambridge, MA: MIT Press; ), 217–234. [Google Scholar]
  9. Bell A. J., Mainen Z. F., Tsodyks M., Sejnowski T. J. (1995). Balancing Conductances may Explain Irregular Cortical Firing. Technical Report, Institute for Neural Computation, San Diego.
  10. Berry M. J., Warland D. K., Meister M. (1997). The structure and precision of retinal spike trains. Proc. Natl. Acad. Sci. U.S.A. 94, 5411–5416. 10.1073/pnas.94.10.5411 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bialek W., Nemenman I., Tishby N. (2001). Predictability, complexity, and learning. Neural Comp. 13, 2409–2463. 10.1162/089976601753195969 [DOI] [PubMed] [Google Scholar]
  12. Bialek W., Ruderman D. L., Zee A. (1991). Optimal sampling of natural images: a design principle for the visual system? in Advances in Neural Information Processing 3, eds Lippman R. P., Moody J. E., Touretzky D. S. (San Mateo, CA: Morgan Kaufmann; ), 363–369. [Google Scholar]
  13. Britten K. H., Newsome W. T., Shadlen M. N., Celebrini S., Movshon J. A. (1996). A relationship between behavioral choice and the visual responses of neurons in macaque MT. Vis. Neurosci. 13, 87–100. 10.1017/S095252380000715X [DOI] [PubMed] [Google Scholar]
  14. Butts D. A., Goldman M. S. (2006). Tuning curves, neuronal variability, and sensory coding. PLoS Biol. 4:e92. 10.1371/journal.pbio.0040092 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Cessac B., Cofre R. (2013). Spike train statistics and Gibbs distributions. Available online at: http://arXiv.org:1302.5007. [DOI] [PubMed]
  16. Costa M., Goldberger A. L., Peng C. K. (2002). Multiscale entropy analysis of complex physiologic time series. Phys. Rev. Lett. 89:068102. 10.1103/PhysRevLett.89.068102 [DOI] [PubMed] [Google Scholar]
  17. Costa M., Goldberger A. L., Peng C. K. (2005). Multiscale entropy analysis of biological signals. Phys. Rev. E 71:021906. 10.1103/PhysRevE.71.021906 [DOI] [PubMed] [Google Scholar]
  18. Cover T. M., Thomas J. A. (2006). Elements of Information Theory, 2nd Edn. New York, NY: Wiley-Interscience. [Google Scholar]
  19. Crutchfield J. P. (1994). The calculi of emergence: Computation, dynamics, and induction. Physica D 75, 11–54. 10.1016/0167-2789(94)90273-9 [DOI] [Google Scholar]
  20. Crutchfield J. P., Ellison C. J., Mahoney J. R. (2009). Time's barbed arrow: Irreversibility, crypticity, and stored information. Phys. Rev. Lett. 103:094101. 10.1103/PhysRevLett.103.094101 [DOI] [PubMed] [Google Scholar]
  21. Crutchfield J. P., Feldman D. P. (2003). Regularities unseen, randomness observed: levels of entropy convergence. CHAOS 13, 25–54. 10.1063/1.1530990 [DOI] [PubMed] [Google Scholar]
  22. Crutchfield J. P., Young K. (1989). Inferring statistical complexity. Phys. Rev. Let. 63, 105–108. 10.1103/PhysRevLett.63.105 [DOI] [PubMed] [Google Scholar]
  23. Dan Y., Atick J. J., Reid R. C. (1996). Efficient coding of natural scenes in the lateral geniculate nucleus: experimental test of a computational theory. J. Neurosci. 16, 3351–3362. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Destexhe A., Rudolph M., Paré D. (2003). The high-conductance state of neocortical neurons in vivo. Nat. Rev. Neurosci. 4, 739–751. 10.1038/nrn1198 [DOI] [PubMed] [Google Scholar]
  25. Deweese M. (1996). Optimization principles for the neural code. Netw. Comp. Neural Sys. 7, 325–331. 10.1088/0954-898X/7/2/013 [DOI] [PubMed] [Google Scholar]
  26. DeWeese M. R., Meister M. (1999). How to measure the information gained from one symbol. Network 10, 325–340. 10.1088/0954-898X/10/4/303 [DOI] [PubMed] [Google Scholar]
  27. DeWeese M. R., Wehr M., Zador A. M. (2003). Binary spiking in auditory cortex. J. Neurosci. 23, 7940–7949. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. DeWeese M. R., Zador A. M. (2006). Non-Gaussian membrane potential dynamics imply sparse, synchronous activity in auditory cortex. J. Neurosci. 26, 12206–12218. 10.1523/JNEUROSCI.2813-06.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Farmer J. D., Ott E., Yorke J. A. (1983). The dimension of chaotic attractors. Physica 7D, 153 10.1007/978-0-387-21830-4/11 [DOI] [Google Scholar]
  30. Gao Y., Kontoyiannis I., Bienenstock E. (2008). Estimating the entropy of binary time series: methodology, some theory and a simulation study. Entropy 10, 71–99. 10.3390/entropy-e10020071 [DOI] [Google Scholar]
  31. Gaspard P., Wang X.-J. (1993). Noise, chaos, and (ϵ, τ)-entropy per unit time. Phys. Rep. 235, 291–343. 10.1016/0370-1573(93)90012-3 [DOI] [Google Scholar]
  32. Gerstner W., Kistler W. M. (2002). Spiking Neuron Models: Single Neurons, Populations, Plasticity. Cambridge, UK: Cambridge University Press. [Google Scholar]
  33. Girardin V. (2005). On the different extensions of the ergodic theorem of information theory, in Recent Advances in Applied Probability Theory, eds Baeza-Yates R., Glaz J., Gzyl H., Husler J., Palacios J. L. (New York, NY: Springer; ), 163–179. [Google Scholar]
  34. Haslinger R., Klinkner K. L., Shalizi C. R. (2010). The computational structure of spike trains. Neural Comp. 22, 121–157. 10.1162/neco.2009.12-07-678 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Jacobs A. L., Fridman G., Douglas R. M., Alam N. M., Latham P. E., Prusky G. T., et al. (2009). Ruling out and ruling in neural codes. Proc. Natl. Acad. Sci. U.S.A. 106, 5937–5941. 10.1073/pnas.0900573106 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. James R. G., Burke K., Crutchfield J. P. (2014). Chaos forgets and remembers: measuring information creation, destruction, and storage. Phys. Lett. A 378, 2124–2127. 10.1016/j.physleta.2014.05.014 [DOI] [Google Scholar]
  37. James R. G., Ellison C. J., Crutchfield J. P. (2011). Anatomy of a bit: information in a time series observation. CHAOS 21:037109. 10.1063/1.3637494 [DOI] [PubMed] [Google Scholar]
  38. Koepsell K., Wang X., Hirsch J. A., Sommer F. T. (2010). Exploring the function of neural oscillations in early sensory systems. Front. Neurosci. 4, 53–61. 10.3389/neuro.01.010.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Laughlin S. B. (1981). A simple coding procedure enhances a neuron's information capacity. Z. Naturforsch. 36c, 910–912. [PubMed] [Google Scholar]
  40. Linsker R. (1989). An application of the principle of maximum information preservation to linear systems, in Advances in Neural Information Processing 1, ed Touretzky D. (San Mateo, CA: Morgan Kaufmann; ), 186–194. [Google Scholar]
  41. London M., Roth A., Beeren L., Häusser M., Latham P. E. (2010). Sensitivity to perturbations in vivo implies high noise and suggests rate coding in cortex. Nature 466, 123–128. 10.1038/nature09086 [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Mackay D. M., McCulloch W. W. (1952). The limiting information capacity of a neuronal link. Bull. Math. Biophys. 14, 127–135. 10.1007/BF02477711 [DOI] [Google Scholar]
  43. Martius G., Der R., Ay N. (2013). Information driven self-organization of complex robotics behaviors. PLoS ONE 8:e63400. 10.1371/journal.pone.0063400 [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Marzen S., Crutchfield J. P. (2014). Information anatomy of stochastic equilibria. Entropy 16, 4713–4748. 10.3390/e16094713 [DOI] [Google Scholar]
  45. Marzen S., Crutchfield J. P. (2015). Informational and causal architecture of discrete-time renewal processes. Entropy 17, 4891–4917. [Google Scholar]
  46. Mayer-Kress G. (ed.). (1986). Dimensions and Entropies in Chaotic Systems: Quantification of Complex Behavior. Berlin: Springer. [Google Scholar]
  47. Meister M., Lagnado L., Baylor D. A. (1995). Concerted signaling by retinal ganglion cells. Science 270, 1207–1210. 10.1126/science.270.5239.1207 [DOI] [PubMed] [Google Scholar]
  48. Nemenman I., Bialek W., de Ruyter van Steveninck R. R. (2004). Entropy and information in neural spike trains: progress on the sampling problem. Phys. Rev. E 69, 1–6. 10.1103/PhysRevE.69.056111 [DOI] [PubMed] [Google Scholar]
  49. Nemenman I., Lewen G. D., Bialek W., de Ruyter van Steveninck R. R. (2008). Neural coding of natural stimuli: Information at sub-millisecond resolution. PLoS Comp. Bio. 4:e1000025. 10.1371/journal.pcbi.1000025 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Nirenberg S., Carcieri S. M., Jacobs A. L., Latham P. E. (2001). Retinal ganglion cells act largely as independent encoders. Nature 411, 698–701. 10.1038/35079612 [DOI] [PubMed] [Google Scholar]
  51. Palmer S. E., Marre O., Berry II, M. J., Bialek W. (2013). Predictive information in a sensory population. Proc. Natl. Acad. Sci. U.S.A. 112, 6908–6913. 10.1073/pnas.1506855112 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Panzeri S., Treves A., Schultz S., Rolls E. T. (1999). On decoding the responses of a population of neurons from short time windows. Neural Comp. 11, 1553–1577. 10.1162/089976699300016142 [DOI] [PubMed] [Google Scholar]
  53. Paz A. (1971). Introduction to Probabilistic Automata. New York, NY: Academic Press. [Google Scholar]
  54. Rabiner L. R. (1989). A tutorial on hidden Markov models and selected applications. IEEE Proc. 77:257. [Google Scholar]
  55. Rabiner L. R., Juang B. H. (1986). An introduction to hidden Markov models. IEEE ASSP Mag. 4–16. 10.1109/MASSP.1986.116534218428778 [DOI] [Google Scholar]
  56. Reinagel P., Reid R. C. (2000). Temporal coding of visual information in the thalamus. J. Neurosci. 20, 5392–5400. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Rieke F., Warland D., de Ruyter van Steveninck R., Bialek W. (1999). Spikes: Exploring the Neural Code. New York, NY: Bradford Book. [DOI] [PubMed] [Google Scholar]
  58. Sakitt B., Barlow H. B. (1982). A model for the economical encoding of the visual image in cerebral cortex. Biol. Cybern. 43, 97–108. 10.1007/BF00336972 [DOI] [PubMed] [Google Scholar]
  59. Schneidman E., Berry M. J., Segev R., Bialek W. (2006). Weak pairwise correlations imply strongly correlated network states in a neural population. Nature 440, 1007–1012. 10.1038/nature04701 [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Schneidman E., Bialek W., Berry M. J. (2003). Synergy, redundancy, and independence in population codes. J. Neurosci. 23, 11539–11553. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Shadlen M. N., Newsome W. T. (1995). Is there a signal in the noise? Curr. Opin. Neurobiol. 5, 248–250. 10.1016/0959-4388(95)80033-6 [DOI] [PubMed] [Google Scholar]
  62. Shadlen M. N., Newsome W. T. (1998). The variable discharge of cortical neurons: Implications for connectivity, computation, and information coding. J. Neurosci. 18, 3870–3896. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Shalizi C. R., Crutchfield J. P. (2001). Computational mechanics: Pattern and prediction, structure and simplicity. J. Stat. Phys. 104, 817–879. 10.1023/A:1010388907793 [DOI] [Google Scholar]
  64. Shannon C. E. (1948). A mathematical theory of communication. Bell Sys. Tech. J. 27, 379–423, 623–656. 10.1002/j.1538-7305.1948.tb00917.x [DOI] [Google Scholar]
  65. Softky W. R., Koch C. (1993). The highly irregular firing of cortical cells is inconsistent with temporal integration of random EPSPs. J. Neurosci. 13, 334–350. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Srinivasan M. V., Laughlin S. B., Dubs A. (1982). Predictive coding: A fresh view of inhibition in the retina. Proc. R. Soc. Lond. Ser. B 216, 427–459. 10.1098/rspb.1982.0085 [DOI] [PubMed] [Google Scholar]
  67. Stein R. B. (1967). The information capacity of neurons using a frequency code. Biophys. J. 7, 797–826. 10.1016/S0006-3495(67)86623-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Stevens C. F., Zador A. (1996). Information through a spiking neuron, in Advance Neural Information Processing System, eds Touretzky D., Mozer M. C., Hasselmo M. E. (Cambridge, MA: MIT Press; ), 75–81. [Google Scholar]
  69. Stevens C. F., Zador A. M. (1998). Input synchrony and the irregular firing of cortical neurons. Nat. Neurosci. 1, 210–217. 10.1038/659 [DOI] [PubMed] [Google Scholar]
  70. Strong S. P., Koberle R., de Ruyter van Steveninck R., Bialek W. (1998). Entropy and information in neural spike trains. Phys. Rev. Lett. 80, 197–200. 10.1103/PhysRevLett.80.197 [DOI] [Google Scholar]
  71. Tchernookov M., Nemenman I. (2013). Predictive information in a nonequilibrium critical model. J. Stat. Phys. 153, 442–459. 10.1007/s10955-013-0833-6 [DOI] [Google Scholar]
  72. Theunissen F. E., Miller J. P. (1991). Representation of sensory information in the cricket cercal sensory system. ii: information theoretic calculation of system accuracy and optimal tuning curve widths of four primary interneurons. J. Neurophys. 66, 1690–1703. [DOI] [PubMed] [Google Scholar]
  73. Tokdar S., Xi P., Kelly R. C., Kass R. E. (2010). Detection of bursts in extracellular spike trains using hidden semi-Markov point process models. J. Comput. Neurosci. 29, 203–212. 10.1007/s10827-009-0182-2 [DOI] [PubMed] [Google Scholar]
  74. Treves A., Panzeri S. (1995). The upward bias in measures of information derived from limited data samples. Neural Comp. 7, 399–407. 10.1162/neco.1995.7.2.399 [DOI] [Google Scholar]
  75. Verdú S., Weissman T. (2006). Erasure entropy, in IEEE International Symposium on Information Theory (ISIT 2006) (Seattle, WA), 98–102. 10.1109/ISIT.2006.261682 [DOI] [Google Scholar]
  76. Victor J. D. (2002). Binless strategies for estimation of information from neural data. Phys. Rev. E 66:051903. 10.1103/PhysRevE.66.051903 [DOI] [PubMed] [Google Scholar]
  77. Vilela R. D., Lindner B. (2009). Are the input parameters of white noise driven integrate and fire neurons uniquely determined by rate and CV? J. Theo. Bio. 257, 90–99. 10.1016/j.jtbi.2008.11.004 [DOI] [PubMed] [Google Scholar]
  78. Yang Y., Zador A. M. (2012). Differences in sensitivity to neural timing among cortical areas. J. Neurosci. 32, 15142–15147. 10.1523/JNEUROSCI.1411-12.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Yeung R. W. (2008). Information Theory and Network Coding. New York, NY: Springer. [Google Scholar]

Articles from Frontiers in Computational Neuroscience are provided here courtesy of Frontiers Media SA

RESOURCES