Skip to main content
Network Neuroscience logoLink to Network Neuroscience
. 2020 Jul 1;4(3):678–697. doi: 10.1162/netn_a_00141

Correlated activity favors synergistic processing in local cortical networks in vitro at synaptically relevant timescales

Samantha P Sherrill 1,*, Nicholas M Timme 2, John M Beggs 3, Ehren L Newman 4
PMCID: PMC7462423  PMID: 32885121

Abstract

Neural information processing is widely understood to depend on correlations in neuronal activity. However, whether correlation is favorable or not is contentious. Here, we sought to determine how correlated activity and information processing are related in cortical circuits. Using recordings of hundreds of spiking neurons in organotypic cultures of mouse neocortex, we asked whether mutual information between neurons that feed into a common third neuron increased synergistic information processing by the receiving neuron. We found that mutual information and synergistic processing were positively related at synaptic timescales (0.05–14 ms), where mutual information values were low. This effect was mediated by the increase in information transmission—of which synergistic processing is a component—that resulted as mutual information grew. However, at extrasynaptic windows (up to 3,000 ms), where mutual information values were high, the relationship between mutual information and synergistic processing became negative. In this regime, greater mutual information resulted in a disproportionate increase in redundancy relative to information transmission. These results indicate that the emergence of synergistic processing from correlated activity differs according to timescale and correlation regime. In a low-correlation regime, synergistic processing increases with greater correlation, and in a high-correlation regime, synergistic processing decreases with greater correlation.

Keywords: Neural information processing, Correlated activity, Computation, In vitro networks, Effective connectivity, Cortex, Information theory, Synergy

Author Summary

In the present work, we address the question of whether correlated activity in functional networks of cortical circuits supports neural computation. To do so, we combined network analysis with information theoretic tools to analyze the spiking activity of hundreds of neurons recorded from organotypic cultures of mouse somatosensory cortex. We found that, at timescales most relevant to direct neuronal communication, neurons with more correlated activity predicted greater computation, suggesting that correlated activity does support computation in cortical circuits. Importantly, this result reversed at timescales less relevant to direct neuronal communication, where even greater correlated activity predicted decreased computation. Thus, the relationship between correlated activity and computation depends on the timescale and the degree of correlation in neuronal interactions.

INTRODUCTION

What role does the correlated activity among cortical neurons play in neural information processing? Correlated activity is ubiquitous throughout the brain, emerging from both external stimuli and internal dynamics. Correlated activity is predictive of information processing (for a review, see Salinas & Sejnowski, 2001). However, the extent to which it is favorable for information processing is not clear. What is needed to better understand the role of neural correlations in information processing is a comparison of how the amount of correlated activity between upstream neurons relates to the amount of resulting information processing in cortical microcircuits.

The view that correlated neural activity is favorable for neural information processing is widely held within the cognitive rhythms community and is based on the idea that correlation facilitates both communication between circuits and the orchestration of processing within circuits. Correlated neural activity, especially synchronous activity, is understood to generate the coherent rhythms that are observed in local field potentials, electrocorticography, and electroencephalography that are theorized to subserve specific computational or cognitive mechanisms (e.g., Fries, 2015; Hasselmo et al., 2002; Hernandez protect et al., 2020; Honey et al., 2017; Lisman & Jensen, 2013; Newman et al., 2014; Norman et al., 2006; Ward, 2003). Ample empirical evidence derived from in vitro, in silico, and in vivo studies supports the importance of synchrony for organizing information transmission in cortical circuits (Averbeck & Lee, 2004; Azouz & Gray, 2003; Fries, 2015; Poulet & Petersen, 2008; Salinas & Sejnowski, 2001; Yu et al., 2008). The synchronization of neuronal spiking is indeed linked to higher order cognitive and behavioral processes (Grammont & Riehle, 1999; Riehle et al., 1997; Vinck et al., 2015). From this perspective, correlation is favorable for information processing. Yet, even within the cognitive rhythms community there is recognition that excess correlation can be unfavorable and that, in some circumstances, desynchronization supports information processing better than synchronization (e.g., Bastos et al., 2015; Hanslmayr et al., 2012; van Winsun et al., 1984). The discrepancy between this and the standard view of the cognitive rhythms community warrants reconciliation (Hanslmayr, Staresina, & Bowman, 2016).

The view that correlated activity is unfavorable for neural information processing is widely held in the sensory processing and artificial neural network communities and is based on the idea that correlation is synonymous with redundancy and thus reduces efficiency (Attneave, 1954; Barlow, 1961; Gutnisky & Dragoi, 2008; Schneidman, Bialek, & Berry, 2003; Shadlen & Newsome, 1998). This bandwidth-limiting effect of correlation motivates the “redundancy-reduction hypothesis” (Atick & Redlich, 1990; Attneave, 1954; Barlow, 1961; Shadlen & Newsome, 1998). In this line of thinking, a normative goal of sensory information processing is to reduce the redundancy of neuronal signals (Atick & Redlich, 1992; Barlow, 1961; Field, 1987; Gutnisky & Dragoi, 2008; Laughlin, 1989; Rieke et al., 1995; van Hateren, 1992). Many works, together, indicate that signal redundancy decreases from lower order sensory areas to higher order sensory areas (Berry et al., 1997; Chechik et al., 2006; Dan et al., 1998; Doi et al., 2012; Montani et al., 2007; Nirenberg et al., 2001; Puchalla et al., 2005; Reich et al., 2001; Reinagel & Reid, 2000; Zohary et al., 1994). From this perspective, correlation is unfavorable for information processing. Yet, as in the cognitive rhythms community, there is an indication that qualification is needed to this standard view, given recent empirical evidence that correlated activity can also increase processing (Nigam, Pojoga, & Dragoi, 2019).

Our aim with the work described here was to determine which of these two perspectives better accounts for the relationship between correlation and synergistic processing—a component of information processing—in local cortical microcircuits at synaptic timescales and beyond. To accomplish this goal, we analyzed the spiking activity of hundreds of neurons recorded simultaneously from each of 25 organotypic cultures of mouse somatosensory cortex (Figure 1). In these recordings we identified hundreds of thousands of triads wherein individual neurons received significant functional input from two other neurons at synaptic timescales (<14 ms). For each triad, we measured the total information about the receiving neurons firing that was carried in the sending neurons activity and then decomposed this into the constituent components: the unique contributions of each sender, the redundancy between the senders (i.e., redundancy), and the synergy between the senders (i.e., synergistic processing). Across triads, we found that the correlation between the senders was low, but that greater correlations were predictive of greater synergistic processing by the receiving neuron. When we extended the timescale of the analysis to consider triads at extrasynaptic windows (up to 3,000 ms), we observed a significant increase in the overall correlation between unique senders’ activity. We also observed a shift to a negative relationship between sender correlation and synergistic processing. Secondary analyses of the redundancy and information transmission indicated that both perspectives are valid but become relevant in different correlation regimes. Specifically, in low-correlation regimes (at synaptic timescales), information transmission and redundancy also increased, but the rate of increase of information transmission was greater than redundancy, thereby enabling greater synergistic processing. In high-correlation regimes (at extrasynaptic timescales), information transmission leveled off but redundancy continued to increase, thereby consuming bandwidth capacity and reducing synergistic processing.

Figure 1. .

Figure 1. 

Methodological approach taken to ask if correlated activity is favorable or unfavorable for synergistic processing in organotypic cultures of mouse cortex. (A) Hour-long recordings of spiking activity were collected in vitro from organotypic cultures of mouse somatosensory cortex. (B) Spike sorting yielded spike trains of hundreds of well-isolated individual neurons. (C) Effective connectivity between neurons was determined by quantifying transfer entropy between spike trains. The resulting effective networks were analyzed to identify all effective triads consisting of two effective connections to a common receiver. (D) For each triad, we quantified how correlated the activity of the senders was with mutual information (gray arrow) and decomposed the total information transmitted to the receiver (measured via multivariate transfer entropy, orange perimeter) into redundancy and synergy components (via partial information decomposition, indicated with purple and green areas, respectively). (E and F) We sought to differentiate between two alternate hypotheses. (E) Predicted relationship between mutual information and synergy if correlated activity is favorable for computation. (F) Predicted relationship between mutual information and synergy if correlated activity is unfavorable for computation.

RESULTS

We asked whether correlated activity between neurons supports computation in cortical microcircuits by analyzing hour-long recordings of spiking activity from organotypic cultures of mouse somatosensory cortex (n = 25) as summarized in Figure 1. Recordings contained between 98 and 594 well-isolated neurons (median = 310). We identified effective connections between neurons in each recording as those that had significant transfer entropy. For every effective triad consisting of one neuron receiving effective connections from two other neurons we quantified the amount of correlated activity between the sending neurons using mutual information. Mutual information was selected for its ability to detect nonlinear as well as linear dependencies. Our findings were, however, not sensitive to the method for quantifying correlated activity (see Supporting Information where we show the same qualitative pattern of results using Pearson correlation as well as other measures). The mutual information between senders was normalized to reflect the proportion of the maximum possible mutual information for that triad given the constituent neuron entropies (pMImax). We quantified the amount of computation performed by the receiver on the inputs using “synergy,” a term derived from partial information decomposition. Synergy was normalized to reflect the proportion of the receiving neuron entropy for which it accounted (pHrec). Neither normalization was required (as shown in the Supporting Information) but were included to control for variability across networks. Across triads, we asked whether synergy was positively or negatively correlated with mutual information. This analysis was repeated at varying timescales as determined by the granularity of the data binning and the delay between bins. All summary statistics are reported as medians followed by 95% bootstrap confidence intervals in brackets.

Correlated Activity Is Associated with More Synergy at Synaptic Timescales

To determine the relationship between mutual information and synergy at timescales relevant for synaptic processing, we analyzed the functional dynamics at three timescales spanning 0.05–14 ms (centered on 3, 5, and 11 ms) for each of the 25 recordings, yielding 75 total functional networks. Each network had many effective triads (count = 806 [448 1,350]). The median mutual information across triads of individual networks was 0.008 [0.005, 0.010], which was ∼0.8% of the maximum possible mutual information. Mutual information was highly variable around this median value within individual networks, varying across ∼7.0 [6.7, 7.2] orders of magnitude within individual networks. We then asked how this variability in mutual information was correlated with variability in synergy.

Plotting synergy as a function of mutual information across triads for each network revealed clear positive relationships (e.g., Figure 2A). To aggregate results across networks with differing numbers of triads and to maintain the ability to observe possible nonmonotonic relationships between mutual information and synergy, we collapsed results across triads based on mutual information deciles for each network (Figure 2B). Mutual information and synergy were strongly positively correlated across deciles (Spearman r = 0.95 [0.93, 0.98]; Zs.r. = 7.53, n = 75 networks, p < 1 × 10−13; Figure 2C). The median slope over deciles in the log-log space was 0.46 [0.41, 0.50] Δlog10(pHrec)/Δlog10(pMImax) (Figure 2D), indicating that synergy increases exponentially as a function of mutual information with an exponent of ∼0.46. This positive relationship indicated that greater synergistic processing emerges from integration of correlated activity at synaptic timescales.

Figure 2. .

Figure 2. 

Increased correlated activity predicts greater synergistic processing at synaptic timescales. (A) Synergy plotted as a function of mutual information for a representative network shows a positive relationship. Gray dots indicate individual triads. Boundaries of deciles of mutual information are depicted via dashed lines. Median synergy for each decile (red dots) increased across deciles. The black line shows the linear least squares fit of the red dots (rho = 0.99, slope = 0.39). (B) Synergy across mutual information deciles for all 75 synaptic timescale networks are shown as red lines. The median synergy and mutual information for each decile is shown by the black points with error bars indicating 95% bootstrap confidence intervals. (C) Histogram of Spearman rank correlation coefficients for synergy versus mutual information shown in B. A solid blue line marks the median correlation coefficient across the 75 networks, and the shaded region around the blue line shows the 95% bootstrap confidence interval around the median. (D) Average regression fit line across the 75 linear regressions of synergy versus mutual information indicates consistently positive relationship. Shaded region indicates the 95% confidence interval.

Correlated Activity Is Associated with Greater Gains in Information Transmission Than Redundancy at Synaptic Timescales

Theories regarding the favorability of correlated activity, or lack thereof, for neural computation are motivated by the impact of increasing correlation on information transmission and redundancy. To address these theories, we examined how information transmission and redundancy covaried with mutual information and how these relationships related to the generation of synergy. Information transmission was quantified as the multivariate transfer entropy (mvTE). We normalized mvTE to reflect the proportion of the entropy of the receiving neuron accounted for (pHrec). Redundancy was quantified using the redundancy term derived from partial information decomposition and was also normalized to reflect the proportion of the entropy of the receiving neuron accounted for (pHrec).

Theories that view correlated activity as favorable predict that correlated activity increases information transmission. Consistent with this, we found that mvTE was reliably positively correlated with mutual information at synaptic timescales (Spearman r = 0.99 [0.98, 1.00]; Zs.r. = 7.58, n = 75 networks, p < 1 × 10−13; slope = 0.35 [0.32, 0.38] Δlog10(pHrec)/Δlog10(pMImax); Figure 3A; Table 1). Theories that view correlated activity as unfavorable predict that correlated activity increases redundancy. Supporting this, we found that redundancy was significantly positively related to mutual information at synaptic timescales (Spearman r = 0.99 [0.98, 0.99], Zs.r. = 7.56, n = 75 networks, p < 1 × 10−13; slope = 0.51 [0.47, 0.57] Δlog10(pHrec)/Δlog10(pMImax); Figure 3A; Table 1). These results indicate that both hypotheses are well motivated with respect to the expected effect of increasing correlated activity.

Figure 3. .

Figure 3. 

Growth of information transmission (mvTE) is greater than growth of redundancy over deciles of mutual information at synaptic timescales. (A) Multivariate transfer entropy (mvTE) and redundancy (Red) increase with mutual information. Synergy data is replotted here from Figure 2B for ease of comparison. (B) Same data as shown in A without log-scaling to show rapid growth of mvTE relative to redundancy. For all plots, lines depict medians and shaded regions depict 95% bootstrap confidence intervals around the median.

Table 1. .

Relationships between information terms at synaptic timescales.

graphic file with name netn-04-678-t01.jpg

Note. Off diagonal: median Spearman rank correlation coefficients (belowdiagonal) and linear regression slopes (above diagonal) for pairwise comparisons of synergy, redundancy, multivariate transfer entropy (mvTE), and the difference between mvTE and redundancy (mvTE-Red). Slopes are in units of Δlog10(pHrec)/Δlog10(pHrec). Along diagonal: median Spearman rank correlation coefficients (left) and linear regression slopes (right) for each information term compared to mutual information. Slopes are in units of Δlog10(pHrec)/Δlog10(pMImax). *Values that are significantly different from zero after Bonferroni–Holm correction (all p values < 0.001).

Next, we related the changes in mvTE and redundancy observed over deciles of mutual information to those observed in synergy. Consistent with the idea that correlated activity can support greater computation by facilitating information transmission, synergy and mvTE were reliably positively correlated across deciles of mutual information (Spearman r = 0.95 [0.92, 0.98]; Zs.r. = 7.53, p < 1 × 10−13, n = 75 networks; Table 1). Interestingly, synergy and redundancy were also reliably positively correlated across deciles of mutual information (Spearman r = 0.99 [0.98, 0.99]; Zs.r. = 7.56, p < 1 × 10−13, n = 75 networks; Table 1). This positive correlation is contrary to the idea that redundancy impedes computation at synaptic timescales.

We then examined the relative influence of correlated activity on information transmission and redundancy. In a non-log-scaled space, the slope relating mutual information to mvTE was significantly greater than the slope relating mutual information to redundancy (1.41 [1.26, 1.58] vs. 0.48 [0.45, 0.55] (ΔpHrec/ΔpMImax) ; Zs.r. = 7.52, n = 75 networks, p < 1 × 10−13; Figure 3B). Note that although the slope relating mutual information to mvTE was less than the slope relating mutual information to redundancy in the log-log space, mvTE was offset vertically over redundancy, corresponding to a significantly larger y-intercept (0.017 vs. 0.003, Zs.r. = 7.52, n = 75 networks, p < 1 × 10−11; Figure 3A). This means that the exponential growth of mvTE was steeper than that of redundancy over the values of mutual information observed at synaptic timescales.This can be seen by comparing the relationship between mutual information and mvTE and redundancy without log-scaling (Figure 3B). Because mvTE increases faster than redundancy as a function of mutual information, the difference between the two also grew as a function of mutual information (slope = 0.13 [0.06, 0.19] Δlog10(pHrec)/Δlog10(pMImax); Zs.r. = 4.19, p < 1 × 10−4, n = 75 networks; Table 1; curve not shown). The growth of this difference could account for how correlated activity supports synergistic processing at synaptic timescales, as it was positively correlated with synergy across deciles of mutual information (Spearman r = 0.88 [0.85, 0.90]; Zs.r. = 7.52, p < 1 × 10−13, n = 75 networks; Table 1).

Activity Becomes More Correlated at Extrasynaptic Timescales, with Negative Returns for Synergy

At synaptic timescales, (normalized) mutual information was generally low (∼0.8% of maximum possible, as described above). From this, we reasoned that the observed positive relationship between mutual information and synergistic processing may have been due to synaptic processing operating in a regime of low overall correlation. To explore this further, we sought to test what happens in a regime of greater overall correlation.

We had observed that the median mutual information increased progressively across the three synaptic timescales (pMImax across timescales: 3 ms = 0.004 [0.002, 0.005]; 5 ms = 0.008 [0.006, 0.010]; and 11 ms = 0.019 [0.013, 0.023]; Figure 4A), suggesting that extending the analysis to longer timescales would yield even greater mutual information. Examining median mutual information in seven additional logarithmically spaced timescales, extending out to 3,000 ms, showed that it indeed continued to increase (pMImax across timescales: 23 ms = 0.03 [0.02, 0.04]; 48 ms = 0.07 [0.05, 0.08]; 104 ms = 0.12 [0.09, 0.17]; 225 ms = 0.16 [0.10, 0.23]; 485 ms = 0.20 [0.14, 0.26]; 1,044 ms = 0.22 [0.17, 0.27]; 2,250 ms = 0.25 [0.21, 0.33]). All data is shown in Figure 4B.

Figure 4. .

Figure 4. 

Greater correlated activity emerges at longer timescales with diminishing returns for synergistic processing. (A) Median mutual information increases across synaptic timescales. (B) Median mutual information continues to increase at extrasynaptic timescales. (C) Median mutual information is significantly larger at extrasynaptic timescales than at synaptic timescales. (D) Plotting synergy as a function of mutual information across timescales shows that the positive relationship between synergy and mutual information only exists at synaptic timescales where mutual information is relatively low. This relationship becomes negative at extrasynaptic timescales where mutual information is high. Note that mutual information, plotted along the x-axis, varies across subpanels. A dashed vertical line where mutual information is 0.07 is included to facilitate visual alignment across panels. (E) Curves from D are replotted on the same axes here to show the overall relationship between synergy and mutual information. (F) Spearman rank correlation coefficients of synergy versus mutual information for all networks across timescales show that synergy and mutual information are positively related at shorter timescales and negatively related at longer timescales. (G) Synergy and mutual information are significantly positively correlated at synaptic timescales and significantly negatively correlated at extrasynaptic timescales. For all plots shown above, bars/solid lines indicate medians, shaded regions/error bars depict 95% bootstrap confidence intervals around the median, and * indicates p < 0.05.

To enable statistical comparison between the patterns observed at longer timescales with those we reported for the synaptic timescale, we defined the timescales that were two orders of magnitude longer than those used to span the synaptic timescales as the “extrasynaptic timescale” (i.e., those centered on 225, 485, and 1,044 ms vs. 3, 5, and 11 ms). Comparing mutual information between these two timescales revealed that correlated activity was significantly greater at extrasynaptic timescales than synaptic timescales (0.20 [0.16, 0.23] vs. 0.008 [0.005, 0.010] pMImax, Zr.s. = 10.51, n = 150 networks, p < 1 × 10−25; Figure 4C).

Analyzing the relationship between mutual information and synergy at longer timescales revealed that the positive correlation observed at synaptic timescales was lost as the median mutual information increased further (Figure 4DG). At timescales beyond the synaptic timescales, the relationship between mutual information and synergy inverted gradually. Numerically, this can be seen in the Spearman rank correlation coefficients relating mutual information to synergy, which started positive and gradually shifted to negative across timescales (Figure 4F; Spearman r across timescales: 3 ms = 0.98 [0.94, 0.99]**; 5 ms = 0.93 [0.88, 0.96]**; 11 ms = 0.95 [0.88, 0.96]**; 23 ms = 0.81 [0.65, 0.95]**; 48 ms = 0.47 [0.09, 0.70]**; 104 ms = 0.15 [−0.24, 0.36]; 225 ms = −0.43 [−0.78, −0.03]**; 485 ms = −0.68 [−0.88, −0.39]**; 1,044 ms = −0.96 [−1.00, −0.92]**; 2,250 ms = −1.00 [−1.00, −0.99]**; ** = p < 0.05). Though the overall trend between mutual information and synergy remained positive at the ∼48-ms timescale, evidence of a saturation point emerged as synergy began to decrease for mutual information values greater than ∼0.07. Across networks, the saturation point was reliably around 0.07 [0.06, 0.09]. At longer timescales, a greater percentile of triads had mutual information values above 0.07. This resulted in diminishing returns of synergy as mutual information grew further (Figure 4D and E). The relationship between mutual information and synergy was significantly negative at timescales including and greater than 225 ms. Whereas mutual information and synergy were positively correlated at synaptic timescales (as described above), mutual information was significantly negatively correlated with synergy at extrasynaptic timescales (Spearman r = −0.77 [−0.88, −0.62]; Zs.r. = −6.57, n = 75 networks, p < 1 × 10−10; slope = −0.16 [−0.21, −0.10] Δlog10(pHrec)/Δlog10(pMImax); Figure 4F and G). The slopes observed at the extrasynaptic timescales were significantly less than those observed at synaptic timescales (Zr.s. = −10.40, n = 150 networks, p < 1 × 10−24).

These results were obtained using mutual information and synergy values that were normalized by the maximum possible values given the firing rates and bin sizes for each triad at each timescale. This was done because information terms such as mutual information and synergy are sensitive to the entropy of the data being analyzed, which is, in turn, sensitive to firing rates and bin sizes. The lognormal distribution of firing rates and the different timescales generated variability in the total entropy and thus information terms. By normalizing, we minimized any effects that may have been due to changes in entropy within and across timescales. Critically, however, none of the qualitative results described here depended on this normalization (see Supporting Information Figure S6 for nonnormalized results).

Correlated Activity Is Associated with Greater Gains in Redundancy Than Information Transmission at Extrasynaptic Timescales

The robust positive relationship between mvTE and mutual information observed at synaptic timescales was lost at longer timescales (Figure 5A and B). This can be seen in the Spearman rank correlation coefficients relating mvTE to mutual information, which started positive and gradually shifted toward zero across timescales (Figure 5C; Spearman r across timescales: 3 ms = 0.99 [0.96, 1.00]**; 5 ms = 0.98 [0.93, 1.00]**; 11 ms = 0.99 [0.96, 1.00]**; 23 ms = 0.98 [0.93, 1.00]**; 48 ms = 0.90 [0.28, 0.98]**; 104 ms = 0.95 [0.89, 0.99]**; 225 ms = 0.99 [0.94, 1.00]**; 485 ms = 0.88 [0.42, 0.95]**; 1,044 ms = 0.27 [−0.39, 0.71]; 2,250 ms =−0.25 [−0.72, 0.54]; **p < 0.05). Consistent with the loss of a relationship between mvTE and mutual information at longer timescales, mvTE ceased to be reliably correlated with synergy across deciles of mutual information at extrasynaptic timescales (Spearman r = −0.19 [−0.41, 0.03]; Zs.r. = −1.88, p = 0.06, n = 75 networks; Figure 5D). The correlation between mvTE and synergy was significantly greater at synaptic compared with extrasynaptic timescales (Zr.s. = 9.99, p < 1 × 10−22, n = 150 networks; Figure 5D). These results indicate that the growth of information transmission with correlated activity that contributed to increased synergistic processing at synaptic timescales was lost at extrasynaptic timescales.

Figure 5. .

Figure 5. 

Redundancy encroaches upon multivariate transfer entropy at longer timescales, accounting for decreasing synergy at those timescales. (A) Log-scaled multivariate transfer entropy (mvTE), redundancy (Red), and synergy (Syn) versus mutual information, for all timescales. Growth of mvTE with respect to mutual information is steep at short timescales, but decreases across timescales. Redundancy grows with mutual information at all timescales. (B) MvTE curves in A plotted on the same axes. (C) Spearman rank correlation coefficients of synergy versus mvTE across timescales indicates that mvTE saturates at longer timescales. (D) Synergy and mvTE are significantly positively related at synaptic timescales, but not at extrasynaptic timescales. (E) Redundancy curves in A plotted on the same axes. (F) Spearman rank correlation coefficients of synergy versus redundancy across timescales shows how the two are positively related at shorter timescales and negatively related at longer timescales. (G) Synergy and redundancy are significantly positively related at synaptic timescales and significantly negatively related at extrasynaptic timescales. (H) Difference of MvTE and redundancy curves in A plotted on the same axes shows that the overall trend looks most similar to that of synergy versus mutual information in Figure 4E. (I) Spearman rank correlation coefficients of synergy versus mvTE redundancy across timescales shows that the two are positively related at all timescales. (J) Synergy and mvTE redundancy are significantly positively related at synaptic and extrasynaptic timescales. Synergy versus mutual information data has been replotted from Figure 4D here in A for ease of comparison. For all plots, solid/bold lines indicate medians and shaded regions/error bars depict 95% bootstrap confidence intervals around the median.

Redundancy remained significantly positively related with mutual information at all timescales (Figure 5E; Spearman r across timescales: 3 ms = 0.98 [0.94, 0.99]**; 5 ms = 0.98 [0.92, 0.99]**; 11 ms = 0.99 [0.95, 1.00]**; 23 ms = 0.99 [0.96, 1.00]**; 48 ms = 0.95 [0.87, 0.99]**; 104 ms = 0.99 [0.96, 1.00]**; 225 ms = 1.00 [0.99, 1.00]**; 485 ms = 1.00 [0.99, 1.00]**; 1,044 ms = 1.00 [0.98, 1.00]**; 2,250 ms = 1.00 [1.00, 1.00]**; **p < 0.05). Though synergy and redundancy had been positively correlated at synaptic timescales (as described above), we found that they were negatively correlated at extrasynaptic timescales (Spearman r = −0.64 [−0.82, −0.53]; Zs.r. = −5.95, p < 1 × 10−8; n = 75 networks; Figure 5G). The correlation between redundancy and synergy was significantly greater at synaptic compared with extrasynaptic timescales (Zr.s. = 10.55, p < 1 × 10−25, n = 150 networks; Figure 5G). This result indicates that, at extrasynaptic timescales, redundancy impeded synergistic processing.

Because redundancy, but not mvTE, continued to increase as a function of mutual information at longer timescales, the difference between the two became smaller. This difference continued to be positively correlated with synergy at extrasynaptic timescales (Spearman r = 0.50 [0.32, 0.66]; Zs.r. = 5.60, p < 1 × 10−7; n = 75 networks; Figure 5J). The correlation between mvTE-redundancy and synergy was significantly greater at synaptic compared with extrasynaptic timescales (Zr.s. = 6.57, p < 1 × 10−10, n = 150 networks; Figure 5J); however, correlations were significantly positive at both timescales. This indicated that reduction of the gap between mvTE and redundancy across timescales could account for why correlated activity stops supporting synergistic processing at extrasynaptic timescales.

DISCUSSION

Our aim with the work described here was to determine whether correlated activity in cortical microcircuits was favorable or unfavorable for synergistic processing. By applying information theoretic analyses to the spiking activity of hundreds of neurons, we showed that correlated activity is favorable for synergistic processing when the overall level of correlations were low. We also showed that correlated activity was unfavorable for synergistic processing when overall correlation levels were an order of magnitude greater. Secondary analyses indicated that the relationship between correlated activity and synergistic processing varied as a function of the relative amount of information transmission and redundancy. When overall correlation levels were low, we observed that incremental increases in correlation were associated with larger increases in information transmission than redundancy. We hypothesize that this effectively left greater bandwidth available for synergistic processing. When overall correlation levels were high, incremental increases in correlation were associated with larger increases in redundancy than information transmission. We hypothesize that this effectively caused redundancy to consume the bandwidth available for synergistic processing. This is summarized in Figure 6.

Figure 6. .

Figure 6. 

Neural computation can be expressed as a function of the correlation between inputs. When the correlation of inputs is low, synergistic processing benefits from increased correlation. In this regime, both information transmission and redundancy increase with greater correlation, but information transmission grows faster than redundancy effectively increasing the bandwidth available for synergistic processing. When the correlation of inputs is high, synergistic processing is hurt by increased correlation. In this regime, redundancy and not information transmission increases with greater correlation, reducing the bandwidth available for synergistic processing.

Correlation and Information Processing Emerge with Cognition

Correlated activity has been shown to increase as cognition emerges. This has been shown, both during neuronal development (Chiappalone et al., 2006) and during the transition out of deep anesthesia (Bettinardi et al., 2015). Recovery from anesthesia also brings about an increase in information transmission between neurons (Fagerholm et al., 2016). Given the positive relationship we describe here, we predict that computation would also increase under these contexts. Further, we predict that it is the increase in computation that marks the emergence of cognition. We propose this because without computation, or information modification events in which the whole is greater than the sum of the parts, all neuronal activity would be passive information transfer, which cannot account for the emergence of higher order brain states and behavior. To establish this more concretely, future studies should examine synergy in the context of cognition and criticality.

Correlation Regimes Parse the Favorability Debate

These findings provide empirical support for each of two seemingly contradictory perspectives regarding the favorability of correlated activity for neural information processing. The first perspective, held by many in the cognitive rhythms community, states that correlated activity is favorable for information processing because correlated activity facilitates information transmission (Averbeck & Lee, 2004; Fries, 2005, 2015; Lisman & Jensen, 2013; Salinas & Sejnowski, 2001). The second perspective, held by many in the sensory processing and artificial neural networks communities, states that correlated activity is unfavorable for information processing because correlated activity is viewed as equivalent to capacity-consuming redundancy (Attneave, 1954; Barlow, 1961). Even within the respective communities there is recognition that the relationship between correlated activity and processing may not be so straightforward. For example, within the cognitive rhythms community, Hanslmayr has explicitly argued that synchronized activity can both facilitate and impede processing (Hanslmayr et al., 2012; Hanslmayr et al., 2016). Likewise, in the sensory processing community, there is evidence that correlated activity can also increase processing (e.g., Nigam, Pojoga, & Dragoi, 2019). Each perspective, our data suggest, is relevant in distinct correlation regimes observed, here, across timescales.

The idea that the favorability of correlated activity for information processing varies nonmonotonically can account for seeming inconsistencies both within and between the separate perspectives. For example, while correlated activity is largely held as benefiting cognitive processing within the cognitive rhythms community (e.g., Fries, 2005; Lisman & Jensen, 2013), it can be that information processing requires decorrelation in higher correlation regimes (e.g., Hanslmayr et al., 2016). Indeed, it has been suggested that correlated activity serves distinct roles in different neural systems such as the cortex and hippocampus (Hanslmayr et al., 2016). Consistent with the possibility for regional differences in the favorability of correlated activity, the present findings indicate that the key differentiating functional factor between areas is the correlation level. Breaking from simple interregional differences, however, this framework would also predict that there can be systematic variability in the favorability of correlated activity even within a region. This is consistent with the finding that correlated activity improves synergistic coding in primary visual cortex (e.g., Nigam, Pojoga, & Dragoi, 2019). As such, the present findings also serve to account for apparent inconsistencies in the sensory processing community.

Correlation Regime As a Covariate of Timescale

Importantly, the variance in correlation regime we observed was confounded with timescale. That is, we observed the low-correlation regime at short timescales and the high-correlation regime at long timescales. While our motivating question was grounded in the importance of correlation rather than timescale, it is not possible to disentangle these two factors fully here, and additional work is needed on this topic.

“Signal” Versus “Noise” Correlations

The analysis of correlated activity used here differs from the approach commonly used in sensory processing and artificial neural networks research in that we did not separately analyze “signal correlations” and “noise correlations.” In such research, in which neural responses to experimentally delivered stimuli are evaluated, it is possible to separate activity into what is stimulus dependent (aka, signal correlations) and what is internally generated (aka, noise correlations) (Cohen & Kohn, 2011; Gawne & Richmond, 1993). Here, no stimuli were delivered, and the analysis integrated activity over the entire hour of the recording, likening our analysis to one of noise correlations only. This does not mean, however, that the nonmonotonic function relating correlated activity to synergistic processing need apply only to noise correlations. Though signal and noise correlations can be analytically separated, the same underlying physiological constraints should be expected to hold for both types of correlations. If there is an important difference between the activity comprising signal and noise correlations, it is, we argue, that correlations induced by extrinsic factors (e.g., delivered stimuli) reflect transient shifts in the statistics of the local processing dynamics. To determine how these results relate to signal and noise correlations, future work must test how synergistic processing relates to the level of correlated activity in response to stimuli delivered in vivo.

Related Works

The present findings are consistent with prior analyses of synergistic processing and, we hypothesize, could account for previously reported topological determinants of synergistic processing. In a previous analysis of synergistic processing in developing organotypic cultures, Wibral et al. (2017) observed a similar pattern of synergy increasing initially, and then decreasing, as a function of culture maturity. Notably, they also found that mutual information among the neurons increased steadily across the same developmental timeline. Thus, the underlying pattern of results is consistent with the pattern observed here. Previous analysis of the topological determinants of synergistic processing found that synergistic processing increased with the number of connections made by upstream neurons (Timme et al., 2016). In line with this observation, we have previously shown that neurons in rich clubs of cortical microcircuits do about twice as much synergistic processing as neurons outside of rich clubs (Faber et al., 2019). Based on our present findings, we predict that (1) upstream neurons with more connections have greater correlated activity, and (2) rich club neurons have greater correlated activity than neurons outside of the rich club.

Significance of Optimal Synergistic Processing

The nonmonotonic curve relating mutual information to synergy revealed by our analyses indicates that the most synergistic processing was observed when the activity of the sending neurons was correlated at ∼7% of the maximum possible. It is unclear whether this reflects a universal optimum or if this value should be expected to vary as a function of the local circuit physiology. Physiology plays an important role in regulating the amount of correlation (e.g., Gutnisky & Dragoi, 2008; Minces et al., 2017). What is not clear based on existing work, however, is whether these changes shift processing relative to a fixed local optimum or whether the optimal correlation also shifts. To address this, future research must compare how the optimal value varies across brain regions, physiological conditions, and behavioral conditions, if at all.

Relevance of In Vitro Preparation

The use of organotypic cultures in the present work facilitated the recording of hundreds of neurons simultaneously. While organotypic cultures naturally differ from intact in vivo tissue, organotypic cultures nonetheless exhibit synaptic structure and electrophysiological activity very similar to that found in vivo (Beggs & Plenz, 2004; Bolz et al., 1990; Caeser et al., 1989; Götz & Bolz, 1992; Ikegaya et al., 2004; Klostermann & Wahle, 1999; Plenz & Aertsen, 1996). For example, the distribution of firing rates observed in cultures is lognormal, as seen in vivo (Nigam et al., 2016), and the strengths of functional connections are lognormally distributed, similar to the distribution of synaptic strengths observed in patch clamp recordings (reviewed in Buzsáki & Mizuseki, 2014; Song et al., 2005). These features indicate that organotypic cortical cultures serve as a reasonable model system for exploring local cortical networks, while offering unique accessibility to large neuron count and high temporal resolution recordings. However, additional work will need to be done to understand how the synergy–mutual information relationship observed in vitro differs from what may exist in vivo, particularly in the context of behavior.

While stimulus-driven activity has been favored in research for its ability to provide insight into neural coding mechanisms, such studies assume that the brain is primarily reflexive and that internal dynamics are not informative with regard to information processing. However, internally driven spontaneous activity of neurons, or activity that does not track external variables in observable ways, has been repeatedly shown to be no less cognitively interesting than stimulus-linked activity (Johnson et al., 2009; Raichle, 2010; for a review see Tozzi et al., 2016; Tsodyks et al., 1999). Not only is spontaneous activity predominant throughout the brain, but it also drives critical processes such as neuronal development (Cang et al., 2005; Chiappalone et al., 2006; Wibral et al., 2017).

Conclusions

We show, for the first time, that the relationship between correlated activity and synergistic processing is positive at synaptic timescales with low levels of overall correlation but negative at extrasynaptic timescales with high levels of overall correlation. We propose that the relationship between correlated activity and synergistic processing varies as a function of the overall correlation regime. The shift from a positive to a negative relationship between regimes in this framework is accounted for by shifts in the effects of increasing correlation on information transmission and redundancy. In the low-correlation regime, increases in correlation increased information transmission faster than redundancy, thereby increasing the effective bandwidth for synergistic processing. In the high-correlation regime, increases in correlation increased redundancy faster than information transmission, thereby decreasing the effective bandwidth for synergistic processing. Both dynamics had been hypothesized to exist previously by separate frameworks within separate research communities. The present work serves to bridge across these disconnected domains of work to synthesize the two perspectives into a unified framework.

MATERIALS and METHODS

To answer the question of how correlation and computation are related in cortical circuits, we combined network analysis with information theoretic tools to analyze the spiking activity of hundreds of neurons recorded from organotypic cultures of mouse somatosensory cortex. Because of space limitations, here we provide an overview of our methods and focus on those steps that are most relevant for interpreting our results. A comprehensive description of all our methods can be found in the Supporting Information.

All procedures were performed in strict accordance with guidelines from the National Institutes of Health, and approved by the Animal Care and Use Committees of Indiana University and the University of California, Santa Cruz.

Electrophysiological Recordings

All results reported here were derived from the analysis of electrophysiological recordings of 25 organotypic cultures prepared from slices of mouse somatosensory cortex. One-hour-long recordings were performed at 20 kHz sampling by using a 512-channel array of 5-μm-diameter electrodes arranged in a triangular lattice with an interelectrode distance of 60 μm (spanning ∼0.9 mm by 1.9 mm). Once the data were collected, spikes were sorted using a PCA approach (Ito et al., 2014; Litke et al., 2004; Timme et al., 2014) to form spike trains of between 98 and 594 (median = 310) well-isolated individual neurons, depending on the recording.

Network Construction

Networks of effective connectivity, representing global activity in recordings, were constructed following the methods described by Timme et al. (2014, 2016). Briefly, weighted effective connections between neurons were established using transfer entropy (TE; Schreiber, 2000). To consider synaptic interactions, we computed TE at three timescales spanning 0.05–14 ms, discretized into overlapping bins of 0.05–3 ms, 1.6–6.4 ms, and 3.5–14 ms. To consider extrasynaptic interactions, we computed additional networks at timescales spanning 7.5–3,000 ms, discretized into overlapping bins of 7.5–30 ms, 16.15–64.6 ms, 34.8–139.2 ms, 75–300 ms, 161.6–646.4 ms, 348.1–1,392.4 ms, and 750–3,000 ms (Supporting Information Table S1). This resulted in an additional 7 networks per recording, or 175 networks (250 networks total). Only significant TE, determined through comparison to the TE values obtained with jittered spike trains (α = 0.001; 5,000 jitters), were used in the construction of the networks. TE values were normalized by the total entropy of the receiving neuron so as to reflect the proportion of the receiver neuron’s capacity that can be accounted for by the transmitting neuron.

Quantifying Correlation

The degree of similarity between all pairs of neurons was quantified as mutual information (Cover & Thomas, 1991; Shannon & Weaver, 1949):

MInorm(IJ)=i,jp(i,j)logp(i,j)p(i)p(j)/min(ipilogpi,jpjlogpj) (1)

Mutual information is a symmetric measure that quantifies how much uncertainty is reduced in one variable (I) by knowing another (J). All mutual information values were normalized by the maximum possible mutual information that could have occurred between each pair of neurons, which is equivalent to the minimum entropy of the two neurons (Gray & Shields, 1977). Using the normalized values allowed us to consider the proportion of the total possible mutual information that occurred between neurons and to control for differences in entropy across pairs of neurons, improving the interpretability of the numbers obtained. The results reported here are robust to the choice of similarity measure, as shown in Supporting Information Figures S4 and S5.

Quantifying Computation, Redundancy, and Information Transmission

Computation was operationalized as synergy. Synergy measures the additional information regarding the future state of the receiver, gained by considering the prior state of the senders jointly, beyond what they offered individually, after accounting for the redundancy between the sending neurons and the past state of the receiver itself. Synergy was calculated according to the partial information decomposition (PID) approach described by Williams and Beer (2011), including use of the Imin term to calculate redundancy (see Supporting Information). PID compares the measured bivariate TE between neurons TE(JI) and TE(KI) with the measured multivariate TE (the triad-level information transmission) among neurons TE({J, K} → I) to estimate terms that reflect the unique information carried by each neuron, the redundancy between neurons, and the synergy (i.e., gain over the sum of the parts) between neurons. Redundancy was computed as per Supporting Information Equations S8S10. Synergy was then computed via:

Synergy({J,K}I)=TE({J,K}I)TE(JI)TE(KI)+Redundancy({J,K}I) (2)

As with bivariate TE, synergy, redundancy and multivariate TE (mvTE) were normalized by the total entropy of the receiving neuron.

Although there are other methods for calculating partial information terms (Bertschinger et al., 2014; Lizier et al., 2018; Pica et al., 2017; Wibral et al., 2017), we chose this measure because it is capable of detecting linear and nonlinear interactions and it has been shown to be effective for our datatype (Timme et al., 2016; Faber et al., 2019). In addition, unlike other methods (Lizier et al., 2011; Stramaglia et al., 2012), PID of mvTE can decompose the interaction into nonnegative and nonoverlapping terms. However, to address previously raised concerns that PID overestimates the redundancy term (Bertschinger et al., 2014; Pica et al., 2017), and consequently synergy, we also used an alternate implementation of PID that estimates synergy based on the lower bound of redundancy. In this implementation, the effective threshold for triads to generate synergy is higher. This approach yielded the same qualitative pattern of results at synaptic timescales (see Supporting Information Figure S7). At extrasynaptic timescales, this alternate version of synergy remained positively related to mutual information. This is likely due to the fact that the increased threshold meant that only triads with larger amounts of synergy were included in this analysis, which also meant that redundancy had not effectively cannibalized the mvTE.

Note, we did not examine interactions larger than triads because of the multifold increase in the computational burden that arises in considering higher order synergy terms. In addition to the combinatorial explosion of increased numbers of inputs, the number of PID terms increases rapidly as the number of variables increases. However, based on bounds calculated for the highest order synergy term by Timme et al. (2016), it was determined that the information gained by including an additional input beyond two either remained constant or decreased. Thus, it was inferred that lower order (two-input) computations dominated. In addition, although we did not consider more than two inputs at a time, because we considered all possible triads in each network, we effectively subsampled the entire space of inputs for each neuron.

Also, note that mutual information (or any type of correlation) between senders does not necessarily imply that sender neurons have redundant information about the receiver neuron; although it does make redundancy more likely. Of course, for large mutual information values, redundancy becomes inevitable, as the majority of the information between the senders is shared, and therefore any information they may have about the receiver is likely to be overlapping.

Statistics

To aggregate results across networks with differing numbers of triads and to maintain the ability to observe possible nonmonotonic relationships between sender mutual information and synergy, we collapsed synergy, redundancy, and mvTE across triads based on deciles of sender mutual information for each network. That is, we calculated the median values of synergy, redundancy, and mvTE for each decile of sender mutual information. These values were used in all analyses. We obtained the same qualitative pattern of results when analyses were performed at the triad-level.

All values were log10 scaled in all analyses unless stated otherwise. This scaling was motivated by the inherent log-normality of these variables, as well as by the goodness-of-fit obtained for relationships in the log10 space compared with the linear space.

All results are reported as medians followed by the 95% bootstrap confidence limits (computed using 10,000 iterations) reported inside of square brackets. Accordingly, figures depict the medians with error bars reflecting the 95% bootstrap confidence limits. Comparisons between conditions or against null models were performed using the nonparametric Wilcoxon signed-rank test, unless specified otherwise. The threshold for significance was set at 0.05, unless indicated otherwise in the text. Bonferroni–Holm corrections were used in cases of multiple comparisons.

ACKNOWLEDGMENTS

We thank Benjamin Dann, Colleen Hughes, and Thomas Kreuz for helpful comments and discussion.

SUPPORTING INFORMATION

Supporting Information for this article is available at https://doi.org/10.1162/netn_a_00141.

ROLE INFORMATION

Samantha Sherrill: Conceptualization; Formal analysis; Funding acquisition; Investigation; Methodology; Project administration; Software; Validation; Visualization; Writing - Original Draft; Writing - Review & Editing. Nicholas Timme: Data curation; Resources; Software; Validation; Writing - Review & Editing. John Beggs: Conceptualization; Data curation; Funding acquisition; Methodology; Project administration; Resources; Software; Supervision; Validation; Writing - Review & Editing. Ehren Newman: Conceptualization; Funding acquisition; Methodology; Project administration; Resources; Supervision; Validation; Visualization; Writing - Original Draft; Writing - Review & Editing.

FUNDING INFORMATION

Ehren Newman, Whitehall Foundation (http://dx.doi.org/10.13039/100001391), Award ID: 17-12-114. John Beggs, National Science Foundation (http://dx.doi.org/10.13039/501100008982), Award ID: 1429500. John Beggs, National Science Foundation, Award ID: 1513779. Samantha Sherrill, National Science Foundation (http://dx.doi.org/10.13039/100000001), Award ID: 1735095. Samantha Sherrill, Indiana Space Grant Consortium.

Supplementary Material

TECHNICAL TERMS

Information processing:

The transfer, storage, and computation (or? modification) of information.

Correlated activity:

The synchronous spiking of two neurons.

Information transmission:

Here we use this terminology to refer to the multivariate transfer entropy obtained when two neurons predict a third.

Redundancy:

A measure that quantifies the overlapping information gained about a third neuron by knowing the spiking of two neurons.

Timescale:

The range of time in which delays between spiking neurons were considered.

Organotypic culture:

A cell culture derived from tissue—that retains many of the structural and functional properties of the intact tissue.

Information:

The reduction in uncertainty, typically measured in bits.

Synergy:

A measure that quantifies the information gained about a third neuron by knowing the spiking of two neurons jointly.

Effective connectivity:

Time-directed statistical dependencies of one spiking neuron on another.

Transfer entropy:

A measure that quantifies the amount of directed information transfer between two spiking neurons.

Mutual information:

A symmetric measure that quantifies the amount of information one spiking neuron has about another.

Computation:

The integration of multiple sources of information to produce a unique output.

Information transfer:

The transfer of unmodified information from one spiking neuron to another.

Contributor Information

Samantha P. Sherrill, Department of Psychological and Brain Sciences and Program in Neuroscience, Indiana University Bloomington, Bloomington, IN, USA.

Nicholas M. Timme, Department of Psychology, Indiana University-Purdue University Indianapolis, Indianapolis, IN, USA

John M. Beggs, Department of Physics & Program in Neuroscience, Indiana University Bloomington, Bloomington, IN, USA

Ehren L. Newman, Department of Psychological and Brain Sciences and Program in Neuroscience, Indiana University Bloomington, Bloomington, IN, USA

REFERENCES

  1. Atick J. J., & Redlich A. N. (1990). Towards a theory of early visual processing. Neural Computation, 2(3), 308–320. [Google Scholar]
  2. Atick J. J., & Redlich A. N. (1992). What does the retina know about natural scenes? Neural Computation, 4(2), 196–210. [Google Scholar]
  3. Attneave F. (1954). Some informational aspects of visual perception. Psychological Review, 61(3), 183. [DOI] [PubMed] [Google Scholar]
  4. Averbeck B. B., & Lee D. (2004). Coding and transmission of information by neural ensembles. Trends in Neurosciences, 27(4), 225–230. [DOI] [PubMed] [Google Scholar]
  5. Azouz R., & Gray C. M. (2003). Adaptive coincidence detection and dynamic gain control in visual cortical neurons in vivo. Neuron, 37(3), 513–523. [DOI] [PubMed] [Google Scholar]
  6. Barlow H. B. (1961). Possible principles underlying the transformation of sensory messages. Sensory Communication, 1, 217–234. [Google Scholar]
  7. Barlow H. (2001). Redundancy reduction revisited. Network: Computation in Neural Systems, 12(3), 241–253. [PubMed] [Google Scholar]
  8. Bastos A. M., Vezoli J., Bosman C. A., Schoffelen J. M., Oostenveld R., Dowdall J. R., De Weerd P., Kennedy H., & Fries P. (2015). Visual areas exert feedforward and feedback influences through distinct frequency channels. Neuron, 85(2), 390–401. [DOI] [PubMed] [Google Scholar]
  9. Beggs J. M., & Plenz D. (2004). Neuronal avalanches are diverse and precise activity patterns that are stable for many hours in cortical slice cultures. Journal of Neuroscience, 24(22), 5216–5229. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Berry M. J., Warland D. K., & Meister M. (1997). The structure and precision of retinal spike trains. Proceedings of the National Academy of Sciences, 94(10), 5411–5416. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bertschinger N., Rauh J., Olbrich E., Jost J., & Ay N. (2014). Quantifying unique information. Entropy, 16(4), 2161–2183. [Google Scholar]
  12. Bettinardi R. G., Tort-Colet N., Ruiz-Mejias M., Sanchez-Vives M. V., & Deco G. (2015). Gradual emergence of spontaneous correlated brain activity during fading of general anesthesia in rats: Evidences from fMRI and local field potentials. NeuroImage, 114, 185–198. 10.1016/j.neuroimage.2015.03.037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Bolz J., Novak N., Götz M., & Bonhoeffer T. (1990). Formation of target-specific neuronal projections in organotypic slice cultures from rat visual cortex. Nature, 346(6282), 359. [DOI] [PubMed] [Google Scholar]
  14. Buzsáki G., & Mizuseki K. (2014). The log-dynamic brain: how skewed distributions affect network operations. Nature Reviews Neuroscience, 15(4), 264. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Caeser M., Bonhoeffer T., & Bolz J. (1989). Cellular organization and development of slice cultures from rat visual cortex. Experimental Brain Research, 77(2), 234–244. [DOI] [PubMed] [Google Scholar]
  16. Cang J., Rentería R. C., Kaneko M., Liu X., Copenhagen D. R., & Stryker M. P. (2005). Development of precise maps in visual cortex requires patterned spontaneous activity in the retina. Neuron, 48(5), 797–809. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Chechik G., Anderson M. J., Bar-Yosef O., Young E. D., Tishby N., & Nelken I. (2006). Reduction of information redundancy in the ascending auditory pathway. Neuron, 51(3), 359–368. [DOI] [PubMed] [Google Scholar]
  18. Chiappalone M., Bove M., Vato A., Tedesco M., & Martinoia S. (2006). Dissociated cortical networks show spontaneously correlated activity patterns during in vitro development. Brain Research, 1093(1), 41–53. [DOI] [PubMed] [Google Scholar]
  19. Cohen M. R., & Kohn A. (2011). Measuring and interpreting neuronal correlations. Nature Neuroscience, 14(7), 811. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Cover T. M., & Thomas J. A. (1991). Entropy, relative entropy and mutual information. Elements of Information Theory, 2, 1–55. [Google Scholar]
  21. Dan Y., Alonso J. M., Usrey W. M., & Reid R. C. (1998). Coding of visual information by precisely correlated spikes in the lateral geniculate nucleus. Nature Neuroscience, 1(6), 501. [DOI] [PubMed] [Google Scholar]
  22. Doi E., Gauthier J. L., Field G. D., Shlens J., Sher A., Greschner M., … Litke A. M. (2012). Efficient coding of spatial information in the primate retina. Journal of Neuroscience, 32(46), 16256–16264. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Faber S. P., Timme N. M., Beggs J. M., & Newman E. L. (2019). Computation is concentrated in rich clubs of local cortical networks. Network Neuroscience, 3(2), 384–404. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Fagerholm E. D., Scott G., Shew W. L., Song C., Leech R., Knöpfel T., & Sharp D. J. (2016). Cortical entropy, mutual information and scale-free dynamics in waking mice. Cerebral Cortex, 26(10), 3945–3952. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Field D. J. (1987). Relations between the statistics of natural images and the response properties of cortical cells. JOSA A, 4(12), 2379–2394. [DOI] [PubMed] [Google Scholar]
  26. Fries P. (2005). A mechanism for cognitive dynamics: Neuronal communication through neuronal coherence. Trends in Cognitive Sciences, 9(10), 474–480. [DOI] [PubMed] [Google Scholar]
  27. Fries P. (2015). Rhythms for cognition: Communication through coherence. Neuron, 88(1), 220–235. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Gawne T. J., & Richmond B. J. (1993). How independent are the messages carried by adjacent inferior temporal cortical neurons? Journal of Neuroscience, 13(7), 2758–2771. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Grammont F., & Riehle A. (1999). Precise spike synchronization in monkey motor cortex involved in preparation for movement. Experimental Brain Research, 128(1–2), 118–122. [DOI] [PubMed] [Google Scholar]
  30. Götz M., & Bolz J. (1992). Formation and preservation of cortical layers in slice cultures. Journal of Neurobiology, 23(7), 783–802. [DOI] [PubMed] [Google Scholar]
  31. Gray R. M., & Shields P. C. (1977). The maximum mutual information between two random processes. Information and Control, 33(4), 273–280. [Google Scholar]
  32. Gutnisky D. A., & Dragoi V. (2008). Adaptive coding of visual information in neural populations. Nature, 452(7184), 220. [DOI] [PubMed] [Google Scholar]
  33. Hanslmayr S., Staudigl T., & Fellner M. C. (2012). Oscillatory power decreases and long-term memory: The information via desynchronization hypothesis. Frontiers in Human Neuroscience, 6, 74. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Hanslmayr S., Staresina B. P., & Bowman H. (2016). Oscillations and episodic memory: Addressing the synchronization/desynchronization conundrum. Trends in Neurosciences, 39(1), 16–25. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Hasselmo M. E., Bodelón C., & Wyble B. P. (2002). A proposed function for hippocampal theta rhythm: Separate phases of encoding and retrieval enhance reversal of prior learning. Neural Computation, 14(4), 793–817. [DOI] [PubMed] [Google Scholar]
  36. Hernandez J., Cooper K., & Newman E. L. (2020). Medial entorhinal cortex activates in a traveling wave in the rat. eLife, 9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Honey C. J., Newman E. L., & Schapiro A. C. (2017). Switching between internal and external modes: A multiscale learning principle. Network Neuroscience, 1(4), 339–356. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Ikegaya Y., Aaron G., Cossart R., Aronov D., Lampl I., Ferster D., & Yuste R. (2004). Synfire chains and cortical songs: temporal modules of cortical activity. Science, 304(5670), 559–564. [DOI] [PubMed] [Google Scholar]
  39. Ito S., Yeh F. C., Hiolski E., Rydygier P., Gunning D. E., Hottowy P., Timme N., Litke A. M., & Beggs J. M. (2014). Large-scale, high-resolution multielectrode-array recording depicts functional network differences of cortical and hippocampal cultures. PloS One, 9(8), e105324. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Johnson A., Fenton A. A., Kentros C., & Redish A. D. (2009). Looking for cognition in the structure within the noise. Trends in Cognitive Sciences, 13(2), 55–64. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Klostermann O., & Wahle P. (1999). Patterns of spontaneous activity and morphology of interneuron types in organotypic cortex and thalamus–cortex cultures. Neuroscience, 92(4), 1243–1259. [DOI] [PubMed] [Google Scholar]
  42. Laughlin S. B. (1989). The role of sensory adaptation in the retina. Journal of Experimental Biology, 146(1), 39–62. [DOI] [PubMed] [Google Scholar]
  43. Lisman J. E., & Jensen O. (2013). The theta-gamma neural code. Neuron, 77(6), 1002–1016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Litke A. M., Bezayiff N., Chichilnisky E. J., Cunningham W., Dabrowski W., Grillo A. A., … Kalmar R. S. (2004). What does the eye tell the brain?: Development of a system for the large-scale recording of retinal output activity. IEEE Transactions on Nuclear Science, 51(4), 1434–1440. [Google Scholar]
  45. Lizier J. T., Heinzle J., Horstmann A., Haynes J., & Prokopenko M. (2011). Multivariate information-theoretic measures reveal directed information structure and task relevant changes in fMRI connectivity. Journal of Computational Neuroscience, 30, 85–107. [DOI] [PubMed] [Google Scholar]
  46. Lizier J. T., Bertschinger N., Jost J., & Wibral M. (2018). Information decomposition of target effects from Multi-Source interactions: Perspectives on previous, current and future work. Entropy, 20(4), 307. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Minces V., Pinto L., Dan Y., & Chiba A. A. (2017). Cholinergic shaping of neural correlations. Proceedings of the National Academy of Sciences, 114(22), 5725–5730. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Montani F., Kohn A., Smith M. A., & Schultz S. R. (2007). The role of correlations in direction and contrast coding in the primary visual cortex. Journal of Neuroscience, 27(9), 2338–2348. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Newman E. L., & Hasselmo M. E. (2014). Grid cell firing properties vary as a function of theta phase locking preferences in the rat medial entorhinal cortex. Frontiers in Systems Neuroscience, 8, 193. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Nigam S., Shimono M., Ito S., Yeh F. C., Timme N., Myroshnychenko M., … Masmanidis S. C. (2016). Rich-club organization in effective connectivity among cortical neurons. Journal of Neuroscience, 36(3), 670–684. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Nigam S., Pojoga S., & Dragoi V. (2019). Synergistic codingof visual information in columnar networks. Neuron. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Nirenberg S., Carcieri S. M., Jacobs A. L., & Latham P. E. (2001). Retinal ganglion cells act largely as independent encoders. Nature, 411(6838), 698. [DOI] [PubMed] [Google Scholar]
  53. Norman K. A., Newman E., Detre G., & Polyn S. (2006). How inhibitory oscillations can train neural networks and punish competitors. Neural Computation, 18(7), 1577–1610. [DOI] [PubMed] [Google Scholar]
  54. Pica G., Piasini E., Chicharro D., & Panzeri S. (2017). Invariant components of synergy, redundancy, and unique information among three variables. Entropy, 19(9), 451. [Google Scholar]
  55. Plenz D., & Aertsen A. (1996). Neural dynamics in cortex-striatum co-cultures—II. Spatiotemporal characteristics of neuronal activity. Neuroscience, 70(4), 893–924. [DOI] [PubMed] [Google Scholar]
  56. Poulet J. F., & Petersen C. C. (2008). Internal brain state regulates membrane potential synchrony in barrel cortex of behaving mice. Nature, 454(7206), 881. [DOI] [PubMed] [Google Scholar]
  57. Puchalla J. L., Schneidman E., Harris R. A., & Berry M. J. (2005). Redundancy in the population code of the retina. Neuron, 46(3), 493–504. [DOI] [PubMed] [Google Scholar]
  58. Raichle M. E. (2010). Two views of brain function. Trends in Cognitive Sciences, 14(4), 180–190. [DOI] [PubMed] [Google Scholar]
  59. Reich D. S., Mechler F., & Victor J. D. (2001). Independent and redundant information in nearby cortical neurons. Science, 294(5551), 2566–2568. [DOI] [PubMed] [Google Scholar]
  60. Reinagel P., & Reid R. C. (2000). Temporal coding of visual information in the thalamus. Journal of Neuroscience, 20(14), 5392–5400. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Riehle A., Grün S., Diesmann M., & Aertsen A. (1997). Spike synchronization and rate modulation differentially involved in motor cortical function. Science, 278(5345), 1950–1953. [DOI] [PubMed] [Google Scholar]
  62. Rieke F., Bodnar D. A., & Bialek W. (1995). Naturalistic stimuli increase the rate and efficiency of information transmission by primary auditory afferents. Proceedings, of the Royal Society of London. Series B: Biological Sciences, 262(1365), 259–265. [DOI] [PubMed] [Google Scholar]
  63. Salinas E., & Sejnowski T. J. (2001). Correlated neuronal activity and the flow of neural information. Nature Reviews Neuroscience, 2(8), 539. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Schneidman E., Bialek W., & Berry M. J. (2003). Synergy, redundancy, and independence in population codes. Journal of Neuroscience, 23(37), 11539–11553. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Schreiber T. (2000). Measuring information transfer. Physical Review Letters, 85, 461–464. [DOI] [PubMed] [Google Scholar]
  66. Shadlen M. N., & Newsome W. T. (1998). The variable discharge of cortical neurons: Implications for connectivity, computation, and information coding. Journal of Neuroscience, 18(10), 3870–3896. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Shannon C. E., & Weaver W. (1949). The mathematical theory of information. [Google Scholar]
  68. Song S., Sjöström P. J., Reigl M., Nelson S., & Chklovskii D. B. (2005). Highly nonrandom features of synaptic connectivity in local cortical circuits. PLoS Biology, 3(3), e68. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Stramaglia S., Wu G. R., Pellicoro M., & Marinazzo D. (2012). Expanding the transfer entropy to identify information circuits in complex systems. Physical Review E, 86. [DOI] [PubMed] [Google Scholar]
  70. Timme N., Ito S., Myroshnychenko M., Yeh F. C., Hiolski E., Litke A. M., & Beggs J. M. (2014). Multiplex networks of cortical and hippocampal neurons revealed at different timescales. BMC Neuroscience, 15(1), P212. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Timme N. M., Ito S., Myroshnychenko M., Nigam S., Shimono M., Yeh F. C., Hottowy P., Litke A. M., & Beggs J. M. (2016). High-degree neurons feed cortical computations. PLoS Computational Biology, 12(5), e1004858. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Tozzi A., Zare M., & Benasich A. A. (2016). New perspectives on spontaneous brain activity: Dynamic networks and energy matter. Frontiers in Human Neuroscience, 10, 247 10.3389/fnhum.2016.00247 [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Tsodyks M., Kenet T., Grinvald A., & Arieli A. (1999). Linking spontaneous activity of single cortical neurons and the underlying functional architecture. Science, 286(5446), 1943–1946. [DOI] [PubMed] [Google Scholar]
  74. van Hateren J. H. (1992). A theory of maximizing sensory information. Biological Cybernetics, 68(1), 23–29. [DOI] [PubMed] [Google Scholar]
  75. van Winsun W., Sergeant J., & Geuze R. (1984). The functional significance of event-related desynchronization of alpha rhythm in attentional and activating tasks. Electroencephalography and Clinical Neurophysiology, 58(6), 519–524. [DOI] [PubMed] [Google Scholar]
  76. Vinck M., Batista-Brito R., Knoblich U., & Cardin J. A. (2015). Arousal and locomotion make distinct contributions to cortical activity patterns and visual encoding. Neuron, 86(3), 740–754. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Ward L. M. (2003). Synchronous neural oscillations and cognitive processes. Trends in Cognitive Sciences, 7(12), 553–559. [DOI] [PubMed] [Google Scholar]
  78. Wibral M., Pampu N., Priesemann V., Siebenhühner F., Seiwert H., Lindner M., Lizier J. T., & Vicente R. (2013). Measuring information-transfer delays. PloS One, 8(2). [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Wibral M., Finn C., Wollstadt P., Lizier J., & Priesemann V. (2017). Quantifying information modification in developing neural networks via partial information decomposition. Entropy, 19(9), 494. [Google Scholar]
  80. Williams P. L., & Beer R. D. (2010). Nonnegative decomposition of multivariate information. arXiv preprint arXiv:1004.2515. [Google Scholar]
  81. Williams P. L., & Beer R. D. (2011). Generalized measures of information transfer. arXiv preprint arXiv:1102.1507. [Google Scholar]
  82. Yu S., Huang D., Singer W., & Nikolić D. (2008). A small world of neuronal synchrony. Cerebral Cortex, 18(12), 2891–2901. [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Zohary E., Shadlen M. N., & Newsome W. T. (1994). Correlated neuronal discharge rate and its implications for psychophysical performance. Nature, 370(6485), 140. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials


Articles from Network Neuroscience are provided here courtesy of MIT Press

RESOURCES