Skip to main content

This is a preprint.

It has not yet been peer reviewed by a journal.

The National Library of Medicine is running a pilot to include preprints that result from research funded by NIH in PMC and PubMed.

bioRxiv logoLink to bioRxiv
[Preprint]. 2025 Jul 19:2025.07.15.664935. [Version 1] doi: 10.1101/2025.07.15.664935

Hierarchical Neural Circuit Theory of Normalization and Inter-areal Communication

Asit Pal 1,2,*, Shivang Rawat 2,3, David J Heeger 4,5,, Stefano Martiniani 1,2,3,5,
PMCID: PMC12338607  PMID: 40791322

Summary

The primate brain exhibits a hierarchical, modular architecture with conserved microcircuits executing canonical computations across reciprocally connected cortical areas. Though feedback connections are ubiquitous, their functions remain largely unknown. To investigate the role of feedback, we present a hierarchical neural circuit theory with feedback connections that dynamically implements divisive normalization across its hierarchy. In a two-stage instantiation (V1V2), increasing feedback from V2 to V1 amplifies responses in both areas, more so in the higher cortical area, consistent with experiments. We analytically derive power spectra (V1) and coherence spectra (V1-V2), and validate them against experimental observations: peaks in both spectra shift to higher frequencies with increased stimulus contrast, and power decays as 1/f4 at high frequencies (f). The theory further predicts distinctive spectral signatures of feedback and input gain modulation. Crucially, the theory offers a unified view of inter-areal communication, with emergent features commensurate with empirical observations of both communication subspaces and inter-areal coherence. It admits a low-dimensional communication subspace, where inter-areal communication is lower-dimensional than within-area communication, and frequency bands characterized by high inter-areal coherence. It further predicts that: i) increasing feedback strength enhances inter-areal communication and diminishes within-area communication, without altering the subspace dimensionality; ii) high-coherence frequencies are characterized by stronger communication (ability to estimate neural activity in one brain area from neural activity in another brain area) and reduced subspace dimensionality. Finally, a three-area (V1V4 and V1V5) instantiation of the theory demonstrates that differential feedback from higher to lower cortical areas dictates their dynamic functional connectivity. Altogether, our theory provides a robust and analytically tractable framework for generating experimentally-testable predictions about normalization, inter-areal communication, and functional connectivity.

In brief

A hierarchical recurrent neural circuit theory with reciprocal connections, that dynamically implements divisive normalization, predicts key experimental findings. The theory offers a unified view of inter-areal communication in which both communication through subspace and communication through coherence are emergent properties, predicting that high-coherence frequencies exhibit stronger inter-areal communication and reduced subspace dimensionality.

Introduction

The primate brain is a complex hierarchical network of interconnected cortical areas that perform computations essential for sensorimotor processing and cognitive functions. Among these, certain computations are repeated across cortical areas and are thus said to be “canonical computations”. Divisive normalization [1, 2] is one such fundamental canonical principle.

Feedback connections are ubiquitous in the primate brain [35], yet their precise roles in neural computation and communication remain largely unknown [68]. They have been hypothesized to serve various functions, including attentional modulation [911], inter-areal communication [12], learning [13, 14], and generative modeling [15]. Although divisive normalization has been extensively studied in feedforward recurrent networks [1619], its implementation and effects in hierarchical networks with feedback connections remain unexplored.

Effective communication between brain areas is essential for complex cognitive tasks, yet the precise mechanisms of communication remain an active area of investigation. One prominent hypothesis, “communication through coherence” (CTC) [20, 21], suggests that synchronized oscillations (coherence) across brain regions during diverse cognitive, perceptual, and behavioral tasks are crucial for effective inter-areal communication [2226]. However, the causal nature of the relationship remains a key unresolved question. It is currently debated whether coherence is a fundamental mechanism enabling interareal communication or rather an epiphenomenon arising from it [2731]. Another prominent alternate hypothesis proposes that brain areas communicate through a low-dimensional “communication subspace (CS)” embedded within the high-dimensional neural activity [3235]. This concept, also identified in artificial neural networks [36], posits that only activity within these shared subspaces between the source and target areas facilitates communication. Currently, a comprehensive theoretical framework capable of simultaneously exploring these different communication mechanisms and explaining their dynamic modulation by factors such as feedback and stimulus contrast is lacking.

To address these unresolved questions, we introduce a novel hierarchical recurrent neural circuit theory (Fig. 1) built on two a priori constraints: i) a hierarchy of brain areas with recurrent, feedforward, and feedback connections, and ii) recurrently implemented normalization within each cortical area. Based on the theory, we derive analytical expressions to compute inter-areal coherence and communication subspaces, offering experimentally testable predictions about the effects of both input gain and feedback strength on inter-areal communication. As a consequence of these two a priori constraints, our theoretical framework gives rise to emergent properties that explain a wide spectrum of neurophysiological phenomena, including modulation of contrast response functions, oscillatory dynamics evident in both power and coherence spectra, communication subspaces, and the dynamic control of functional connectivity. A central contribution of this work is to present a unified view of inter-areal communication. The theory simultaneously predicts empirical observations of both communication subspaces and interareal coherence, and reveals that coherence shapes the communication subspace.

Figure 1: Schematic of the two-stage recurrent circuit model.

Figure 1:

Solid and dashed lines represent excitatory and inhibitory connections, respectively. Principal excitatory neurons are denoted by y, modulatory excitatory neurons by u, and modulatory inhibitory neurons by a and q. Subscripts 1 and 2 designate neurons in visual areas V1 and V2, respectively. z1 denotes the input drive from the LGN to V1, a weighted sum of the responses of LGN neurons. Parameters β1 and γ1 modulate the input and feedback gain to V1, respectively.

Note that throughout our analysis, we neither adjust model parameters (such as time constants, weight matrices, and noise amplitudes) to reproduce specific experimental trends nor fit data when making direct comparisons. All results presented were obtained using a single set of model parameters.

Results

Theory

Initially conceived to explain the responses in the primary visual cortex (V1) [1], the normalization model has emerged as a canonical computation across sensory and cognitive domains [2], including higher level vision [37, 38], olfaction [39], audition [40, 41], somatosensation [42], attention [4355], multisensory integration [5660], working memory [6164], and decision-making [6568]. At its core, the normalization model proposes that the response of an individual neuron is scaled by the collective activity of its neighboring neurons (viz., the normalization pool), analogous to normalizing the length of a vector. Functionally, this computation serves as an automatic gain control mechanism, enabling neurons to operate effectively over a broad dynamic range of inputs [1, 69, 70]. It also simplifies the interpretation of neural signals by downstream brain regions [7072] and maintains response consistency across different input conditions [1, 39, 69, 70, 73]. In addition, normalization decorrelates population responses to reduce redundancy [39, 7477], prevents excessive recurrent excitation to maintain network stability, [18, 19], and fosters efficient learning [78, 79].

Normalization is a computational-level description of neural responses that can be implemented using various mechanisms. Experimental evidence suggests that normalization operates via recurrent amplification [8083], amplifying weak inputs more than strong inputs, consonant with evidence that cortical circuits are dominated by recurrent excitation [8489]. ORGaNICS is one such theoretical framework that dynamically implements divisive normalization by modulating recurrent amplification [16, 17]. Previous work on normalization models (including previous work on ORGaNICs) has focused on a single neural circuit within a brain region, such as V1, ignoring the abundant top-down feedback connections. Here, we develop a new hierarchical ORGaNICs theory with feedback in addition to feedforward and recurrent connections.

The hierarchical ORGaNICs theory describes neuronal population responses as dynamic processes evolving in a recurrent circuit, modeled by nonlinear differential equations (see Fig. 1 for illustration and Methods for equations). The circuit comprises distinct cell types: principal (excitatory), modulator (excitatory), and modulator (inhibitory) cells (n.b., “modulator ” refers to a multiplicative computation, not necessarily implemented with neuromodulators). Output firing rates of principal cells depend on three terms: 1) feedforward input drive modulated (viz., scaled) by input gain; 2) feedback drive modulated by feedback gain; 3) the sum of recurrent drive and scaled feedback drive, collectively modulated by recurrent gain. Unlike generic E-I models in which all neurons perform the same computation, ORGaNICs assign unique computations to each cell type. Because normalization ensures circuit stability over a broad range of parameters [18, 19], the theory is uniquely well suited to analytical treatment, e.g., to derive closed-form solutions for the power spectral density, coherence, and communication subspaces (see Methods and Appendix).

Each neuron’s input drive is a weighted sum of responses from a population of neurons at lower stages of the hierarchy. For example, LGN responses provide input drive to V1 (z1 in Fig. 1), and V1 responses similarly drive V2 (solid arrow from y1 to y2 in Fig. 1, representing principal neuron responses in each area). This weighting is defined by an embedding weight matrix composed of positive synaptic weights, which specify each neuron’s stimulus preferences, such as orientation selectivity (Fig. 1, V1 tuning curves), analogous to a standard receptive field model. Input gain is determined independently by input modulator cells (β1 and β2 in Fig.1).

The feedback drive is a weighted sum of the responses of a population of neurons in one or more brain areas from a higher stage in the hierarchy, e.g., V2 provides feedback to V1 (Fig. 1, solid arrow from y2 to y1). The feedback drive is specified by a feedback weight matrix. For simplicity, the feedback weights from V2 to V1 were chosen to be the transpose of the feedforward weight matrix from V1 to V2 [15]. For the exact structure of the different matrices, see Fig. 7. The feedback gain from V2 to V1 is denoted by γ1 in Fig. 1.

Figure 7: Connectivity matrices in the hierarchical recurrent circuit model.

Figure 7:

a, The V1 encoding matrix (Wzx), each row represents the tuning curve of a V1 neuron. b, The recurrent connectivity matrix for V1 (W11). The V2 recurrent connectivity matrix W22 has a similar structure. c, The Feedback connectivity matrix from V2 to V1 (W12). The corresponding feedforward matrix from V1 to V2, (W21), is its transpose. Since W21 is symmetric, the feedforward and feedback matrices are identical (W12 = W21).

The recurrent drive is a weighted sum of principal cell responses (Fig. 1, curved arrows from y1 onto itself and from y2 onto itself). For the results reported in this paper, the recurrent weights were selected to have a center-surround architecture in which the closest recurrent connections are excitatory and the more distant ones are inhibitory (Fig. 1, recurrent synaptic weights). The recurrent gain depends on the responses of recurrent modulator cells (a1 and a2 in Fig. 1), which in turn depend on: 1) the principal cell responses (y1 and y2 in Fig. 1) via excitatory (u1 and u2 in Fig. 1) and 2) inhibitory (q1 and q2 in Fig. 1) interneurons; 3) feedback from principal cells at a higher stage (solid arrow from y2 to a1 in Fig. 1). Normalization is achieved dynamically through the action of the modulator cells, which carry the “normalization signal”.

Figure 2: Comparison of theoretical predictions (left column) and experimental observations (right column) in V1 and V2.

Figure 2:

All theoretical predictions are generated using the same baseline parameters (Table 2). a, Theoretical mean firing rates as a function of stimulus contrast (V1: dashed line, V2: solid line). b, Experimental mean firing rates of V1 and V2 (data from [90] and [91], replotted for comparison). The shaded area with a solid border indicates the 25th to 75th percentile range for V1, and the one with the dashed border indicates the same for V2. c, Theoretical V1 power spectra at various stimulus contrast levels. Power spectra were normalized using the equation (powerbaseline)(power+baseline), where baseline power is taken to be spontaneous activity at 0% contrast. The peak power shifts towards higher frequency with increasing contrast. The inset shows the 1f4 power law decay at high frequency. d, Experimental power spectra from macaque V1 (data from [22], replotted for comparison). e, Theoretical coherence spectra between maximally firing neurons in V1 and V2. The peak coherence shifts towards higher frequency with increasing contrast. f, Experimental coherence spectra from macaque V1-V2 (data from [22], replotted for comparison). g, Theoretical communication subspaces, prediction performance as a function of dimensionality for inter-area (circles) and within-area (squares) communication. Simulation results indicate that the inter-area communication subspace is lower dimensional than the within-area communication. h, Experimental communication subspaces from macaque V1-V2 (data from [32], replotted for comparison). The plotted prediction performance in both panels g and h is an average across different subsets of the source and target neural populations, and the shaded areas represent the standard error of the mean (SEM).

Contrast response functions

The two-stage recurrent model with feedback implements divisive normalization exactly, across multiple cortical areas simultaneously (i.e., the model’s steady-state responses align precisely with the analytical steady-state solutions given by the normalization equation Eq. 2 when all the gains are set to γ = β = 1 and the recurrent weight matrix is identity, see Methods and Appendix). As shown in Fig. 2a, both V1 (dashed) and V2 (solid) exhibit the characteristic saturating, sigmoidal dependence on contrast. The model predicts a lower semisaturation constant (the contrast at which the firing rate reaches half-saturation) and a steeper slope for cortical area V2 compared to V1. This prediction is commensurate with experimental observations in the literature that used optimal stimulus sizes (Fig. 2b) [90, 91]. Similar hierarchical trends, specifically an increase in slope and a decrease in the semisaturation constant, have also been observed across other brain areas like the LGN, V1, and MT with progression up the visual hierarchy [92].

Increasing the feedback gain, γ1, enhances neural responses across cortical areas (Fig. 3a). However, this enhancement is not uniform across cortical areas. The gain change is notably higher for neurons in the higher cortical area (V2). Elevating input gain, β1, also enhances neural responses across both cortical areas (Fig. 3b). However, unlike the effects observed with feedback gain modulation, the response enhancement from changing input gain is nearly identical in both cortical areas.

Figure 3: Theoretical predictions: Modulating feedback (left) and input gain (right).

Figure 3:

a,b, Firing rates as a function of contrast. Increasing either feedback gain or input gain enhances neural responses, with feedback gain showing a greater impact in the higher cortical area V2. c,d, V1 power spectra: 3% contrast. An alpha peak is observed for low feedback gain, but the peak diminishes with increasing feedback gain. The alpha peak is absent with input gain changes. e,f, V1 power spectra: 50% contrast. A consistent gamma peak is observed, which shifts toward higher frequencies with increasing feedback gain (e) and input gain (f). g,h, Coherence spectra: 3% contrast. A broad peak in the beta band is observed at low feedback gain, which vanishes at high feedback gain. No such peak is observed with changes in input gain. i,j, Coherence spectra: 50% contrast. A beta peak is observed for low contrasts, which shifts toward higher (gamma) frequencies with increasing feedback gain. No beta peak is observed for changes in input gain, but the gamma peak shifts toward higher frequencies with increasing input gain. k,l, Communication subspaces. Increasing feedback gain enhances inter-areal (circles) communication while decreasing within-area (squares) communication. Conversely, increasing input gain decreases both inter- and within-area communication.

Power Spectra

Power spectra analysis of electroencephalographic (EEG) and local field potential (LFP) signals has been used extensively to identify and quantify oscillatory activity in various frequency bands (e.g., alpha, beta, gamma, theta), hypothesized to be linked to various cognitive functions, behavioral states, and neurological conditions [9396]. Given that our model has a fixed point solution and is analytically tractable, we can calculate the power and coherence spectra analytically, given a model of the noise in the circuit (see Methods).

We adopted an analytically tractable noise model that adds low-pass filtered synaptic noise to the membrane potential and multiplicative Gaussian white noise to the firing rate dynamics. Furthermore, we explicitly accounted for long-range correlated noise in LFP measurements. This type of noise, propagating through brain tissue, affects LFPs with negligible impact on spiking activity and firing rates [9799]. Further details are provided in the Appendix (“Intrinsic neural circuit noise” and “Extrinsic noise in the local field potential”).

Our theory successfully replicates key experimental observations of V1 power spectra across varying contrast levels. These theoretical V1 predictions (Fig. 2c) are compared with experimental spectra from macaque V1 (Fig. 2d) under identical contrast conditions [22]. The model accurately captures three critical experimental trends: 1) the power spectrum of the noise is shaped by recurrent normalization to exhibit a resonant peak in the gamma-frequency range (≈ 40-60 Hz); 2) the peak frequency shifts rightward with increasing contrast; and 3) the peak power decreases after reaching a maximum around 40% contrast [100102]. The oscillatory activity in the gamma-frequency range depends on normalization; removing normalization eliminates the peaks (see Suppl. Fig. 11e). The peak frequencies depend on the values of the intrinsic time constants (see Table 1); larger time constants would shift the peaks to lower frequencies (see Discussion).

Table 1:

Default values of the parameters

τ y1,u1,a1,q1 τ y2,u2,a2,q2 γ 1 β {1,2} α {1,2} σ {1,2}
1 ms 1 ms 1.0 1.0 10 0.07

The theory also reproduces the steep fall-off in power reported at higher frequencies [103], where power approximately scales as 1/f4 with increasing frequency ( f ) (Fig. 2c, inset). This occurs because the synaptic noise added to the membrane potentials is first low-pass filtered by the synapse’s own time constant, and then by the network model itself that acts as a low-pass filter (each step contributing a factor of 1/f2).

Modulation of feedback and input gains elicits distinct changes in the power spectra for a constant stimulus contrast. Figures 3cf illustrate these distinct effects on the normalized power spectra when feedback gain (left panels) or input gain (right panels) is varied. These comparisons are presented for both a low (3%; Figs. 3c,d) and a high (50%; Figs. 3e,f) stimulus contrast. See Supplementary Material for comprehensive results across other contrast levels (Suppl. Fig. 8).

The model predicts oscillatory activity in the alpha-frequency range (≈ 10 Hz) for low stimulus contrasts and small feedback gain values (Fig. 3c). As the feedback gain to the excitatory neurons increases, this alpha peak diminishes and ultimately vanishes at high feedback gain levels, while a broad peak concurrently emerges in the gamma frequency range (Figs. 3c,e). No such oscillatory activity in the alpha-frequency range is observed by manipulating the input gain, for any contrast (Figs. 3d,f).

The power spectra are dominated by oscillatory activity in the gamma-frequency range at high stimulus contrast (Figs. 3e,f). Increasing the feedback gain causes this gamma peak to shift towards higher frequencies. A similar shift of the gamma peak to higher frequencies is observed when the input gain is increased. However, a key distinction arises in their impact on the peak’s bandwidth: the gamma peak bandwidth narrows with increasing input gain but broadens with increasing feedback gain.

Coherence

Coherence measures the statistical correlation between two signals (e.g., LFPs in two different brain areas) as a function of frequency. Neuronal synchrony is hypothesized to contribute to the dynamic selection and routing of information both within and between brain areas [21, 104, 105].

Our model’s predictions for V1-V2 coherence spectra accurately capture key experimental trends observed across various stimulus contrasts. We analytically determined the coherence between the maximally firing neurons of V1 and V2 (see Methods). Figure 2e displays the theoretically predicted V1-V2 coherence spectra for different contrasts. These model predictions are compared with experimental V1-V2 coherence data from macaque monkeys, obtained under identical contrast conditions (Fig. 2f; [22]). The model accurately captures several crucial experimental observations: 1) high coherence at low frequencies; 2) a shift of the peak coherence toward higher frequencies with increasing contrast; and 3) a decrease in peak coherence after reaching a maximum around 40% contrast.

Feedback and input gain differentially modulate V1-V2 coherence across contrasts (Fig. 3gj). At low contrast (3%), a coherence peak in the beta band (12-30 Hz) is observed with low feedback gain (Fig. 3g); this peak disappears as feedback gain is increased. Notably, at low feedback gain, a dip in V1-V2 coherence occurs in the alpha band, corresponding to the frequencies where peak power is observed in V1. This suggests that alpha band activity, despite its prominence in V1 power, may not effectively propagate to V2 or participate in inter-areal communication. No such peaks are observed with changes in the input gain for low contrast (Fig. 3h). At high contrast (50%), increasing feedback gain shifts the peak coherence towards higher frequencies and increases both its magnitude and width (Fig. 3i). A similar frequency shift in peak coherence occurs with increasing input gain, but its magnitude and width decrease (Fig. 3j).

Communication subspaces

Neural communication is hypothesized to be channeled through “communication subspaces (CS)” [3235]. Although neural activity space is highdimensional, neural communication may not utilize this vast space. CS are lower-dimensional manifolds within the broader neural activity space that preferentially engage in inter-areal communication. In contrast, within-area communication utilizes a distinct, higher-dimensional “private” subspace [32].

Our theory successfully captures key characteristics of the CS (Figs. 2g,h). We derived an analytical model to examine both inter-areal (V1-V2) and within-area (V1-V1) communication (see Methods). We found that the CS is low-dimensional, much lower than the number of neurons (30 neurons for the analysis) in each brain area. In addition, the inter-areal communication (V1-V2) subspace is lowerdimensional than the within-area (V1-V1) communication subspace. Figure 2g illustrates the model prediction performance (i.e., the ability to predict activity fluctuations of a target subpopulation from the activity fluctuations of a source subpopulation using a linear model) as a function of the dimensionality of the CS. These theoretical predictions align well with experimental data (Fig. 2h; [32]).

Feedback and input gain differentially modulate communication subspaces (Fig. 3k,l). Increasing feedback gain from the higher cortical area (V2) to the lower cortical area (V1) enhanced inter-areal communication while diminishing within-area communication (Fig. 3k), without substantially altering the subspace dimensionality (see Supplementary Material, Fig. 10a,b). Conversely, increasing input gain reduced both the inter-area and within-area communication (Fig. 3l). These results are for 100% contrast; see Supplementary Material (Fig. 10c) for other contrasts. The observed enhancement of inter-areal communication with increased feedback suggests a key role for top-down feedback, to dynamically modulate the efficiency of information transfer between cortical areas without altering the underlying structure of the communication subspace (see below, Dynamic modulation of functional connectivity through feedback), a finding consistent with previous work on subspace dimensionality [33].

Frequency decomposition of communication subspaces

The theory predicts that the effectiveness of communication between brain areas varies systematically with frequency and contrast (Fig. 4). We quantified the efficacy of communication by analytically calculating the prediction performance for each frequency component (see Methods). To achieve this, we transformed the noise correlations of the source and target populations into power spectra using the Wiener-Khinchin theorem. This conversion enabled us to predict the power of the target population from the source population (using a linear readout) at each frequency, thus providing a measure of communication efficacy (prediction performance) as a function of frequency.

Figure 4: Theoretical prediction: Frequency-dependence of communication.

Figure 4:

a, V1-V2 coherence spectra for different contrasts; same as in Fig. 2e. b, Prediction performance versus frequency at different contrasts, averaged across different subsets of the source and target populations. At low stimulus contrast (≈ 10 %), a peak in communication efficacy is observed around 20 Hz. As stimulus contrast increases, the preferred frequency of communication shifts towards higher frequencies (40 Hz). The magnitude of communication reaches a maximum around 40% contrast, before declining at higher contrast levels. Under conditions of very low contrast (< 10%), communication is mostly concentrated at very low frequencies. These frequency-specific trends in communication parallel the V1-V2 coherence patterns shown in panel a. c, Dimensionality of the communication subspace versus frequency, averaged across different subsets of the source and target populations. A notable dip in dimensionality occurs at frequencies corresponding to peaks in communication (panel b) and coherence (panel a). The shaded areas in panels b and c represent the standard error of the mean (SEM).

V1-V2 communication is dynamically modulated by stimulus contrast in a frequency-dependent manner. At low stimulus contrasts (e.g, approximately 10%), a distinct peak in communication is observed at ≈ 20 Hz (Fig. 4b). As stimulus contrast increased, this preferred frequency for communication shifts toward higher frequencies (Fig. 4b). Furthermore, the magnitude of this communication peak exhibits a non-monotonic relationship with contrast, reaching its maximum at around 40% contrast before declining at higher contrasts. This particular effect is due to the normalization; removing normalization eliminates the peaks in both coherence and prediction performance (see Suppl. Fig. 11fg). Under conditions of very low contrast, no distinct peak in communication is observed; instead, communication is mostly concentrated at very low frequencies (<10 Hz). Notably, these observed frequency-specific trends in communication bear a striking resemblance to the V1-V2 coherence patterns (Fig. 4a). This parallelism indicates a relationship between the degree of synchronization of neural activity between brain areas and the effectiveness of communication between these brain regions, a prediction that can be tested experimentally (see Discussion).

The dimensionality of the inter-areal (V1-V2) CS also varies with frequency and stimulus contrast (Fig. 4c). The dimensionality of the communication subspace decreases at frequencies where inter-areal coherence and communication are maximal, for each contrast (Fig. 4c).

The dimensionality of the CS is given by the rank of the covariance matrix between the source (V1) and target (V2) subpopulations (see Methods: Interpretation of the results). The dimensionality of the covariance matrix is influenced by three main factors. First, the dimensionality remains low in the absence of normalization (Suppl. Fig. 11h). However, including normalization specifically reduces dimensionality at coherent frequencies (Fig. 4c). Second, the dimensionality is increased for a fully interconnected (all-to-all) connectivity between V1 and V2 (Suppl. Fig. 11l). Third, delocalized inputs (e.g., noise images or naturalistic images) lead to a higher dimensionality than localized inputs (e.g., gratings) (Suppl. Fig. 11p) as observed in experimental results [32].

Dynamic modulation of functional connectivity through feedback

Anatomical connections influence neural communication but do not completely determine it. Rather, it has been observed experimentally that functional connectivity between brain areas changes dynamically [106108].

Our theory predicts that functional connectivity is controlled by modulating feedback gain (Fig. 5). We expanded our model to encompass three cortical areas: V1, V4, and V5 (also called MT). In this expanded model, V1 has reciprocal connections with both V4 and V5, representing the parallel dorsal and ventral visual pathways, respectively [109111]. For simplicity, we assume there are no direct connections between V4 and V5 in the model. Analysis of this three-area model provides compelling evidence for the role of top-down feedback in controlling functional connectivity. When the feedback strength from V5 → V1 is greater than the feedback from V4 → V1, we observe that V1 strengthens communication with V5 (Fig. 5a). Conversely, when the feedback from V4 to V1 is stronger, V1 shows enhanced communication with V4 (Fig. 5b). This suggests that the relative strength of feedback from the higher cortical area (V5 or V4) can dictate the information flow from the lower cortical area (V1). One possible way this could be achieved is by selectively enhancing the neural activity in either V5 or V4, which will, in turn, increase the feedback to V1. This is commensurate with experimental evidence that feature-based attention and task demands may selectively boost activity in one or another cortical area [112118], and specifically that attending to motion increases the transfer of motion information (communication) between V1 and V5 [119121]. Therefore, top-down signals can dynamically modulate functional connectivity from lower to higher cortical areas, simply by adjusting the gain of neural activity in one higher cortical area (e.g., V4) relative to another (e.g., V5). This mechanism enables flexible cortical information routing, which is essential for cognitive flexibility.

Figure 5: Theoretical prediction: Feedback-dependent modulation of functional connectivity.

Figure 5:

a, Feedback from V5 → V1 is stronger than feedback from V4 → V1, and V1 preferentially enhances communication with V5. b, Feedback from V4 → V1 is stronger than feedback from V5 → V1, and V1 preferentially enhances communication with V4. The plotted prediction performance is an average across different subsets of the source and target populations, and the shaded areas represent the standard error of the mean (SEM). These results demonstrate that top-down feedback can dynamically route information flow between cortical areas, i.e., modulating functional connectivity.

Note that, to isolate the variable of interest, we modeled V4 and V5 with identical parameters for their recurrent, feedforward (V1 → V4/V5) and feedback (V1 ← V4/V5) connectivity. This ensures that the observed increase in communication with either V4 or V5 is a direct consequence of the stronger feedback from that specific area.

Robustness with respect to model parameters

All the results in this paper are obtained from the same set of parameters (Table 1) unless otherwise specified. To ensure the robustness of our findings, we verified them against a wide range of model parameters.

We systematically repeated the calculations and simulations described above, varying each parameter individually while holding others at their baseline values. These parameter ranges are specified in Table 2. Overall, our key findings remain robust across a wide range of parameter regimes without requiring precise parameter tuning:

  • Contrast-response functions (Fig. 2a). Higher cortical area (V2) consistently exhibits greater contrast gain, reflected by a lower semisaturation contrast and a steeper exponent than lower cortical area (V1).

  • Gamma oscillations (Fig. 2c,e). Both the power spectra and coherence spectra exhibit prominent gamma-band peaks that shift to higher frequencies as contrast increases.

  • Alpha oscillations (Fig. 3c). Low contrast inputs and weak feedback gain produce clear alphaband peaks in the power spectra.

  • Communication subspaces (Fig. 2g). The V1-V2 subspace is lower dimensional than the V1-V1 subspace.

  • Communication subspaces and feedback gain (Fig. 3k). The V1-V2 communication strength consistently increases with increasing feedback gain.

  • Communication subspaces and input gain (Fig. 3l). V1-V1 and V1-V2 communication strength consistently decrease with increasing input gain.

  • Frequency-dependence of communication subspaces (Fig. 4). The communication strength peaks at the same frequencies as coherence for each contrast level, and subspace dimensionality is minimized at those frequencies.

  • Dynamic functional connectivity (Fig. 5). Enhancing feedback from V4 to V1, selectively boosts communication between V1 and V4, and likewise, enhancing feedback from V5 to V1, selectively enhances communication between V1 and V5.

  • Oscillatory activity depends on normalization (Suppl. Fig. 11e,f). Removing normalization abolishes the gamma frequency peaks in both power and coherence spectra.

  • Dimensionality of inter-areal communication subspaces (Suppl. Fig. 11l,p). The dimensionality is low when the input drive is localized (e.g., gratings). However, delocalized input drives (e.g., naturalistic images) result in a higher dimensionality. Fully all-to-all connections between V1 and V2 also produce a higher subspace dimensionality.

Table 2:

Valid parameter ranges

Parameter Tested stability range
γ 1 [0.0 - 1.1]
β {1,2} [0.0 - 2.5]
α {1,2} [7.5 - 50]
σ {1,2} [0.035 - 0.14]

Discussion

We introduce an analytically tractable theory of neural circuit function that implements divisive normalization dynamically across multiple interconnected brain areas. Emergent properties of the theory account for a diverse array of neurophysiological phenomena, including oscillatory activity (measurable as peaks in LFP power and coherence spectra) and communication subspaces. Based on the theory, we propose a unified view of inter-areal communication with the following experimentally-testable predictions: i) increasing feedback strength should enhance inter-areal communication and diminish within-area communication, without altering the subspace dimensionality; and ii) frequencies exhibiting high coherence should correspond to both stronger communication and reduced subspace dimensionality. Furthermore, the theory proposes a mechanism for dynamic functional connectivity: top-down feedback signals enhance communication between brain areas simply by adjusting the gain of neural activity in one or more cortical areas. This modulation of neural gain offers a broader mechanism for dynamically changing the functional connections between brain areas over time.

Neural circuit function

The original ORGaNICs model [16, 17], in its initial form without feedback connections, serves as a biophysically-plausible counterpart to Long Short Term Memory units (LSTMs) [122], a class of artificial recurrent neural networks (RNNs) prevalent in machine-learning applications [123126]. Unlike LSTMs or any other previously proposed RNNs, ORGaNICs have built-in recurrent normalization, which offers several advantages, including stability in the network even in circuits with exuberant recurrent connections [18, 19].

The hierarchical version of the ORGaNICs theory introduces feedback connections, drawing a partial analogy to “hierarchical predictive coding” models of sensory and perceptual processing [127132]. However, the hierarchical model diverges from predictive coding in its fundamental mechanisms. In predictive coding, it is proposed that forward-propagating signals represent prediction errors, while feedback pathways convey predictions to lower levels of hierarchy. The goal is to minimize prediction error by matching sensory input with top-down predictions. In hierarchical ORGaNICs, instead of simply propagating prediction errors, the interplay between feedforward and feedback drives adjusts neuronal responses to a value that balances the feedforward processing of its inputs with the feedback (see Ref. [15] for details). This mechanism is more consistent with neurophysiological and psychophysical phenomena than predictive coding models. For instance, neurons maintain sustained activity in response to a predictable stimulus [133], whereas predictive coding theories would suggest that the activity, representing prediction error, should be diminished or “explained away” as the stimulus becomes predictable.

Normalization

The proposed theory has an analytical fixed point that follows the normalization equation (Eq. 2), for arbitrary (non-negative) normalization weights, when the input drive is constant over time (see Appendix for derivation) [18, 19]. The model achieves normalization dynamically via recurrent amplification, which is inversely regulated by the activity of inhibitory modulator cells. When the input is weak, the modulator cells have a small response, resulting in strong recurrent amplification. Conversely, a strong input generates a large response from modulator cells, which suppresses the amplification. This process modulates both excitatory and inhibitory recurrent signals, which is consistent with the experimental observation that normalization involves a decrease in both recurrent excitatory and inhibitory conductances [134]. Because the theory successfully implements normalization, it is commensurate with a wide range of physiological and psychophysical phenomena previously explained by the normalization model (see Introduction for references). The functional importance of this mechanism is further validated by our model. When divisive normalization is removed from the model, oscillatory peaks in both the power and coherence spectra are eliminated (Suppl. Fig. 11e,f). Also, the relationship between coherence, prediction performance, and subspace dimensionality is lost (Suppl. Fig. 11g,h).

Role of feedback

This paper provides an experimentally testable computational model of cortical function where neural activity is shaped by a combination of feedforward drive (bottom-up) and feedback drive (top-down). Feedback in the model may serve a variety of functions. First, feedback provides a means for controlling inter-areal communication (functional connectivity) simply by changing the gain of neural activity in one or more cortical areas (see below, Attention and functional connectivity). Second, the feedback drive may contribute to selective attention (see below, Attention and functional connectivity). Third, the relative strength of the input gain and feedback gain determines whether neural responses are driven bottom-up, top-down, or a combination of the two. Fourth, feedback may contribute to learning to minimize prediction error [15, 135]. In addition, feedback to inhibitory neurons is essential for maintaining network stability.

The interplay between the input gain and the feedback gain determines the primary driver of neural responses. If the input gain is strong and the feedback is weak, the system behaves like traditional feedforward models. Conversely, with strong feedback and weak input, the network is similar to generative models [136138], creating sensory representations from higher-level abstract knowledge, akin to recalling an image from memory (see Ref. [15] for details). When input and feedback are balanced, the framework integrates prior beliefs with sensory evidence, akin to models based on Bayesian inference [139145]. The model predicts differential effects of changing input gain versus feedback gain on the power, coherence spectra, and inter-areal communication (see Fig. 3). Hence, these metrics may be used as a signature for distinguishing changes in input gain from changes in feedback.

Unified view of inter-areal communication

Our theoretical framework offers a novel perspective on inter-areal communication. Rather than viewing “communication through coherence” (CTC) and “communication subspace” (CS) as alternative hypotheses, our theory unifies them, showing that coherence and communication subspaces are complementary empirical phenomena emerging naturally from the same underlying mechanism. We found that the frequency decomposition of communication subspaces closely mirrors the patterns of inter-areal coherence at different contrasts. Specifically, frequencies exhibiting high coherence are also those for which the communication efficacy within the subspace is maximal, and the subspace dimensionality is minimized. Consequently, coherence might dynamically shape the communication subspaces, for instance, by reducing their dimensionality at preferred frequencies [34]. Moreover, the predicted frequency-dependence of communication subspaces might serve as a marker for testing neural circuitry theories (such as ours).

Attention and functional connectivity

The theory provides a possible explanation for the efficient routing of sensory information between cortical areas, i.e., for changing functional connectivity. Lower cortical areas, such as V1, can flexibly relay information to specific higher cortical areas (e.g., V4 vs. V5) based on the relative strength of the feedback from each of the higher cortical areas. The strength of the feedback can be controlled either by changing the feedback gain or (perhaps more simply) by changing the gain of the neural responses in each of the higher cortical areas. We hypothesize that this feedback-mediated process may account for changes in functional connectivity [108, 146, 147] and feature-based attention [112116]. Consistent with this hypothesis, experimental evidence demonstrates that feature-based attention is spatially global, i.e., that neural activity is boosted in feature-selective neurons with receptive fields covering the visual field [113, 148].

The theory may also provide insights into the neural processes underlying spatially-selective attention. According to the theory, there are two processes by which attention may modulate the gain of neural activity: i) attention may modulate the input gain of the sensory neurons, following earlier work on the normalization model of attention [53]; ii) attention may modulate the activity of the sensory neurons via the feedback drive. One or the other of these two proposed mechanisms for attentional modulation (e.g., changing the input gain or the feedback drive) may offer a better fit to electrophysiology and behavioral data. Alternatively, one of the mechanisms (e.g., changing the input gain) may offer a better explanation of exogenous attention, whereas the other mechanism (e.g., the feedback drive) may offer a better explanation of endogenous attention. Importantly, the theory predicts different effects of changing input gain versus feedback gain on the power and coherence spectra, as noted above, so these metrics may be used as a signature for distinguishing changes in input gain from changes in feedback.

Thalamo-cortical model

Alpha oscillations are a defining characteristic of thalamo-cortical dynamics, particularly prominent within the visual system during states of “relaxed wakefulness” [149151]. Specifically, alpha oscillations are hypothesized to be generated and modulated by thalamo-cortical loops [151, 152], suggesting that alpha rhythms signify a state of functional inhibition or “idling” in sensory pathways, potentially gated by corticothalamic feedback [152, 153]. Our theory predicts the emergence of oscillatory activity in the alpha band (8-12 Hz) under conditions of low stimulus contrast and low feedback to excitatory neurons, and high feedback to inhibitory neurons (γ < 1, Figs. 3c,g). This aligns well with the existing literature demonstrating the role of alpha oscillations in mediating cortical inhibition [153, 154]. This positions the theory as a potential tool for modeling the thalamocortical interactions underlying alpha waves.

Furthermore, we hypothesize that the input gain modulator responses depend in part on thalamocortical loops, hence positioning the theory as a means for modeling the interplay between bottom-up sensory input and top-down feedback within the thalamocortical circuit, the gating of visual information flow, thalamic control of functional cortical connectivity [155] and thalamocortical loops in other neural systems [156160].

Limitations

We assume the feedback to be instantaneous, but there is an associated feedback delay [161, 162]. The delay in receiving feedback signals from higher cortical areas is due to the finite speed of signal conduction, processing, and computation. These delays arise from factors like axonal conduction times and synaptic transmission. It will be interesting to investigate how the synchronization between cortical areas changes as a function of feedback delay time.

The feedforward, feedback, and recurrent weight matrices may comprise positive and negative weights. Hence, we posit additional inhibitory interneurons to invert the sign of the responses, corresponding to negative weights in the synaptic weight matrices. These inhibitory neurons need not be one-to-one with their excitatory inputs. Rather, each may compute a partial weighted sum of their inputs corresponding to the negative terms in the matrix multiplications. Regardless, these inhibitory interneurons will introduce an additional delay.

Most of the computations posited by the theory (see Methods, Eqs. 1a, 1c, 1b, and 1d) are biophysically plausible (see below and Ref. [17]), but there are some elements of computation for which the underlying mechanisms are unclear. For example, the weighted sums may be implemented as a push-pull balanced increase in excitatory and decrease in inhibitory synaptic inputs (or vice versa) that injects current without changing conductance [163] and the gain control may be implemented with a push-push balanced increase in excitatory and inhibitory synaptic inputs that changes conductance without injecting current [163, 164]. However, we do not have a hypothesis for the cellular mechanism underlying the element-by-element product of firing rates y1+u1+2 in Eq. 1b.

The intrinsic time constants are unrealistically short for individual neurons (see Table 1). If more realistic time constants were used, the oscillatory activity and coherence in the gamma-frequency band would shift toward a lower frequency. Instead, one might interpret the time constants in the theory as effective time constants of a population of neurons rather than interpreting them literally as the intrinsic time constants (capacitance/conductance) of individual neurons.

Biology plausibility

We hypothesize a correspondence between the variables in the equations that describe the theory and cortical microcircuits composed of identified cell types. Unlike generic E-I models (e.g., Wilson and Cowan [165]), we propose that different cell types (different morphology and biochemistry) perform different computations, each with a different output nonlinearity (i.e., F-I curve) transforming membrane potentials to firing rates (see Appendix). Specifically, principal cells y may correspond to pyramidal cells, inhibitory modulatory cells a may correspond to a combination of parvalbumin-expressing (PV+) cells and somatostatin-expressing (SST+) cells, and inhibitory interneurons q may correspond to vasoactive intestinal polypeptide (VIP) cells. Principal cells y provide excitatory input to both inhibitory modulatory cells a (via u) and inhibitory interneurons q, inhibitory interneurons q provide inhibitory input to inhibitory modulatory cells a, and inhibitory modulatory cells a provide inhibitory input to principal cells y (Fig. 1). This arrangement is commensurate with empirical observations of cortical circuits: pyramidal cells provide excitatory input to PV+, SST+, and VIP cells, VIP cells provide inhibitory input to SST+ cells, and both PV+ and SST+ cells provide inhibitory input to pyramidal cells [166170]. The inhibitory modulator cells a are hypothesized to have large RFs and broad orientation-selectivity (reflecting properties of the normalization pool), consistent with the response properties of SST+ and PV+ neurons, respectively [167, 171, 172]. We also hypothesize excitatory feedback from neurons in higher cortical areas to both principal cells y and inhibitory modulatory cells a (Fig. 1). Feedback to principal cells may be received by the distal apical dendrites of pyramidal cells in layer 1 (see below). Feedback to inhibitory modulator cells is commensurate with experimental evidence that feedback from higher cortical areas targets SST+ neurons [173175], and that normalization depends in part on loops through higher visual cortical areas [176179]. A possible discrepancy between the model and the actual cortical circuit is that the excitatory input from principal cells y to inhibitory modulatory cells a passes through another excitatory cell type u whereas pyramidal cells provide direct excitatory input to PV+ and SST+ cells. One could, however, rewrite the equations for the theory (see Methods, Eqs. 1c and 1b) such that principal cells provide direct excitatory input to inhibitory modulatory cells (specifically by substituting the expression for u from Eq. 1b into the expression for a in Eq. 1c).

We hypothesize a correspondence between the computations performed by the principal cells (y) and the computational processes occurring within the dendritic compartments of pyramidal cells. The equation governing the responses of the principal cells (see Fig. 6 and Methods, Eq. 1a) expresses the following synaptic current: (input gain) (input drive) + (recurrent gain) [recurrent drive + (feedback gain) (feedback drive)]. The input drive may be computed as a weighted sum of feedforward inputs in the distal basal dendrites, commensurate with evidence that feedforward synaptic inputs contact basal dendrites. The input gain may correspond to inhibitory neurons that contact the most proximal trunk of the basal dendritic tree. The feedback drive may be computed as a weighted sum of feedback inputs received by distal apical dendrites in layer 1, which are known to receive signals from higher cortical areas. The feedback gain might be attributed to the amplification of the feedback drive within the apical trunk of pyramidal cells [180]. The recurrent drive may correspond to a weighted sum of lateral inputs at more proximal apical dendrites, and the recurrent gain may correspond to inhibitory inputs that contact the most proximal trunk of the apical dendritic tree.

Figure 6: Hypothesized mapping of computational components in the model onto the dendritic compartments of a pyramidal cell.

Figure 6:

The input drive is hypothesized to be the weighted sum of feedforward inputs arriving at the distal basal dendrites. This is modulated by an input gain, corresponding to inhibition at the proximal basal dendritic trunk. In the apical tuft, the feedback drive represents a weighted sum of feedback inputs from higher cortical areas. This signal is amplified by a feedback gain within the distal apical trunk. The recurrent drive corresponds to the sum of lateral inputs on the proximal apical dendrites. The combined recurrent and feedback signals are then modulated by a recurrent gain at the proximal apical trunk. The overall synaptic current is a combination of these gain-modulated drives. The figure is taken from [185].

Experimental evidence suggests that acetylcholine (ACh) selectively adjusts the balance between bottom-up sensory inputs and top-down signals related to expectancy and uncertainty [181]. It has also been hypothesized that ACh signals when bottom-up sensory inputs are known to be reliable [182, 183]. Consequently, it is reasonable to hypothesize that the input gain and/or feedback gain might be controlled (at least in part) by ACh. Although ACh is released broadly throughout the cortex, its effect can be regionally specific [184], possibly offering a mechanism for how the values of the state parameters can differ across the hierarchy of brain areas.

Methods

The model

We propose a multistage hierarchical recurrent circuit that incorporates feedforward connections from the primary visual cortex (V1) to higher visual cortical areas (e.g., V2, V4, MT) and feedback connections from these higher brain areas to V1 (Fig. 1). We refer to V2 throughout this paper for concreteness, but it is merely an example of a higher brain area that is reciprocally connected with V1. This model builds on the ORGaNICs framework developed by Heeger et al. [16, 17]. Divisive Normalization (DN) is achieved in each brain area by local inhibitory neuronal populations.

The response of principal excitatory neurons in the model is a sum of three key components: 1) input drive, a weighted sum of neural responses from a brain area that provides feedforward input (e.g., LGN → V1 and V1 → V2), modulated by input gain; 2) feedback drive, a weighted sum of neural responses from the higher cortical area (e.g., V1 ← V2 or V1 ← V4), modulated (viz., scaled) by feedback gain; and 3) the sum of recurrent drive—a weighted sum of neuronal responses from other neurons in the same brain area (e.g., V1 → V1 and V2 → V2)—and scaled feedback drive, collectively modulated by recurrent gain.

The proposed neural circuit comprises four distinct cell types, the principal excitatory neurons (y), the modulatory excitatory neurons (u), the modulatory inhibitory neurons (a and q), with the following dynamics:

τy1dy1dt=y1+β12z1+(11+a1+)(W11y1++γ12W12y2+) (1a)
τu1du1dt=u1+σ122+N1y1+u1+2 (1b)
τa1da1dt=a1+W12y2+2q1++u1++a1u1++α1du1dt (1c)
τq1dq1dt=q1+y1+. (1d)

The above equations are for V1; V2 follows a similar set of equations (see Appendix). The membrane potentials of neurons in cortical areas V1 and V2 are denoted by y1 and y2, respectively, while their corresponding firing rates are represented by y1+ and y2+. The symbol represents the Hadamard product, element-wise multiplication of vectors. To estimate the firing rates y1+ and y2+ from their respective membrane potentials, we employ a modified Gaussian Rectification (GR) model [186, 187]. The variable z1 denotes the input drive to V1 from the preceding brain area (the lateral geniculate nucleus of the thalamus, LGN), computed as a weighted (Wzx) sum of the LGN firing rates. τk is the intrinsic time constant for each subtype of neuron, where k=y,u,a,q. W11y1+ and W12y2+ are the recurrent and feedback drive, respectively, where W11 is the recurrent weight matrix for V1. W11 has a center-surround structure where the closer connections are excitatory and the more distant connections are inhibitory (Fig. 7b), same as in [17] where the feedforward ORGaNICs was introduced. The recurrent weight matrix for V2, W22, has a similar structure. The maximum eigenvalue for these recurrent weight matrices is set to unity to ensure stability. However, we have shown that this need not be the case because normalization ensures stability [18, 19]. W12 represents the interareal feedback connectivity matrix from V2 to V1, its entries are all non-negative, and the matrix is diagonally dominant (Fig. 7c). The feedforward matrix from V1 to V2, W21, is taken to be the transpose of W12. Given that W12 is symmetric, it follows that W12 = W21. The normalization matrix in V1 (N1) and V2 (N2) is taken to be all ones, meaning each neuron contributes equally to the normalization. The input gain and feedback gain for V1 are modulated by β1 and γ1, respectively. The recurrent and feedback amplifications in V1 are modulated dynamically by the inhibitory neuron population a1.

Both the V1 and V2 neurons follow the normalization equation exactly when β and γ are set to unity (i.e. when there is identical feedback to excitatory (y1) and inhibitory (a1) neurons) and W11 is the identity matrix (i.e., each neuron recurrently amplifies itself), but we have shown that the circuit closely approximates normalization for a wide range of recurrent weight matrices (W11) [19]. For V1 neurons, normalization is expressed as:

y1+=z12σ2+N1z12, (2)

and likewise for V2

y2+=z22σ2+N2z22, (3)

where z2=W21y1+ is the drive from V1 to V2, σ is the semi-saturation constant. Normalization is achieved dynamically via recurrent amplification (amplifying weak inputs but not strong inputs) with the three modulatory neurons, u1, a1 and q1 whose dynamics are governed by Eq. 1b, Eq. 1c, and Eq. 1d respectively. The equations for modulatory neurons are derived by substituting the normalization equation (Eq. 2) in the equation for the principal neurons (Eq. 1a) (see Appendix for derivation). An increase in the activity of principal neurons (y1) leads to enhanced activation in u1 and q1. This, in turn, drives a net increase of activity in modulatory neurons (a1). The elevated a1 activity subsequently exerts an inhibitory effect on y1, thereby implementing normalization by the recurrent interactions between the neural populations (see Appendix for derivation).

The parameter γ1 modulates the feedback gain to the principal excitatory neuron population (y1). This determines the relative feedback to the excitatory neurons (y1) compared to the inhibitory neurons (a1). As a rule of thumb, γ1 < 1 indicates greater feedback inhibition, γ1 = 1 denotes a balance between feedback excitation and inhibition, and γ1 > 1 signifies more feedback excitation. For simplicity, we have set the feedback to the inhibitory neurons (a1) at a fixed value of 0.50 (hence, the factor of 12 in equation Eq. 1c and Eq. 1a, see Appendix for details). Similarly, the parameter β1 modulates the relative input gain to the principal excitatory neuron population (y1) compared to the modulatory excitatory neuron population (u1). For simplicity, we set the input gain to the modulatory neurons (u1) to 0.50 (hence, the factor of 12 in equation Eq. 1a and Eq. 1b, see Appendix). Note that when β1 and γ1 are both set to unity and the recurrent matrix is identity, the network achieves excitatory-inhibitory balance exactly, and the normalization (Eq. 2) is followed precisely.

We modeled each brain area using a ring model, where principal neurons are arranged in a circle to represent continuous features like stimulus orientation. The recurrent connectivity between these neurons is determined by the distance between them (Fig. 7b). Each modeled area consists of 72 neurons for each of the four specified types (y1,u1,a1,q1). To simulate sensory input, a visual stimulus with a defined contrast and orientation is presented. The stimulus is then convolved with the tuning curves of the V1 neurons to provide the input drive to the V1 principal neurons. The tuning curves are modeled as raised cosine functions, each peaking at the neuron’s preferred orientation (Fig. 7a).

Power and Coherence Spectra

The time evolution of neural population activity in V1 and V2 can be described by a general stochastic differential equation (SDE) with an additive noise term LdW:

dx=f(x,u)dt+LdW, (4)

where xn is a vector representing the activity of neurons; f(x,u)n represents the deterministic part of the dynamics explained above (Eq. 1); dW is a vector of independent Gaussian increments with correlation matrix D𝔼dWdW, and Ln×n is the dispersion matrix defining how noise enters the system. Given the analytical solution for the fixed point of the dynamical system (Eq. 2), we can linearize the system around the fixed point (x*):

dx=Jx*xx*dt+LdW, (5)

where Jx* is the Jacobian computed at the fixed point x*. This linearization leads to a wide-sense stationary Gaussian process, x(t), whose power spectral density matrix, 𝓢(ω)n×n, is given by [188]:

𝓢(ω)=𝔼|xˆ(ω)|2=(ıωI+J)1LDL(ıωI+J). (6)

From 𝓢(ω), we can directly compute the coherence, κij, between any two i and j neurons in the two areas:

κij=𝒮ij2𝒮ii𝒮jj. (7)

Communication subspaces

Here we present the method used to characterize the communication subspaces between two subpopulations (source and target) of neurons [32, 162, 189, 190]. These subpopulations may comprise distinct neurons in the same brain area or across different brain areas. The basic idea is to determine how well the fluctuations in responses of a target subpopulation can be predicted using a linear transform (a weighted sum) of the fluctuations in responses of the source subpopulation, and to determine the dimensionality of this linear transform.

The fluctuations in the responses and their covariability are fully determined by the correlation matrix at steady-state when the strength of the noise is small. Therefore the correlation matrix, denoted as C(0), for the responses x(t) can be computed by solving the Lyapunov equation [188]:

JC(0)+C(0)J=LDL. (8)

Once C(0) is obtained, we follow the method proposed by Semedo et al. [32] to construct an analytical model of the communication subspace. To achieve this, we divide the mean-subtracted responses of the entire neuronal population, denoted as x, into two subsets: s (responses of the source neurons, e.g., in V1) and t (responses of the target neurons, e.g., in V2 or V1). We consider a linear model for predicting target responses using the source responses, tˆ=Bs, where the mean squared error (MSE) is given by:

ε=𝔼tBs2=TrC2+BC1B2BC3, (9)

where C1=𝔼ss,C2=𝔼tt, and C3=𝔼st. These matrices can be directly obtained from C(0). To find the optimal readout matrix Bopt, we minimize ε with respect to B, leading to the Ordinary Least Squares solution:

Bopt=C11C3, (10)

First, we define the covariance of the target activity predicted by the optimal model (note the “hats”), which we denote as Cˆ2:

Cˆ2=𝔼[tˆtˆ]=BoptC1Bopt (11)

By substituting the expression for Bopt, we arrive at a key relationship:

Cˆ2=C11C3C1C11C3=C3C11C3 (12)

The principal dimensions of this predicted covariance matrix, Cˆ2, define the communication subspace. These are found by computing the eigen-decomposition:

Cˆ2=VΛV (13)

Here, the columns of the orthogonal matrix V are the eigenvectors of Cˆ2 and Λ is a diagonal matrix of the corresponding eigenvalues, sorted in descending order. The prediction performance for a reduced-rank approximation of dimension i is defined as: A rank-i approximation, BRRR, is constructed by projecting the optimal solution Bopt onto the subspace spanned by the top i eigenvectors of Cˆ2.

BRRR(i)=BoptV(i)V(i)=BoptPi (14)

where V(i) is a matrix containing first i columns of V. Here Pi=V(i)V(i) is the projection matrix. Note that Pi is symmetric (Pi=Pi) and idempotent (Pi2=Pi). We define the prediction performance as:

Performance(i)=1εiε0, (15)

where εi is the MSE for the rank-i approximation

εi=TrC2+BRRR(i)C1BRRR(i)2BRRR(i)C3. (16)

ε0 is the MSE for the case where the readout matrix is a null matrix, corresponding to the total variance of the target population ε0=TrC2.

To simplify the expressions for εi. We substitute BRRR(i)=BoptPi into the expression:

εi=TrC2Tr2PiBoptC3PiBoptC1BoptPi (17)

Now, we substitute Bopt=C3C11 and recognize that the term BoptC1Bopt is Cˆ2. Using the cyclic property of the trace (Tr(AB)=Tr(BA)) and the idempotent property of the projection matrix (Pi2=Pi), we can simplify further:

εi=TrC2TrCˆ2Pi (18)

By applying the cyclic property and expanding Pi, we can see that the second term is the sum of the top i eigenvalues λj of Cˆ2.

TrCˆ2V(i)V(i)=TrV(i)Cˆ2V(i)=j=1iλj (19)

Now the expression for prediction performance simplifies to:

Performance(i)=j=1iλjTrC2, (20)

Which can also be interpreted as:

Performance(i)=Explained variance(idimensions)Total target variance. (21)

The maximum prediction performance (full-rank approximation) is given by:

Performance=TrCˆ2TrC2=TrC3C11C3TrC2 (22)

Interpretation of the results:

The prediction performance (Eq. 22) is governed by the interplay between the source (C1), target (C2), and cross-population (C3) covariances. The effectiveness of communication (indicated by a large predictive performance) can be understood through three key factors: 1) Source-target correlation: Prediction performance is directly enhanced by a stronger correlation between the source and target populations (i.e., larger C3 norm). 2) Signal reliability: Prediction performance is boosted by the reliability of the signals. Less noisy activity in source and target populations, indicated by smaller diagonal entries in C1 and C2, leads to better prediction performance. 3) Alignment: Prediction performance is maximized when the principal directions of source covariance (eigenvectors of C1) align with the principal directions of source-target covariance (left singular vectors of C3).

This framework also explains the results in the frequency domain. The key metric here is the cross-power spectrum, which is the Fourier transform of the cross-covariances, and coherence, which is the squared cross-power spectrum normalized by the product of the auto-spectra, Eq. 7. Frequencies that show high coherence, indicating high cross-covariances between source and target, the prediction performance is highest at these specific frequencies. Note that the dimensionality of the communication subspace is given by the rank of the matrix Cˆ2=C3C11C3. Since the source covariance matrix C1 is a full-rank matrix, the dimensionality of the subspace is determined exactly by the rank of the cross-covariance matrix, C3. In the frequency domain, this means the dimensionality at a specific frequency is given by the rank of the cross-power spectral density matrix. Our results show that the extent of normalization from surrounding neurons plays a critical role in shaping the dimensionality of the communication subspace. With surround normalization, dimensionality decreases specifically at frequencies where coherence is strongest (Fig. 4c). In contrast, without this normalization, no such correlation between coherence and dimensionality is observed (Fig. 11h).

Frequency decomposition of the communication subspace

To analyze how communication between different brain areas varies across frequencies, we consider the frequency decomposition of the neural responses. The Wiener-Khinchin theorem states that the autocorrelation function C(τ) is the inverse Fourier transform of the power spectral density (PSD) matrix (𝓢(ω)):

C(τ)=𝓢(ω)eıωτdω (23)

To isolate C(τ) at a particular frequency (ω*), we apply an idealized narrow band-pass filter centered at ω*. Specifically, the filter’s transfer function H(ω) is given by,

H(ω)=1ifωA0otherwise (24)

where A=ω*ϵ,ω*+ϵω*ϵ,ω*+ϵ and ϵ is a small positive value defining the narrow bandwidth 2ϵ around ω*. The spectral density matrix of this filtered signal is given by,

𝓢filtered(ω)=|H(ω)|2𝓢(ω) (25)

The resulting autocorrelation of the filtered signal is expressed as:

Cfiltered(τ)=𝓢filtered(ω)eıωτdω=|H(ω)|2𝓢(ω)eıωτdω. (26)

Using the definition of H(ω) and assuming ϵ to be sufficiently small such that 𝓢(ω) and eıωτ are approximately constant over the integration intervals of width 2ϵ, the frequency-specific autocorrelation at ω* is:

C(τ)ω*2ϵ𝓢ω*eıω*τ+𝓢ω*eıω*τ. (27)

At zero time lag (τ=0),

C(0)ω*2ϵ𝓢ω*+𝓢ω*. (28)

Since the PSD is Hermitian, i.e., 𝓢(ω)=𝓢*(ω)), we have

C(0)ω*=4ϵRe𝓢ω* (29)

This frequency-specific autocorrelation matrix (C(0)ω* can then be used in place of C(0) in the communication subspace analysis described above to evaluate the prediction performance at any frequency. As our analysis does not depend on the absolute scaling of this matrix, the prefactor 4ϵ can be disregarded, and we effectively use Re𝓢ω* as the frequency-specific correlation matrix.

Supplementary Material

1

Highlights.

  • Hierarchical neural circuit with reciprocal connections implementing divisive normalization.

  • Unified theory of inter-areal communication.

  • Emergent properties of the theory explain key experimental findings.

  • Feedback from higher cortical areas dictates functional connectivity.

Acknowledgements

We thank Pascal Fries and Adam Kohn for generously providing the experimental data in Fig 2. We acknowledge Michael Halassa, Hillel Adesnik, Adam Kohn and Flaviano Morone for their valuable feedback on the manuscript, and Ruben Coen-Cagli, John Maunsell, and Jason MacLean for insightful discussions. This work was initiated in part at the Aspen Center for Physics, which is supported by National Science Foundation grant PHY-2210452. This work was supported by the National Institute of Health under award numbers R01MH137669 and R01EY035242. S.M. also acknowledges support from the Simons Center for Computational Physical Chemistry (Simons Foundation grant 839534, MT). This work was supported in part through the NYU IT High-Performance Computing resources, services, and staff expertise.

Appendix. Derivation of the General Hierarchical Recurrent Normalization Circuit

This appendix provides a detailed derivation of the general dynamical equations for the hierarchical recurrent normalization model with reciprocal connections. We begin with the primary visual cortex (V1). Assuming N1 is the normalization weight matrix for V1 and z1 is the input drive to V1, we can write the normalization equations for the V1 principal neurons with complementary receptive fields as:

y1s+=z1nσn+N1(z1n+z1n)y1s=z1nσn+N1(z1n+z1n) (30)

Where y1s+ and y1s represent the steady-state firing rates of principal neurons corresponding to positive z1n and negative z1n input drives, respectively. Here, σ represents the semi-saturation constant. The exponent n is typically observed to be n2 for the primary visual cortex, it can be different for different brain areas. Mathematically:

z1n+z1n=z1n (31)

Our goal is to derive the dynamical equations for a general n, that yield the fixed points defined by Eq. 30. First, we distinguish between the membrane potentials and the firing rates of neurons. The membrane potential of a given type of neuron is denoted by vectors such as y1, u1, a1, and q1. Their corresponding firing rates are y1+ (and y1 for complementary responses), u1+,a1+, and q1+. The instantaneous firing rates are related to membrane potentials by applying rectification (denoted by ⌊⋅⌋) followed by power-law activation [191],

x+=xβ. (32)

The values of β are taken to be 2,12,1,1 for y1,u1,a1,q1, respectively. Combining the firing rates of principal neurons with the complementary receptive fields, y1s+ and y1s, Eq. 30 can be alternatively written as,

y1n=y1s++y1s=z1nσn+N1z1n (33)

To proceed with our derivation, we need to isolate the z1n term. From Eq. 33, we have:

z1n=σny1n+y1nN1z1n (34)

where denotes element-wise multiplication. Using the identity x1x2=Dx1x2, where Dx1 is a diagonal matrix with the elements of vector x1 on its diagonal, we can rewrite and solve for z1n:

z1n=σny1n+Dy1nN1z1n (35)

Factoring out z1n:

IDy1nN1z1n=σny1n (36)

Multiplying both sides with [I – D(|y1|n)N1]1 from the left yields:

z1n=IDy1nN11σny1n (37)

This expression for z1n will be used in subsequent steps. Now, we introduce the dynamical equation for the membrane potential y1 of the principal V1 neurons, consistent with biological computation in pyramidal cells (see Section ‘Biology plausibility’):

τy1dy1dt=y1+b1z1+11+a1+W11y1+1ny11n+γ1W12y2+1n. (38)

Here, τy1 is the time constant, b1 is the input gain, a1+ represents firing rates of the inhibitory neuron population. W11 is the recurrent V1 connectivity matrix, W12 is the connectivity matrix containing the synaptic weights from V2 to V1. W12y2+1n represents the feedback signal from V2 to V1, calculated as the weighted sum of the activities of V2 neurons. γ1 is the feedback gain modulation parameter. y2+ is the firing rate of V2 principal neurons, defined as y2n.

For simplicity, we assume that the recurrent weight matrix W11=I. Given that y1+1/py11/p=y1y1=y1 and y2+1/n=y2, the equation for y1 simplifies to:

τy1dy1dt=y1+b1z1+11+a1+y1+γ1W12y2 (39)

Now at steady state, dy1dt=0,

0=y1+b1z1+11+a1+y1+γ1W12y2 (40)

Expanding the terms:

0=y1+b1z1+11+a1+y1+11+a1+γ1W12y2 (41)

Rearranging terms:

a1+1+a1+y1=b1z1+11+a1+γ1W12y2 (42)

Moving the feedback term to the left-hand side:

a1+1+a1+y111+a1+γ1W12y2=b1z1 (43)

Factoring out y1:

y1a1+1+a1+11+a1+γ1W12y2y1=b1z1 (44)

We define a modulatory term u1+, such that:

y1u1+=b1z1 (45)

where:

u1+=a1+1+a1+11+a1+γ1W12y2y1 (46)

Next, we derive the dynamical equation for a1, from the definition of u1+:

u1+1+a1+=a1+γ1W12y2y1 (47)

Upon rearranging,

0=a1++γ1W12y2y1+1+a1+u1+ (48)

This equation describes the steady state of a1. It exhibits a form analogous to that of y1, comprising an input drive and a recurrent drive term. Therefore, we propose the following dynamical equation for the membrane potential a1:

τa1da1dt=a1+γ1W12y2y1+1+a1+u1+ (49)

According to Dale’s Law, excitatory neurons, such as y1, cannot directly inhibit other neurons. Therefore, the division by y1 implies the existence of an intermediate inhibitory interneuron. We introduce an inhibitory interneuron population q1, that tracks y1:

τq1dq1dt=q1+y1+1/n (50)

With this interneuron, the dynamic equation for a1 becomes:

τa1da1dt=a1+γ1W12y2+1/nq1++1+a1+u1+ (51)

Now, we return to Eq. 45 to derive the dynamical equation for the modulatory neuron u1. Starting with y1u1+=b1z1, we take the modulus on both sides and raise to the power of n:

y1nu1+n=b1nz1n (52)

Substituting the expression for z1n from Eq. 37:

y1nu1+n=b1nIDy1nN11σny1n (53)

We assume that b1 has equal components, i.e., b1=b11. Multiplying both sides by IDy1nN1 and rearranging:

IDy1nN1u1+ny1nb1n=σny1n (54)

Multiplying both sides by D1y1n:

D1y1nIDy1nN1u1+ny1nb1n=D1y1nσny1n (55)

This simplifies to:

u1+nb1nN1u1+ny1nb1n=σn (56)

Since we assumed that b1 has equal components, we get:

u1+nN1u1+ny1n=σnb1n (57)

Elementwise multiplying both sides by u1/u1+n from left and rearrange:

0=u1+u1u1+nb1nσn+N1u1+ny1n (58)

This equation describes the steady state of u1. We propose the following dynamical equation for u1,

τu1du1dt=u1+u1u1+nb1nσn+N1u1+ny1n (59)

These equations (Eq. 39, Eq. 59, Eq. 51, Eq. 50) describe the dynamics of principal excitatory neurons, modulatory inhibitory neurons, and modulatory excitatory neurons in V1. They exhibit the fixed point solution given by the normalization by design. These equations are general for any choice of n and β. We get a particularly simplified form when we set n=2 and β=2,12,1,1 for y1,u1,a1,q1, respectively.

Finally we rewrite the set of four dynamical equations for V1 as:

τy1dy1dt=y1+b1y1z1+11+a1+W11y1++γ1y1W12y2+ (60)
τu1du1dt=u1+b1u12σ2+N1y1+u1+2 (61)
τa1da1dt=a1+γ1a1W12y2+q1++u1++a1+u1++α1du1dt (62)
τq1dq1dt=q1+y1+ (63)

where we have decoupled the input gain b1 and feedback gain γ1 across each equation. When b1y1=b1u1 and γ1y1=γ1a1, and W11=I we exactly recover the normalization equation, Eq. 2, at steady state for y1. Then to reduce the number of parameters we re-define the input gain as β1b1y1/b1u1 and fix b1u1=1/2. Similarly, we re-define the feedback gain as γ1γ1y1/γ1a1 and fix γ1a1=1/2. We thus arrive at Eq. 1 of the main text. Finally, note that the term α1du1dt is added to the dynamics of a1 to achieve correct oscillatory behavior. This term does not impact the steady state solution but influences the dynamics.

The dynamical equations for V2 can be derived following a similar procedure:

τy2dy2dt=y2+b2y2z2+11+a2+W22y2+ (64)
τu2du2dt=u2+b2u22σ2+N2y2+u2+2 (65)
τa2da2dt=a2+u2++a2+u2++α2du2dt (66)
τq2dq2dt=q2+y2+ (67)

where, z2=W21y1+; and, like for y1, we have decoupled the input gain b2 across each equation. When b2y2=b2u2 and W22=I we exactly recover the normalization equation, Eq. 2, at steady state for y2. To reduce the number of parameters, we redefine the input gain as β2b2y2/b2u2 and fix b2u2=b2y2=1/2. Note that in principle y2 can receive feedback from higher cortical areas, in which case the equations would look identical to Eq. 1 but the subscripts change as 12 and 23.

In addition to the membrane potential, the corresponding firing rates of V1 and V2 neurons are also simulated. These are modeled as low-pass filtered versions of their target rectified power-law transformations. For V1, as an example:

τy1+dy1+dt=y1++y12 (68)
τu1+du1+dt=u1++u1 (69)
τa1+da1+dt=a1++a1 (70)
τq1+dq1+dt=q1++q1 (71)

Intrinsic neural circuit noise

The influence of noise on the deterministic neural dynamics is treated analytically in this model. We consider two primary sources of noise: firing rate noise originating from the spiking process and synaptic noise.

i) Firing rate noise is added to the firing rate vector k+, where k+ represents various neural population firing rates (e.g., excitatory like y1+,u1+ or inhibitory like a1+,q1+). This noise is attributed to the spiking mechanism, whose strength is found by the modified Gaussian Rectification (GR) model [186, 187]. Due to the high degree of redundancy in neural circuits, the strength of this noise is quenched due to the averaging of these redundant signals. The result is that the noise has a negligible impact on the firing rate dynamics [192, 193]. For a given firing rate vector k+, with an associated activation exponent β, the noise dynamics are described by:

τk+dk+dt=k++kβ+σkη(t), (72)

where τk+ is the time constant for the firing rate dynamics of population k+,σk represents the noise strengths for different neuron types, and η(t) is a vector of independent Gaussian white noise process, each with zero mean and unit variance.

ii) Synaptic noise is modeled as exponentially low-pass filtered Gaussian white noise (f), given by the following SDE:

τfdfdt=f+σfη(t), (73)

where τf is the noise filtering time constant (τf=1msec, approximating the time constant of synaptic transmission), and σf represents the noise strength. This filtered noise (f) is then added to the dynamical equations for the membrane potentials of each neuron, Eq. 6067, and it is the main source of intrinsic noise.

Upon incorporating these sources of noise, we get a high-dimensional nonlinear stochastic dynamical system with the following variables:

  • Variables representing the membrane potential for different neuron types implementing normalization (Eq. 6067).

  • Firing rates (Eq. 72) for each membrane potential variable.

  • Filtered synaptic noise (Eq. 73) for each membrane potential variable.

We then linearize this dynamical system about the fixed point found by simulating the deterministic part of the dynamical system, which gives us the Jacobian matrix J. We also construct the dispersion matrix L and the diffusion matrix D using the strength of the noise added in various dynamical variables. The definition of these matrices allows us to compute the covariance matrix and power spectra analytically as discussed in the Methods section of the main text.

Extrinsic noise in the local field potential

In addition to the noise sources detailed above, we also account for additional sources of noise that contribute to LFP power and coherence spectra. We consider two components of extrinsic noise: i) Correlated noise (ζcorrelated), assumed to stem from the volume conduction of electric fields generated by transmembrane current through the extracellular brain tissue, with negligible impact on spiking activity and firing rates [9799]. ii) Uncorrelated noise (ζuncorrelated), that encompasses noise from various other sources including motion artifact, instrumentation noise, thermal noise, and biological sources of noise [194196]. This extrinsic noise is added to the simulated neural responses (membrane potentials for different neuron types, denoted collectively by x):

x(t)=x(t)+ζuncorrelated(t)+ζcorrelated(t) (74)

Since these noise sources are uncorrelated with the responses of the neurons, the power spectrum of the measured signals 𝓢(ω) is the sum of the spectrum of the neurons’ activity 𝓢(ω) and the spectra of these noise components:

𝓢(ω)=𝓢(ω)+𝓢uncorrelated(ω)+𝓢correlated(ω) (75)

The correlated noise is modeled to produce a 1ω4 power law scaling, which is achieved by doubly low-pass filtered white noise. Its power spectrum for each element is given by:

𝓢correlated, element(ω)=Σcorr1+τnoise2ω22 (76)

The uncorrelated power spectrum 𝓢uncorrelated(ω) is diagonal, with each diagonal term having a similar filtered form:

𝓢uncorrelated(ω)=DΣuncorr1+τnoise2ω22 (77)

In our simulation, the ratio of power spectral densities of the correlated noise to the uncorrelated noise (i.e, Σcorr/Σuncorr for corresponding element) is set to 10 for all the results. The noise parameters used in the simulation are as follows,

Table 3:

Noise parameters

τx τf τnoise σf Σcorr Σuncorr
1 ms 1 ms 50 ms 0.01 9 × 10−4 9 × 10−5

References

  • [1].Heeger David J. “Normalization of cell responses in cat striate cortex”. In: Visual neuroscience 9.2 (1992), pp. 181–197. [DOI] [PubMed] [Google Scholar]
  • [2].Carandini Matteo and Heeger David J. “Normalization as a canonical neural computation”. In: Nature reviews neuroscience 13.1 (2012), pp. 51–62. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Salin Paul-Antoine and Bullier Jean. “Corticocortical connections in the visual system: structure and function”. In: Physiological reviews 75.1 (1995), pp. 107–154. [DOI] [PubMed] [Google Scholar]
  • [4].Shipp Stewart and Zeki Semir. “The organization of connections between areas V5 and V2 in macaque monkey visual cortex”. In: European Journal of Neuroscience 1.4 (1989), pp. 333– 354. [DOI] [PubMed] [Google Scholar]
  • [5].Kennedy Hand Bullier J. “A double-labeling investigation of the afferent connectivity to cortical areas V1 and V2 of the macaque monkey”. In: Journal of Neuroscience 5.10 (1985), pp. 2815–2830. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [6].Vanduffel Wim et al. “Functional impact of cerebral connections”. In: Proceedings of the National Academy of Sciences 94.14 (1997), pp. 7617–7620. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [7].Alonso JM et al. “Influence of layer V of area 18 of the cat visual cortex on responses of cells in layer V of area 17 to stimuli of high velocity”. In: Experimental brain research 93 (1993), pp. 363–366. [DOI] [PubMed] [Google Scholar]
  • [8].Sandell Julie H and Schiller Peter H. “Effect of cooling area 18 on striate cortex cells in the squirrel monkey.” In: Journal of Neurophysiology 48.1 (1982), pp. 38–48. [DOI] [PubMed] [Google Scholar]
  • [9].Hopfinger Joseph B, Buonocore Michael H, and Mangun George R. “The neural mechanisms of top-down attentional control”. In: Nature neuroscience 3.3 (2000), pp. 284–291. [DOI] [PubMed] [Google Scholar]
  • [10].Awh Edward, Belopolsky Artem V, and Theeuwes Jan. “Top-down versus bottom-up attentional control: A failed theoretical dichotomy”. In: Trends in cognitive sciences 16.8 (2012), pp. 437–443. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [11].Baluch Farhan and Itti Laurent. “Mechanisms of top-down attention”. In: Trends in neurosciences 34.4 (2011), pp. 210–224. [DOI] [PubMed] [Google Scholar]
  • [12].Quax Silvan, Jensen Ole, and Tiesinga Paul. “Top-down control of cortical gamma-band communication via pulvinar induced phase shifts in the alpha rhythm”. In: PLoS computational biology 13.5 (2017), e1005519. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Makino Hiroshi and Komiyama Takaki. “Learning enhances the relative impact of top-down processing in the visual cortex”. In: Nature neuroscience 18.8 (2015), pp. 1116–1122. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [14].Li Wu, Piëch Valentin, and Gilbert Charles D. “Perceptual learning and top-down influences in primary visual cortex”. In: Nature neuroscience 7.6 (2004), pp. 651–657. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [15].Heeger David J. “Theory of cortical function”. In: Proceedings of the National Academy of Sciences 114.8 (2017), pp. 1773–1782. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [16].Heeger David J and Zemlianova Klavdia O. “A recurrent circuit implements normalization, simulating the dynamics of V1 activity”. In: Proceedings of the National Academy of Sciences 117.36 (2020), pp. 22494–22505. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [17].Heeger David J and Mackey Wayne E. “Oscillatory recurrent gated neural integrator circuits (ORGaNICs), a unifying theoretical framework for neural dynamics”. In: Proceedings of the National Academy of Sciences 116.45 (2019), pp. 22783–22794. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [18].Rawat Shivang, Heeger David, and Martiniani Stefano. “Unconditional stability of a recurrent neural circuit implementing divisive normalization”. In: Advances in Neural Information Processing Systems 37 (2024), pp. 14712–14750. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [19].Morone Flaviano et al. “Stabilization of recurrent neural networks through divisive normalization”. In: bioRxiv (2025), pp. 2025–05. [Google Scholar]
  • [20].Bastos Andre M, Vezoli Julien, and Fries Pascal. “Communication through coherence with inter-areal delays”. In: Current opinion in neurobiology 31 (2015), pp. 173–180. [DOI] [PubMed] [Google Scholar]
  • [21].Fries Pascal. “A mechanism for cognitive dynamics: neuronal communication through neuronal coherence”. In: Trends in cognitive sciences 9.10 (2005), pp. 474–480. [DOI] [PubMed] [Google Scholar]
  • [22].Roberts Mark J et al. “Robust gamma coherence between macaque V1 and V2 by dynamic frequency matching”. In: Neuron 78.3 (2013), pp. 523–536. [DOI] [PubMed] [Google Scholar]
  • [23].Salinas Emilio and Sejnowski Terrence J. “Correlated neuronal activity and the flow of neural information”. In: Nature reviews neuroscience 2.8 (2001), pp. 539–550. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [24].Varela Francisco et al. “The brainweb: phase synchronization and large-scale integration”. In: Nature reviews neuroscience 2.4 (2001), pp. 229–239. [DOI] [PubMed] [Google Scholar]
  • [25].Fries Pascal et al. “Synchronization of oscillatory responses in visual cortex correlates with perception in interocular rivalry”. In: Proceedings of the National Academy of Sciences 94.23 (1997), pp. 12699–12704. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [26].Fries Pascal et al. “Oscillatory neuronal synchronization in primary visual cortex as a correlate of stimulus selection”. In: Journal of Neuroscience 22.9 (2002), pp. 3739–3754. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [27].Greenwood Priscilla E and Ward Lawrence M. “Attentional selection and communication through coherence: Scope and limitations”. In: PLOS Computational Biology 20.8 (2024), e1011431. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [28].Rolls Edmund T and Treves Alessandro. “The neuronal encoding of information in the brain”. In: Progress in neurobiology 95.3 (2011), pp. 448–490. [DOI] [PubMed] [Google Scholar]
  • [29].Rolls Edmund T, Webb Tristan J, and Deco Gustavo. “Communication before coherence”. In: European Journal of Neuroscience 36.5 (2012), pp. 2689–2709. [DOI] [PubMed] [Google Scholar]
  • [30].Schneider Marius et al. “A mechanism for inter-areal coherence through communication based on connectivity and oscillatory power”. In: Neuron 109.24 (2021), pp. 4050–4067. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [31].Ray Supratim and Maunsell John HR. “Do gamma oscillations play a role in cerebral cortex?” In: Trends in cognitive sciences 19.2 (2015), pp. 78–85. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [32].Semedo João D et al. “Cortical areas interact through a communication subspace”. In: Neuron 102.1 (2019), pp. 249–259. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [33].Srinath Ramanujan, Ruff Douglas A, and Cohen Marlene R. “Attention improves information flow between neuronal populations without changing the communication subspace”. In: Current Biology 31.23 (2021), pp. 5299–5313. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [34].Young Ryan A et al. “Hippocampal-prefrontal communication subspaces align with behavioral and network patterns in a spatial memory task”. In: bioRxiv (2024), pp. 2024–07. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [35].MacDowell CJ et al. Multiplexed Subspaces Route Neural Activity Across Brain-wide Networks. bioRxiv, 2023.02. 08.527772. 2023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [36].Peter G Poggi Stefan Mihalas, and Mastrovito Dana. “Communication subspaces align with training in ANNs”. In: bioRxiv (2024), pp. 2024–11. [Google Scholar]
  • [37].Kay Kendrick N et al. “A two-stage cascade model of BOLD responses in human visual cortex”. In: PLoS computational biology 9.5 (2013), e1003079. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [38].Doostani Narges, Hossein-Zadeh Gholam-Ali, and Vaziri-Pashkam Maryam. “The normalization model predicts responses in the human visual cortex during object-based attention”. In: Elife 12 (2023), e75726. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [39].Olsen Shawn R, Bhandawat Vikas, and Wilson Rachel I. “Divisive normalization in olfactory population codes”. In: Neuron 66.2 (2010), pp. 287–299. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [40].Ahrens Misha B, Linden Jennifer F, and Sahani Maneesh. “Nonlinearities and contextual influences in auditory cortical responses modeled with multilinear spectrotemporal methods”. In: Journal of Neuroscience 28.8 (2008), pp. 1929–1942. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [41].Rabinowitz Neil C et al. “Contrast gain control in auditory cortex”. In: Neuron 70.6 (2011), pp. 1178–1191. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [42].Brouwer Gijs Joost et al. “Normalization in human somatosensory cortex”. In: Journal of Neurophysiology 114.5 (2015), pp. 2588–2599. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [43].Denison Rachel N, Carrasco Marisa, and Heeger David J. “A dynamic normalization model of temporal attention”. In: Nature Human Behaviour 5.12 (2021), pp. 1674–1685. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [44].Ni Amy M and Maunsell John HR. “Neuronal effects of spatial and feature attention differ due to normalization”. In: Journal of Neuroscience 39.28 (2019), pp. 5493–5505. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [45].Zhang Xilin et al. “A normalization frame-work for emotional attention”. In: PLoS biology 14.11 (2016), e1002578. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [46].Schwedhelm Philipp, Krishna B Suresh, and Treue Stefan. “An extended normalization model of attention accounts for feature-based attentional enhancement of both response and coherence gain”. In: PLoS computational biology 12.12 (2016), e1005225. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [47].Smith Philip L, Sewell David K, and Lilburn Simon D. “From shunting inhibition to dynamic normalization: Attentional selection and decision-making in brief visual displays”. In: Vision Research 116 (2015), pp. 219–240. [DOI] [PubMed] [Google Scholar]
  • [48].Maunsell John HR. “Neuronal mechanisms of visual attention”. In: Annual review of vision science 1.1 (2015), pp. 373–391. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [49].Beuth Frederik and Hamker Fred H. “A mechanistic cortical microcircuit of attention for amplification, normalization and suppression”. In: Vision research 116 (2015), pp. 241–257. [DOI] [PubMed] [Google Scholar]
  • [50].Herrmann Katrin, Heeger David J, and Carrasco Marisa. “Feature-based attention enhances performance by increasing response gain”. In: Vision research 74 (2012), pp. 10–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [51].Ni Amy M and Maunsell John HR. “Spatially tuned normalization explains attention modulation variance within neurons”. In: Journal of neurophysiology 118.3 (2017), pp. 1903–1913. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [52].Herrmann Katrin et al. “When size matters: attention affects performance by contrast or response gain”. In: Nature neuroscience 13.12 (2010), pp. 1554–1559. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [53].Reynolds John H and Heeger David J. “The normalization model of attention”. In: Neuron 61.2 (2009), pp. 168–185. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [54].Lee Joonyeol and Maunsell John HR. “A normalization model of attentional modulation of single unit responses”. In: PloS one 4.2 (2009), e4651. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [55].Boynton Geoffrey M. “A framework for describing the effects of attention on visual responses”. In: Vision research 49.10 (2009), pp. 1129–1143. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [56].Sugihara Tadashi et al. “Integration of auditory and visual communication information in the primate ventrolateral prefrontal cortex”. In: Journal of Neuroscience 26.43 (2006), pp. 11138–11147. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [57].Avillac Marie, Hamed Suliann Ben, and Duhamel Jean-René. “Multisensory integration in the ventral intraparietal area of the macaque monkey”. In: Journal of Neuroscience 27.8 (2007), pp. 1922–1932. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [58].Ohshiro Tomokazu, Angelaki Dora E, and DeAngelis Gregory C. “A normalization model of multisensory integration”. In: Nature neuroscience 14.6 (2011), pp. 775–782. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [59].Ohshiro Tomokazu, Angelaki Dora E, and DeAngelis Gregory C. “A neural signature of divisive normalization at the level of multisensory integration in primate cortex”. In: Neuron 95.2 (2017), pp. 399–411. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [60].Diehl Maria M and Romanski Lizabeth M. “Responses of prefrontal multisensory neurons to mismatching faces and vocalizations”. In: Journal of Neuroscience 34.34 (2014), pp. 11233–11243. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [61].Taylor Robert and Bays Paul M. “Efficient coding in visual working memory accounts for stimulus-specific variations in recall”. In: Journal of Neuroscience 38.32 (2018), pp. 7132–7142. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [62].Schneegans Sebastian and Bays Paul M. “Drift in neural population activity causes working memory to deteriorate over time”. In: Journal of Neuroscience 38.21 (2018), pp. 4859–4869. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [63].Ma Wei Ji, Husain Masud, and Bays Paul M. “Changing concepts of working memory”. In: Nature neuroscience 17.3 (2014), pp. 347–356. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [64].Bays Paul M. “Noise in neural populations accounts for errors in working memory”. In: Journal of Neuroscience 34.10 (2014), pp. 3632–3645. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [65].Louie Kenway, Grattan Lauren E, and Glimcher Paul W. “Reward value-based gain control: divisive normalization in parietal cortex”. In: Journal of Neuroscience 31.29 (2011), pp. 10627–10639. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [66].Louie Kenway, Khaw Mel W, and Glimcher Paul W. “Normalization is a general neural mechanism for context-dependent decision making”. In: Proceedings of the National Academy of Sciences 110.15 (2013), pp. 6139–6144. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [67].Louie Kenway et al. “Dynamic divisive normalization predicts time-varying value coding in decision-related circuits”. In: Journal of Neuroscience 34.48 (2014), pp. 16046–16057. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [68].Khaw Mel W, Glimcher Paul W, and Louie Kenway. “Normalized value coding explains dynamic adaptation in the human valuation process”. In: Proceedings of the National Academy of Sciences 114.48 (2017), pp. 12696–12701. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [69].Heeger David J, Simoncelli Eero P, and Movshon J Anthony. “Computational models of cortical visual processing.” In: Proceedings of the National Academy of Sciences 93.2 (1996), pp. 623–627. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [70].Simoncelli Eero P and Heeger David J. “A model of neuronal responses in visual area MT”. In: Vision research 38.5 (1998), pp. 743–761. [DOI] [PubMed] [Google Scholar]
  • [71].Simoncelli Eero P. “Local analysis of visual motion”. In: The visual neurosciences 109 (2003), pp. 1616–1623. [Google Scholar]
  • [72].Beck Jeffrey M, Latham Peter E, and Pouget Alexandre. “Marginalization in neural circuits with divisive normalization”. In: Journal of Neuroscience 31.43 (2011), pp. 15310–15319. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [73].Ringach Dario L. “Population coding under normalization”. In: Vision research 50.22 (2010), pp. 2223–2232. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [74].Schwartz Odelia and Simoncelli Eero P. “Natural signal statistics and sensory gain control”. In: Nature neuroscience 4.8 (2001), pp. 819–825. [DOI] [PubMed] [Google Scholar]
  • [75].Westrick Zachary M, Heeger David J, and Landy Michael S. “Pattern adaptation and normalization reweighting”. In: Journal of Neuroscience 36.38 (2016), pp. 9805–9816. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [76].Lyu Siwei and Simoncelli Eero P. “Nonlinear extraction of independent components of natural images using radial gaussianization”. In: Neural computation 21.6 (2009), pp. 1485–1519. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [77].Luo Shawn X, Axel Richard, and Abbott Larry F. “Generating sparse and selective third-order responses in the olfactory system of the fly”. In: Proceedings of the National Academy of Sciences 107.23 (2010), pp. 10713–10718. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [78].Ba Jimmy Lei, Kiros Jamie Ryan, and Hinton Geoffrey E. “Layer normalization”. In: arXiv preprint arXiv:1607.06450 (2016). [Google Scholar]
  • [79].Ren Mengye et al. “Normalizing the normalizers: Comparing and extending network normalization schemes”. In: arXiv preprint arXiv:1611.04520 (2016). [Google Scholar]
  • [80].Bolding Kevin A and Franks Kevin M. “Recurrent cortical circuits implement concentrationinvariant odor coding”. In: Science 361.6407 (2018), eaat6904. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [81].Adesnik Hillel and Scanziani Massimo. “Lateral competition for cortical space by layer-specific horizontal circuits”. In: Nature 464.7292 (2010), pp. 1155–1160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [82].Sato Tatsuo K et al. “An excitatory basis for divisive normalization in visual cortex”. In: Nature neuroscience 19.4 (2016), pp. 568–570. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [83].Adesnik Hillel. “Synaptic Mechanisms of Feature Coding in the Visual Cortex of Awake Mice”. In: Neuron 95.5 (2017), pp. 1147–1159. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [84].Douglas Rodney J et al. “Recurrent excitation in neocortical circuits”. In: Science 269.5226 (1995), pp. 981–985. [DOI] [PubMed] [Google Scholar]
  • [85].Kapfer Christoph et al. “Supralinear increase of recurrent inhibition during sparse activity in the somatosensory cortex”. In: Nature neuroscience 10.6 (2007), pp. 743–753. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [86].Callaway Edward M. “Local circuits in primary visual cortex of the macaque monkey”. In: Annual review of neuroscience 21.1 (1998), pp. 47–74. [DOI] [PubMed] [Google Scholar]
  • [87].Yoshimura Yumiko, Dantzker Jami L, and Callaway Edward M. “Excitatory cortical neurons form fine-scale functional networks”. In: Nature 433.7028 (2005), pp. 868–873. [DOI] [PubMed] [Google Scholar]
  • [88].Douglas Rodney J and Martin Kevan AC. “Recurrent neuronal circuits in the neocortex”. In: Current biology 17.13 (2007), R496–R500. [DOI] [PubMed] [Google Scholar]
  • [89].Peron Simon et al. “Recurrent interactions in local cortical circuits”. In: Nature 579.7798 (2020), pp. 256–259. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [90].Levitt Jonathan B, Kiper Daniel C, and Movshon J Anthony. “Receptive fields and functional architecture of macaque V2”. In: Journal of neurophysiology 71.6 (1994), pp. 2517–2542. [DOI] [PubMed] [Google Scholar]
  • [91].Geisler Wilson S and Albrecht Duane G. “Visual cortex neurons in monkeys and cats: detection, discrimination, and identification”. In: Visual neuroscience 14.5 (1997), pp. 897–919. [DOI] [PubMed] [Google Scholar]
  • [92].Sclar Gary, Maunsell John HR, and Lennie Peter. “Coding of image contrast in central visual pathways of the macaque monkey”. In: Vision research 30.1 (1990), pp. 1–10. [DOI] [PubMed] [Google Scholar]
  • [93].Kahana Michael J. “The cognitive correlates of human brain oscillations”. In: Journal of Neuroscience 26.6 (2006), pp. 1669–1672. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [94].Buzsaki Gyorgy and Draguhn Andreas. “Neuronal oscillations in cortical networks”. In: science 304.5679 (2004), pp. 1926–1929. [DOI] [PubMed] [Google Scholar]
  • [95].Siegel Markus, Donner Tobias H, and Engel Andreas K. “Spectral fingerprints of large-scale neuronal interactions”. In: Nature Reviews Neuroscience 13.2 (2012), pp. 121–134. [DOI] [PubMed] [Google Scholar]
  • [96].Uhlhaas Peter et al. “Neural synchrony in cortical networks: history, concept and current status”. In: Frontiers in integrative neuroscience 3 (2009), p. 543. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [97].Buzsáki György, Anastassiou Costas A, and Koch Christof. “The origin of extracellular fields and currents—EEG, ECoG, LFP and spikes”. In: Nature reviews neuroscience 13.6 (2012), pp. 407–420. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [98].Claude Bédard Helmut Kröger, and Destexhe Alain. “Model of low-pass filtering of local field potentials in brain tissue”. In: Physical Review E—Statistical, Nonlinear, and Soft Matter Physics 73.5 (2006), p. 051911. [DOI] [PubMed] [Google Scholar]
  • [99].Einevoll Gaute T et al. “Modelling and analysis of local field potentials for studying the function of cortical circuits”. In: Nature Reviews Neuroscience 14.11 (2013), pp. 770–785. [DOI] [PubMed] [Google Scholar]
  • [100].Jia Xiaoxuan, Xing Dajun, and Kohn Adam. “No consistent relationship between gamma power and peak frequency in macaque primary visual cortex”. In: Journal of Neuroscience 33.1 (2013), pp. 17–25. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [101].Ray Supratim and Maunsell John HR. “Differences in gamma frequencies across visual cortex restrict their possible use in computation”. In: Neuron 67.5 (2010), pp. 885–896. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [102].Henrie J Andrew and Shapley Robert. “LFP power spectra in V1 cortex: the graded effect of stimulus contrast”. In: Journal of neurophysiology 94.1 (2005), pp. 479–490. [DOI] [PubMed] [Google Scholar]
  • [103].Miller Kai J et al. “Power-law scaling in the brain surface electric potential”. In: PLoS computational biology 5.12 (2009), e1000609. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [104].Engel Andreas K, Fries Pascal, and Singer Wolf. “Dynamic predictions: oscillations and synchrony in top–down processing”. In: Nature Reviews Neuroscience 2.10 (2001), pp. 704–716. [DOI] [PubMed] [Google Scholar]
  • [105].Fries Pascal. “Rhythms for cognition: communication through coherence”. In: Neuron 88.1 (2015), pp. 220–235. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [106].Hutchison R Matthew et al. “Dynamic functional connectivity: promise, issues, and interpretations”. In: Neuroimage 80 (2013), pp. 360–378. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [107].Allen Elena A et al. “Tracking whole-brain connectivity dynamics in the resting state”. In: Cerebral cortex 24.3 (2014), pp. 663–676. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [108].Calhoun Vince D et al. “The chronnectome: time-varying connectivity networks as the next frontier in fMRI data discovery”. In: Neuron 84.2 (2014), pp. 262–274. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [109].Choi Sang-Han et al. “Proposal for human visual pathway in the extrastriate cortex by fiber tracking method using diffusion-weighted MRI”. In: Neuroimage 220 (2020), p. 117145. [DOI] [PubMed] [Google Scholar]
  • [110].Mishkin Mortimer, Ungerleider Leslie G, and Macko Kathleen A. “Object vision and spatial vision: two cortical pathways”. In: Trends in neurosciences 6 (1983), pp. 414–417. [Google Scholar]
  • [111].Goodale Melvyn A and Milner A David. “Separate visual pathways for perception and action”. In: Trends in neurosciences 15.1 (1992), pp. 20–25. [DOI] [PubMed] [Google Scholar]
  • [112].Corbetta Maurizio et al. “Selective and divided attention during visual discriminations of shape, color, and speed: functional anatomy by positron emission tomography”. In: Journal of neuroscience 11.8 (1991), pp. 2383–2402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [113].Serences John T and Boynton Geoffrey M. “Feature-based attentional modulations in the absence of direct visual stimulation”. In: Neuron 55.2 (2007), pp. 301–312. [DOI] [PubMed] [Google Scholar]
  • [114].Maunsell John HR and Treue Stefan. “Feature-based attention in visual cortex”. In: Trends in neurosciences 29.6 (2006), pp. 317–322. [DOI] [PubMed] [Google Scholar]
  • [115].Treue Stefan and Trujillo Julio C Martinez. “Feature-based attention influences motion processing gain in macaque visual cortex”. In: Nature 399.6736 (1999), pp. 575–579. [DOI] [PubMed] [Google Scholar]
  • [116].Cohen Marlene R and Maunsell John HR. “Using neuronal populations to study the mechanisms underlying spatial and feature attention”. In: Neuron 70.6 (2011), pp. 1192–1204. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [117].Liu Taosheng et al. “Cortical mechanisms of feature-based attentional control”. In: Cerebral cortex 13.12 (2003), pp. 1334–1343. [DOI] [PubMed] [Google Scholar]
  • [118].Martinez-Trujillo Julio C and Treue Stefan. “Feature-based attention increases the selectivity of population responses in primate visual cortex”. In: Current biology 14.9 (2004), pp. 744–751. [DOI] [PubMed] [Google Scholar]
  • [119].Saproo Sameer and Serences John T. “Attention improves transfer of motion information between V1 and MT”. In: Journal of Neuroscience 34.10 (2014), pp. 3586–3596. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [120].Al-Aidroos Naseem, Said Christopher P, and Turk-Browne Nicholas B. “Top-down attention switches coupling between low-level and high-level areas of human visual cortex”. In: Proceedings of the National Academy of Sciences 109.36 (2012), pp. 14675–14680. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [121].Saalmann Yuri B, Pigarev Ivan N, and Vidyasagar Trichur R. “Neural mechanisms of visual attention: how top-down feedback highlights relevant locations”. In: Science 316.5831 (2007), pp. 1612–1615. [DOI] [PubMed] [Google Scholar]
  • [122].Hochreiter Sepp and Schmidhuber Jürgen. “Long short-term memory”. In: Neural computation 9.8 (1997), pp. 1735–1780. [DOI] [PubMed] [Google Scholar]
  • [123].Graves Alex. “Generating sequences with recurrent neural networks”. In: arXiv preprint arXiv:1308.0850 (2013). [Google Scholar]
  • [124].Graves Alex, Mohamed Abdel-rahman, and Hinton Geoffrey. “Speech recognition with deep recurrent neural networks”. In: 2013 IEEE international conference on acoustics, speech and signal processing. Ieee. 2013, pp. 6645–6649. [Google Scholar]
  • [125].Cho Kyunghyun et al. “Learning phrase representations using RNN encoder-decoder for statistical machine translation”. In: arXiv preprint arXiv:1406.1078 (2014). [Google Scholar]
  • [126].Sutskever Ilya, Vinyals Oriol, and Le Quoc V. “Sequence to sequence learning with neural networks”. In: Advances in neural information processing systems 27 (2014). [Google Scholar]
  • [127].Rao Rajesh PN and Ballard Dana H. “Dynamic model of visual recognition predicts neural response properties in the visual cortex”. In: Neural computation 9.4 (1997), pp. 721–763. [DOI] [PubMed] [Google Scholar]
  • [128].Wacongne Catherine, Changeux Jean-Pierre, and Dehaene Stanislas. “A neuronal model of predictive coding accounting for the mismatch negativity”. In: Journal of Neuroscience 32.11 (2012), pp. 3665–3678. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [129].Rao Rajesh PN and Ballard Dana H. “Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects”. In: Nature neuroscience 2.1 (1999), pp. 79–87. [DOI] [PubMed] [Google Scholar]
  • [130].Lee Tai Sing and Mumford David. “Hierarchical Bayesian inference in the visual cortex”. In: Journal of the Optical Society of America A 20.7 (2003), pp. 1434–1448. [DOI] [PubMed] [Google Scholar]
  • [131].Friston Karl. “A theory of cortical responses”. In: Philosophical transactions of the Royal Society B: Biological sciences 360.1456 (2005), pp. 815–836. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [132].Bastos Andre M et al. “Canonical microcircuits for predictive coding”. In: Neuron 76.4 (2012), pp. 695–711. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [133].Burns Samuel P, Xing Dajun, and Shapley Robert M. “Comparisons of the dynamics of local field potential and multiunit activity signals in macaque visual cortex”. In: Journal of Neuroscience 30.41 (2010), pp. 13739–13749. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [134].Ozeki Hirofumi et al. “Inhibitory stabilization of the cortical network underlies visual surround suppression”. In: Neuron 62.4 (2009), pp. 578–592. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [135].Senn Walter et al. “A neuronal least-action principle for real-time learning in cortical circuits”. In: ELife 12 (2024), RP89674. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [136].Heeger David J and Bergen James R. “Pyramid-based texture analysis/synthesis”. In: Proceedings of the 22nd annual conference on Computer graphics and interactive techniques. 1995, pp. 229–238. [Google Scholar]
  • [137].Portilla Javier and Simoncelli Eero P. “A parametric texture model based on joint statistics of complex wavelet coefficients”. In: International journal of computer vision 40 (2000), pp. 49–70. [Google Scholar]
  • [138].Hinton Geoffrey E et al. “The” wake-sleep” algorithm for unsupervised neural networks”. In: Science 268.5214 (1995), pp. 1158–1161. [DOI] [PubMed] [Google Scholar]
  • [139].Heeger David J. “Optical flow using spatiotemporal filters”. In: International journal of computer vision 1 (1988), pp. 279–302. [Google Scholar]
  • [140].Heeger David J and Simoncelli Eero P. “Model of visual motion sensing”. In: Spatial vision in humans and robots 19 (1993), pp. 367–392. [Google Scholar]
  • [141].Ma Wei Ji et al. “Bayesian inference with probabilistic population codes”. In: Nature neuroscience 9.11 (2006), pp. 1432–1438. [DOI] [PubMed] [Google Scholar]
  • [142].Knill David C and Pouget Alexandre. “The Bayesian brain: the role of uncertainty in neural coding and computation”. In: TRENDS in Neurosciences 27.12 (2004), pp. 712–719. [DOI] [PubMed] [Google Scholar]
  • [143].Landy Michael S et al. “Measurement and modeling of depth cue combination: in defense of weak fusion”. In: Vision research 35.3 (1995), pp. 389–412. [DOI] [PubMed] [Google Scholar]
  • [144].Ernst Marc O and Banks Martin S. “Humans integrate visual and haptic information in a statistically optimal fashion”. In: Nature 415.6870 (2002), pp. 429–433. [DOI] [PubMed] [Google Scholar]
  • [145].Fetsch Christopher R et al. “Neural correlates of reliability-based cue weighting during multisensory integration”. In: Nature neuroscience 15.1 (2012), pp. 146–154. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [146].Honey Christopher J et al. “Predicting human resting-state functional connectivity from structural connectivity”. In: Proceedings of the National Academy of Sciences 106.6 (2009), pp. 2035–2040. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [147].Bastos André M and Schoffelen Jan-Mathijs. “A tutorial review of functional connectivity analysis methods and their interpretational pitfalls”. In: Frontiers in systems neuroscience 9 (2016), p. 175. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [148].Saenz Melissa, Buracas Giedrius T, and Boynton Geoffrey M. “Global effects of feature-based attention in human visual cortex”. In: Nature neuroscience 5.7 (2002), pp. 631–632. [DOI] [PubMed] [Google Scholar]
  • [149].Jones Edward G. “Thalamic circuitry and thalamocortical synchrony”. In: Philosophical Transactions of the Royal Society of London. Series B: Biological Sciences 357.1428 (2002), pp. 1659–1673. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [150].Hughes Stuart W and Crunelli Vincenzo. “Thalamic mechanisms of EEG alpha rhythms and their pathological implications”. In: The Neuroscientist 11.4 (2005), pp. 357–372. [DOI] [PubMed] [Google Scholar]
  • [151].Bollimunta Anil et al. “Neuronal mechanisms and attentional modulation of corticothalamic alpha oscillations”. In: Journal of Neuroscience 31.13 (2011), pp. 4935–4943. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [152].Halgren Mila et al. “The generation and propagation of the human alpha rhythm”. In: Proceedings of the National Academy of Sciences 116.47 (2019), pp. 23772–23782. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [153].Jensen Ole and Mazaheri Ali. “Shaping functional architecture by oscillatory alpha activity: gating by inhibition”. In: Frontiers in human neuroscience 4 (2010), p. 186. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [154].Klimesch Wolfgang, Sauseng Paul, and Hanslmayr Simon. “EEG alpha oscillations: the inhibition–timing hypothesis”. In: Brain research reviews 53.1 (2007), pp. 63–88. [DOI] [PubMed] [Google Scholar]
  • [155].Schmitt L Ian et al. “Thalamic amplification of cortical connectivity sustains attentional control”. In: Nature 545.7653 (2017), pp. 219–223. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [156].Destexhe Alain. “Modelling corticothalamic feedback and the gating of the thalamus by the cerebral cortex”. In: Journal of Physiology-Paris 94.5-6 (2000), pp. 391–410. [DOI] [PubMed] [Google Scholar]
  • [157].Sherman S Murray and Usrey W Martin. “Exploring Thalamocortical Interactions: Circuitry for Sensation, Action, and Cognition”. In: (2021). [Google Scholar]
  • [158].Rikhye Rajeev V, Gilra Aditya, and Halassa Michael M. “Thalamic regulation of switching between cortical representations enables cognitive flexibility”. In: Nature neuroscience 21.12 (2018), pp. 1753–1763. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [159].Rikhye Rajeev V, Wimmer Ralf D, and Halassa Michael M. “Toward an integrative theory of thalamic function”. In: Annual review of neuroscience 41.1 (2018), pp. 163–183. [DOI] [PubMed] [Google Scholar]
  • [160].Guo Zengcai V et al. “Maintenance of persistent activity in a frontal thalamocortical loop”. In: Nature 545.7653 (2017), pp. 181–186. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [161].Schmolesky Matthew T et al. “Signal timing across the macaque visual system”. In: Journal of neurophysiology 79.6 (1998), pp. 3272–3278. [DOI] [PubMed] [Google Scholar]
  • [162].Semedo João D et al. “Feedforward and feedback interactions between visual cortical areas use different population activity patterns”. In: Nature communications 13.1 (2022), p. 1099. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [163].Carandini Matteo and Heeger David J. “Summation and division by neurons in primate visual cortex”. In: Science 264.5163 (1994), pp. 1333–1336. [DOI] [PubMed] [Google Scholar]
  • [164].Chance Frances S, Abbott Larry F, and Reyes Alex D. “Gain modulation from background synaptic input”. In: Neuron 35.4 (2002), pp. 773–782. [DOI] [PubMed] [Google Scholar]
  • [165].Wilson Hugh R and Cowan Jack D. “Excitatory and inhibitory interactions in localized populations of model neurons”. In: Biophysical journal 12.1 (1972), pp. 1–24. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [166].Riedemann Therese. “Diversity and function of somatostatin-expressing interneurons in the cerebral cortex”. In: International journal of molecular sciences 20.12 (2019), p. 2952. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [167].Adesnik Hillel et al. “A neural circuit for spatial summation in visual cortex”. In: Nature 490.7419 (2012), pp. 226–231. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [168].Pfeffer Carsten K et al. “Inhibition of inhibition in visual cortex: the logic of connections between molecularly distinct interneurons”. In: Nature neuroscience 16.8 (2013), pp. 1068–1076. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [169].Campagnola Luke et al. “Local connectivity and synaptic dynamics in mouse and human neocortex”. In: Science 375.6585 (2022), eabj5861. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [170].Lee Jung Hoon and Mihalas Stefan. “Visual processing mode switching regulated by VIP cells”. In: Scientific Reports 7.1 (2017), p. 1843. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [171].Atallah Bassam V et al. “Parvalbumin-expressing interneurons linearly transform cortical responses to visual stimuli”. In: Neuron 73.1 (2012), pp. 159–170. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [172].Sanzeni Alessandro et al. “Inhibition stabilization is a widespread property of cortical networks”. In: Elife 9 (2020), e54875. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [173].Studer Florian, Zeppillo Tommaso, and Barkat Tania Rinaldi. “Top-down inputs are controlled by somatostatin-expressing interneurons during associative learning”. In: bioRxiv (2024), pp. 2024–12. [Google Scholar]
  • [174].Naka Alexander et al. “Complementary networks of cortical somatostatin interneurons enforce layer specific control”. In: Elife 8 (2019), e43696. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [175].Shen Shan et al. “Distinct organization of two cortico-cortical feedback pathways”. In: Nature communications 13.1 (2022), p. 6389. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [176].Angelucci Alessandra, Levitt Jonathan B, and Lund Jennifer S. “Anatomical origins of the classical receptive field and modulatory surround field of single neurons in macaque visual cortical area V1”. In: Progress in brain research 136 (2002), pp. 373–388. [DOI] [PubMed] [Google Scholar]
  • [177].Angelucci Alessandra et al. “Circuits for local and global signal integration in primary visual cortex”. In: Journal of Neuroscience 22.19 (2002), pp. 8633–8646. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [178].Nassi Jonathan J, Lomber Stephen G, and Born Richard T. “Corticocortical feedback contributes to surround suppression in V1 of the alert primate”. In: Journal of Neuroscience 33.19 (2013), pp. 8504–8517. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [179].Nassi Jonathan J et al. “Corticocortical feedback increases the spatial extent of normalization”. In: Frontiers in systems neuroscience 8 (2014), p. 105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [180].Zhu J Julius. “Maturation of layer 5 neocortical pyramidal neurons: amplifying salient layer 1 and layer 4 inputs by Ca2+ action potentials in adult rat tuft dendrites”. In: The Journal of physiology 526.3 (2000), pp. 571–587. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [181].Sarter Martin and Bruno John P. “Cognitive functions of cortical acetylcholine: toward a unifying hypothesis”. In: Brain Research Reviews 23.1-2 (1997), pp. 28–46. [DOI] [PubMed] [Google Scholar]
  • [182].Angela J Yu and Dayan Peter. “Uncertainty, neuromodulation, and attention”. In: Neuron 46.4 (2005), pp. 681–692. [DOI] [PubMed] [Google Scholar]
  • [183].Marshall Louise et al. “Pharmacological fingerprints of contextual uncertainty”. In: PLoS Biology 14.11 (2016), e1002575. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [184].Pafundo Diego E et al. “Top-down-mediated facilitation in the visual cortex is gated by subcortical neuromodulation”. In: Journal of Neuroscience 36.10 (2016), pp. 2904–2914. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [185].Bekkers John M. “Pyramidal neurons”. In: Current biology 21.24 (2011), R975. [DOI] [PubMed] [Google Scholar]
  • [186].Rawat Shivang, Heeger David, and Martiniani Stefano. “A Comprehensive Large-Scale Model of the Primary Visual Cortex (V1)”. In: Bulletin of the American Physical Society (2024). [Google Scholar]
  • [187].Carandini Matteo. “Amplification of trial-to-trial response variability by neurons in visual cortex”. In: PLoS biology 2.9 (2004), e264. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [188].Rawat Shivang and Martiniani Stefano. “Element-wise and recursive solutions for the power spectral density of biological stochastic dynamical systems at fixed points”. In: Physical Review Research 6.4 (2024), p. 043179. [Google Scholar]
  • [189].Kohn Adam et al. “Principles of corticocortical communication: proposed schemes and design considerations”. In: Trends in Neurosciences 43.9 (2020), pp. 725–737. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [190].Semedo João D et al. “Statistical methods for dissecting interactions between brain areas”. In: Current opinion in neurobiology 65 (2020), pp. 59–69. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [191].Heeger David J. “Half-squaring in responses of cat striate cells”. In: Visual neuroscience 9.5 (1992), pp. 427–443. [DOI] [PubMed] [Google Scholar]
  • [192].Allen Christina and Stevens Charles F. “An evaluation of causes for unreliability of synaptic transmission.” In: Proceedings of the National Academy of Sciences 91.22 (1994), pp. 10380–10383. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [193].Mainen Zachary F and Sejnowski Terrence J. “Reliability of spike timing in neocortical neurons”. In: Science 268.5216 (1995), pp. 1503–1506. [DOI] [PubMed] [Google Scholar]
  • [194].Nelson Matthew J and Pouget Pierre. Do electrode properties create a problem in interpreting local field potential recordings? 2010. [DOI] [PubMed] [Google Scholar]
  • [195].Ludwig Kip A et al. “Using a common average reference to improve cortical neuron recordings from microelectrode arrays”. In: Journal of neurophysiology 101.3 (2009), pp. 1679–1689. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [196].Whitmore Nathan W and Lin Shih-Chieh. “Unmasking local activity within local field potentials (LFPs) by removing distal electrical signals using independent component analysis”. In: Neuroimage 132 (2016), pp. 79–92. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

1

Articles from bioRxiv are provided here courtesy of Cold Spring Harbor Laboratory Preprints

RESOURCES