Skip to main content
Frontiers in Network Physiology logoLink to Frontiers in Network Physiology
. 2025 Oct 31;5:1693772. doi: 10.3389/fnetp.2025.1693772

Population coding and self-organized ring attractors in recurrent neural networks for continuous variable integration

Roman Kononov 1,2, Vasilii Tiselko  1,3,4, Oleg Maslennikov  1,2,*, Vladimir Nekorkin  1,2
PMCID: PMC12615411  PMID: 41245775

Abstract

Representing and integrating continuous variables is a fundamental capability of the brain, often relying on ring attractor circuits that maintain a persistent bump of activity. To investigate how such structures can self-organize, we trained a recurrent neural network (RNN) on a ring-based path integration task using population-coded velocity inputs. The network autonomously developed a modular architecture: one subpopulation formed a stable ring attractor to maintain the integrated position, while a second, distinct subpopulation organized into a dissipative control unit that translates velocity into directional signals. Furthermore, systematic perturbations revealed that the precise topological alignment between these modules is essential for reliable integration. Our findings illustrate how functional specialization and biologically plausible representations can emerge from a general learning objective, offering insights into neural self-organization and providing a framework for designing more interpretable and robust neuromorphic systems for navigation and control.

Keywords: recurrent neural networks, bump attractors, population coding, continuous variable integration, nonlinear dynamics, network physiology, neural representation

1 Introduction

A central challenge in physiology is to uncover how complex neural and physiological systems achieve robust, flexible information processing through the structured interaction of distributed components—a phenomenon deeply rooted in the principles of self-organization. In recent years, the rapidly growing field of network physiology has emphasized understanding the coordinated dynamics and functional connectivity within and across distinct subsystems, with the goal of elucidating mechanisms underlying adaptive behavior, resilience, and nonequilibrium phase transitions in living systems (Bartsch et al., 2015; Ivanov and Bartsch, 2014). Neural networks, in particular, serve as canonical models of such emergent dynamics, in which collective behaviors—ranging from oscillations to discrete or continuous attractor states—arise from recurring patterns of connectivity and population-level coding (Vladimirovich Maslennikov et al., 2022).

Within this framework, population coding and attractor dynamics have been recognized as fundamental organizing principles that underpin neural computation (Haken, 1983). Paradigmatic examples include bump and ring attractor networks, which enable the representation and integration of continuous variables. These networks are not confined to a single species or brain region but represent a species-agnostic neural motif for computation, found in contexts ranging from spatial orientation in mammals (Zhang, , 1996; Moser et al., 2008) to the internal compass of insects like the fruit fly (Kim et al., 2017), birds (Ben-Yishay et al., 2021), and even fish (Vinepinsky et al., 2020). This evolutionary convergence implies that by studying these motifs, we can gain insight into the general physiological principles of brain function across diverse species (Khona and Fiete, 2022; Basu et al., 2024). These distributed representations also exemplify physiological robustness, permitting reliable encoding in the presence of noise and fluctuating inputs (Averbeck et al., 2006; Pouget et al., 2000). In the hippocampal-entorhinal circuit, for instance, place and grid cells collectively generate a dynamic map of the environment (Hafting et al., 2005; McNaughton et al., 2006). Theoretically, such 2D spatial representations can be constructed by combining multiple 1D ring attractors, each encoding orientation along a different axis, highlighting the role of the ring attractor as a fundamental computational building block (Bush and Burgess, 2014).

Crucially, network-based models such as ring attractors not only capture these population-level coding schemes, but also provide a theoretical framework—rooted in the self-organization of collective variables—for understanding how internal states can be flexibly updated, maintained, and read out by downstream systems (Skaggs et al., 1994; Fiete et al., 2008; Seeholzer et al., 2019). In head direction systems and cortical integration circuits, the spontaneous emergence of ring attractors as solutions to path integration and spatial memory tasks exemplifies how nonequilibrium transitions and bifurcations in network structure give rise to functionally specialized modules (Georgopoulos et al., 1986; Salinas and Abbott, 1994; Ganguli and Sompolinsky, 2012; Knierim et al., 1995; Heinze et al., 2018). Such processes underscore a general principle, highlighted in the Synergetics tradition: physiological networks can self-organize connectivity patterns and explicit coding strategies to achieve both specialization and adaptive coordination among functional subunits.

Despite considerable progress, a central open question in both physiology and artificial intelligence remains: How can such modular, interpretable architectures—capable of robust continuous integration and flexibly dealing with circular variables—arise autonomously through learning mechanisms? Here, we use the term “autonomously” to describe the emergence of structured connectivity and dynamics driven by a high-level functional objective, rather than through handcrafted design. And how do the emergent patterns of organization constrain or enhance physiological function? Building bridges between biological plausibility and artificial network design is thus essential for advancing our theoretical understanding and for informing translational neuromorphic engineering (Banino et al., 2018; Izzo et al., 2023; Ganguli and Sompolinsky, 2012).

Motivated by these themes, and inspired by the pioneering insights of Hermann Haken, we here examine how a recurrent neural network trained with explicit population coding can self-organize into functional, physiologically congruent subpopulations that support robust, interpretable continuous variable integration. By dissecting the network’s emergent architecture, dynamical coding strategies, and the response to systematic perturbations, we aim to illuminate general organizing principles that underlie both neural computation and the broader dynamics of self-organized physiological networks.

2 Materials and methods

We address the problem of continuous navigation on a ring—a canonical task in theoretical and systems neuroscience—where an agent receives, at each timestep t , a velocity signal vt[1,1] as input and must integrate it over time to estimate its angular position xt0,2π on the ring. The velocity input vt is generated synthetically using a diverse range of motion profiles, which ensures robust training and testing of the network’s ability to perform under both stochastic and structured input conditions. Specifically, vt is drawn from both smoothed random walks v~t and deterministic regimes ut such as linear ramps and fixed-velocity intervals. The full vt is constructed as:

vt+1=vt+v~t+ut (1)
v~t+1=0.8v~t+αmηt+βv~t, (2)

where ηtU(1,1),αU(0,2),mU(10,625),βU(0.05,0.45) with vt values clipped to [1,1] to preserve realistic, bounded motion dynamics. This mixture of inputs produces rich and biologically plausible trajectories, challenging the neural network to generalize continuous integration across varied conditions.

To mimic the manner in which biological circuits represent continuous variables, we employ population coding for the input: the scalar velocity vt is encoded as a distributed neural activation pattern ItRn , with each neuron having a Gaussian tuning curve centered on a specific preferred velocity value. As illustrated in Figure 1, this encoding transforms a dynamic, time-varying scalar signal into a spatiotemporal pattern of activity across the input population, which then serves as the input to the recurrent network:

P:vtItRn. (3)

FIGURE 1.

Graph (a) depicts a velocity curve showing a sinusoidal pattern fluctuating between -1.0 and 1.0 over time from 0 to 100. Graph (b) represents a population code heat map, with neuron activation indicated by colors from black (0.0) to yellow (1.0), mirroring the velocity curve.

Population coding of a dynamic input signal. Panel (a) shows an example of a time-varying velocity signal, here one period of a sinusoid, presented to the network over time. Panel (b) illustrates how this scalar profile is encoded as a spatiotemporal activity pattern across the input neuron population. Each neuron (y-axis) is tuned to a preferred velocity. As the input velocity changes over time (x-axis), the peak of the activity bump shifts across the population, creating a dynamic representation of the signal shown in (a).

Such population codes enable robust integration and flexible transformation of noisy or ambiguous sensory signals, paralleling mechanisms observed in the brain’s sensory and motor systems. Notably, our input population encodes only velocity, not angular position, which differs from some biological head-direction systems where conjunctive coding is observed (Yoder et al., 2015). This choice allows us to specifically investigate how a network can autonomously learn to transform a pure velocity signal into a stable angular representation through its recurrent dynamics.

The target coordinate at time t , denoted xt , reflects the agent’s true angular position as obtained by integrating the velocity input:

xt=0tvτdτmod2π. (4)

In our discrete-time simulation, this integral is approximated using a second-order Euler method. During training, this serves as the supervisory signal. The “target neuron” refers to the output neuron whose preferred position x~n is closest to xt at each step, a construct central to defining the network loss. We acknowledge that the use of an external, ground-truth supervisory signal is a simplification. In a biological context, such a signal would not be explicitly available. However, it can be interpreted as an abstraction of learning guided by other sensory modalities (e.g., visual landmarks) or by corrective feedback loops during development and exploration (Yoder et al., 2015; Levenstein et al., 2024). This supervised framework, also used in similar computational studies (Cueva and Wei, 2018), allows us to efficiently probe the types of network architectures and dynamics that are effective solutions for the task, revealing principles of self-organization that may be achieved through more implicit, biologically plausible learning mechanisms.

We arrange the neurons of the input and output populations to uniformly tile the relevant ranges, [1,1] and 0,2π respectively, defining each neuron’s preferred value by x~ . Population codes prevent the angular discontinuity that would otherwise occur at the 2π boundary, providing continuity for the network’s internal representations—a property essential for handling circular variables.

The overall task and network architecture, which consists of functionally distinct input and output populations coupled through a recurrent weight matrix, is schematized in Figure 2. The artificial recurrent neural network we trained evolves according to:

ht+1=ReLUWhhht+It, (5)

where Whh is the recurrent weight matrix. We employ the ReLU activation function due to its computational efficiency and qualitative biological plausibility (Andreevic et al., 2025). While our model utilizes non-spiking neurons to maintain computational tractability, its emergent dynamics offer insights into the functional organization that may be implemented by more biophysically detailed, spiking networks (Song and Wang, 2005). Network stability is supported by orthogonal initialization, gradient clipping, and a curriculum learning protocol that increases trial length over training. The weights are optimized with Adam gradient descent using a population-based error function (loss) described below.

FIGURE 2.

Illustration of a neural network diagram. A waveform on the left represents an input signal over time (t) with vertical axis labeled V. Arrows point to a network of connected neurons represented by numbered circles, linked by arrows labeled \(W_{hh}\) within a dashed box. On the right, a circular sequence labeled "target" and "n-th neuron" shows additional numbered nodes with directional arrows indicating connections. Additional label \(d\) with dashed arrow suggests a measurement or parameter.

Schematic of the ring navigation task and network architecture. The model performs path integration by processing a time-varying velocity profile, V(t) , as input (left panel). The network architecture (middle panel) consists of a recurrent layer with two functionally distinct populations: an input population that encodes the incoming velocity and an output population that integrates this signal to produce an activity pattern representing the angular coordinate. Both populations are coupled through the recurrent weight matrix, Whh . The output neurons represent a circular coordinate space, x (right panel). During training, the network learns to align its activity with a target coordinate. The parameter d represents the minimal angular distance between the preferred positions of adjacent neurons, defining the spatial resolution of the population code.

Our loss function is designed to reflect the graded, population-level readout present in biological codes. The neuron nearest to the target coordinate is treated as the “class” with the correct label, with adjacent neurons given smaller weights:

L=γt=1Tn=1Nht,n1Tt=1T1+tTαn=1N/2ωt,not,n, (6)

with

ot,n=logexpht,nn=1Nexpht,n, (7)

and

ωt,n=expdt,nn=1Nexpdt,n, (8)

where dt,n is the minimal circular distance between the target position xt and the preferred position x~n of the n -th output neuron. The hyperparameter α accentuates the importance of accurate integration later in the trial, while γ regularizes total network activity.

To prevent units in the output population from becoming permanently silent (a common issue in ReLU networks), we implemented a simple homeostatic mechanism. In addition to adding a small, loss-adaptive noise to the gradients during each training step, we also periodically reinitialized inactive neurons. At the end of several batches, any neuron whose activity remained below a small threshold (ϵ=106) for the entire trial duration was considered inactive. The weights of these inactive neurons are periodically reinitialized (using the average of their neighbors’ weights) and their optimizer state reset. This is analogous to homeostatic regulation in biological circuits and ensures all computational resources are utilized. 1

3 Results

We designed our network with a pre-defined modular structure to investigate functional specialization. The total population of N = 800 neurons was partitioned into two equally sized groups: the first 400 neurons were designated as the “output population” and received no direct external input, while the remaining 400 neurons were the “input population” and received the external velocity signal. Both populations were uniformly tiled over their respective domains, creating a structured basis for analysis of functional specialization and connectivity.

To dissect the computational dynamics, we express the network’s full state at time t as the concatenation of input and output population activities:

ht=htohti (9)

with hti and hto denoting the activities of input and output populations. The recurrent weight matrix Whh naturally decomposes as:

Whh=WooWoiWioWii (10)

corresponding to intra-population (diagonal blocks Woo,Wii ) and inter-population (off-diagonal blocks Woi,Wio ) connectivity.

The evolution of each sub-population is governed by:

ht+1i=ReLUWiihti+Wiohto+It (11)
ht+1o=ReLUWoohto+Woihti (12)

which capture the continuous integration, stabilization, and transformation of sensory input in a modular, population-based format.

A successfully trained network exhibits clear functional specialization, as shown in Figure 3. While the input population activity directly mirrors the incoming velocity signal (Figure 3a), the output population integrates this signal to maintain a stable, localized bump of activity representing the agent’s angular position (Figure 3b). This functional division is a direct result of the self-organized recurrent connectivity matrix (Figure 3c), where the output-to-output block (Woo) develops a robust ring-shaped attractor structure along its diagonal, which enables stable representations of position. We use the term ‘attractor’ to describe the functional behavior of the network, which creates quasi-stable states that are robust to noise and persist over timescales relevant for the task, even if they are not attractors in the strict mathematical sense of having infinite stability. In contrast, the input population’s recurrent weights (Wii) show a more localized structure. These learned weights facilitate, respectively, the persistence of spatial memory and the rapid encoding of transient velocity inputs.

FIGURE 3.

Four-panel figure depicting neural network activity and weight matrices. Panel (a) shows input and output neuron activity over time, with lines representing variables X and V. Panel (b) illustrates neuron activity pattern evolution. Panels (c) and (d) display weight matrices \(W_{hh}\) with different densities, using a color gradient to indicate weight magnitude from blue (negative) to red (positive). Time and activity scales are marked, with labels for neurons and time on the axes.

Self-organized network structure and dynamics during path integration. The network consists of an input population (neurons 1–400) and an output population (neurons 401–800). (a) Spatiotemporal activity of the population-coded input signal (It) fed to the network, with the original scalar velocity profile superimposed (purple curve). (b) The resulting spatiotemporal activity (ht) of the fully trained recurrent network, with the decoded output coordinate superimposed (red curve). The output population (top half) maintains a persistent bump of activity that integrates the input, while the input population (bottom half) dynamically reflects the signal shown in (a). (c) The learned recurrent weight matrix (Whh) from the fully trained network. The matrix reveals a clear self-organized modular structure, where the output-to-output block ( Woo , top-left) has formed a ring attractor with strong weights along the main diagonal. (d) The learned recurrent weight matrix (Whh) from a network that was trained with the feedback block from output to input (Wio) permanently nullified.

To further probe the specialization of network modules, we simulated the autonomous activity of the input and output populations in the absence of inter-population coupling. Here, each population was initialized with Gaussian bumps at multiple, distinct positions. This setup allows us to differentiate persistent from transient attractor dynamics. In this isolated condition, the activity of the input (h~i) and output (h~o) populations evolves according to:

h~t+1i=ReLUWiih~ti (13)
h~t+1o=ReLUWooh~to (14)

The results of these autonomous simulations, shown in Figure 4, confirm the functional roles of the two modules. The output population sustains the initial activity bumps for an extended period (Figure 4a), demonstrating a memory-like capability. By contrast, the input population shows rapid decay (Figure 4b), consistent with its role as a transient encoder. We note that the specific network instance visualized here shows some heterogeneity in its response, with stronger activity in one region of the ring. This is a feature of this particular trained network, as the effect varies across different training runs and is not a systematic bias of the model. The central finding demonstrated here—the clear functional contrast between the persistent output and dissipative input populations—is a robust result observed consistently across all successful networks.

FIGURE 4.

Heat maps representing neural activations over time for four panels labeled a, b, c, and d. Panels a and c display activations for 400 neurons over time intervals reaching 20 units, while panels b and d show activations for 400-800 neurons within 10-unit intervals. Activation ranges from -1.00 (blue) to 1.00 (red).

Autonomous dynamics reveal functional specialization of input and output populations. To test the intrinsic properties of the learned modules, we simulated their autonomous activity driven only by their internal recurrent connections (i.e., without external input or inter-population signals). Each population was initialized with five distinct Gaussian bumps of activity. (a,c) The output populations exhibit transiently persistent activity, maintaining the bumps for an extended period before eventually decaying to the zero state. This demonstrates a clear memory-like capability, essential for integration. Notably, the persistence in the fully connected network (a) is significantly more robust, sustaining the localized activity much longer than in the feedforward-only network (c). (b,d) In sharp contrast, the input populations from both the fully connected network (b) and the feedforward-only network (d) show dissipative dynamics, where the initial activity bumps decay rapidly. These simulations correspond to evolving the activity according to (Pouget et al., 2000; Hafting et al., 2005).

To investigate the role of modular connectivity, we selectively ablated the feedback pathway (Wio) from the output to the input population. This manipulation tests the importance of top-down signals for stabilizing network dynamics. We observed that while the purely feedforward architecture could still perform integration to some degree, the activity in the output population became less stable and exhibited significant drift over time, leading to a rapid degradation in positional accuracy.

The underlying reason for this instability is revealed by comparing the autonomous dynamics of the two networks in Figure 4. The output module from the network trained without feedback is intrinsically less stable, showing a much faster decay of activity (Figure 4c) compared to its counterpart from the fully connected network (Figure 4a). This highlights that the network learns to sustain activity for a prolonged but finite duration, a property we term transient persistence. This suggests that in the full network, the feedback pathway allows the input population to participate in a larger recurrent circuit that actively stabilizes the activity bump on the output ring. Without this top-down connection, the entire burden of maintaining persistent activity falls solely upon the internal recurrence of the output population (Woo) , rendering the memory trace more susceptible to decay and noise. This demonstrates the critical role of the complete recurrent structure, including feedback loops, for robust memory maintenance.

We then explored the mapping and transfer of control signals by delivering a velocity input that increased linearly from 1 to 1. The signal propagation from input to output was visualized and further tested by permuting the ordering of output neurons (breaking the topological alignment between population code and ring structure).

As illustrated in Figure 5b, a properly trained network successfully integrates the velocity ramp (Figure 5a). However, network performance collapses when the population code-to-ring mapping is disrupted by permuting the output neurons (Figure 5c), confirming that the encoded structure and network topology must remain aligned for successful integration. The output population in the unperturbed circuit receives a smoothly propagating population control signal, but this signal can no longer synchronize correctly with a shuffled arrangement, revealing the codependency of circuitry and coding motif.

FIGURE 5.

Heatmap panels showing neural activation patterns. Panel a depicts \(I_t\) with a gradient along neurons. Panel b shows \(W_{hh}h_t\) with complex patterns. Panel c is a permuted version of \(W_{hh}h_t\). Panels d to g display \(W_{oi}h_t^i\) with velocities on the x-axis and neurons on the y-axis. Activation levels range from -3 to 3, indicated by a color bar from blue to red.

Topological alignment of control signals is critical for integration. We probed the network’s integration mechanism using a linearly increasing velocity input. (a) The input velocity ramp from −1 to 1 over the trial. (b) The activity of the fully trained network (Whhht) shows successful integration, evidenced by the smoothly moving bump in the output population (neurons 1–400). (c) However, randomly permuting the order of neurons within the output population disrupts the learned topological mapping. This single change completely abolishes integration, resulting in chaotic network activity. (d–g) Visualization of the feedforward control signal (Woihti) from the input to the output population, plotted as a function of the input velocity. (d) For the fully trained network, the raw control signal appears unstructured. (e) However, when the output neurons are sorted by their peak response velocity, a highly structured diagonal pattern emerges, revealing a precise, topographically organized ‘push’ that systematically drives the activity bump. (f,g) Remarkably, the same analysis for the network trained without feedback reveals a nearly identical underlying control structure (g) when sorted. This striking similarity demonstrates that the dynamic mechanism for controlling the coordinate is learned via the feedforward pathway in both networks. This supports the hypothesis that the feedforward projection (Woi) is responsible for driving the movement of the representation, while the recurrent connections within the output population (Woo) are primarily responsible for stabilizing it.

Further analysis of the feedforward control signal (Woihti) provides a key insight into the network’s modular strategy. As shown in Figures 5d–g, while the raw control signal appears noisy (Figures 5d,f), sorting the output neurons by their peak response velocity reveals a highly organized diagonal structure in both the full network (Figure 5e) and the feedforward-only network (Figure 5g). The remarkable similarity between these two sorted signals strongly suggests that the fundamental mechanism for converting velocity into a directional “push” on the output population is implemented by the feedforward pathway (Woi) . This confirms that the role of the feedback pathway (Wio) is not to shape the control signal itself, but rather to contribute to the overall dynamic stability of the network, a conclusion consistent with the autonomous dynamics shown in Figure 4.

Taken together, these analyses show that population-coded recurrent networks can naturally self-organize into specialized modules for fast encoding and persistent memory. Faithful function relies not only on the learned synaptic weights, but also on the precise and consistent internal mapping between neural populations and their target representations. These features are hallmarks of modular, robust physiological computation as observed in biological navigation and memory systems.

4 Discussion and conclusion

In this study, we demonstrated that a recurrent neural network trained on a continuous integration task can autonomously self-organize into a modular architecture with functionally distinct and physiologically congruent subpopulations. Here, self-organization refers to the emergence of structured connectivity as a result of a supervised learning process, rather than arising from unsupervised, local update rules in the physical sense. Our findings contribute to bridging the gap between the dynamics of artificial neural networks and the principles of neural computation observed in biological systems.

Our central finding is the emergent division of labor within the network. The output subpopulation develops a ring attractor, a canonical structure for encoding circular variables like head direction and spatial orientation (Zhang, , 1996; Kim et al., 2017). This structure supports persistent, localized activity, enabling it to function as a robust memory module for the integrated position. Concurrently, the input subpopulation forms a dissipative, segment-like architecture that acts as a transient control unit, transforming velocity signals into directional commands that drive the movement of the activity bump on the ring. This modular separation—separating memory from control—is a key organizational principle in the brain, allowing for flexible and robust computation (Salinas and Abbott, 1994; Ganguli and Sompolinsky, 2012). Unlike models where ring connectivity is hardwired, here it emerges solely from the learning objective, suggesting that attractor dynamics are a natural and efficient solution for continuous variable integration.

A key distinction of our work, however, is the demonstration of the critical importance of topological alignment between these emergent modules. As shown in our permutation experiment (Figures 5a–c), the network’s function is not merely a product of its component parts but depends fundamentally on the learned, ordered mapping between the control signals from the input population and the spatial layout of the output ring attractor. This highlights that for distributed neural codes to be computationally effective, the wiring” must respect the coding”. Disruptions to this alignment, analogous to developmental disorders or brain injury, can lead to a catastrophic failure of function, even if the individual modules remain intact. Our perturbation analyses thus underscore the role of feedback and precise inter-module connectivity, echoing experimental findings where disrupting specific pathways compromises memory and integration (Seeholzer et al., 2019; McNaughton et al., 2006; Bonnevie et al., 2013).

Our results also inform the broader field of network physiology by providing a concrete computational example of how specialized subsystems can arise and coordinate within a larger, interconnected system. The balance between the persistent dynamics of the memory module and the dissipative dynamics of the control module illustrates how networks can achieve both stability and adaptability. This emergent coordination within a complex neural network serves as a key example of physiological resilience and the principles of system-level self-organization that are central to the study of network physiology (Bartsch et al., 2015; Ivanov and Bartsch, 2014; Ivanov, 2021).

4.1 Limitations and future directions

We acknowledge several limitations that open avenues for future research. First, our model employs non-spiking neurons and a supervised, gradient-based learning rule. While this framework is computationally powerful, future work should explore how these functional architectures could emerge using more biologically plausible spiking neurons and local, Hebbian-like learning rules (Song and Wang, 2005; Levenstein et al., 2024; Pugavko et al., 2023; Maslennikov et al., 2024). Second, the supervisory signal, while justifiable as an abstraction, could be replaced with reinforcement learning or unsupervised learning objectives to better model autonomous discovery in biological agents (Banino et al., 2018; Ivanov et al., 2025). Furthermore, training the network on more complex and physiologically grounded velocity profiles, such as those derived from animal tracking data (Sargolini et al., 2006), could reveal how network solutions are shaped by naturalistic input statistics. Exploring the network’s resilience to transient perturbations, such as temporary loss of connectivity between modules (Cooper and Mizumori, 2001), would also provide deeper insights into the robustness of these self-organized circuits.

On a translational level, our work illustrates how interpretable, modular architectures can be learned rather than handcrafted, offering a path toward more explainable AI and robust autonomous systems. This is particularly relevant for neuromorphic engineering and robotics, where many existing applications of ring attractors rely on hand-crafted weights (Rivero-Ortega et al., 2023). Our approach, where functional weights are learned, offers a promising route to developing more adaptive and flexible controllers. For mobile and field robotics, key considerations include not only robustness and interpretability but also low power consumption, a primary goal of neuromorphic systems (Izzo et al., 2023; Robin et al., 2022). Furthermore, the ring attractor motif is not limited to 1D orientation but serves as a foundational component for more complex spatial representations, such as modeling the 2D planar motion of a robot, linking back to the principles of grid cell computation (Knowles et al., 2023). The increasing availability of specialized neuromorphic hardware, such as Intel’s Loihi processors (Davies et al., 2018), and associated software frameworks like LAVA (Lava Framework Authors, 2022), makes these brain-inspired models increasingly viable for real-world, embedded applications where online learning and energy efficiency are paramount.

4.2 Conclusion

In summary, we have shown that explicit population coding guides a recurrent network to self-organize into a modular system comprising a ring attractor for memory and a dissipative controller for input processing. This emergent structure, highly reminiscent of biological circuits for navigation, depends critically on the precise topological alignment between its functional modules. Our findings underscore how general learning principles can give rise to specialized, interpretable, and physiologically plausible neural computations, advancing our understanding of both natural and artificial intelligence.

Funding Statement

The author(s) declare that financial support was received for the research and/or publication of this article. The work was supported by the Russian Science Foundation, grant No 23-72-10088.

Footnotes

1

The code used to generate the results in this study is available at https://github.com/p0mik13/PopulCoding_RingNavigationTask_F.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

RK: Formal Analysis, Investigation, Software, Visualization, Writing – original draft, Writing – review and editing. VT: Conceptualization, Methodology, Writing – review and editing, Investigation. OM: Conceptualization, Data curation, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Writing – original draft, Writing – review and editing. VN: Conceptualization, Methodology, Project administration, Supervision, Writing – review and editing.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  1. Averbeck B. B., Latham P. E., Pouget A. (2006). Neural correlations, population coding and computation. Nat. Rev. Neurosci. 7 (5), 358–366. 10.1038/nrn1888 [DOI] [PubMed] [Google Scholar]
  2. Banino A., Barry C., Uria B., Blundell C., Lillicrap T., Mirowski P., et al. (2018). Vector-based navigation using grid-like representations in artificial agents. Nature 557 (7705), 429–433. 10.1038/s41586-018-0102-6 [DOI] [PubMed] [Google Scholar]
  3. Bartsch R. P., Liu K. K. L., Bashan A., Ivanov P.Ch. (2015). Network physiology: how organ systems dynamically interact. PloS One 10 (11), e0142143. 10.1371/journal.pone.0142143 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Basu J., Nagel K. (2024). Neural circuits for goal-directed navigation across species. Trends Neurosci. 47 (11), 904–917. 10.1016/j.tins.2024.09.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Ben-Yishay E., Krivoruchko K., Ron S., Ulanovsky N., Derdikman D., Gutfreund Y. (2021). Directional tuning in the hippocampal formation of birds. Curr. Biol. 31 (12), 2592–2602.e4. 10.1016/j.cub.2021.04.029 [DOI] [PubMed] [Google Scholar]
  6. Bonnevie T., Dunn B., Fyhn M., Hafting T., Derdikman D., Kubie J. L., et al. (2013). Grid cells require excitatory drive from the hippocampus. Nat. Neurosci. 16 (3), 309–317. 10.1038/nn.3311 [DOI] [PubMed] [Google Scholar]
  7. Bush D., Burgess N. (2014). A hybrid oscillatory interference/continuous attractor network model of grid cell firing. J. Neurosci. 34 (14), 5065–5079. 10.1523/JNEUROSCI.4017-13.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Cooper B. G., Mizumori S. J. Y. (2001). Temporary inactivation of the retrosplenial cortex causes a transient reorganization of spatial coding in the hippocampus. J. Neurosci. 21 (11), 3986–4001. 10.1523/JNEUROSCI.21-11-03986.2001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Cueva C. J., Wei X.-X. (2018). Emergence of grid-like representations by training recurrent neural networks to perform spatial localization. arXiv Prepr. arXiv:1803.07770. 10.48550/arXiv.1803.07770 [Google Scholar]
  10. Davies M., Srinivasa N., Lin T.-H., Chinya G., Cao Y., Choday S. H., et al. (2018). Loihi: a neuromorphic manycore processor with on-chip learning. Ieee Micro 38 (1), 82–99. 10.1109/mm.2018.112130359 [DOI] [Google Scholar]
  11. Fiete I. R., Burak Y., Brookings T. (2008). What grid cells convey about rat location. J. Neurosci. 28 (27), 6858–6871. 10.1523/JNEUROSCI.5684-07.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Ganguli S., Sompolinsky H. (2012). Compressed sensing, sparsity, and dimensionality in neuronal information processing and data analysis. Annu. Rev. Neurosci. 35 (1), 485–508. 10.1146/annurev-neuro-062111-150410 [DOI] [PubMed] [Google Scholar]
  13. Georgopoulos A. P., Schwartz A. B., Kettner R. E. (1986). Neuronal population coding of movement direction. Science 233 (4771), 1416–1419. 10.1126/science.3749885 [DOI] [PubMed] [Google Scholar]
  14. Hafting T., Fyhn M., Molden S., Moser M.-B., Moser E. I. (2005). Microstructure of a spatial map in the entorhinal cortex. Nature 436 (7052), 801–806. 10.1038/nature03721 [DOI] [PubMed] [Google Scholar]
  15. Haken H. (1983). Synergetics: an introduction. 3rd edition. Berlin: Springer. [Google Scholar]
  16. Heinze S., Narendra A., Cheung A. (2018). Principles of insect path integration. Curr. Biol. 28 (17), R1043–R1058. 10.1016/j.cub.2018.04.058 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Ivanov P.Ch. (2021). The new field of network physiology: building the human physiolome. Front. Netw. physiology 1, 711778. 10.3389/fnetp.2021.711778 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Ivanov P.Ch, Bartsch R. P. (2014). “Network physiology: mapping interactions between networks of physiologic networks,” in Networks of networks: the last frontier of complexity (Springer; ), 203–222. [Google Scholar]
  19. Ivanov D. A., Larionov D. A., Maslennikov O. V., Voevodin V. V. (2025). Neural network compression for reinforcement learning tasks. Sci. Rep. 15 (1), 9718. 10.1038/s41598-025-93955-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Izzo D., Hadjiivanov A., Dold D., Meoni G., Blazquez E. (2023). “Neuromorphic computing and sensing in space,” in Artificial intelligence for space: AI4SPACE (Boca Raton, FL: CRC Press; ), 107–159. 10.48550/arXiv.2212.05236 [DOI] [Google Scholar]
  21. Khona M., Fiete I. R. (2022). Attractor and integrator networks in the brain. Nat. Rev. Neurosci. 23 (12), 744–766. 10.1038/s41583-022-00642-0 [DOI] [PubMed] [Google Scholar]
  22. Kim S. S., Rouault H., Druckmann S., Jayaraman V. (2017). Ring attractor dynamics in the drosophila central brain. Science 356 (6340), 849–853. 10.1126/science.aal4835 [DOI] [PubMed] [Google Scholar]
  23. Knierim J. J., Kudrimoti H. S., McNaughton B. L. (1995). Place cells, head direction cells, and the learning of landmark stability. J. Neurosci. 15 (3), 1648–1659. 10.1523/JNEUROSCI.15-03-01648.1995 [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Knowles T. C., Summerton A. G., Whiting J. G. H., Pearson M. J. (2023). Ring attractors as the basis of a biomimetic navigation system. Biomimetics 8 (5), 399. 10.3390/biomimetics8050399 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Kononov R. A., Maslennikov O. V., Nekorkin V. I. (2025). Dynamics of recurrent neural networks with piecewise linear activation function in the context-dependent decision-making task. Izv. VUZ. Appl. Nonlinear Dyn. 33 (2), 249–265. 10.18500/0869-6632-003147 [DOI] [Google Scholar]
  26. Lava Framework Authors (2022). Lava: a framework for neuromorphic computing. [Google Scholar]
  27. Levenstein D., Efremov A., Eyono R. H., Peyrache A., Richards B. (2024). Sequential predictive learning is a unifying theory for hippocampal representation and replay. bioRxiv, 2024–04. 10.1101/2024.04.28.591528 [DOI] [Google Scholar]
  28. Maslennikov O., Perc M., Nekorkin V. (2024). Topological features of spike trains in recurrent spiking neural networks that are trained to generate spatiotemporal patterns. Front. Comput. Neurosci. 18, 1363514. 10.3389/fncom.2024.1363514 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Maslennikov O. V., Pugavko M. M., Shchapin D. S., Nekorkin V. I. (2022). Nonlinear dynamics and machine learning of recurrent spiking neural networks. Physics-Uspekhi 65 (10), 1020–1038. 10.3367/ufne.2021.08.039042 [DOI] [Google Scholar]
  30. McNaughton B. L., Battaglia F. P., Jensen O., Moser E. I., Moser M.-B. (2006). Path integration and the neural basis of the’cognitive map. Nat. Rev. Neurosci. 7 (8), 663–678. 10.1038/nrn1932 [DOI] [PubMed] [Google Scholar]
  31. Moser E. I., Kropff E., Moser M. B. (2008). Place cells, grid cells, and the brain's spatial representation system. Annu. Rev. Neurosci. 31, 69–89. 10.1146/annurev.neuro.31.061307.090723 [DOI] [PubMed] [Google Scholar]
  32. Pouget A., Dayan P., Zemel R. S. (2000). Information processing with population codes. Nat. Rev. Neurosci. 1 (2), 125–132. 10.1038/35039062 [DOI] [PubMed] [Google Scholar]
  33. Pugavko M. M., Maslennikov O. V., Nekorkin V. I. (2023). Multitask computation through dynamics in recurrent spiking neural networks. Sci. Rep. 13 (1), 3997. 10.1038/s41598-023-31110-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Rivero-Ortega J. D., Mosquera-Maturana J. S., Pardo-Cabrera J., Hurtado-López J., Hernández J. D., Romero-Cano V., et al. (2023). Ring attractor bio-inspired neural network for social robot navigation. Front. Neurorobotics 17, 1211570. 10.3389/fnbot.2023.1211570 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Robinson B. S., Norman-Tenazas R., Cervantes M., Symonette D., Johnson E. C., Joyce J., et al. (2022). Online learning for orientation estimation during translation in an insect ring attractor network. Sci. Rep. 12 (1), 3210. 10.1038/s41598-022-05798-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Salinas E., Abbott L. F. (1994). Vector reconstruction from firing rates. J. Comput. Neurosci. 1 (1), 89–107. 10.1007/BF00962720 [DOI] [PubMed] [Google Scholar]
  37. Sargolini F., Fyhn M., Hafting T., McNaughton B. L., Witter M. P., Moser M.-B., et al. (2006). Conjunctive representation of position, direction, and velocity in entorhinal cortex. Science 312 (5774), 758–762. 10.1126/science.1125572 [DOI] [PubMed] [Google Scholar]
  38. Seeholzer A., Moritz D., Gerstner W. (2019). Stability of working memory in continuous attractor networks under the control of short-term plasticity. PLoS Comput. Biol. 15 (4), e1006928. 10.1371/journal.pcbi.1006928 [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Skaggs W., Knierim J., Kudrimoti H., McNaughton B. (1994). A model of the neural basis of the rat’s sense of direction. Adv. Neural Inf. Process. Syst. 7. Available online at: https://pubmed.ncbi.nlm.nih.gov/11539168/ . [PubMed] [Google Scholar]
  40. Song P., Wang X.-J. (2005). Angular path integration by moving “hill of activity”: a spiking neuron model without recurrent excitation of the head-direction system. J. Neurosci. 25 (4), 1002–1014. 10.1523/JNEUROSCI.4172-04.2005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Vinepinsky E., Cohen L., Perchik S., Ben-Shahar O., Donchin O., Segev R. (2020). Representation of edges, head direction, and swimming kinematics in the brain of freely-navigating fish. Sci. Rep. 10 (1), 14762. 10.1038/s41598-020-71217-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Yoder R. M., Peck J. R., Taube J. S. (2015). Visual landmark information gains control of the head direction signal at the lateral mammillary nuclei. J. Neurosci. 35 (4), 1354–1367. 10.1523/JNEUROSCI.1418-14.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Zhang K. (1996). Representation of spatial orientation by the intrinsic dynamics of the head-direction cell ensemble: a theory. J. Neurosci. 16 (6), 2112–2126. 10.1523/JNEUROSCI.16-06-02112.1996 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.


Articles from Frontiers in Network Physiology are provided here courtesy of Frontiers Media SA

RESOURCES