Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2017 Oct 2;13(10):e1005782. doi: 10.1371/journal.pcbi.1005782

A single-cell spiking model for the origin of grid-cell patterns

Tiziano D’Albis 1, Richard Kempter 1,2,3,*
Editor: Volker Steuber4
PMCID: PMC5638623  PMID: 28968386

Abstract

Spatial cognition in mammals is thought to rely on the activity of grid cells in the entorhinal cortex, yet the fundamental principles underlying the origin of grid-cell firing are still debated. Grid-like patterns could emerge via Hebbian learning and neuronal adaptation, but current computational models remained too abstract to allow direct confrontation with experimental data. Here, we propose a single-cell spiking model that generates grid firing fields via spike-rate adaptation and spike-timing dependent plasticity. Through rigorous mathematical analysis applicable in the linear limit, we quantitatively predict the requirements for grid-pattern formation, and we establish a direct link to classical pattern-forming systems of the Turing type. Our study lays the groundwork for biophysically-realistic models of grid-cell activity.

Author summary

When an animal explores an environment, grid cells activate at multiple spatial locations that form a strikingly-regular triangular pattern. Grid cells are believed to support high-level cognitive functions such as navigation and spatial memory, yet the origin of their activity remains unclear. Here we focus on the hypothesis that grid patterns emerge from a competition between persistent excitation by spatially-selective inputs and the reluctance of a neuron to fire for long stretches of time. Using a computational model, we generate grid-like activity by only spatially-irregular inputs, Hebbian synaptic plasticity, and neuronal adaptation. We study how the geometry of the output patterns depends on the spatial tuning of the inputs and the adaptation properties of single cells. The present work sheds light on the origin of grid-cell firing and makes specific predictions that could be tested experimentally.

Introduction

Grid cells are neurons of the medial entorhinal cortex (mEC) tuned to the position of the animal in the environment [1, 2]. Unlike place cells, which typically fire in a single spatial location [3, 4], grid cells have multiple receptive fields that form a strikingly-regular triangular pattern in space. Since their discovery, grid cells have been the object of a great number of experimental and theoretical studies, and they are thought to support high-level cognitive functions such as self-location [e.g. 5, 6], spatial navigation [e.g. 79], and spatial memory [10, 11]. Nevertheless, to date, the mechanisms underlying the formation of grid spatial patterns are yet to be understood [12, 13].

The attractor-network theory proposes that grid fields could arise from a path-integrating process, where bumps of neural activity are displaced across a low-dimensional continuous attractor by self-motion cues [1421]. The idea that self-motion inputs could drive spatial firing is motivated by the fact that mammals can use path integration for navigation [22], that speed and head-direction signals have been recorded within the mEC [23, 24], and that, in the rat [1, 25] but not in the mouse [26, 27], grid firing fields tend to persist in darkness. However, grid-cell activity may rely also on non-visual sensory inputs—such as olfactory or tactile cues—even in complete darkness [28]. Additionally, the attractor theory alone cannot explain how grid fields are anchored to the physical space, and how the properties of the grid patterns relate to the geometry of the enclosure [2931].

A different explanation for the formation of grid-cell activity is given by the so-called oscillatory-interference models [3236]. In those models, periodic spatial patterns are generated by the interference between multiple oscillators whose frequencies are controlled by the velocity of the animal. Speed-modulated rhythmic activity is indeed prominent throughout the hippocampal formation in rodents and primates [3740], particularly within the theta frequency band (4-12 Hz). Additionally, reduced theta rhythmicity disrupts grid-cell firing [41, 42], and grid-cell phase precession [43] is intrinsically generated by interference models; but see [44]. Despite their theoretical appeal, however, these models cannot explain grid-cell activity in the absence of continuous theta oscillations in the bat [45], and they are inconsistent with the grid-cell membrane-potential dynamics as measured intracellularly [46, 47]; see [48] for a hybrid oscillatory-attractor model.

Here we focus on the idea that grid-cell activity does not originate from self-motion cues, but rather from a learning process driven by external sensory inputs. In particular, it was proposed that grid patterns could arise from a competition between persistent excitation by spatially-selective inputs and the reluctance of a neuron to fire for long stretches of time [4953]. In this case, Hebbian plasticity at the input synapses could imprint a periodic pattern in the output activity of a single neuron. Spatially-selective inputs, i.e., inputs with significant spatial information, are indeed abundant within the mEC [5456] and its afferent structures [5761] And spike-rate adaptation, which is ubiquitous in the brain [62], could hinder neuronal firing in response to persistent excitation.

Kropff and Treves [49] explored this hypothesis by means of a computational model; see also [6367] and Sec Related models for similar works. The emergence of grid-like patterns was demonstrated with theoretical arguments and with numerical simulations of a rate-based network. However, because of a relatively abstract level of description, the outcomes of the model could not be easily confronted with experimental data. Specifically, the simulations included a network-level normalization mechanism that constrained the mean and the sparseness of the output activity, and it remained unsettled whether grid patterns could emerge in a single-cell scenario. Additionally, the synaptic weights did not obey Dale’s law. And the robustness of the model was not tested against shot noise due to stochastic spiking. Finally, the link between the numerical simulations and the underlying mathematical theory remained rather loose.

To overcome these issues, we propose here a single-cell spiking model based on similar principles as the model by Kropff and Treves [49], but that is, on the one hand, more biologically realistic, and on the other hand, better suited for mathematical treatment. Importantly, we show that grid patterns can emerge from a single-cell feed-forward mechanism needless of any network-level interaction (although recurrent dynamics may be still required to explain the coherent alignment of grid patterns [1]). To increase biological plausibility, we consider a stochastic spiking neuron model, and we constrain the synaptic weights to non-negative values (Dale’s law). Finally, by studying the model analytically, we quantitatively predict the requirements for grid-pattern formation, and we establish a direct link to classical pattern-forming systems via the Turing instability [68].

Results

Model of neural activity

We consider a single cell that receives synaptic input from N spatially-tuned excitatory neurons. Input spike trains Siin(t)kδ(t-ti,kin) for i = 1, 2, …, N are generated by independent inhomogeneous Poisson processes with instantaneous rates riin(t) where δ(t) is the Dirac delta function, and ti,kin is the timing of the kth input spike at synapse i. Similarly, the output spike train Sout(t)kδ(t-tkout) is generated by an inhomogeneous Poisson process with instantaneous rate rout(t) where tkout denotes the timing of the kth output spike.

We assume that inputs are integrated linearly at the output, and that the output neuron is equipped with an intrinsic spike-rate adaptation mechanism, that is,

rout(t)r0+0dτK(τ)i=1NwiSiin(t-τ) (1)

where r0 is a baseline rate, wi is the synaptic weight of input neuron i, and the function K is a temporal filter modeling the spike-rate adaptation dynamics. Note that the instantaneous output rate rout depends only on the temporal history of the input spikes and that there is no reset mechanism after the emission of an output spike.

The impulse response of the adaptation kernel K is the sum of two exponential functions:

K(t){1τSexp(-tτS)-μτLexp(-tτL)fort00fort<0 (2)

where τS and τL are the short and long filter time constants (0 < τS < τL), and the parameter μ > 0 sets the filter integral 0dtK(t)=1-μ (Fig 1A). Intuitively, at the arrival of an input spike, the firing probability of the output neuron is first increased for a short time that is controlled by the time constant τS, and then decreased for a longer time that is controlled by the time constant τL. This second hyper-polarization dynamics effectively hinders the neuron to fire at high rates for long stretches of time, mimicking a spike-rate adaptation mechanism [6971]. From a signal-processing perspective, the adaptation kernel K performs a temporal band-pass filtering of the input activity (Fig 1B), and the two time constants τS and τL control the resonance frequency kres at which the filter response is maximal. Note that in Sec Pattern formation with after-spike potentials we study a variant of the present model where neuronal adaptation is obtained though after-spike hyperpolarizing potentials associated to the output activity of the neuron.

Fig 1. The temporal adaptation kernel K.

Fig 1

(A) Impulse response of the filter (Eq 2). A positive peak with amplitude K(0) = 1/τSμ/τL ≈ 3.4 spikes/s is followed by a slow negative response. Note that the kernel is small for t > τmax, i.e., |K(t)| < 0.01|K(0)| for t > τmax, with τmax = 5τL = 0.8 s. (B) Frequency response of the filter. The dashed vertical line indicates the filter’s resonance frequency kres = 1.23 s−1. Parameter values: τS = 0.1 s, τL = 0.16 s, μ = 1.06. The integral of the filter is 1 − μ = −0.06.

Model of synaptic plasticity

We assume spike-timing dependent plasticity (STDP) at the input synapses [e.g. 7276]. Input and output spikes trigger weight changes Δwi according to the following rule:

  1. For each pair of a post-synaptic spike and a pre-synaptic spike at synapse i, we set
    Δwi=ηW(Δt) (3)
  2. For each pre-synaptic spike at synapse i, we set
    Δwi=η(β-αwi) (4)

where η ≪ 1 is a small learning rate, and the STDP learning window Wt) sets the weight change as a function of the time difference Δttpretpost between pre- and post-synaptic spikes. We consider a symmetric STDP learning window [77]

W(Δt)Wtot2τWexp(-|Δt|/τW) (5)

where the time constant τW > 0 controls the maximal time lag at which plasticity occurs, and Wtot=-dtW(t) is the integral of the learning window. The first part of the learning rule (Eq 3) is the classical Hebbian term whereas the second part (Eq 4) is a local normalization term that stabilizes the average synaptic strength wav=N-1i=1Nwi and prevents the individual weights to grow unbounded. This normalization term mimics local homoeostatic processes observed experimentally [7880]; see also [81] for a review. The parameters α > 0 and β set, respectively, the rate of weight decay and the target average weight wav (Sec Weight normalization). Importantly, the synaptic weights are constrained to non-negative values by imposing the hard bounds

wi0i. (6)

Model of input spatial tuning

We consider excitatory inputs with firing rates riin that are tuned to the spatial position of a virtual rat exploring a square arena of side-length L, i.e.,

riin(t)Ψiin(xt) (7)

where xt is the position of the virtual rat at time t, and Ψiin is a spatial tuning curve. We characterize the spatial tuning curves Ψiin in two alternative scenarios:

  1. spatially-regular inputs, i.e., each input has a single spatial receptive field;

  2. spatially-irregular inputs, i.e., each input has multiple spatial receptive fields at random locations.

The first scenario, which is reminiscent of hippocampal place-cell activity [3, 82, 83], is easier to study analytically and cheaper to simulate numerically. The second scenario, which is reminiscent of parasubicular activity [5761], is motivated by the anatomy of the entorhinal circuit (Sec Input spatial tuning and the origin of grid-cell patterns). In both cases, we consider circularly-symmetric receptive fields that cover the arena evenly. Indeed, place fields in open environments do not show systematic shape biases, and, in the absence of specific reward or goal locations, their centres are roughly homogeneously distributed [3, 5761, 82, 83]. Note, however, that border-like inputs [84, 85]—which are not radially-symmetric—are present in the real system, but not explicitly modeled here. Finally, for simplicity, we assume periodic boundaries at the edges of the arena.

Spatially-regular inputs

In the case of spatially-regular inputs, we assume tuning curves of the form

Ψiin(x)G(|x-ri|)withi=1,2,,N (8)

where ri is the receptive-field center of neuron i and G is a Gaussian function:

G(r)L2rav2πσ2exp(-r22σ2). (9)

The parameter σ > 0 sets the width of the receptive field, and rav is the average firing rate in the environment. We assume that the input receptive-field centers ri cover the entire arena evenly.

This input scenario is considered for the mathematical derivations in Sec Weight dynamics for spatially-regular inputs and for the numerical simulations in Secs Emergence of grid spatial patterns and Geometrical properties of the grid patterns.

Spatially-irregular inputs

In the case of spatially-irregular inputs, each tuning curve Ψiin is the sum of M > 1 Gaussian receptive fields with random amplitudes Aij and random receptive-field centers rij with i = 1, 2, …, N and j = 1, 2, …, M, that is,

Ψiin(x)1βij=1MAijG(|x-rij|). (10)

The scaling factors βi=j=1MAij normalize the inputs Ψiin to the same average rate rav, and all the superimposed fields share the same field size σ (Eq 9). The field amplitudes Aij are uniformly distributed in the range (0, 1), and the receptive-field centers rij are uniformly distributed in the environment.

This input scenario is considered for the mathematical derivations in Sec Eigenvalue spectrum for spatially-irregular inputs and for the numerical simulations in Sec Pattern formation with spatially-irregular inputs.

Model of spatial exploration

The movement of the virtual rat follows a smooth random walk that satisfies the following three assumptions: (i) the movement speed v is constant in time; (ii) the random walk is isotropic and ergodic with respect to the auto-covariance; (iii) the virtual-rat trajectories are smooth within time stretches shorter than the time length τmax = 5τL of the adaptation kernel K (Fig 1A). Note that assumption (i) is obviously not valid in general. However, because synaptic plasticity acts on a time scale that is much slower than behaviour, the relevant variable for pattern formation is the rat running speed averaged over long stretches of time (e.g. minutes), which can be considered approximately constant. We assume an average running speed of 25 cm/s, which is experimentally plausible [86]. Assumptions (ii) and (iii) hold by ignoring directional anisotropies deriving from the geometry of the environment, and by observing that experimental rat trajectories are approximately straight over short running distances (e.g, over distances shorter than 25 cm) [86].

Mathematically, the two-dimensional virtual-rat trajectories xt are sampled from the stochastic process

dXtdtv[cos(θt),sin(θt)]withθt=σθWt, (11)

where the angle θt sets the direction of motion and Wt is a standard Wiener process (Fig 2). The parameters v and σθ control the speed of motion and the tortuosity of the trajectory. Note that we also perform simulations with variable running speeds. In this case, the speed is sampled from an Ornstein-Uhlenbeck process with long-term mean v¯=v.

Fig 2. Example virtual-rat trajectories.

Fig 2

Colored lines denote example virtual-rat trajectories obtained by integrating Eq 11 starting at the center of the gray disk. Filled dots indicate the position of the virtual rat at time τmax = 5τL = 0.8 s. Note that the trajectories are smooth within time stretches shorter than τmax. Parameter values: v = 0.25 m/s, θσ = 0.7. The disk radius is max = 20 cm.

Mathematical results on grid-pattern formation

The grid-cell model presented above is studied both analytically and numerically. In this section, we obtain an equation for the average dynamics of the synaptic weights, and we derive the requirements for spatial pattern formation. In Sec Numerical results on grid-pattern formation we demonstrate the emergence of grid-like activity by simulating both the detailed spiking model and the averaged system. The analytical results presented here may be skipped by the less mathematically-inclined reader.

We study structure formation in the activity of an output cell by averaging the weight dynamics resulting from the stochastic activation of input and output neurons (Sec Model of neural activity) and the STDP learning rule (Sec Model of synaptic plasticity), while a virtual rat explores a two-dimensional enclosure and the inputs are spatially tuned (Secs Model of input spatial tuning and Model of spatial exploration). We take both ensemble averages across spike-train realizations and temporal averages within a time window of length T. The averaging time length T separates the time scale of neural activation (of the order of the width τW of the learning window W) from the time scale τstr of structure formation, i.e., τWTτstr. Because τstr is inversely proportional to the learning rate η (Eq 29), such averaging is always possible provided that the learning rate η is small enough. In other words, we assume that within a time T, the virtual rat has roughly explored the entire environment, but the synaptic weights did not change considerably. In this case, the dynamics of the synaptic weights wi is approximated by a drift-diffusion process, where the deterministic drift term reads [74]

η-1dw¯idt=(β-αw¯i)Siin(t)¯+-dsW(s)Siin(t+s)Sout(t)¯ (12)

with w¯i0. The functions Siin and Sout denote input and output spike trains (Sec Model of neural activity), the angular brackets denote ensemble averages over input and output spike trains, and the overbars denote temporal averages, i.e., f¯(t)T-1t-Ttdsf(s). Following Kempter et al. [74] we derive

Siin(t+s)Sout(t)¯=Siin(t+s)Sout(t)¯+w¯iSiin(t)¯K(-s), (13)

where the ensemble averages read

Siin(t)=riin(t) (14)
Sout(t)=rout(t)=(1)r0+0dτK(τ)j=1Nwjrjin(t-τ). (15)

Finally, from Eqs 1215 we obtain

η-1ddtw¯i=j=1NCijw¯j-aw¯i+bwithw¯i0 (16)

where we defined

Cij0dτK(τ)-dsW(s)riin(t+s)rjin(t-τ)¯ (17)
arav[α--dsW(s)K(-s)] (18)
brav(Wtotr0+β). (19)

Note that in deriving Eq 16 we approximated the temporal average of the input rates riin¯ with the spatial average rav of the input tuning curves Ψiin. This approximation holds with the assumption that in a time T the virtual rat roughly covers the entire space evenly.

By ignoring the non-linear weight constraints w¯i0, the average weight dynamics is described by a linear system with coupling terms Cij (Eq 16). The coefficients Cij are given by the temporal correlations of the input rates riin and rjin, filtered by the adaptation kernel K and the STDP learning window W (Eq 17).

To further simplify the calculations, we assume that the low-pass filtering introduced by the STDP learning window can be neglected for the purpose of studying pattern formation. In particular, we assume that the learning window W decays much faster than the changes in the input correlations riin(t+s)rjin(t-τ)¯ (Eq 17), which holds for τWσ/v. In this case, we obtain

CijWtot0dτK(τ)riin(t)rjin(t-τ)¯ (20)

where Wtot is the integral of the learning window (Eq 5).

Finally, by assuming smooth virtual-rat trajectories at constant speed v, the correlation matrix Cij can be estimated solely from the input tuning curves Ψiin and the adaptation kernel K (Sec Input correlation for general inputs, Eq 47):

CijWtotL20dτK(τ)|z|=τvdzΨiinΨjin|z (21)

where L2 is the area explored by the virtual rat. In Eq 21, the matrix element Cij is obtained by integrating the spatial cross-correlation of the input tuning curves ΨiinΨjin over circles of radius τv, and by weighting each integral with the amplitude of the adaptation kernel K at time τ. Note that Eq 21 holds for generic spatial tuning curves Ψiin.

Weight dynamics for spatially-regular inputs

To study the emergence of spatial patterns, we now consider the simplified scenario of spatially-regular inputs (Sec Spatially-regular inputs). That is, the input tuning curves Ψiin are circularly-symmetric Gaussian functions that cover the entire space evenly (Eqs 8 and 9). This input representation is particularly useful because it establishes a direct mapping between the neuron identity (the index i) and a position in physical space (the receptive-field center ri). Therefore, studying pattern formation in the activity of the output neuron is reduced to studying pattern formation in the space of the synaptic weights. Note, however, that such a simple input scenario is not necessary for the formation of grid-cell patterns in general, as shown in Sec Pattern formation with spatially-irregular inputs.

With spatially-regular inputs, the average weight dynamics in Eq 16 can be rewritten by labeling the synaptic weights according to the corresponding receptive-field centers ri:

η-1ddtw¯(ri)=i=1NC(ri,rj)w¯(rj)-aw¯(ri)+b (22)

where w¯(ri)=w¯i and C(ri, rj) = Cij. Additionally, in the limit of a large number N ≫ 1 of input neurons and receptive fields that cover the environment with constant density ρ = N/L2, the sum in Eq 22 can be replaced by an integral over all the receptive-field centers r′:

η-1ddtw¯(r)=ρdrC(r,r)w¯(r)-aw¯(r)+b (23)

where the correlation function C(r, r′) is the continuous extension of the correlation matrix Cij = C(ri, rj). Because the inputs are translation invariant (Eq 8), the correlation function C is also translation invariant, i.e., C(r, r′) = C(rr′, 0) = C(rr′), where we omit the second argument 0 ≔ (0, 0) for readability. In this case, the integral in Eq 23 can be expressed as a two-dimensional convolution in space:

η-1ddtw¯(r)=ρdrC(r-r)w¯(r)-aw¯(r)+b. (24)

Fig 3 shows the correlation C as a function of the input receptive-field distance |rr′| for the adaptation kernel K in Fig 1 and Gaussian input fields with size σ = 6.25 cm (Eq 9). The function C has the shape of a typical Mexican-hat kernel, i.e., it is positive for short receptive-field distances (attraction domain), negative for intermediate distances (repulsion domain), and zero otherwise. In this case, the synaptic weights of close-by input fields grow together whereas the synaptic weights of input fields that are further apart are repelled from each other (Eq 24). Such a competitive Mexican-hat interaction is at the basis of many pattern-forming systems found in nature, and it is directly related to diffusion-driven instabilities of the Turing type [see e.g. 87].

Fig 3. Input correlation function C for spatially-regular inputs.

Fig 3

The function is circularly symmetric, i.e., it depends only on the distance |rr′| between the receptive-field centers r and r′ (Eq 54). In the attraction domain (red shaded area) the correlation is positive and the synaptic weights grow in the same direction. In the repulsion domain (blue shaded area) the correlation is negative and the synaptic weights grow in opposite directions. Parameter values: σ = 6.25 cm, rav = 0.4 s−1, τS = 0.1 s, τL = 0.16 s, μ = 1.06, Wtot = 1 s, L = 1 m, v = 0.25 m/s.

Eigenvalue spectrum for spatially-regular inputs

To study spatially-periodic solutions, we take the two-dimensional Fourier transform with respect to r at both sides of Eq 24:

η-1ddtw^(k)=(ρC^(k)-a)w^(k)+δ(k)b (25)

where we defined the Fourier transform pair

w^(k)drw¯(r)exp(-2πjk·r),w¯(r)=1(2π)2dkw^(k)exp(2πjk·r), (26)

k is a two-dimensional wave vector, and j=-1 is the imaginary unit. Solving Eq 25 for k ≠ (0, 0), we obtain

w^(k)=w^0(k)exp(ηλ(k)t) (27)

where w^0(k) denotes the weight spectrum at time t = 0, and we defined

λ(k)ρC^(k)-afork(0,0). (28)

The function λ(k) defines the eigenvalue spectrum of the dynamical system in Eq 24, and the corresponding eigenfunctions are the elements of the Fourier basis exp(2πj kr). Eq 28 is also called the dispersion relation of the system. Note that solving Eq 25 for k = (0, 0) one obtains the dynamics of the total synaptic weight, which is kept normalized by the learning rule (Sec Weight normalization).

From Eq 27, the Fourier modes of the synaptic weights w^(k) grow or decay exponentially with rates proportional to the eigenvalues λ(k). Therefore, a structure in the synaptic weights emerges on a time scale

τstr1ηλmax (29)

where λmax ≔ maxk[λ(k)] is the largest eigenvalue in the system.

Importantly, the eigenvalues λ(k) are linearly related to the Fourier transform of the input-correlation function C^(k) (Eq 28), which is circularly-symmetric for circularly-symmetric inputs. In this case, in Sec Input correlation for spatially-regular inputs (Eq 62) we derive

C^(k)WtotL24π2G˜2(k)K˜sp(k)withk|k| (30)

where G˜ and K˜sp (Eqs 63 and 64) are the zeroth-order Hankel transforms (Eq 59) of the input tuning curve G (Eq 9) and of the equivalent adaptation kernel in space

Ksp(r)(61)1rvK(rv). (31)

Finally, by plugging Eqs 30 into 28, we obtain

λ(k)ρWtotL24π2G˜2(k)K˜sp(k)-awithk0. (32)

From Eqs 27 and 32 we recognize a necessary condition for spatial patterns to emerge: the eigenvalue spectrum λ(k) = λ(k) shall have a global maximum λmax > 0 at a frequency kmax > 0. In this case, all the Fourier modes k at the critical frequency |k| = kmax are unstable (Eq 25), and spatially-periodic patterns could emerge.

Fig 4 shows the critical frequency kmax (panels A1 and B1) and the largest eigenvalue λmax (panels A2 and B2) as a function of the parameters of the adaptation kernel K, i.e., the short time constant τS, the long time constant τL, and the kernel integral 1 − μ (Eq 2). The input receptive-field width σ is kept constant. In panels A1 and B1, the green-shaded regions correspond to parameter values where periodic grid-like patterns could emerge (kmax > 0). Conversely, the white regions denote parameter values where place-cell-like receptive fields could emerge (kmax = 0) [88]. We note that the spatial scale of the periodic patterns depends on the long adaptation time constant τL (panel A1), but is largely unaffected by the short time constant τS (panel B1). Additionally, the largest spatial frequencies are obtained for small values of τL and negative kernel integrals (panel A1). This leads us to the following predictions: the grid scale shall depend on the long temporal dynamics of the adaptation kernel, and the smallest grid scales require adaptation kernels with an overall inhibitory effect on the activity of the output neuron. We also note that larger values of τL correspond to larger values of λmax (panel A2). Thus, we predict that grids at larger scales shall develop faster than grids at smaller scales (Eq 29).

Fig 4. Impact of the adaptation kernel on grid-pattern formation.

Fig 4

(A1-A2) Critical spatial frequency kmax (A1) and largest eigenvalue λmax (A2) as a function of the kernel integral 1 − μ and the long kernel time constant τL. The short time constant is τS = 0.1 s. The black lines are iso-levels (see annotated values). Regions enclosed by two adjacent iso-lines are colored uniformly (darker colors denote larger values). Within the black region in A1 we obtain λmax ≤ 0 s−1 (see white region in A2). Within the black region in A2 we obtain kmax = 0 m−1 (see white region in A1). The dashed horizontal line indicates zero-integral kernels. The star denotes the parameter values τS = 0.1 s, τL = 0.16 s, μ = 1.06 of the kernel in Fig 1. (B1-B2) Same as in A but varying the short kernel time constant τS. The long time constant is τL = 0.16 s. The eigenvalue spectrum is estimated from Eq 32. Further parameter values: σ = 6.25 cm, rav = 0.4 s−1, Wtot = 1 s, ρ = 900 m−2, L = 1 m, v = 0.25 m/s, a = 1.1 s−1.

Importantly, the formation of grid-like patterns also requires a nonlinearity in the system. Indeed, for triangular lattices to emerge, only three wave vectors k of the same length |k| shall survive. But this cannot be achieved in a linear system where all Fourier modes develop independently from each other (Eq 27). Yet the non-linear weight constraints imposed in our model (Eq 6) are sufficient to generate triangular patterns (Sec Emergence of grid spatial patterns).

In summary, the theory presented here gives necessary conditions for spatial pattern formation, and it predicts how the shape of the adaptation kernel K influences the scale of the grids and the relative time required for their formation. The theory remains however agnostic about the specific two-dimensional periodicity of the resulting patterns, i.e., it cannot predict whether the final solutions are, e.g., planar waves, square, rhomboidal, or triangular lattices. Further mathematical insights on this topic could be obtained by using perturbation methods [see e.g. 89], but this is beyond the scope of the current manuscript.

Numerical results on grid-pattern formation

In Sec Mathematical results on grid-pattern formation we derived an equation for the average dynamics of the synaptic weights wi, under the STDP learning rule and the stochastic activation of input and output neurons (Eq 16). In the case of spatially-regular inputs, we then computed the systems eigenvalue spectrum λ(k) in terms of the Gaussian input tuning curve G and the temporal adaptation kernel K (Eq 32). We showed that periodic spatial patterns could emerge if the eigenvalue spectrum λ(k) had a global maximum λmax > 0 at a frequency kmax > 0 (Fig 4).

Fig 5A shows the eigenvalue spectrum λ(k) for a choice of the parameter values such that this condition is satisfied. With adaptation time constants τS = 0.1 s and τL = 0.16 s (Eq 2, Fig 1, star in Fig 4), and Gaussian input receptive fields of size σ = 6.25 cm (Eq 9), the eigenvalue spectrum peaks at the critical frequency kmax = 3 m−1. In the following, we simulate the emergence of grid-like patterns in this scenario.

Fig 5. Grid-pattern formation with spatially-regular inputs.

Fig 5

(A) Eigenvalue spectrum λ(k) of the averaged weight dynamics (Eq 32). The black solid line shows the continuous spectrum in the limit of infinite-size environments; the red dots show the discrete eigenvalues for a square arena of side length L = 1 m with periodic boundaries. The horizontal dashed line separates positive and negative eigenvalues. The vertical gray line indicates the critical spatial frequency kmax = 3 m−1. The eigenvalue at frequency k = 0 is not shown. Parameter values: τS = 0.1 s, τL = 0.16 s, σ = 6.25 cm. (B) Time-resolved distribution of N = 900 synaptic weights updated according to the STDP rule in Eqs 36. Red triangles indicate the time points shown in C. Inset: fraction of weights close to the lower saturation bound (wi < 5 ⋅ 10−3). (C) Top row: evolution of the synaptic weights over time. Weights are sorted according to the two-dimensional position of the corresponding input receptive-field centers. Note that each panel has a different color scale (maximum weight at the bottom-left corner, see B for distributions). Bottom row: Fourier amplitude of the synaptic weights at the top row. The red circle indicates the frequency kmax = 3 m−1 of the largest eigenvalue (see panel A). (D) Time evolution of weights' Fourier amplitudes |w^(k)| for wave vectors k at the critical frequency |k| = kmax. Wave vector angles (color coded) are relative to the largest mode at the end of the simulation (t = 106 s). The black triangles indicate time points in C. (E) Gridness score of the weight pattern over time. The gridness score quantifies the degree of triangular periodicity. See Sec Numerical simulations for further details and parameter values.

Emergence of grid spatial patterns

First, we simulate the detailed spiking model with spatially-regular inputs (Sec Spatially-regular inputs). The results are shown in Fig 5B–5E. In line with the theory, a structure emerges in the synaptic weights (Fig 5B and 5C) on a time scale of τstr = 1/(ηλmax) ≈ 5 ⋅ 104 s (Eq 29) where η = 2 ⋅ 10−5 is the learning rate and λmax ≈ 1 s−1 is the largest eigenvalue in the system. Additionally, the weight spectrum is quickly dominated by the critical frequency kmax = 3 m−1 (Fig 5C, bottom row) at which the eigenvalue spectrum has a global maximum (Fig 5A).

Importantly, the synaptic weights also develop a periodic triangular symmetry, which is reminiscent of grid-cell patterns. Such triangular symmetry emerges after a substantial fraction of weights has hit the low saturation bound (Eq 6, Fig 5B, inset). Periodic pattern formation is indeed a strictly non-linear phenomenon, and excluding the spike generation process, weight saturation is the only non-linearity present in the system. In the linear regime, all Fourier modes k with frequency |k| = kmax exponentially grow with equal rate ηλmax and independently from each other (Eq 25). In this case, the random weight pattern at time t = 0 s is amplified at the frequency kmax, but no periodic structure emerges. In the non-linear regime, instead, the exponentially growing modes are mutually coupled, and a spontaneous symmetry breaking occurs: only three Fourier modes with wave vectors that are 60 degrees apart survive in our simulations (see Fig 5C and 5D).

In the example of Fig 5, a triangular symmetry starts to emerge after 2 ⋅ 105 s, i.e., about 50 hours of exploration of the virtual rat. Yet the time scale of learning depends on the learning rate η and on the largest eigenvalue λmax (Eq 29), which are under-constrained by experimental data (see Eq 32 for the dependence of λmax on other model parameters). From a theoretical standpoint, the speed of learning is limited by the noise in the system, which is due to the virtual-rat random walk and the stochastic spiking of the neurons. To theoretically explore this limit and test the robustness of the model against noisy initial conditions, we simulated the development of the synaptic weights for different values of the learning rate and multiple random initializations of the synaptic weights. The results are reported in Fig 6A. With larger learning rates, grid-like patterns emerge faster. However, if the learning rate is too large, e.g., η = 10 ⋅ 10−5 in our simulations, the gridness score fluctuates at low levels and no stable grid pattern emerges (yellow line in Fig 6A). Therefore, our results suggest that tens of hours of spatial exploration are required for stable grid patterns to emerge. Finally, the model is robust to random initializations of the synaptic weights, and to variations of the running speed of the virtual rat (Fig 6B).

Fig 6. Time scales of learning.

Fig 6

(A) Median gridness scores of the input synaptic weights for 40 random weight initializations and different learning-rate values, i.e., η = (2, 3, 5, 10) ⋅ 10−5. The weight development is simulated with the detailed spiking model with spatially-regular inputs and constant virtual-rat speed (see also Fig 5). (B) Median gridness scores of the input synaptic weights simulated with constant (black line) and variable (green line) virtual-rat speeds for 40 random weight initializations. Variable running speeds are obtained by sampling from an Ornstein-Uhlenbeck process with long-term mean v¯=0.25 m/s, volatility σv = 0.1 m ⋅ s−1.5 and mean-reversion speed θv = 10 s−1. The inset shows the distribution of running speeds (mean: 0.25 m/s std: 0.02 m/s). Note that the long-term mean v¯ of the process equals the speed v in constant-speed simulations. See Sec Numerical simulations for further details and additional parameter values.

Geometrical properties of the grid patterns

We now discuss the geometrical properties of the simulated grid patterns. A periodic triangular grid is characterized by three fundamental properties: i) the grid scale, i.e., the distance between two neighboring peaks; ii) the grid spatial phase, i.e., the spatial offset of the grid peaks with respect to a reference point; and iii) the grid orientation, i.e, the angle between one of the three grid axes and a reference direction.

Grid scale

In our model, the grid scale is set by the critical frequency kmax at which the eigenvalue spectrum has a global maximum (Eq 32 and Fig 5). This critical frequency depends only on the movement speed v of the virtual rat, the width σ of the input tuning curve G, and the temporal dynamics of the adaptation kernel K (Fig 4). Therefore, grid patterns at different scales are obtained, for example, by varying the width σ of the input receptive fields or the long time scale τL of the adaptation kernel (Fig 7, see also Fig 4). This theoretical result is consistent with the facts that spatial tuning in the hippocampal formation is typically broader ventrally than dorsally [9092], and that grid scales vary in the same direction [1, 93]. Additionally, we predict that the adaptation time scale may also have a dorso-ventral gradient, similarly to other intrinsic cellular properties in the mEC [e.g. 71, 9497].

Fig 7. Spatial scale of the grid patterns.

Fig 7

Example grid patterns obtained with different adaptation kernels K (Eq 2, top row) and different input tuning curves G (Eq 9, left-most column). For each choice of the functions K and G, the synaptic weights (left) and their corresponding Fourier spectra (right) at the end of the simulation are shown (t = 106 s). The synaptic-weight maps have different color scales (maximal values at the bottom-left corner). The red circles indicate the spatial frequency kmax of the weight patterns. Synaptic weights were obtained by simulating the average weight dynamics in Eq 16. Note that we used a larger enclosure (L = 2 m) as compared to the one in Figs 5 and 6 (L = 1 m). See Sec Numerical simulations for further details and parameter values.

Grid spatial phase

With evenly-distributed input fields and periodic boundaries, the spatial phases of the grid patterns depend only on the initial condition of the synaptic weights, i.e., random weight initializations result in uniformly-distributed grid phases (Fig 8A1 and 8B1). This result is in line with the phases of nearby grid cells being roughly evenly distributed in experimental data [1], but see also [98]. Yet it remains unclear whether the same results would be obtained in the case of non-periodic boundaries.

Fig 8. Geometric properties of the grid patterns.

Fig 8

(A) Distribution of grid spatial phases (A1) and grid orientations (A2) for patterns at frequency kmax = 3 m−1 in an arena of side-length L = 2 m (σ = 6.25 cm, τL = 0.16 s; see also Fig 7, bottom-left panel). Distributions were obtained from the average weight dynamics in Eq 16 for 200 random initializations of the synaptic weights (t = 106 s). Only patterns with gridness scores larger than 0.5 were considered (197/200). Panel A3 shows example weight patterns for the two most common orientations in A2 (maximal values at the bottom-left corner). (B) Same as in A but for patterns at spatial frequency kmax = 2 m−1 in an arena of side-length L = 2 m (σ = 6.25 cm, τL = 0.35 s; see also Fig 7, bottom-right panel). A fraction of 182/200 grids had a gridness score larger than 0.5. See Sec Numerical simulations for further details and parameter values.

Grid orientation

With periodic boundary conditions, our model produces grid orientations that are distributed non-uniformly. Precisely, the distribution of grid orientations depends on the scale of the pattern relative to the size of the environment, e.g., in the same environment patterns at different scales tend to align differently (compare panels A2 and B2 in Fig 8). In the examples of Fig 8A3, one of the grid axes tends to align to a border of the arena whereas in the examples of Fig 8B3 one of the grid axes tends to align to a diagonal of the arena. Similar results are obtained by keeping the grid scale fixed and varying the size of the environment, e.g., compare Fig 8A2 (kmax = 3 m−1 and L = 2 m) and Fig 9F (kmax = 3 m−1 and L = 1 m). In general, we expect grid orientations to be uniformly distributed only in infinite-sized environments, or in environments that are much larger than the pattern size. Nevertheless, because grid orientation depends on the boundary conditions, it remains difficult to compare the distributions obtained here with the ones observed experimentally [1, 93, 99, 100]. Finally, in order to explain grid alignment across cells and/or environments [1, 93], collateral interactions between developing grid cells may be required [50, 51, 101] (see also Sec Recurrent dynamics).

Fig 9. Grid-pattern formation with spatially-irregular inputs.

Fig 9

(A) Four examples of irregular input firing-rate maps (top row) and the corresponding Fourier spectra (bottom row). The maximal firing rate (spikes/s) is reported at the bottom-left corner. The red circles indicate the spatial frequency kmax = 3 m−1. (B) Four examples of output firing-rate maps (top row) and the corresponding Fourier spectra (bottom row). The gridness score is reported at the bottom-right corner. Output firing-rate maps were estimated from the average weight dynamics in Eq 16 (t = 106 s) for four different realizations of the spatial inputs. (C-F) Distribution of gridness scores (C), grid spatial frequencies (D), grid spatial phases (E), and grid orientations (F) for 100 random realizations of the spatial inputs. The red vertical line in C indicates the mean score (0.77). See Sec Numerical simulations for further details and parameter values.

Pattern formation with spatially-irregular inputs

We demonstrated the emergence of grid-like patterns in the case of spatially-regular inputs, i.e., for each input cell having a single Gaussian receptive field in space (Sec Spatially-regular inputs). We now show that similar results are obtained in the case of spatially-irregular inputs (Sec Spatially-irregular inputs). We generate spatially-irregular inputs Ψiin by superimposing M > 1 Gaussian receptive fields with equal width σ, but random centers and random amplitudes (Eq 10, see Fig 9A for examples). The functions Ψiin are normalized such that their average firing rate rav is constant for all input neurons and independent from the number M of superimposed receptive fields.

We test grid-pattern formation in this scenario by simulating the average dynamics of the synaptic weights (Eqs 16 and 21) for random realizations of the input tuning curves Ψiin, with N = 3600 input neurons and M = 10 receptive fields per neuron. We then estimate output firing-rate maps from the synaptic weights at the end of the simulations (t = 106 s). The results are shown in Fig 9B and 9C. In the majority of the cases (73/100) a regular grid-like pattern emerges at the output.

Like in the case of spatially-regular inputs, the spatial scale of the output patterns depends on the long adaptation time constant τL and on the width σ of the input receptive fields. Indeed, for σ = 6.25 cm and τL = 0.16 s, we obtain output grid patterns with spatial frequency kmax = 3 m−1 (Fig 9D), which is equal to the one obtained for spatially-regular inputs with the same parameter values (Figs 5 and 7, bottom-left panel). This can be understood by the fact that the expected eigenvalue spectrum for spatially-irregular inputs 〈λirr(k)〉 is qualitatively similar to the eigenvalue spectrum λ(k) for spatially-regular inputs (Sec Eigenvalue spectrum for spatially-irregular inputs, Eq 72):

λirr(k)Φλ(k)+const. (33)

where 0 ≤ Φ ≤ 1 is a scale factor. We also find that the scale factor Φ depends on the number M of superimposed fields, i.e., Φ ≈ 4/(3M) for M > 3 (Eq 82), meaning that structure formation is slower for larger numbers of superimposed fields (Eq 29).

Finally, like in the case of spatially-regular inputs, with periodic boundary conditions the spatial phases of the simulated grids distribute evenly in the arena (Fig 9E), and the grid orientations tend to cluster according to the grid scale and the size of the environment (Fig 9F, see also Fig 8A2 for the same grid scale in a larger environment).

Discussion

We studied the origin of grid-cell patterns in a single-cell spiking model relying solely on 1) spatially-tuned feed-forward inputs, 2) spike-rate adaptation, 3) and synaptic plasticity at the input synapses. We considered two input scenarios: spatially-regular inputs (reminiscent of place-cell activity), and spatially-irregular inputs (reminiscent of parasubicular activity). First, we studied the average dynamics of the system analytically, and we derived necessary conditions for the emergence of spatially-periodic solutions (Sec Mathematical results on grid-pattern formation). We then simulated the model numerically, and showed that grid-like patterns emerge both with spatially-regular and spatially irregular inputs (Sec Numerical results on grid-pattern formation). In the following, we discuss the main assumptions and predictions of our model.

Input spatial tuning and the origin of grid-cell patterns

We assumed that the feed-forward input activity is spatially tuned. Such spatial tuning could be provided by hippocampal place cells, or by other cortical or sub-cortical structures with less regular spatial firing. From a theoretical point of view, we find that grid patterns emerge faster with place-cell-like inputs, i.e., with inputs having a single receptive field in space. From an anatomical point of view, both scenarios seem plausible. On the one hand, grid-cell activity requires excitatory drive from the hippocampus [102], which projects to the deep layers of the mEC [103, 104] where grid cells are found [23, 60]. On the other hand, parasubicular inputs target layer II of the mEC [61, 105108] where grid cells are most abundant [23, 60]. Although a small fraction of parasubicular cells already shows grid-like tuning [60, 61], the activity in parasubiculum is often characterized by multiple spatially-irregular fields [5761] similar to those assumed in our model (Fig 9A).

That grid-cell activity could originate from parasubicular inputs is further supported by the detailed layout of the entorhinal circuit. Layer II principal neurons segregate into stellate and pyramidal cells, which are distinguished by their morphology, intrinsic properties [69], and immunoreactivity [109111]. Interestingly, pyramidal-cell somata cluster into anatomical patches [110, 111], which are preferentially targeted by parasubicular axons [61]; and the spiking activity in parasubiculum precedes the activity of layer II pyramidal cells by a few degrees in the theta cycle [61]. Such a network configuration suggests that grid patterns may originate in the layer II pyramidal cells via parasubicular inputs, and be inherited by the stellate cells via feed-forward projections. Consistent with this view is that both stellate and pyramidal cells show grid spatial tuning [55], and that direct intra-laminar connections are found from pyramidal onto stellate cells and not vice-versa [112, 113]; but see [114].

In summary, our model is consistent with entorhinal grid-cell activity originating either in the superficial layers via parasubicular input or in the deep layers via hippocampal input. It is also possible that multiple sites of origin exist, and that grid-like tuning is inherited—and even sharpened—via feed-forward projections from the deep to the superfical layers [115119] or from the superficial to the deep layers [104].

Spike-rate adaptation

Our grid-cell model relies on the presence of a spike-rate adaptation mechanism. Spike-rate adaptation has been observed throughout the cortex [62], and is prominent in layer II of the mEC, in both stellate and pyramidal neurons [69, 70]. Yoshida et al. [71] also reported a dorso-ventral gradient in the adaptation strength of layer II entorhinal cells. However, because adaptation was found to be stronger ventrally than dorsally, Yoshida et al. [71] interpreted their results as evidence against grid-cell models based on adaptation. Yet the critical variable controlling the grid scale is not the strength of adaptation, but rather its temporal dynamics (Fig 7), which was not systematically analyzed [71]; see also [101] for a similar discussion on this point.

We modeled spike-rate adaptation by applying a temporal kernel K to the input spike trains (Eq 1). The kernel K, was composed of a brief depolarization peak and a slower hyper-polarizing potential (on a time scale of hundreds of milliseconds). Such a slow hyper-polarizing potential reduced the output firing rate in response to persistent excitation, and it filtered the input activity in a low-frequency band (i.e. with a resonance frequency of about 1 Hz, see Fig 1). The shape of the kernel was motivated by long-lasting hyper-polarizing potentials following excitatory post-synaptic potentials found in hippocampal CA1 pyramidal neurons [120], although similar responses have not been observed in the mEC yet.

However, the formation of grid-cell patterns could rely on any other cellular or synaptic mechanism that effectively acts as a band-pass filter on the input activity. A candidate mechanism is the after-spike hyperpolarizing potential (AHP). AHPs are indeed observed in the superficial layers of the mEC where single action potentials are followed by both a fast (2-5 ms) and a medium AHP (20-100 ms) [69, 97, 121]. To assess whether such hyperpolarizing potentials could underlie grid-pattern formation, we extended our model to account for AHPs (Sec Pattern formation with after-spike potentials). However, we found that grids at typical spatial scales cannot be obtained by AHPs alone. Yet after-spike potentials could amplify the effects of a band-pass filtering mechanism that is already present at the input.

Spike-rate adaptation could also rely on hyperpolarization-activated cation currents (Ih), which depend on HCN channels [122, 123]. Fast Ih currents (mediated by HCN1 channels) have been shown to control the theta-frequency resonance of entorhinal stellate cells in vitro [95, 97, 124126]. Instead, slower Ih currents (mediated by HCN2-4 channels) could generate in entorhinal cells the low-frequency resonance assumed by our model (Fig 1B).

Synaptic plasticity

We propose that grid-cell patterns emerge from a synaptic reorganization of the mEC network, which is assumed to be plastic. This is in line with both LTP and LTD being reported in the entorhinal cortex [121, 127130], but see also [131]. Additionally, asymmetric STDP was observed in the mEC [76]. Although we used a symmetric learning window in our model, the exact window shape has little effect on grid-pattern formation, provided that its temporal width (on the order of tens of milliseconds) is much shorter than the correlation length of the input activities (on the order of hundreds of milliseconds).

Structure formation via Hebbian learning is typically a slow process. In our model, grid-like patterns emerge on a time scale that is inversely proportional to the learning rate η and to the maximal eigenvalue λmax (Eq 29). The latter depends on the spatial density ρ = N/L2 of input receptive fields, on the integral Wtot of the learning window, on the shapes of the input-tuning curves G, and on the dynamics of the adaptation kernel K (Eq 32). Because most of these quantities are under-constrained by empirical data, a direct comparison with experimental time scales remains difficult. Yet learning shall be slow enough such that the input correlations that drive structure formation dominate over random fluctuations of the synaptic weights, which are due to the random walk of the virtual rat and the shot noise of the stochastic spiking. In our simulations, we find that this requires tens of hours of spatial exploration (Fig 6A).

Such slow process may seem in contrast with grid-cell activity appearing immediately in a novel environment [1, 132]. However, grid-like tuning may not need to be learned in each environment anew, but rather recalled—and possibly refined—from the experience of similar environments explored in the past. Although hippocampal place cells [133, 134] and entorhinal non-grid spatial cells [56] seem to remap completely in novel spaces, pattern formation could still leverage on residual correlations across environments that are hardly observable from the simultaneous recordings of only a few tens of neurons. Additionally, grid-cell learning could generalize across spatial contexts through border and boundary-vector inputs [84, 85], which are invariant across environments.

We suggest that a structure in the synaptic weights may be formed during the animal’s ontogenetic development, i.e., within a two-week period after the animal leaves the nest [135137]. Consistent with this hypothesis is that stable spatial firing is observed before grid-cell maturation, e.g., hippocampal place cells develop prior to grid cells [135, 136] and irregular spatial cells are present before grid cells [137].

Recurrent dynamics

We studied the emergence of grid patterns in a purely single-cell model, ignoring any network-level interaction between the neurons. However, because excitatory and inhibitory recurrent circuits have been described in the mEC [19, 20, 112, 113, 138], grid cells are likely to be mutually coupled [139, 140]. Such recurrent connections could explain the modular organization of grid-cell properties [93, 101] and their coherent responses to environmental changes [139]. Feedback interactions within a module may also amplify an initially broad grid-tuning given by the feed-forward inputs, similarly to the sharpening of receptive fields in visual cortex [141, 142]. Finally, recurrent dynamics may sustain grid-like activity when the feed-forward inputs are temporally untuned, like in attractor models [14]. Still, spatially-tuned feed-forward inputs could be required for the initial formation of grid-like patterns [see e.g. 21].

Related models

Our work—and the one by Kropff and Treves [49]—belong to a broad category of grid-cell models based on spatially-tuned feed-forward inputs and Hebbian synaptic plasticity [6367]. In all these models, periodic spatial patterns arise via a common underlying principle: the input correlations that drive the dynamics of the synaptic weights have the form of a Mexican-hat kernel (Fig 3). What distinguishes the models among each other—and generates distinct predictions—is the specific mechanism by which such Mexican-hat interactions are obtained.

In our model, a Mexican-hat kernel results from the intrinsic adaptation dynamics of the output neuron, which controls the grid scale directly (Fig 7).

By contrast, in the models by Castro and Aguiar [63] and Stepanyuk [64], Mexican-hat correlations arise from the learning rule itself, i.e., by assuming that synaptic plasticity switches between LTP and LTD based on pre- and post-synaptic activities [143]. In this case, the grid spatial scale shall be affected by interfering with the learning rule.

In a different model, Dordek et al. [65] obtain Mexican-hat correlations by constraining the input activity to be effectively zero-mean. The authors discuss that such a zero-mean constraint could originate either from lateral inhibition or from a zero-mean temporal filter controlling the output activity of the neuron. In the latter case, the model by Dordek et al. [65] is analogous to the present one. We note, however, that effectively zero-mean inputs are neither necessary nor sufficient for grid patterns to emerge. Instead, pattern formation depends on the dynamics of the temporal filter and on the shape of the input tuning curves, but not on their means. This can be easily understood by considering the system’s eigenvalue spectrum in Fourier space (Eq 30), where the zero-frequency mode (k = 0) is not relevant for the emergence of spatially-periodic patterns. Also note that the smallest grid scales in our model are obtained with negative-mean temporal filters (Fig 4). Yet our results agree with the ones of Dordek et al. [65] in that the non-linearity introduced by imposing non-negative synaptic weights is sufficient for a triangular symmetry to emerge.

Alternatively, Mexican-hat correlations could emerge from phase-precessing feed-forward inputs [66]. In this case, grid-cell activity shall be impaired when phase precession is disrupted.

Finally, Weber and Sprekeler [67] proposed a model where the interplay between spatially-narrow feed-forward excitation and spatially-broad feed-forward inhibition generates a Mexican-hat kernel. This model predicts that the grid scale shall be affected by manipulating inhibitory inputs to the mEC.

Model predictions and conclusion

We presented a single-cell model for the origin of grid-cell activity based on Hebbian synaptic plasticity and spike-rate adaptation. Our work builds upon the model by Kropff and Treves [49] and improves its original formulation in several aspects: 1) grid-like patterns emerge form a purely single-cell mechanism independently of any network-level interaction; 2) neuronal activities are spike-based and stochastic; 3) the input synaptic weights are purely excitatory; 4) the dynamics of the synaptic weights is studied analytically and linked to classical Turing-like patterns.

The present model makes the following experimental predictions. First, grid-cell patterns shall be affected by disrupting synaptic plasticity during ontogenetic development, which is consistent with preliminary data from Dagslott et al. [144]. Second, adult grid-cell activity shall be influenced by systematic behavioral or environmental biases in the first weeks of spatial exploration, e.g., by rising animals in environments without boundaries or with non-zero surface curvature [52, 145]. Third, the grid scale shall be affected by three factors: 1) the spatial tuning-width of the feed-forward inputs; 2) the average speed of the rat during ontogenetic development; 3) the time constant of the recovery from spike-rate adaptation. Fourth, grids at larger scales shall develop faster as compared to grids at smaller scales (Fig 4).

We believe that manipulations of the intrinsic adaptation properties of single cells are key to distinguish our model from other feed-forward models based on Hebbian learning (Sec Related models). To this end, further experimental work shall be devoted to pinpoint the biophysical mechanisms underlying adaptation in the mEC. Extensions of the present model could also explain how the geometry of the enclosure affects grid-cell symmetry [99], and how grid-like tuning emerges in non-spatial contexts [146, 147].

To conclude, our study contributes to a better understanding of the fundamental principles governing grid-cell activity, and lays the groundwork for more biophysically-realistic grid-cell models.

Materials and methods

Weight normalization

Here we derive the dynamics of the mean synaptic weight wav=N-1i=1Nw¯i for a neuron with N synapses and temporally-averaged weights w¯i. We recall the weight dynamics in Eq 16

η-1ddtw¯i=j=1NCijw¯j-aw¯i+bwithw¯i0. (34)

By taking the average over the index i at both sides of Eq 34 we obtain

η-1ddtwav=(NCav-a)wav+b (35)

where we defined the mean correlation CavN−2ij Cij. Note that we used the property ∑j Cij = NCav for all i, which holds true for translation-invariant inputs. Therefore, for NCav < a, the mean weight wav decays exponentially with time constant

τav1η(a-NCav) (36)

to the normalization level

wavba-NCav. (37)

Input correlation for general inputs

In this section we estimate the input correlation matrix

Cij(20)Wtot0dτK(τ)riin(t)rjin(t-τ)¯withi,j=1,,N (38)

for general spatial tuning curves Ψiin and smooth movement trajectories of the virtual rat (Sec Model of spatial exploration). We start by computing the temporal average riin(t)rjin(t-τ)¯ of the product between the input activities riin(t) and the delayed input activities rjin(t-τ). We assume that the stochastic process Xt controlling the virtual-rat trajectory (Eq 11) is ergodic with respect to the auto-covariance, i.e.,

1T0Tdtxtxt-τ=Xt,Xt-τforT (39)

where the angular brackets denote statistical expectation. By using this ergodicity property (Eq 39) and the spatial tuning of the inputs (Eq 7), we derive

riin(t)rjin(t-τ)¯=Ψiin(xt)Ψjin(xt-τ)¯Ψiin(Xt)Ψjin(Xt-τ). (40)

Note that Eq 40 is only valid in an approximate sense because Eq 39 assumes T → ∞, but the averaging time window has finite length Tτstr where τstr is structure-formation time constant (Eq 29). From Eq 40 follows

riin(t)rjin(t-τ)¯Ψiin(Xt)Ψjin(Xt-τ) (41)
dxdxΨiin(x)Ψjin(x)p(x,t,x,t-τ) (42)
=dxdxΨiin(x)Ψjin(x)p(x,t-τ|x,t)p(x,t) (43)
=1L2dxdxΨiin(x)Ψjin(x)p(x,t-τ|x,t) (44)

where the integrals in Eqs 4244 run over all positions in the environment (a square arena of side-length L), and p(x, t, x′, tτ) is the joint probability density of the virtual rat being at position x at time t and at position x′ at time tτ. From Eqs 43 to 44, we used the fact that, for large times t, the virtual rat has equal probability of being in any position x, i.e., p(x, t) = 1/L2.

Eq 44 shows that the temporal average riin(t)rjin(t-τ)¯ can be estimated from the input tuning curves Ψiin and Ψjin, and the conditional probability density p(x′, tτ|x, t). This conditional probability density has not yet been solved for correlated random walks in two dimensions [148]. Nevertheless, an additional approximation is possible. Because the temporal average riin(t)rjin(t-τ)¯ is weighted by the adaptation adaptation kernel K(τ) (Eq 38), and K(τ) is negligible for τ > τmax ≈ 5τL (Eq 2), we are interested in the conditional probability p(x′, tτ|x, t) only at lags τ < τmax. In this case, for movement trajectories that are sufficiently smooth, we can assume that in a time τ the virtual rat has moved to a position x at distance |xx′| = τv from the initial position x′, that is

p(x,t-τ|x,t)δ(|x-x|-τv)2πτv (45)

where v is the speed of the virtual rat (Eq 11), and the denominator ensures that dxp(x′, tτ|x, t) = 1; see also Fig 2 for exemplary virtual-rat trajectories in this scenario. We now use Eq 45 in Eq 44, and let zx′ − x:

riin(t)rjin(t-τ)¯1L2dzδ(|z|-τv)2πτv|z|=τvdzdxΨiin(x)Ψjin(x+z)ΨiinΨjin|z. (46)

From Eq 46, the temporal average riin(t)rjin(t-τ)¯ is approximated by the integral of the spatial cross-correlation ΨiinΨjin over a circle of radius τv. Finally, by using Eq 46 in Eq 20, we obtain

CijWtotL20dτK(τ)|z|=τvdzΨiinΨjin|z. (47)

Input correlation for spatially-regular inputs

In this section we compute the input correlation function C and its Fourier spectrum C^ in the case of spatially-regular inputs (see Sec Weight dynamics for spatially-regular inputs). First, we rewrite the input correlation matrix Cij in Eq 21 as a continuous function C(r, r′) by labeling neurons according to their receptive-field centers r and r′:

C(r,r)WtotL20dτK(τ)|z|=τvdzΨrinΨrin|z (48)

where Ψrin(x)G(|x-r|) is a Gaussian input tuning curve centered at position r (Eq 9). Because the inputs are translation invariant, the correlation function C depends only on the translation vector urr′:

C(r,r)=C(u,0)=C(u)WtotL20dτK(τ)|z|=τvdzΨuinΨ0in|z (49)
=WtotL20dτK(τ)|z|=τvdzΨ0inΨ0in|u+z (50)

where Ψ0in(x)G(|x|) is the tuning curve centered at the origin 0 = (0, 0). Next, we substitute in Eq 50 the definition of the integral operator in Eq 46:

C(u)WtotL20dτK(τ)dzδ(|z|-τv)2πτvΨ0inΨ0in|u+z. (51)

It is easy to see that the auto-correlation of a Gaussian is still a Gaussian:

Ψ0inΨ0in|u=L4rav24πσ2exp(-|u|24σ2) (52)

from which we derive

Ψ0inΨ0in|u+z=L4rav24πσ2[exp(-|u|2+|z|24σ2)exp(-|u||z|cos(φ)2σ2)] (53)

where φ is the angle between the vectors u and z. Finally, by expressing in polar coordinates the vector z ≔ |z|[cos(φ), sin(φ)], from Eqs 51 and 53 we obtain

C(u)WtotL2rav24πσ20dτK(τ)exp(-|u|2+(τv)24σ2)I0(-|u|τv2σ2) (54)

where I0(x)1/(2π)02πdφexp(xcos(φ)) is the zeroth-order modified Bessel function of the first kind.

Fourier spectrum of the input correlation function

Here we compute the Fourier spectrum of the correlation function C in Eq 51. First, we observe that the second integral in Eq 51 is a two-dimensional cross-correlation in the variable z between the functions δ(|z| − τv) and Ψ0inΨ0in|z evaluated at point u. Therefore, by taking the two-dimensional Fourier transform with respect to u at both sides of Eq 51 yields

C^(k)WtotL2|Ψ^0in(k)|20dτK(τ)dzδ(|z|-τv)2πτvexp(2πjz·k)=J0(τv|k|) (55)

where we defined the Fourier transform pair:

C^(k)duC(u)exp(-2πjk·u);C(u)=1(2π)2dkC^(k)exp(2πjk·u) (56)

with ku = |k||u| cos(θ), and we used the definition of the zeroth-order Bessel function

J0(k)12π02πdθexp(2πjkcos(θ)). (57)

Because the tuning function Ψ0in(x)G(|x|) is circularly symmetric, its two-dimensional Fourier transform Ψ^0in(k) is proportional to the zeroth-order Hankel transform of G:

Ψ^0in(k)=2πG˜(k)withk|k|, (58)

where we defined the zeroth-order Hankel transform pair:

G˜(k)0drrG(r)J0(kr)andG(r)=0dkkG˜(k)J0(kr). (59)

By using Eq 58 in Eq 55 we obtain

C^(k)=Wtot4π2L2G˜2(k)0dτK(τ)J0(τvk) (60)

and by defining the equivalent adaptation kernel in space

Ksp(r)1rvK(rv) (61)

we find

C^(k)=Wtot4π2L2G˜2(k)K˜sp(k). (62)

Finally, the zeroth-order Hankel transforms of the Gaussian tuning curve G (Eq 9) and of the adaptation kernel in space Ksp (Eqs 61 and 2) read

G˜(k)=L2rav2πexp(-k2σ22) (63)
K˜sp(k)=1τSv[k2+(τSv)2]0.5μτLv[k2+(τLv)2]0.5. (64)

Eigenvalue spectrum for spatially-irregular inputs

In this section we estimate the expected eigenvalue spectrum 〈λirr(k)〉 for spatially-irregular inputs (Secs Spatially-irregular inputs and Pattern formation with spatially-irregular inputs). We recall that, for spatially-regular inputs, in Sec Mathematical results on grid-pattern formation we obtained (Eq 32):

λ(k)ρWtotL24π2G˜2(k)=(58)|Ψ^0in(k)|2K˜sp(k)-awithk|k|0 (65)

where G˜ and K˜sp are the zeroth-order Hankel transforms of the input tuning curve G (Eq 9) and of the equivalent adaptation kernel in space Ksp (Eqs 31 and 61). Note that the parameters ρ, L, Wtot, and a do not depend on k. From Eq 65, the eigenvalue spectrum λ(k) is linearly-related to the input power spectrum |Ψ^0in(k)|2 where Ψ0in(x)G(|x|) is an input tuning curve centered at the origin 0 ≔ (0, 0) (Sec Input correlation for spatially-regular inputs).

Here, in analogy to Eq 65, we assume that the expected eigenvalue spectrum 〈λirr(k)〉 for spatially-irregular inputs is linearly-related to the expected input power |Ψ^pin(k)|2, that is,

λirr(k)ρWtotL2|Ψ^pin(k)|2K˜sp(k)-awithk0 (66)

where Ψ^pin(k) is the two-dimensional Fourier transform of the spatially-irregular tuning curve Ψpin(x), and the angular brackets denote statistical expectation across input realizations (see Eq 56 for a definition of the two-dimensional Fourier transform). The validity of this assumption is confirmed numerically at the end of this section.

Let us compute the expected input power spectrum |Ψ^pin(k)|2. We recall that the input maps Ψpin(x) are obtained by the superimposing M Gaussian receptive fields (Eq 10)

Ψpin(x)1βpm=1MApmG(|x-rpm|)forp=1,2,,N (67)

with

G(r)=(9)L2rav2πσ2exp(-r22σ2)andβpm=1MApm. (68)

The field amplitudes Apm ≥ 0 are uniformly distributed in the range (0, 1), and the receptive field centers rpm are uniformly distributed in the environment (see Fig 9A for examples). From Eq 67 we derive

|Ψ^pin(k)|=2πβpG˜(k)|m=1MApmexp(2πjrpm·k)|αp (69)

where G˜(k) is the zeroth-order Hankel transform of the Gaussian function G(r). In deriving Eq 69, we used the shift property of the Fourier transform and the equivalence between the Fourier and the zeroth-order Hankel transforms for circularly-symmetric functions (Eq 58). Finally, from Eq 69 we obtain

|Ψ^pin(k)|2=4π2G˜2(k)ΦwithΦαp2βp2. (70)

Therefore, for spatially-irregular inputs, the expected power spectrum |Ψ^pin(k)|2 is proportional to the power spectrum 4π2G˜2(k) of a single Gaussian G with scale factor Φ ≥ 0. Note that for |k| = 0 we obtain Φ = 1 (Eqs 69 and 70), which means that the average rate rav is independent of the number M of input receptive fields and their specific spatial arrangement. Using Eq 70 in Eq 66 yields

λirr(k)ρWtotL24π2G˜2(k)K˜sp(k)Φ-awithk0. (71)

Finally, from Eqs 65 and 71 we find (Eq 33)

λirr(k)Φλ(k)+a(1-Φ). (72)

In the next section we estimate the scale factor Φ for |k| > 0.

Approximation of the scale factor Φ

The scale factor

Φ(70)αp2βp2 (73)

is the second moment of the ratio of the random variables

αp(69)|m=1MApmexp(-2πjrpm·k)|andβp(68)m=1MApm (74)

where the field amplitudes Apm ≥ 0 are independently and uniformly distributed in the range (0, 1) and the field centers rpm are independently and uniformly distributed in a square of side-length L.

In general, for two random variables x and y, the first order Taylor expansion of the ratio f(z) = x/y around the expected value μ ≔ (〈x〉, 〈y〉) is

f(z)=f(μ)+fx(μ)Δx+fy(μ)Δy+o(Δx2)+o(Δy2)+o(ΔxΔy) (75)

where z ≔(x, y), Δxx − 〈x〉, Δyy − 〈y〉, and fx and fy are the derivatives of f with respect to x and y. Therefore

Var(xy)=[f(z)-f(μ)]2=fx2(μ)Var(x)+fy2(μ)Var(y)+2fx(μ)fy(μ)Cov(x,y)+k=04o(ΔxkΔy4-k). (76)

By neglecting the higher-order joint moments k=04o(ΔxkΔy4-k) and substituting fx(μ) = 1/〈y〉 and fy(μ) = −〈x〉/〈y2 we obtain

Var(xy)x2y2[Var(x)x2+Var(y)y2-2Cov(x,y)xy] (77)

and

x2y2x2y2[x2x2+y2y2-2Cov(x,y)xy-1]. (78)

In the following, we use Eq 78 to approximate the scale factor Φ (Eq 73).

We start by giving an intuitive interpretation of the random variables αp and βp. Consider a M-steps random walk on the complex plane with random directions rpmk and random step sizes Apm. The coefficients αp measure the total distance traveled by the random walker, and the coefficients βp measure the total length of the path (Eq 74). Note that the larger the number of steps M, the smaller is the correlation between the distance traveled αp and the total path length βp, i.e., |Cov(αp, βp)| ≪ 1 for M ≫ 1. In this case we can neglect the covariance term in Eq 78, and the factor Φ is approximated by knowing only the first two moments of the distributions of αp and βp.

For |k| > 1/L, the random directions rpmk (mod 1) are approximately uniformly distributed in the range (0, 1). In this case, the traveled distance αp follows a Rayleigh distribution with density [149]

f(αp)=2αpMApm2exp(-αp2MApm2) (79)

where Apm2=1/3 for Apm uniformly distributed in interval (0, 1). Therefore, the first two moments of αp read

αp=Mπ12andαp2=M3. (80)

The total path length βp is the sum of M random variables uniformly distributed in (0, 1), which follows an Irwin-Hall distribution. Therefore, the first two moments of βp are

βp=M2andβp2=M+3M212. (81)

Finally, by using Eqs 80 and 81 in Eq 73 we obtain

Φ(M)π3M(4π+13M)forM>1and|k|>1/L (82)
43MforM>3and|k|>1/L. (83)

Fig 10A shows the scale factor Φ as a function of the number M of superimposed Gaussian fields (Eq 82). Note that the approximation is more accurate for large values of M, which correspond to lower values of |Cov(αp, βp)|. Fig 10B shows the largest eigenvalue in the system as a function of M. The good match between the theoretical curve and the numerical estimations supports the validity of Eq 66. Additionally, Eq 66 predicts that, irrespectively of the value of M, the largest eigenvalue λmax = λ(kmax) is always at the critical frequency of kmax = 3 m−1 for σ = 6.25 cm and τL = 0.16 s, which matches the numerical results in Fig 9.

Fig 10. Scale factor Φ and largest eigenvalue λmax for spatially-irregular inputs.

Fig 10

(A) The scale factor Φ for M > 1 superimposed fields (Eq 70). The black dots are obtained by estimating the power spectrum at frequency |k| = 1 m−1 for 3600 input realizations. The red line is the theoretical curve in Eq 82. (B) The largest eigenvalue λmax as a function of the number of superimposed fields M. The black dots are obtained by computing the eigenvalues of the correlation matrix Cijij for N = 3600 inputs, where δij is the Kronecker delta (Eq 21). The red line is obtained from Eqs 71 and 82. Note that, according to Eq 71, the largest eigenvalue is always at the critical frequency kmax = 3 m−1 for any value of M. Parameter values as in Fig 9 (see Sec Numerical simulations).

Pattern formation with after-spike potentials

Here we study whether grid-like patterns could emerge by means of after-spike hyperpolarizing potentials (see discussion in Sec Spike-rate adaptation). To this end, we consider a model of the output neural activity that is alternative to the one presented in the main text (Sec Model of neural activity, Eq 1). We model input post-synaptic potentials (PSPs) with a kernel Kin applied to the input spike trains Sjin, and we model output after-spike hyperpolarizing potentials (AHPs) with a kernel Kout applied to the output spike train Sout:

rout(t)r0+0dsKout(s)Sout(t-s)+0dτKin(τ)j=1NwjSjin(t-τ) (84)

where r0 ≥ 0 is a baseline firing rate.

First, we show that the average dynamics of Eq 84 can be rewritten in terms of an equivalent kernel Keq applied to the input spikes only. We average Eq 84 across input and output spike train realizations:

rout(t)=r0+0dsKout(s)rout(t-s)+0dτKin(τ)j=1Nwjrjin(t-τ). (85)

And by taking the Fourier transform

f^(ω)dtf(t)exp(-jωt);f(t)=12πdωf^(ω)exp(jωt) (86)

at both sides of Eq 85 we obtain

r^out(ω)=r0δ(ω)+K^in(ω)1-K^out(ω)j=1Nwjr^jin(ω). (87)

From Eqs 85 to 87 we assumed that the input and the output kernels are causal, i.e., Kin,out(t) = 0 for t < 0, and that the output kernel has integral different from 1, i.e., K^out(0)=0dtKout(t)1. Finally, by defining the equivalent filter

K^eq(ω)K^in(ω)1-K^out(ω), (88)

the inverse Fourier transform of Eq 87 reads

rout(t)=r0+0dτKeq(τ)j=1Nwjrjin(t-τ), (89)

which is equivalent to Eq 15 with Keq = K.

Next, we compute the equivalent filter Keq for a simple choice of the input and output kernels

Kin(t){1τinexp(-tτin)fort00fort<0 (90)

and

Kout(t){-μoutτoutexp(-tτout)fort00fort<0 (91)

where τin, τout > 0 are decay time constants, and the parameter μout > 0 scales the integral of the output kernel 0dtKout(t)=-μout. We assume that the input kernel Kin (modeling an incoming PSP) decays faster than the output kernel Kout (modeling an output AHP), i.e., τin < τout. From the definition of the filter Keq in Eq 88 we obtain

K^eq(ω)=1/τin1/τin-(1+μout)/τoutH[1/τin-1/τout1/τin+jω-μout/τout(1+μout)/τout+jω] (92)

where we used

K^in(ω)=1/τin1/τin+jωandK^out(ω)=-μout/τout1/τout+jω. (93)

Finally, the inverse Fourier transform of Eq 92 reads

Keq(t)=H·[(1τin-1τout)exp(-tτin)-μoutτoutexp(-tτout/(1+μout))] (94)

for t ≥ 0 and Keq(t) = 0 for t < 0. Eq 94 shows that the equivalent filter Keq is a difference of two exponentials, similarly to the kernel K in Eq 2. Note however that the two exponentials are scaled differently as compared to the original filter K. Additionally, if the integral of the output kernel is negative, the integral of the equivalent filter is always positive (Eq 88 with ω = 0).

To test whether spatially-periodic patterns could still emerge in this scenario, we compute the eigenvalue spectrum λ(k) and the critical spatial frequency kmax by using Eqs 31 and 32 with K = Keq. Surprisingly, we find that typical grid scales (e.g., kmax > 2 m−1) are obtained for output-kernel time constants of the order of seconds, which seem biologically unrealistic (Fig 11). Therefore, we conclude that AHPs alone are not sufficient to generate grid-like patterns. Nevertheless, AHPs could still support structure formation by amplifying the effects of a band-pass filter that is already present at the input.

Fig 11. Grid scale with after-spike hyperpolarizing potentials.

Fig 11

The critical spatial frequency kmax is plotted as a function of the output-kernel integral −μout and the output-kernel time constant τout (Eqs 31 and 32 with K = Keq). The black lines are iso-levels (see annotated values). Regions enclosed by two adjacent iso-lines are colored uniformly (darker colors denote larger values). The input-kernel time constant is τin = 5 ms. Similar results are obtained with different values of τin < τout. Parameter values: σ = 6.25 cm, v = 0.25 m/s, L = 1 m. rav = 0.4 s−1.

Numerical simulations

Model parameters and derived quantities are summarized in Tables 1 and 2.

Table 1. Model parameters.

Neural activity
N Number of synaptic inputs
r0 [s−1] Baseline rate of the output neuron
τS [s] Adaptation kernel short time constant
τL [s] Adaptation kernel long time constant
μ Adaptation kernel scaling parameter
Spatial exploration
L [m] Side-length of the arena
v [m/s] Running speed of the virtual rat
σθ Standard deviation of running directions
Input spatial tuning
rav [s−1] Average input rate in the arena
σ [m] Width of the input receptive fields
M Number of receptive fields per neuron (spatially-irregular inputs)
Synaptic plasticity
η Learning rate
τW [s] Decay time constant of the learning window W
Wtot [s] Integral of the learning window W
α Multiplicative weight-normalization constant
β Additive weight-normalization constant
Derived quantities
a [s−1] Multiplicative weight-normalization rate
b [s−1] Additive weight-normalization rate
λmax [s−1] Maximal eigenvalue
wav Average synaptic weight
τav [s] Weight normalization time scale
τstr [s] Structure formation time scale

Table 2. Default parameter values for the numerical simulations.

See also Table 1 for short descriptions of the parameters. TL: top-left, TR: top-right, BL: bottom-left, BR: bottom-right. Note that in Fig 6A the learning rate η is varied from 2 ⋅ 10−5 to 10 ⋅ 10−5 and that in Fig 6B the virtual-rat running speed is sampled from an Ornstein-Uhlenbeck process with long-term mean v¯=v.

Unit Fig 5
Fig 6
Fig 7 (TL) Fig 7 (TR) Fig 7 (BL)
Fig 8A1–8A4
Fig 7 (BR)
Fig 8B1–8B4
Fig 9
Fig 10
N 900 3600 3600 3600 3600 3600
r0 [s−1] 10 4 4 4 4 4
τS [s] 0.1 0.1 0.1 0.1 0.1 0.1
τL [s] 0.16 0.16 0.35 0.16 0.35 0.16
μ 1.06 1.06 1.06 1.06 1.06 1.06
L [m] 1 2 2 2 2 1
v [m/s] 0.25 0.25 0.25 0.25 0.25 0.25
σθ 0.7 0.7 0.7 0.7 0.7 0.7
rav [s−1] 0.4 0.21 0.085 0.3 0.1 0.8
σ [m] 0.0625 0.045 0.045 0.0625 0.0625 0.0625
M 10
η 2 ⋅ 10−5 5 ⋅ 10−5 5 ⋅ 10−5 5 ⋅ 10−5 5 ⋅ 10−5 5 ⋅ 10−5
τW [s] 0.05
Wtot [s] 1 1 1 1 1 1
α 3.56
β −8.78
a [s−1] 1.1 4 4 4 4 2.5
b [s−1] 0.49 0.69 0.28 1.23 0.31 2.8
λmax [s−1] 1 0.90 0.82 0.80 0.85 1.75
wav 0.05 0.05 0.05 0.05 0.05 0.02
τav [s] 5.13 ⋅ 103 1.44 ⋅ 103 3.57 ⋅ 103 8.12 ⋅ 102 3.18 ⋅ 103 1.42 ⋅ 102
τstr [s] 5 ⋅ 104 2.23 ⋅ 104 2.42 ⋅ 104 2.50 ⋅ 104 2.36 ⋅ 104 1.14 ⋅ 104

Simulation of the detailed spiking model

The detailed spiking model (Figs 5 and 6) is simulated using the Brian2 simulation software [150]. Neural and synaptic variables are integrated with a time step of 1 ms. The random walk of the virtual rat that is updated every 10 ms. The physical space explored by the virtual rat is discretized in 2002 square bins.

Simulation of the averaged weight dynamics

The average weight dynamics (Eq 16) is integrated by using the forward Euler method with integration time step of 50 s (Figs 79). The input correlation matrix C is computed using Eq 54 for spatially-regular inputs, and using Eq 21 for spatially-irregular inputs.

Initialization of the synaptic weights

At the initial condition the synaptic weights are normally distributed around the target normalization level wav=5·10-3. The standard deviation is 10−4 for the spiking simulations and 10−3 for the average weight dynamics.

Data analysis

Grid properties

We compute the grid spatial scale from the two-dimensional Fourier amplitude of the grid pattern. We estimate the radial amplitude profile by averaging over the angular dimension. We then define the grid scale as the frequency where the amplitude profile has a global maximum.

The grid orientation is estimated from the spatial auto-correlogram of the grid pattern. We detect the peak closest to the center in the first quadrant of the auto-correlogram. We then define the grid orientation as the angle between the detected peak and the horizontal axis.

We define the grid spatial phase as the position of the closest peak to the center in the cross-correlation between the grid pattern and a reference grid at the same scale.

Gridness score

We estimate the gridness score similarly to Langston et al. [135]. First, we compute the spatial auto-correlogram of the weight (or firing-rate) pattern and we retain only points within a ring of outer radius Ri and inner radius Ri/2. We then compute the gridness score gi as

gi12[ρi(60)+ρi(120)]-13[ρi(30)+ρi(90)+ρi(150)] (95)

where ρi(φ) is the Pearson’s correlation coefficient between the original ring (of outer radius Ri) and the same ring rotated by φ degrees. The final gridness score is defined as the maximum gi by varying the outer radius Ri between 0.7/kmax and 2.5/kmax where kmax is the spatial frequency of the pattern.

Estimation of output firing-rate maps

The output firing-rate maps Ψout in Fig 9B are computed as follows:

Ψout(x)=r0+dyKsp(|y|)i=1NwiΨiin(x-y) (96)

where r0 is the baseline firing rate, wi are the synaptic weights at the end of the simulation, Ψiin are the input spatial maps, and Ksp is the equivalent adaptation kernel in space (Eq 31). The convolution with the filter Ksp accounts for the average effect of the temporal kernel K on the output firing rate.

Acknowledgments

We are indebited to Jan Herding and Manuel Beiran for previous work on the topic of this manuscript. We thank Thomas McColgan and Eric Reifenstein for helpful feedback and inspiring discussions.

Data Availability

All relevant data are within the paper.

Funding Statement

This work was supported by the Deutsche Forschungsgemeinschaft (DFG) (grant numbers GRK 1589/2, KE 788/3-1) and the Bundesministerium für Bildung und Forschung (BMBF) (grant numbers 01GQ1001A, 01GQ0972) for RK. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Hafting T., Fyhn M., Molden S., Moser M.-B., and Moser E. I. Microstructure of a spatial map in the entorhinal cortex. Nature, 436(7052):801–806, 2005. 10.1038/nature03721 [DOI] [PubMed] [Google Scholar]
  • 2. Rowland D. C., Roudi Y., Moser M.-B., and Moser E. I. Ten years of grid cells. Annu. Rev. Neurosci., 39:19–40, 2016. 10.1146/annurev-neuro-070815-013824 [DOI] [PubMed] [Google Scholar]
  • 3. O’Keefe J. Place units in the hippocampus of the freely moving rat. Exp. Neurol., 51(1):78–109, 1976. ISSN 0014-4886. 10.1016/0014-4886(76)90055-8 [DOI] [PubMed] [Google Scholar]
  • 4. Moser E. I., Kropff E., and Moser M.-B. Place cells, grid cells, and the brain’s spatial representation system. Annu. Rev. Neurosci., 31:69–89, 2008. 10.1146/annurev.neuro.31.061307.090723 [DOI] [PubMed] [Google Scholar]
  • 5. Fiete I. R., Burak Y., and Brookings T. What grid cells convey about rat location. J. Neurosci., 28(27):6858–6871, 2008. 10.1523/JNEUROSCI.5684-07.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Mathis A., Herz A. V., and Stemmler M. Optimal population codes for space: grid cells outperform place cells. Neural Comput., 24(9):2280–2317, 2012. 10.1162/NECO_a_00319 [DOI] [PubMed] [Google Scholar]
  • 7. Bush D., Barry C., Manson D., and Burgess N. Using grid cells for navigation. Neuron, 87(3):507–520, 2015. 10.1016/j.neuron.2015.07.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Stemmler M., Mathis A., and Herz A. V. Connecting multiple spatial scales to decode the population activity of grid cells. Sci. Adv., 1(11):e1500816, 2015. 10.1126/science.1500816 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Horner A. J., Bisby J. A., Zotow E., Bush D., and Burgess N. Grid-like processing of imagined navigation. Curr Biol., 26(6):842–847, 2016. 10.1016/j.cub.2016.01.042 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Ólafsdóttir H. F., Carpenter F., and Barry C. Coordinated grid and place cell replay during rest. Nat. Neurosci., 19(6):792–794, 2016. 10.1038/nn.4291 [DOI] [PubMed] [Google Scholar]
  • 11. O’Neill J., Boccara C., Stella F., Schoenenberger P., and Csicsvari J. Superficial layers of the medial entorhinal cortex replay independently of the hippocampus. Science, 355(6321):184–188, 2017. 10.1126/science.aag2787 [DOI] [PubMed] [Google Scholar]
  • 12. Giocomo L. M., Moser M.-B., and Moser E. I. Computational models of grid cells. Neuron, 71(4):589–603, 2011. 10.1016/j.neuron.2011.07.023 [DOI] [PubMed] [Google Scholar]
  • 13. Zilli E. A. Models of grid cell spatial firing published 2005–2011. Front. Neural Circuits, 6:16, 2012. 10.3389/fncir.2012.00016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. McNaughton B. L., Battaglia F. P., Jensen O., Moser E. I., and Moser M.-B. Path integration and the neural basis of the ‘cognitive map’. Nat. Rev. Neurosci., 7(8):663–678, 2006. 10.1038/nrn1932 [DOI] [PubMed] [Google Scholar]
  • 15. Fuhs M. C. and Touretzky D. S. A spin glass model of path integration in rat medial entorhinal cortex. J. Neurosci., 26(16):4266–4276, 2006. 10.1523/JNEUROSCI.4353-05.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Guanella A., Kiper D., and Verschure P. A model of grid cells based on a twisted torus topology. Int. J. Neural Syst., 17(04):231–240, 2007. 10.1142/S0129065707001093 [DOI] [PubMed] [Google Scholar]
  • 17. Burak Y. and Fiete I. R. Accurate path integration in continuous attractor network models of grid cells. PLoS Comput. Biol., 5(2):e1000291, 2009. 10.1371/journal.pcbi.1000291 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Navratilova Z., Giocomo L. M., Fellous J.-M., Hasselmo M. E., and McNaughton B. L. Phase precession and variable spatial scaling in a periodic attractor map model of medial entorhinal grid cells with realistic after-spike dynamics. Hippocampus, 22(4):772–789, 2012. 10.1002/hipo.20939 [DOI] [PubMed] [Google Scholar]
  • 19. Couey J. J., Witoelar A., Zhang S.-J., Zheng K., Ye J., Dunn B., Czajkowski R., Moser M.-B., Moser E. I., Roudi Y., et al. Recurrent inhibitory circuitry as a mechanism for grid formation. Nat. Neurosci., 16(3):318–324, 2013. 10.1038/nn.3310 [DOI] [PubMed] [Google Scholar]
  • 20. Pastoll H., Solanka L., van Rossum M. C., and Nolan M. F. Feedback inhibition enables theta-nested gamma oscillations and grid firing fields. Neuron, 77(1):141–154, 2013. 10.1016/j.neuron.2012.11.032 [DOI] [PubMed] [Google Scholar]
  • 21. Widloski J. and Fiete I. R. A model of grid cell development through spatial exploration and spike time-dependent plasticity. Neuron, 83(2):481–495, 2014. 10.1016/j.neuron.2014.06.018 [DOI] [PubMed] [Google Scholar]
  • 22. Etienne A. S. and Jeffery K. J. Path integration in mammals. Hippocampus, 14(2):180–192, 2004. 10.1002/hipo.10173 [DOI] [PubMed] [Google Scholar]
  • 23. Sargolini F., Fyhn M., Hafting T., McNaughton B. L., Witter M. P., Moser M.-B., and Moser E. I. Conjunctive representation of position, direction, and velocity in entorhinal cortex. Science, 312(5774):758–762, 2006. 10.1126/science.1125572 [DOI] [PubMed] [Google Scholar]
  • 24. Kropff E., Carmichael J. E., Moser M.-B., and Moser E. I. Speed cells in the medial entorhinal cortex. Nature, 523(7561):419–424, 2015. 10.1038/nature14622 [DOI] [PubMed] [Google Scholar]
  • 25. Barry C., Ginzberg L. L., O’Keefe J., and Burgess N. Grid cell firing patterns signal environmental novelty by expansion. Proc. Nat. Acad. Sci. U.S.A., 109(43):17687–17692, 2012. 10.1073/pnas.1209918109 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Chen G., Manson D., Cacucci F., and Wills T. J. Absence of visual input results in the disruption of grid cell firing in the mouse. Curr Biol., 26(17):2335–2342, 2016. 10.1016/j.cub.2016.06.043 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Pérez-Escobar J. A., Kornienko O., Latuske P., Kohler L., and Allen K. Visual landmarks sharpen grid cell metric and confer context specificity to neurons of the medial entorhinal cortex. eLife, 5:e16937, 2016. 10.7554/eLife.16937 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Maaswinkel H. and Whishaw I. Q. Homing with locale, taxon, and dead reckoning strategies by foraging rats: sensory hierarchy in spatial navigation. Behav. Brain Res., 99(2):143–152, 1999. 10.1016/S0166-4328(98)00100-4 [DOI] [PubMed] [Google Scholar]
  • 29. Barry C., Hayman R., Burgess N., and Jeffery K. J. Experience-dependent rescaling of entorhinal grids. Nat. Neurosci., 10(6):682–684, 2007. 10.1038/nn1905 [DOI] [PubMed] [Google Scholar]
  • 30. Savelli F., Yoganarasimha D., and Knierim J. J. Influence of boundary removal on the spatial representations of the medial entorhinal cortex. Hippocampus, 18(12):1270–1282, 2008. 10.1002/hipo.20511 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31. Krupic J., Bauza M., Burton S., and O’Keefe J. Framing the grid: effect of boundaries on grid cells and navigation. The Journal of physiology, 594(22):6489–6499, 2016. 10.1113/JP270607 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. O’Keefe J. and Burgess N. Dual phase and rate coding in hippocampal place cells: theoretical significance and relationship to entorhinal grid cells. Hippocampus, 15(7):853–866, 2005. 10.1002/hipo.20115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Burgess N., Barry C., and O’Keefe J. An oscillatory interference model of grid cell firing. Hippocampus, 17(9):801–812, 2007. 10.1002/hipo.20327 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Hasselmo M. E., Giocomo L. M., and Zilli E. A. Grid cell firing may arise from interference of theta frequency membrane potential oscillations in single neurons. Hippocampus, 17(12):1252–1271, 2007. 10.1002/hipo.20374 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Burgess N. Grid cells and theta as oscillatory interference: theory and predictions. Hippocampus, 18(12):1157–1174, 2008. 10.1002/hipo.20518 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Zilli E. A. and Hasselmo M. E. Coupled noisy spiking neurons as velocity-controlled oscillators in a model of grid cell spatial firing. J. Neurosci., 30(41):13850–13860, 2010. 10.1523/JNEUROSCI.0547-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Vanderwolf C. Hippocampal electrical activity and voluntary movement in the rat. Electroencephalogr. Clin. Neurophysiol., 26(4):407–418, 1969. ISSN 0013-4694. 10.1016/0013-4694(69)90092-3 [DOI] [PubMed] [Google Scholar]
  • 38. Winson J. Patterns of hippocampal theta rhythm in the freely moving rat. Electroencephalogr. Clin. Neurophysiol., 36:291–301, 1974. 10.1016/0013-4694(74)90171-0 [DOI] [PubMed] [Google Scholar]
  • 39. McFarland W. L., Teitelbaum H., and Hedges E. K. Relationship between hippocampal theta activity and running speed in the rat. J. Comp. Physiol. Psychol., 88:324–328, 1975. 10.1037/h0076177 [DOI] [PubMed] [Google Scholar]
  • 40. Sławińska U. and Kasicki S. The frequency of rat’s hippocampal theta rhythm is related to the speed of locomotion. Brain Res., 796(1):327–331, 1998. [DOI] [PubMed] [Google Scholar]
  • 41. Brandon M. P., Bogaard A. R., Libby C. P., Connerney M. A., Gupta K., and Hasselmo M. E. Reduction of theta rhythm dissociates grid cell spatial periodicity from directional tuning. Science, 332(6029):595–599, 2011. 10.1126/science.1201652 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Koenig J., Linder A. N., Leutgeb J. K., and Leutgeb S. The spatial periodicity of grid cells is not sustained during reduced theta oscillations. Science, 332(6029):592–595, 2011. 10.1126/science.1201685 [DOI] [PubMed] [Google Scholar]
  • 43. Hafting T., Fyhn M., Bonnevie T., Moser M.-B., and Moser E. I. Hippocampus-independent phase precession in entorhinal grid cells. Nature, 453(7199):1248–1252, 2008. 10.1038/nature06957 [DOI] [PubMed] [Google Scholar]
  • 44. Reifenstein E., Stemmler M., Herz A. V., Kempter R., and Schreiber S. Movement dependence and layer specificity of entorhinal phase precession in two-dimensional environments. PLoS One, 9(6):e100638, 2014a. 10.1371/journal.pone.0100638 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. Yartsev M. M., Witter M. P., and Ulanovsky N. Grid cells without theta oscillations in the entorhinal cortex of bats. Nature, 479(7371):103–107, 2011. 10.1038/nature10583 [DOI] [PubMed] [Google Scholar]
  • 46. Domnisoru C., Kinkhabwala A. A., and Tank D. W. Membrane potential dynamics of grid cells. Nature, 495(7440):199–204, 2013. 10.1038/nature11973 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Schmidt-Hieber C. and Häusser M. Cellular mechanisms of spatial navigation in the medial entorhinal cortex. Nat. Neurosci., 16(3):325–331, 2013. 10.1038/nn.3340 [DOI] [PubMed] [Google Scholar]
  • 48. Bush D. and Burgess N. A hybrid oscillatory interference/continuous attractor network model of grid cell firing. J. Neurosci., 34(14):5065–5079, 2014. 10.1523/JNEUROSCI.4017-13.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49. Kropff E. and Treves A. The emergence of grid cells: Intelligent design or just adaptation? Hippocampus, 18(12):1256–1269, 2008. 10.1002/hipo.20520 [DOI] [PubMed] [Google Scholar]
  • 50. Si B., Kropff E., and Treves A. Grid alignment in entorhinal cortex. Biol. Cybern., 106(8–9):483–506, 2012. 10.1007/s00422-012-0513-7 [DOI] [PubMed] [Google Scholar]
  • 51. Si B. and Treves A. A model for the differentiation between grid and conjunctive units in medial entorhinal cortex. Hippocampus, 23(12):1410–1424, 2013. 10.1002/hipo.22194 [DOI] [PubMed] [Google Scholar]
  • 52. Stella F., Si B., Kropff E., and Treves A. Grid cells on the ball. J. Stat. Mech. Theor. Exp., 2013(03):P03013, 2013. 10.1088/1742-5468/2013/03/P03013 [DOI] [Google Scholar]
  • 53. Stella F. and Treves A. The self-organization of grid cells in 3D. eLife, 4:e05913, 2015. 10.7554/eLife.05913 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Tang Q., Burgalossi A., Ebbesen C. L., Ray S., Naumann R., Schmidt H., Spicher D., and Brecht M. Pyramidal and stellate cell specificity of grid and border representations in layer 2 of medial entorhinal cortex. Neuron, 84(6):1191–1197, 2014. 10.1016/j.neuron.2014.11.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55. Sun C., Kitamura T., Yamamoto J., Martin J., Pignatelli M., Kitch L. J., Schnitzer M. J., and Tonegawa S. Distinct speed dependence of entorhinal island and ocean cells, including respective grid cells. Proc. Nat. Acad. Sci. U.S.A., 112(30):9466–9471, 2015. 10.1073/pnas.1511668112 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56. Diehl G. W., Hon O. J., Leutgeb S., and Leutgeb J. K. Grid and nongrid cells in medial entorhinal cortex represent spatial location and environmental features with complementary coding schemes. Neuron, 2017. 10.1016/j.neuron.2017.03.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57. Cacucci F., Lever C., Wills T. J., Burgess N., and O’Keefe J. Theta-modulated place-by-direction cells in the hippocampal formation in the rat. J. Neurosci., 24(38):8265–8277, 2004. 10.1523/JNEUROSCI.2635-04.2004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58. Hargreaves E. L., Rao G., Lee I., and Knierim J. J. Major dissociation between medial and lateral entorhinal input to dorsal hippocampus. Science, 308(5729):1792–1794, 2005. 10.1126/science.1110449 [DOI] [PubMed] [Google Scholar]
  • 59. Hargreaves E. L., Yoganarasimha D., and Knierim J. J. Cohesiveness of spatial and directional representations recorded from neural ensembles in the anterior thalamus, parasubiculum, medial entorhinal cortex, and hippocampus. Hippocampus, 17(9):826–841, 2007. 10.1002/hipo.20316 [DOI] [PubMed] [Google Scholar]
  • 60. Boccara C. N., Sargolini F., Thoresen V. H., Solstad T., Witter M. P., Moser E. I., and Moser M.-B. Grid cells in pre-and parasubiculum. Nat. Neurosci., 13(8):987–994, 2010. 10.1038/nn.2602 [DOI] [PubMed] [Google Scholar]
  • 61. Tang Q., Burgalossi A., Ebbesen C. L., Sanguinetti-Scheck J. I., Schmidt H., Tukker J. J., Naumann R., Ray S., Preston-Ferrer P., Schmitz D., and Brecht M. Functional architecture of the rat parasubiculum. J. Neurosci., 36(7):2289–2301, 2016. 10.1523/JNEUROSCI.3749-15.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62. La Camera G., Rauch A., Thurbon D., Lüscher H.-R., Senn W., and Fusi S. Multiple time scales of temporal response in pyramidal and fast spiking cortical neurons. J. Neurphysiol., 96(6):3448–3464, 2006. 10.1152/jn.00453.2006 [DOI] [PubMed] [Google Scholar]
  • 63. Castro L. and Aguiar P. A feedforward model for the formation of a grid field where spatial information is provided solely from place cells. Biol. Cybern., 108(2):133–143, 2014. 10.1007/s00422-013-0581-3 [DOI] [PubMed] [Google Scholar]
  • 64. Stepanyuk A. Self-organization of grid fields under supervision of place cells in a neuron model with associative plasticity. Biol. Insp. Cogn. Archit., 13:48–62, 2015. [Google Scholar]
  • 65. Dordek Y., Soudry D., Meir R., and Derdikman D. Extracting grid cell characteristics from place cell inputs using non-negative principal component analysis. eLife, 5:e10094, 2016. 10.7554/eLife.10094 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66. Monsalve-Mercado M. M. and Leibold C. Hippocampal spike-timing correlations lead to hexagonal grid fields. Phys. Rev. Lett., 119(3):038101, 2017. 10.1103/PhysRevLett.119.038101 [DOI] [PubMed] [Google Scholar]
  • 67.S. N. Weber and H. Sprekeler. Learning place cells, grid cells and invariances: A unifying model. bioRxiv, page 102525, 2017. [DOI] [PMC free article] [PubMed]
  • 68. Turing A. M. The chemical basis of morphogenesis. Phil. Trans. R. Soc. B, 237(641):37–72, 1952. 10.1098/rstb.1952.0012 [DOI] [Google Scholar]
  • 69. Alonso A. and Klink R. Differential electroresponsiveness of stellate and pyramidal-like cells of medial entorhinal cortex layer II. J. Neurphysiol., 70(1):128–143, 1993. [DOI] [PubMed] [Google Scholar]
  • 70. van der Linden S. and da Silva F. H. L. Comparison of the electrophysiology and morphology of layers III and II neurons of the rat medial entorhinal cortex in vitro. Eur. J. Neurosci., 10(4):1479–1489, 1998. 10.1046/j.1460-9568.1998.00162.x [DOI] [PubMed] [Google Scholar]
  • 71. Yoshida M., Jochems A., and Hasselmo M. E. Comparison of properties of medial entorhinal cortex layer II neurons in two anatomical dimensions with and without cholinergic activation. PLoS One, 8(9):e73904, 2013. 10.1371/journal.pone.0073904 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72. Gerstner W., Kempter R., Van Hemmen J., and Wagner H. A neuronal learning rule for sub-millisecond temporal coding. Nature, 383(6595):76–78, 1996. 10.1038/383076a0 [DOI] [PubMed] [Google Scholar]
  • 73. Markram H., Lübke J., Frotscher M., and Sakmann B. Regulation of synaptic efficacy by coincidence of postsynaptic APs and EPSPs. Science, 275(5297):213–215, 1997. 10.1126/science.275.5297.213 [DOI] [PubMed] [Google Scholar]
  • 74. Kempter R., Gerstner W., and Van Hemmen J. L. Hebbian learning and spiking neurons. Phys. Rev. E, 59(4):4498–4514, 1999. 10.1103/PhysRevE.59.4498 [DOI] [Google Scholar]
  • 75. Song S., Miller K. D., and Abbott L. F. Competitive Hebbian learning through spike-timing-dependent synaptic plasticity. Nature Neurosci., 3(9):919–926, 2000. 10.1038/78829 [DOI] [PubMed] [Google Scholar]
  • 76. Zhou Y.-D., Acker C. D., Netoff T. I., Sen K., and White J. A. Increasing Ca2+ transients by broadening postsynaptic action potentials enhances timing-dependent synaptic depression. Proc. Nat. Acad. Sci. U.S.A., 102(52):19121–19125, 2005. 10.1073/pnas.0509856103 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77. Mishra R. K., Kim S., Guzman S. J., and Jonas P. Symmetric spike timing-dependent plasticity at CA3-CA3 synapses optimizes storage and recall in autoassociative networks. Nat. Commun., 7(11552), 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78. O’Brien R. J., Kamboj S., Ehlers M. D., Rosen K. R., Fischbach G. D., and Huganir R. L. Activity-dependent modulation of synaptic AMPA receptor accumulation. Neuron, 21(5):1067–1078, 1998. 10.1016/S0896-6273(00)80624-8 [DOI] [PubMed] [Google Scholar]
  • 79. Davis G. W. and Goodman C. S. Synapse-specific control of synaptic efficacy at the terminals of a single neuron. Nature, 392(6671):82–86, 1998. 10.1038/32176 [DOI] [PubMed] [Google Scholar]
  • 80. Turrigiano G. G., Leslie K. R., Desai N. S., Rutherford L. C., and Nelson S. B. Activity-dependent scaling of quantal amplitude in neocortical neurons. Nature, 391(6670):892–896, 1998. 10.1038/36103 [DOI] [PubMed] [Google Scholar]
  • 81. Turrigiano G. G. and Nelson S. B. Homeostatic plasticity in the developing nervous system. Nat. Rev. Neurosci., 5(2):97–107, 2004. 10.1038/nrn1327 [DOI] [PubMed] [Google Scholar]
  • 82. Keefe J. O. and Burgess N. Geometric determinants of the place fields of hippocampal neurons. Nature, 381(6581):425–428, 1996. 10.1038/381425a0 [DOI] [PubMed] [Google Scholar]
  • 83. Mizuseki K., Royer S., Diba K., and Buzsáki G. Activity dynamics and behavioral correlates of CA3 and CA1 hippocampal pyramidal neurons. Hippocampus, 22(8):1659–1680, 2012. 10.1002/hipo.22002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84. Solstad T., Boccara C. N., Kropff E., Moser M.-B., and Moser E. I. Representation of geometric borders in the entorhinal cortex. Science, 322(5909):1865–1868, 2008. 10.1126/science.1166466 [DOI] [PubMed] [Google Scholar]
  • 85. Lever C., Burton S., Jeewajee A., O’Keefe J., and Burgess N. Boundary vector cells in the subiculum of the hippocampal formation. J. Neurosci., 29(31):9771–9777, 2009. 10.1523/JNEUROSCI.1319-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86. Reifenstein E., Stemmler M., Herz A. V., Kempter R., and Schreiber S. Movement dependence and layer specificity of entorhinal phase precession in two-dimensional environments. PloS One, 9(6):e100638, 2014b. 10.1371/journal.pone.0100638 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87. Murray J. D. Mathematical Biology I: an introduction, Vol. 17 of Interdisciplinary Applied Mathematics. Springer, New York, NY, USA, 2002. [Google Scholar]
  • 88. D’Albis T., Jaramillo J., and Kempter R. Inheritance of place fields in the hippocampus through Hebbian learning. Neural Comput., 27:1624–1672, 2015. 10.1162/NECO_a_00752 [DOI] [PubMed] [Google Scholar]
  • 89. Ermentrout G. B. and Cowan J. D. A mathematical theory of visual hallucination patterns. Biol. Cybern., 34(3):137–150, 1979. 10.1007/BF00336965 [DOI] [PubMed] [Google Scholar]
  • 90. Jung M. W., Wiener S. I., and McNaughton B. L. Comparison of spatial firing characteristics of units in dorsal and ventral hippocampus of the rat. J. Neurosci., 14(12):7347–7356, 1994. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91. Fyhn M., Molden S., Witter M. P., Moser E. I., and Moser M.-B. Spatial representation in the entorhinal cortex. Science, 305(5688):1258–1264, 2004. 10.1126/science.1099901 [DOI] [PubMed] [Google Scholar]
  • 92. Kjelstrup K. B., Solstad T., Brun V. H., Hafting T., Leutgeb S., Witter M. P., Moser E. I., and Moser M.-B. Finite scale of spatial representation in the hippocampus. Science, 321(5885):140–143, 2008. 10.1126/science.1157086 [DOI] [PubMed] [Google Scholar]
  • 93. Stensola H., Stensola T., Solstad T., Frøland K., Moser M.-B., and Moser E. I. The entorhinal grid map is discretized. Nature, 492(7427):72–78, 2012. 10.1038/nature11649 [DOI] [PubMed] [Google Scholar]
  • 94. Giocomo L. M., Zilli E. A., Fransén E., and Hasselmo M. E. Temporal frequency of subthreshold oscillations scales with entorhinal grid cell field spacing. Science, 315(5819):1719–1722, 2007. 10.1126/science.1139207 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95. Giocomo L. M. and Hasselmo M. E. Time constants of h current in layer II stellate cells differ along the dorsal to ventral axis of medial entorhinal cortex. J. Neurosci., 28(38):9414–9425, 2008. 10.1523/JNEUROSCI.3196-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 96. Garden D. L., Dodson P. D., O’Donnell C., White M. D., and Nolan M. F. Tuning of synaptic integration in the medial entorhinal cortex to the organization of grid cell firing fields. Neuron, 60(5):875–889, 2008. 10.1016/j.neuron.2008.10.044 [DOI] [PubMed] [Google Scholar]
  • 97. Pastoll H., Ramsden H., and Nolan M. F. Intrinsic electrophysiological properties of entorhinal cortex stellate cells and their contribution to grid cell firing fields. Front. Neural Circuits, 6:17, 2012. 10.3389/fncir.2012.00017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98. Heys J. G., Rangarajan K. V., and Dombeck D. A. The functional micro-organization of grid cells revealed by cellular-resolution imaging. Neuron, 84(5):1079–1090, 2014. 10.1016/j.neuron.2014.10.048 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99. Krupic J., Bauza M., Burton S., Barry C., and O’Keefe J. Grid cell symmetry is shaped by environmental geometry. Nature, 518(7538):232, 2015. 10.1038/nature14153 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 100. Stensola T., Stensola H., Moser M.-B., and Moser E. I. Shearing-induced asymmetry in entorhinal grid cells. Nature, 518(7538):207–212, 2015. 10.1038/nature14151 [DOI] [PubMed] [Google Scholar]
  • 101. Urdapilleta E., Si B., and Treves A. Self-organization of modular activity of grid cells. Hippocampus, pages n/a–n/a, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 102. Bonnevie T., Dunn B., Fyhn M., Hafting T., Derdikman D., Kubie J. L., Roudi Y., Moser E. I., and Moser M.-B. Grid cells require excitatory drive from the hippocampus. Nat. Neurosci., 16(3):309–317, 2013. 10.1038/nn.3311 [DOI] [PubMed] [Google Scholar]
  • 103. Tamamaki N. and Nojyo Y. Preservation of topography in the connections between the subiculum, field CA1, and the entorhinal cortex in rats. J. Comp. Neurol., 353(3):379–390, 1995. 10.1002/cne.903530306 [DOI] [PubMed] [Google Scholar]
  • 104. Sürmeli G., Marcu D. C., McClure C., Garden D. L., Pastoll H., and Nolan M. F. Molecularly defined circuitry reveals input-output segregation in deep layers of the medial entorhinal cortex. Neuron, 88(5):1040–1053, 2015. 10.1016/j.neuron.2015.10.041 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 105. van Groen T. and Wyss J. M. The connections of presubiculum and parasubiculum in the rat. Brain Res., 518(1):227–243, 1990. 10.1016/0006-8993(90)90976-I [DOI] [PubMed] [Google Scholar]
  • 106. Caballero-Bleda M. and Witter M. P. Regional and laminar organization of projections from the presubiculum and parasubiculum to the entorhinal cortex: an anterograde tracing study in the rat. J. Comp. Neurol., 328(1):115–129, 1993. 10.1002/cne.903280109 [DOI] [PubMed] [Google Scholar]
  • 107. Caballero-Bleda M. and Witter M. P. Projections from the presubiculum and the parasubiculum to morphologically characterized entorhinal-hippocampal projection neurons in the rat. Exp. Brain Res., 101(1):93–108, 1994. 10.1007/BF00243220 [DOI] [PubMed] [Google Scholar]
  • 108. Canto C. B., Koganezawa N., Beed P., Moser E. I., and Witter M. P. All layers of medial entorhinal cortex receive presubicular and parasubicular inputs. J. Neurosci., 32(49):17620–17631, 2012. 10.1523/JNEUROSCI.3526-12.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109. Varga C., Lee S. Y., and Soltesz I. Target-selective GABAergic control of entorhinal cortex output. Nature Neurosci., 13(7):822–824, 2010. 10.1038/nn.2570 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 110. Ray S., Naumann R., Burgalossi A., Tang Q., Schmidt H., and Brecht M. Grid-layout and theta-modulation of layer 2 pyramidal neurons in medial entorhinal cortex. Science, 343(6173):891–896, 2014. 10.1126/science.1243028 [DOI] [PubMed] [Google Scholar]
  • 111. Kitamura T., Pignatelli M., Suh J., Kohara K., Yoshiki A., Abe K., and Tonegawa S. Island cells control temporal association memory. Science, 343(6173):896–901, 2014. 10.1126/science.1244634 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 112. Fuchs E. C., Neitz A., Pinna R., Melzer S., Caputi A., and Monyer H. Local and distant input controlling excitation in layer II of the medial entorhinal cortex. Neuron, 89(1):194–208, 2016. 10.1016/j.neuron.2015.11.029 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 113. Winterer J., Maier N., Wozny C., Beed P., Breustedt J., Evangelista R., Peng Y., D’Albis T., Kempter R., and Schmitz D. Excitatory microcircuits within superficial layers of the medial entorhinal cortex. Cell Rep., 2017. 10.1016/j.celrep.2017.04.041 [DOI] [PubMed] [Google Scholar]
  • 114. Donato F., Jacobsen R. I., Moser M.-B., and Moser E. I. Stellate cells drive maturation of the entorhinal-hippocampal circuit. Science, 3(6330):eaai8178, 2017. 10.1126/science.aai8178 [DOI] [PubMed] [Google Scholar]
  • 115. Iijima T., Witter M. P., Ichikawa M., Tominaga T., Kajiwara R., and Matsumoto G. Entorhinal-hippocampal interactions revealed by real-time imaging. Science, 272(5265):1176, 1996. 10.1126/science.272.5265.1176 [DOI] [PubMed] [Google Scholar]
  • 116. van Haeften T., Baks-te Bulte L., Goede P. H., Wouterlood F. G., and Witter M. P. Morphological and numerical analysis of synaptic interactions between neurons in deep and superficial layers of the entorhinal cortex of the rat. Hippocampus, 13(8):943–952, 2003. 10.1002/hipo.10144 [DOI] [PubMed] [Google Scholar]
  • 117. Kloosterman F., Van Haeften T., Witter M. P., and Lopes da Silva F. H. Electrophysiological characterization of interlaminar entorhinal connections: an essential link for re-entrance in the hippocampal–entorhinal system. Eur. J. Neurosci., 18(11):3037–3052, 2003. 10.1111/j.1460-9568.2003.03046.x [DOI] [PubMed] [Google Scholar]
  • 118. Beed P., Bendels M. H., Wiegand H. F., Leibold C., Johenning F. W., and Schmitz D. Analysis of excitatory microcircuitry in the medial entorhinal cortex reveals cell-type-specific differences. Neuron, 68(6):1059–1066, 2010. 10.1016/j.neuron.2010.12.009 [DOI] [PubMed] [Google Scholar]
  • 119. Tocker G., Barak O., and Derdikman D. Grid cells correlation structure suggests organized feedforward projections into superficial layers of the medial entorhinal cortex. Hippocampus, 25(12):1599–1613, 2015. 10.1002/hipo.22481 [DOI] [PubMed] [Google Scholar]
  • 120. Nicoll R. and Alger B. Synaptic excitation may activate a calcium-dependent potassium conductance in hippocampal pyramidal cells. Science, 212(4497):957–959, 1981. 10.1126/science.6262912 [DOI] [PubMed] [Google Scholar]
  • 121. Alonso A., De Curtis M., and Llinas R. Postsynaptic Hebbian and non-Hebbian long-term potentiation of synaptic efficacy in the entorhinal cortex in slices and in the isolated adult guinea pig brain. Proc. Nat. Acad. Sci. U.S.A., 87(23):9280–9284, 1990. 10.1073/pnas.87.23.9280 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 122. Accili E. A., Proenza C., Baruscotti M., and DiFrancesco D. From funny current to HCN channels: 20 years of excitation. Physiology, 17(1):32–37, 2002. [DOI] [PubMed] [Google Scholar]
  • 123. Robinson R. B. and Siegelbaum S. A. Hyperpolarization-activated cation currents: from molecules to physiological function. Annu. Rev. Physiol., 65(1):453–480, 2003. 10.1146/annurev.physiol.65.092101.142734 [DOI] [PubMed] [Google Scholar]
  • 124. Richter H., Heinemann U., and Eder C. Hyperpolarization-activated cation currents in stellate and pyramidal neurons of rat entorhinal cortex. Neurosci. Lett., 281(1):33–36, 2000. 10.1016/S0304-3940(00)00794-1 [DOI] [PubMed] [Google Scholar]
  • 125. Dickson C. T., Magistretti J., Shalinsky M. H., Fransén E., Hasselmo M. E., and Alonso A. Properties and role of Ih in the pacing of subthreshold oscillations in entorhinal cortex layer II neurons. J. Neurphysiol., 83(5):2562–2579, 2000. [DOI] [PubMed] [Google Scholar]
  • 126. Nolan M. F., Dudman J. T., Dodson P. D., and Santoro B. HCN1 channels control resting and active integrative properties of stellate cells from layer II of the entorhinal cortex. J. Neurosci., 27(46):12440–12451, 2007. 10.1523/JNEUROSCI.2358-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 127. de Curtis M. and Llinas R. R. Entorhinal cortex long-term potentiation evoked by theta-patterned stimulation of associative fibers in the isolated in vitro guinea pig brain. Brain Res., 600(2):327–330, 1993. 10.1016/0006-8993(93)91391-5 [DOI] [PubMed] [Google Scholar]
  • 128. Yun S. H., Mook-Jung I., and Jung M. W. Variation in effective stimulus patterns for induction of long-term potentiation across different layers of rat entorhinal cortex. J. Neurosci., 22(5):RC214, 2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 129. Solger J., Wozny C., Manahan-Vaughan D., and Behr J. Distinct mechanisms of bidirectional activity-dependent synaptic plasticity in superficial and deep layers of rat entorhinal cortex. Eur. J. Neurosci., 19(7):2003–2007, 2004. 10.1111/j.1460-9568.2004.03292.x [DOI] [PubMed] [Google Scholar]
  • 130. Yang S., Lee D., Chung C., Cheong M., Lee C.-J., and Jung M. Long-term synaptic plasticity in deep layer-originated associational projections to superficial layers of rat entorhinal cortex. Neurosci., 127(4):805–812, 2004. 10.1016/j.neuroscience.2004.06.001 [DOI] [PubMed] [Google Scholar]
  • 131. Deng P.-Y. and Lei S. Long-term depression in identified stellate neurons of juvenile rat entorhinal cortex. J. Neurphysiol., 97(1):727–737, 2007. 10.1152/jn.01089.2006 [DOI] [PubMed] [Google Scholar]
  • 132. Fyhn M., Hafting T., Treves A., Moser M.-B., and Moser E. I. Hippocampal remapping and grid realignment in entorhinal cortex. Nature, 446(7132):190–194, 2007. 10.1038/nature05601 [DOI] [PubMed] [Google Scholar]
  • 133. Muller R. U. and Kubie J. L. The effects of changes in the environment on the spatial firing of hippocampal complex-spike cells. J. Neurosci., 7(7):1951–1968, 1987. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 134. Bostock E., Muller R. U., and Kubie J. L. Experience-dependent modifications of hippocampal place cell firing. Hippocampus, 1(2):193–205, 1991. 10.1002/hipo.450010207 [DOI] [PubMed] [Google Scholar]
  • 135. Langston R. F., Ainge J. A., Couey J. J., Canto C. B., Bjerknes T. L., Witter M. P., Moser E. I., and Moser M.-B. Development of the spatial representation system in the rat. Science, 328(5985):1576–1580, 2010. 10.1126/science.1188210 [DOI] [PubMed] [Google Scholar]
  • 136. Wills T. J., Cacucci F., Burgess N., and O’Keefe J. Development of the hippocampal cognitive map in preweanling rats. Science, 328(5985):1573–1576, 2010. 10.1126/science.1188224 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 137. Wills T. J., Barry C., and Cacucci F. The abrupt development of adult-like grid cell firing in the medial entorhinal cortex. Front. Neural Circuits, 6:21, 2012. 10.3389/fncir.2012.00021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 138. Dhillon A. and Jones R. S. Laminar differences in recurrent excitatory transmission in the rat entorhinal cortex in vitro. Neurosci., 99(3):413–422, 2000. 10.1016/S0306-4522(00)00225-6 [DOI] [PubMed] [Google Scholar]
  • 139. Yoon K., Buice M. A., Barry C., Hayman R., Burgess N., and Fiete I. R. Specific evidence of low-dimensional continuous attractor dynamics in grid cells. Nat. Neurosci., 16(8):1077–1084, 2013. 10.1038/nn.3450 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 140. Dunn B., Mørreaunet M., and Roudi Y. Correlations and functional connections in a population of grid cells. PLoS Comput. Biol., 11(2):e1004052, 2015. 10.1371/journal.pcbi.1004052 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 141. Ben-Yishai R., Bar-Or R. L., and Sompolinsky H. Theory of orientation tuning in visual cortex. Proc. Nat. Acad. Sci. U.S.A., 92(9):3844–3848, 1995. 10.1073/pnas.92.9.3844 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 142. Carandini M. and Ringach D. L. Predictions of a recurrent model of orientation selectivity. Vision Res., 37(21):3061–3071, 1997. 10.1016/S0042-6989(97)00100-4 [DOI] [PubMed] [Google Scholar]
  • 143. Bienenstock E. L., Cooper L. N., and Munro P. W. Theory for the development of neuron selectivity: orientation specificity and binocular interaction in visual cortex. J. Neurosci., 2(1):32–48, 1982. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 144.N. Dagslott, F. Donato, Ø. A. Høydal, T. Waaga, M.-B. Moser, and E. I. Moser. Disrupted spatial representation following knock-out of NMDA receptors in the medial entorhinal cortex. In SFN Abstracts (183.08), 2016.
  • 145.I. U. Kruge, T. Waaga, T. Wernle, M.-B. Moser, and E. I. Moser. Development of grid cells requires only minimal experience with geometric boundaries. In SFN Abstracts (183.12), 2016.
  • 146. Constantinescu A. O., O’Reilly J. X., and Behrens T. E. Organizing conceptual knowledge in humans with a gridlike code. Science, 352(6292):1464–1468, 2016. 10.1126/science.aaf0941 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 147. Aronov D., Nevers R., and Tank D. W. Mapping of a non-spatial dimension by the hippocampal–entorhinal circuit. Nature, 543(7647):719–722, 2017. 10.1038/nature21692 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 148. Codling E. A., Plank M. J., and Benhamou S. Random walk models in biology. J. R. Soc. Interface, 5(25):813–834, 2008. 10.1098/rsif.2008.0014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 149. Beckmann P. Statistical distribution of the amplitude and phase of a multiply scattered field. J. Res. Natl. Bur. Stand., 66D:231–240, 1962. [Google Scholar]
  • 150. Stimberg M., Goodman D. F., Benichoux V., and Brette R. Equation-oriented specification of neural models for simulations. Front. Neuroinform., 8:6, 2014. 10.3389/fninf.2014.00006 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

All relevant data are within the paper.


Articles from PLoS Computational Biology are provided here courtesy of PLOS

RESOURCES