Skip to main content
Journal of Mathematical Neuroscience logoLink to Journal of Mathematical Neuroscience
. 2020 Dec 1;10:20. doi: 10.1186/s13408-020-00097-6

Stability analysis of a neural field self-organizing map

Georgios Detorakis 1,✉,#, Antoine Chaillet 2,3,#, Nicolas P Rougier 4,5
PMCID: PMC7708616  PMID: 33259016

Abstract

We provide theoretical conditions guaranteeing that a self-organizing map efficiently develops representations of the input space. The study relies on a neural field model of spatiotemporal activity in area 3b of the primary somatosensory cortex. We rely on Lyapunov’s theory for neural fields to derive theoretical conditions for stability. We verify the theoretical conditions by numerical experiments. The analysis highlights the key role played by the balance between excitation and inhibition of lateral synaptic coupling and the strength of synaptic gains in the formation and maintenance of self-organizing maps.

Keywords: Self-organizing maps, Neural fields, Lyapunov function, Asymptotic stability, Neural networks

Introduction

Self-organizing maps (SOMs) are neural networks mapping a high-dimensional space to a low-dimensional one through unsupervised learning. They were first introduced by Grossberg [14] and later by Kohonen [19]. SOMs are widely used in computer science and data analysis for quantization and visualization of high-dimensional data [25, 38]. They also constitute a suitable tool in computational neuroscience to study the formation and maintenance of topographic maps in primary sensory cortices such as the visual cortex [24, 31] and the somatosensory cortex [13, 34]. Many variations and applications of Kohonen’s SOM algorithm can be found in [16] and [27].

A type of self-organizing map based on neural fields theory has been introduced in [8], where neural fields are used to drive the self-organizing process. Neural fields are integrodifferential equations that describe the spatiotemporal dynamics of a cortical sheet [35]. The SOM proposed in [8] describes the topographic organization of area 3b of the primary somatosensory cortex of monkeys [21, 28]. The model relies on an earlier work [29] known as the dynamic SOM (DSOM) algorithm. DSOM provides an online SOM learning algorithm, where the Kohonen’s SOM time-dependent learning rate and neighborhood function have been replaced by time-invariant ones. The DSOM neighborhood function and learning rate solely depend on the distance of the winner unit (i.e., the most active neuron) from the input. The model proposed in [8, 9] combines the DSOM time-invariant learning rate and neighborhood function with Oja’s learning rule [26]. As thoroughly described in [8, 9], the model is compatible with anatomical evidence of how area 3b in monkeys develops, maintains, and reorganizes topographic representations of a skin patch of the index finger.

In this work, we provide theoretical insights on the stability and convergence of the neural field SOM algorithm proposed in [8, 9] by studying a more general class of systems than that originally proposed in [8]. We use Lyapunov’s stability theory adapted to neural field dynamics [11]. Since typical activation functions employed in the model (such as absolute values or rectification functions) are not necessarily differentiable, we do not rely on linearization techniques but rather directly assess the stability of the original nonlinear dynamics. Yet, the obtained results are local, meaning that they are valid only for initial conditions in the vicinity of the considered equilibrium. Nonetheless, we show that they agree with numerical simulations. The stability conditions derived in this work can be used toward the direction of tuning neural field models such that they achieve the best possible results in developing self-organizing maps and thus more generalized representations. Moreover, the conditions we propose indicate that the balance between lateral excitation and inhibition keeps the system stable, thus ruling out possible configurations in which learning does not take place properly. These findings are in line with both experimental observations [18, 30] and computational modeling [3537].

The paper is organized as follows. In Sect. 2, we recall the SOM model under concern and its basic mechanisms. In Sect. 3, we present our main theoretical results, which we confront to numerical simulations in Sect. 4. A discussion on the obtained results is provided in Sect. 5. Mathematical proofs are given in Sect. 6.

Self-organizing neural fields

Neural population dynamics

We consider the following neural fields equation:

τut(r,t)=u(r,t)+Ωwl(|rr|)rect(u(r,t))dr+I, 1

where Ω is a connected compact subset of Rq (q=1,2,3). For q=2, the integral of a function g:Ω=Ω1×Ω2R is to be understood as Ωg(r)dr=Ω1Ω2g(r1,r2)dr2dr1 with r=(r1,r2), and similarly for q=3; u(r,t) represents the mean membrane potential at position rΩ and time t0, τ is a positive decay time constant, I denotes an external input, and wl is a function that represents the strength of lateral synaptic coupling. It is given by

wl(x)=we(x)wi(x), 2

where the excitation and inhibition synaptic weights are typically given by

we(x)=Keex2/2σe2 3a

and

wi(x)=Kiex2/2σi2 3b

with Ke,Ki,σe,σi>0. In [8, 9] the input is provided through a two-dimensional skin model. The skin model is composed of a two-dimensional grid and receptors. The receptors are points distributed on the surface of the grid (uniformly). When a stimulus is applied on the grid, the receptors sample the input signal and convey the information to the cortical model. The skin stimulus is a noisy Gaussian-like function, and the input to the neural fields model is provided by the following function:

I(r,p,t)=1|wf(r,t)s(p)|1m, 4

where ||1 denotes the 1-norm: |x|1=i=1m|xi|, and s:R2[0,1]m is a function that maps the raw input from the two-dimensional skin space to [0,1]m. For instance, for a tactile stimulus at position pR2 on the skin, s(p)Rm could be defined as the normalized distance from p to the location of each receptor, thus potentially of much higher dimension than 2. For a more detailed description of receptor model, see [9]. The function wf:Ω×R0Rm represents feed-forward synaptic weights with value updated according to

wft(r,t)=γ(s(p)wf(r,t))Ωwe(|rr|)rect(u(r,t))dr, 5

where γ is a positive constant that represents the learning rate, and rect(x)=max{x,0}. It is worth observing that since s(p)[0,1]m, wf(r,t)[0,1]m for all rΩ and t0 given any initial conditions satisfying wf(r,0)[0,1]m for all rΩ (this can be seen by observing that the entries of wft(r,t) are negative as soon as the corresponding entries of wf(r,t) become greater than 1; similarly, they are positive when the corresponding entries of wf(r,t) get below 0: see (5)). Hence |wf(r,t)s(p)|1m[0,1] at all times. Thus expression (4) can be interpreted as a high input when the feedforward weights are close to s(p) and as a lower input when these are more distant.

The overall model (1), (4), (5) reflects the dynamics of a cortical neural population in combination with a learning rule of the feed-forward connections wf, which convey information from receptors to the cortical sheet. As described in [8, 9], this model can express a variety of different behaviors, depending on the lateral connectivity kernels we and wi.

The main advantage of the learning rule given by Eq. (5) is that it is a biologically plausible modification of the DSOM learning rule [29]. In DSOM the learning rate and neighborhood function are time-invariant and can adapt to the input according to one single parameter, called elasticity. This particular modification leads to the following behavior: if the winner neuron (i.e., the neuron that has the shortest distance from the input stimulus to its corresponding codebook–weight) is close to the stimulus, then the neighborhood function shrinks around it. This results in making the weights of neurons within the dynamic neighborhood stronger and the weights of the other units weaker. However, when the winning unit is very far from the current input, the neighborhood function exhibits a broad activity pattern, promoting learning of every unit in the network. Therefore in [8] the neighborhood function has been replaced by the term Ωwe(|rr|)rect(u(r,t))dr, providing a more realistic and biological plausible learning algorithm for self-organizing maps in the context of neuroscience.

Self-organizing maps

We start by briefly describing how the SOM model introduced in [8] and [9] works. The algorithm starts by initializing the feed-forward weights randomly (usually uniformly), and the neural field activity u(r,0) is set to zero. The second step is sampling the input space by randomly drawn samples of dimension m from an input distribution. At every epoch, one sample is given to the neural field (1) and (5) through Eq. (4). This first step is depicted in Fig. 1(A), where a two-dimensional point p=(p1,p2) is sampled from a uniform distribution, p1,p2U(0,1). The samples are mapped to the neural space through the function s and then are passed to Eq. (4). At this point, we should point out that there are two ways of presenting stimuli while training a self-organizing map. The first is predetermining an amount of input samples and present one at each epoch (online learning) and the second is collecting all the input samples into a batch and giving all of them at once to the network (batch learning). In this work, we use the former (online learning) since it is biologically plausible.

Figure 1.

Figure 1

Neural field self-organizing map. Graphical representation of the learning algorithm introduced in [8, 9]. (A) Tactile bidimensional stimuli are mapped to an m-dimensional space through a function s that involves the skin receptors. This function is used to update the codebooks, which are then mapped to the neural space Ω of lower dimension. The input I receives the mapped input sample and provides input to the neural field and codebook equations. (B) The numerical steady-state solution of Eq. (1) (i.e., bump) defines the neighborhood (the group of neurons) that will have its neurons updating their codebooks based on Eq. (5)

Then the algorithm proceeds with computing the numerical solution of Eqs. (1) and (5). To that aim, Eqs. (1) and (5) are discretized and solved numerically using Euler’s forward method. The numerical solution of Eq. (1) is typically a bell-shaped curve (bump) centered on the neuron that is the closest unit to the input sample and therefore is called the winner neuron or best matching unit (BMU). In Fig. 1(B), this is depicted as a black disc on a discrete lattice. The lattice represents a discretization of the field where each tile corresponds to a neuron. Neurons that lie within the vicinity (within the black disc in Fig. 1(B)) defined by the solution of Eq. (1) update their weights based on Eq. (5). The rest of the neurons feed-forward weights remain in their previous state. Once the temporal integration of Eqs. (1) and (5) is complete, the activity of the field is reset to its baseline activity. Then another input sample is drawn, and the whole process repeats itself. Once the number of epochs is exhausted, the learning stops, and the mapping process is completed.

To make the aforementioned algorithm directly comparable to Kohonen SOM [19], we provide some insights. First, in Kohonen’s SOM, we compute the distance between the input and the codebooks. Here we do the same using Eq. (4). The neighborhood function that Kohonen’s SOM uses to update the feed-forward weights is replaced here by the numerical solution of the neural field (Eq. (1)) and more precisely by the term Ωwe(|rr|)rect(u(r,t))dr. Both the learning rate and the width of the neighborhood function are time-independent in our case, as opposed to Kohonen’s SOM, where they are both time-dependent. Our learning rule is different since we use a modified Oja rule [26], which is based on Hebbian learning [15], and it is therefore biologically plausible [1]. The dimensionality reduction in both models, the Kohonen and ours, takes place at the level of the learning rule. This means that Eq. (5) is responsible for learning the representations and mapping the input distribution (of dimensions m) on a manifold of lower dimension q{1,2,3}.

Explicit conditions for stability

The most important question when one trains a self-organizing map is: Will the learning process converge and properly map the input space to the neural one? In most of the cases, it is not possible to predict this. However, in the specific case of the self-organizing algorithm provided by [8], here we show that it is possible to obtain an analytical condition that guarantees the stability of the equilibrium point of system (1)–(5). Stability during learning is a prerequisite to generate a meaningful mapping and thus a proper topographic map. Moreover, a byproduct of deriving such a stability condition is providing some insights on how to properly tune model parameters.

To this end, we now proceed to the mathematical analysis of the model. For generality, the adopted mathematical framework is slightly wider than merely Eqs. (1), (4), (5) and encompasses more general classes of activation functions and synaptic kernels. We start by introducing the considered class of systems and then provide sufficient conditions for its stability and convergence.

Model under study

The self-organizing neural field (1), (4), (5) is a particular case of the more general dynamics

τut(r,t)=u(r,t)+Ωwl(r,r)fl(u(r,t))dr+fs(wf(r,t)s(p)), 6a
wft(r,t)=γ(s(p)wf(r,t))Ωwe(r,r)fe(u(r,t))dr, 6b

where τ,γ>0, wl,weL2(Ω2,R), the set of all square-integrable functions from Ω2 to R, and fe, fl and fs are Lipschitz continuous functions.

Existence of equilibrium patterns

Assuming that infrΩΩwe(r,r)fe(u(r))dr0, any equilibrium pattern (u,wf) of (6a) and (6b) satisfies the following equations:

u(r)=fs(0)+Ωwl(r,r)fl(u(r))dr, 7a
wf(r)=s(p). 7b

Since ωlL2(Ω2,R), [12, Theorem 3.6] ensures the existence of at least one such equilibrium pattern.

Stability analysis of Eq. (6a) and (6b)

We recall that an equilibrium x of a system x˙(t)=f(x(t)), where x(t):ΩRn for each fixed t0, is called globally exponentially stable if there exist k,ε>0 such that, for all admissible initial conditions,

x(t)xkx(0)xeεt,t0, 8

where denotes the spatial L2-norm. This property ensures that all solutions go to the equilibrium configuration x in the L2 sense (global convergence) and that the transient overshoot is proportional to the L2-norm of the distance between the initial configuration and the equilibrium (stability). The equilibrium pattern x is said to be locally exponentially stable if (8) holds only for solutions starting sufficiently near from it (in the L2 sense). We refer the reader to [11] for a deeper discussion on the stability analysis of neural fields.

Our main result proposes a sufficient condition for the local exponential stability of Eq. (6a) and (6b). Its proof is given in Sect. 6.1.

Theorem 1

Let Ω be a compact connected set of Rq, let wlL2(Ω2,R), and let we:Ω2R be a bounded function. Assume further that fl, fs, and fe are Lipschitz continuous functions, and let l denote the Lipschitz constant of fl. Let (u,wf) denote any equilibrium of Eq. (6a) and (6b), as defined in Eq. (7a) and (7b). Then, under the conditions

ΩΩwl(r,r)2drdr<1l 9

and

infrΩΩwe(r,r)fe(u(r))dr>0, 10

the equilibrium pattern (u,wf) is locally exponentially stable for Eq. (6a) and (6b).

Condition (9) imposes that the synaptic weights of the lateral coupling wl are sufficiently small: stronger lateral synaptic weights can be tolerated if the maximum slope l of the activation function fl is low enough, meaning that the system given by Eq. (6a) is less self-excitable. Recall that if fl is a differentiable function, then l can be picked as the maximum value of its derivative. Nonetheless, Theorem 1 does not impose such a differentiability requirement, thus allowing us to consider nonsmooth functions such as absolute values, saturations, or rectification functions. Note that it was shown in [33] that condition (9) ensures that the system owns a single equilibrium pattern. It is also worth stressing that the slopes of the functions fs and fe do not intervene in the stability conditions.

Condition (10) requires a sufficient excitation in the vicinity of the equilibrium u. Roughly speaking, it imposes that the considered equilibrium pattern u does not lie in a region where fe is zero.

Stability analysis of the SOM neural fields

Theorem 1 provides a stability condition for the model described by Eq. (6a) and (6b). We next apply it to the model given in [8] to derive more explicit and testable stability conditions. More precisely, the self-organizing neural fields (1), (4), (5) can be put in the form of Eq. (6a) and (6b) by letting fe(x)=fl(x)=rect(x), fs(x)=1|x|1m, and

we(r,r)=Kee|rr|2/2σe2, 11a
wi(r,r)=Kie|rr|2/2σi2, 11b
wl(r,r)=we(r,r)wi(r,r). 11c

In view of (7a) and (7b), the equilibrium patterns of Eqs. (1), (4), (5) are given by

u(r)=1+Ωwl(r,r)rect(u(r))dr, 12a
wf(r)=s(p). 12b

The Lipschitz constant of fl is l=1. Based on this, we can also derive the following corollary, whose proof is provided in Sect. 6.2.

Corollary 1

Assume that Ω is a compact connected set of Rq, and let we, wi, and wl be as in (11a)(11c). Then, under the condition that

ΩΩ(Kee|rr|2/2σe2Kie|rr|2/2σi2)2drdr<1, 13

the equilibrium (u,wf), as defined in Eq. (12a) and (12b), is locally exponentially stable for Eqs. (1)(5).

A particular case for which local exponential stability holds is when the excitation and inhibition weight functions are sufficiently balanced. Indeed, it appears clearly that Eq. (13) is fulfilled if KeKi and σeσi. See the discussion in Sect. 5 for further physiological insights on this condition.

The integral involved in (13) can be solved explicitly. For instance, in the two-dimensional case (q=2) the condition boils down to the following:

Corollary 2

Let Ω=[a,b]×[a,b] for some a,bR with ba, and let we, wi, and wl be as in (11a)(11c). Define

ξa,b(σ):=(2σ2(e(ab)22σ21)+σ2π(ab)Erf(abσ2))2,σ>0, 14

where Erf:R(1,1) denotes the Gauss error function. Then, under the condition

Ke2ξa,b(σe/2)+Ki2ξa,b(σi/2)2KeKiξa,b(σeσiσe2+σi2)<1, 15

the equilibrium (u,wf), as defined in Eq. (12a) and (12b), is locally exponentially stable for Eq. (1)(5).

Plenty of approximations are available for the Erf function in the literature. For instance, the following expression approximates it with a 5.10−4 error:

Erf(x)11(1+a1x+a2x2+a3x3+a4x4)4

with a1=0.278393, a2=0.230389, a3=0.000972, and a4=0.078108; see, for instance, [2]. The Erf function is also commonly implemented in mathematical software, thus making Eq. (15) easily testable in practice.

Numerical assessment on a two-dimensional map

To numerically assess whether the above stability condition correctly predicts the performance of the learning process, we focus on a simple example of a two-dimensional map (q=2) and a two-dimensional input space (n=2). Furthermore, we choose s(p) to be the identity function since we do not consider any receptors: the position of the tactile stimuli is assumed to be directly available. This choice is motivated by the fact that the presence or absence of a receptors grid does not affect the theoretical results of the current work. We refer to [8, 9] for a more complex application of the neural field self-organizing algorithm.

We sample two-dimensional inputs from a uniform distribution. Therefore we have si(p)=(p1,p2), where i indicates the ith sample, and p1,p2U(0,1). In all our simulations, we use 7000 sample points and train the self-organizing map over each of them (7000 epochs). It is worth stressing the difference between the training time (epochs) and the simulation time. The former refers to the iterations over all the input samples (stimuli): one such input is presented to the model at each epoch. The latter is attributed to the numerical temporal integration of Eqs. (1)–(5). Thus each epoch corresponds to a predefined number of simulation steps. At the end of each epoch the activity of the neural field is reset to baseline activity before proceeding to the next epoch.

Parameters and simulation details

The neural fields equations are discretized using k=40×40 units. Accordingly, the two-dimensional model (1)–(5) is simulated over a spatial uniform discretization Ωd of the spatial domain Ω=[0,1]×[0,1], namely Ωd=i,j=140(i40,j40). The considered input space, over which the stimuli are uniformly distributed, is also [0,1]×[0,1] (two-dimensional input vectors). The temporal integration is performed using the forward Euler method, whereas the spatial convolution in Eqs. (1)–(5) is computed via the fast Fourier transform (FFT). The learning process runs for 7000 epochs. The components of the feed-forward weights are initialized from a uniform distribution U(0,0.01), and the neural field activity is set to zero. At each epoch, we feed a stimulus to Eqs. (1)–(5), and the system evolves according to its dynamics, whereas the feed-forward weights are being updated. Then we reset the neural fields activity to zero. We run each experiment ten times using a different pseudorandom number generator (PRNG) seed each time (the PRNG seeds are given in Appendix 7: the same initial conditions and the set of PRNG seeds were used in each experimental condition).

The source code is written in Python (Numpy-, Numba-, Sklearn, and Matplotlib-dependent) and are freely distributed under the GPL 3-Clause License (https://github.com/gdetor/som_stability). All the parameters used in numerical simulations are summarized in Table 1. All simulations ran on an Intel NUC machine equipped with an Intel i7-10th generation processor and 32 GB of physical memory, running Ubuntu Linux (20.04.1 LTS, Kernel: 5.4.0-47-generic). The simulation of one self-organizing map consumes 493 MB of physical memory, and it took 2671 seconds to run the 7000 epochs.

Table 1.

Simulation parameters. Ke and Ki are the amplitudes of excitatory and inhibitory lateral connections, respectively; σe and σi are the variances of excitatory and inhibitory lateral connections, respectively; τ is the decay time constant, dt is the integration time step in ms, t is the simulation time in seconds, and γ is the learning rate. In each epoch, one stimulus is presented to the model

Ke σe Ki σi τ dt t γ epochs
Figure 2 0.90 0.11 0.86 1.0 1.0 0.015 25.0 0.002 7000
Figure 3 3.0 0.11 2.80 1.0 1.0 0.015 25.0 0.002 7000

SOM quality measures

We measure the quality of the self-organizing maps using two performance indicators, the distortion D [6] and the δxδy representation [7]. We recall here that Ωd is the spatial uniform discretization of Ω=[0,1]×[0,1] and k=40×40 is the number of nodes (neurons). Furthermore, for each j{1,,k}, wfj(t) denotes the steady-state value of the feed-forward weights at the jth node of the spatial discretization, and t corresponds to the time at the end of an epoch.

The distortion assesses the quality of a self-organizing map. It measures the loss of information over the learning process. In other words, it indicates how good a reconstruction of an input will be after the mapping of all inputs to a lower-dimensional neural map. In a sense, distortion measures how well a SOM algorithm “compresses” the input data with respect to the neighborhood structure. Mathematically, the distortion is computed according to its discrete approximation:

D=1ni=1nminj{1,,k}|si(p)wfj(t)|2, 16

where n is the number of samples we use during the training of the self-organizing map.

Distortion is essentially an indicator of the map convergence, but it is not a reliable tool for assessing its quality. To gauge the quality of the map, we use the δxδy representation [7]. It shows when a map preserves the topology of the input space and hence how well a topographic map is formed. To estimate the δxδy, we compute all the pairwise distances between the feed-forward weights, δx=δx(i,j)=|wfi(t)wfj(t)|, and all the distances between the nodes of the uniform discretization of the input space [0,1]2, δy(i,j)=|yiyj| for i,j=1,,k, where yi are the discrete nodes of Ωd. We plot the δxδy (i.e., δx is the ordinate, and δy the abscissa) along with a straight line, named Lδxδy, that crosses the origin and the mean of δx points. If the point cloud representation of δxδy closely follows the line Lδxδy, then the map is considered well-formed and preserves the topology of the input space.

To quantify the δxδy representation through a scalar performance index, we perform a linear regression on the point cloud of δxδy without fitting the intercept (magenta line in figures), and we get a new line named LΔ. Then we define the measure P=i=1k(aibi)2, where aiLδxδy and biLΔ. Naturally, P should approach zero as the two lines are getting closer, indicating that the self-organizing map respects the topology of the input space, and thus it is well-formed.

Stable case

We start by simulating the model described by Eqs. (1)–(5) with the parameters given in the first line of Table 1. With these parameters, condition (15) is fulfilled (0.47<1), and Corollary 2 predicts that the equilibrium is exponentially stable over each epoch. Accordingly, the model succeeds in building up a self-organizing map as shown in panel (A) of Fig. 2. The white discs indicate the feed-forward weights after learning, and the black dots indicate the input data points (two-dimensional rectangular uniform distribution).

Figure 2.

Figure 2

Two-dimensional SOM performance in the stable case. (A) Feed-forward weights (white discs) as they have been organized into a topographic map after 7000 epochs. The input in this case is a two-dimensional rectangular uniform distribution (black dots). (B) δxδy representation (black cloud), mean of δx (red line), and the linear regression of the δxδy representation (magenta line). The fact that the cloud is aligned around the red line indicates that the topographic map is well organized, as confirmed by a good index performance P=0.01. (C) Distortion indicates that the loss of information during the learning process decreases, and the mapping of the input data to a two-dimensional self-organizing map respects the structure of the neighborhoods. (D) Temporal evolution of norm-2 of feed-forward weights of three neurons placed at r=(0.25,0.25), (0.1,0.225), and (0.35,0.075)). Condition (15) is fulfilled, and therefore the weights converge to an equilibrium giving rise to a well-formed topographic map

Panels (B) and (C) show the δxδy representation and the distortion, respectively. We observe that the δxδy representation indicates a correlation between the feed-forward weights and the rectangular grid points (aligned with the mean of δx–red line). This means that the self-organizing map is well-formed and conserves the topology of the input. Moreover, the distortion declines and converges toward 0.0025, pointing out first that the loss of information during learning is low and that the structure in the self-organizing map is preserved. However, the boundary effects (the density of points is higher at the boundary of the map in panel (A)) affect both the distortion (it does not converge to zero; see panel (C) in Fig. 2) and the δxδy representation (it is not perfectly aligned with the red line; see panel (B) in Fig. 2). In spite of these boundary effects, the obtained δxδy performance indicator is good (P=0.01).

The evolution of the norm-2 of feed-forward weights of three randomly chosen units (r=(0.25,0.25), (0.1,0.225), (0.35,0.075)) is shown in the panel (D) of Fig. 2. This implies that the weights converge to an equilibrium after a transient period of about 2000 epochs. The oscillations around the equilibrium are due to a repeated alteration of the input stimulus, which causes a shift to the feed-forward weights values of each winner neuron (see [8] for more detail).

Unstable case

The second line of Table 1 provides parameters for which Condition (15) is violated (5.25>1). According to our theoretical predictions, the model might not be stable and thus may not be able to develop any self-organizing map at all. To make sure that this is the case (and not merely a transient effect), we have let the training take more epochs (20,000). Nevertheless, here we present only the 7000 first epochs for consistency with the rest of our experiments. This situation is illustrated in Fig. 3, where the self-organizing process has failed to generate a well-formed map (panel (A)). In this case, it is apparent that self-organization process has failed to generate a topographic map.

Figure 3.

Figure 3

Two-dimensional SOM performance in the unstable case. (A) Feed-forward weights (white discs) as they have failed to organize into a topographic map after 7000 epochs. The input in this case is a two-dimensional rectangular uniform distribution (black dots). (B) δxδy representation (black cloud), mean of δx (red line), and the linear regression of the δxδy representation (magenta line). The fact that the cloud looks diffused indicates that the topographic map is not well organized, as confirmed by a high value of P=0.41. (C) Distortion indicates that the loss of information during the learning process never drops (converges to an equilibrium) instead it oscillates. This means that the mapping of the input data to a two-dimensional map has failed. (D) Temporal evolution of norm-2 of feed-forward weights of three neurons placed at r=(0.25,0.25), (0.1,0.225), and (0.35,0.075)). Condition (15) is violated, and accordingly the weights do not converge to an equilibrium

The δxδy representation in panel (B) of Fig. 3 looks like a diffused cloud, indicating that there is no correlation between the grid points and the feed-forward weights, meaning that there is no preservation of the topology of the input space. Accordingly, the performance index reaches the value P=0.41, thus higher than the stable case. Moreover, the distortion in panel (C) of Fig. 3 oscillates without converging to an equilibrium, pointing out that the loss of information remains high and therefore the mapping is not successful. Finally, the norm-2 of feed-forward weights of three units (r=(0.25,0.25), (0.1,0.225), (0.35,0.075)) are shown in panel (D): it is apparent that they do not converge to an equilibrium. Instead, they oscillate violently and never stabilize around an equilibrium configuration.

Numerical assessment of Corollary 2

Finally, we numerically tested condition (15) of Corollary 2 for different values of the parameters Ke and Ki (all other parameters remained the same as in Table 1). For each pair (Ke,Ki), we computed the left-hand side of Eq. (15), the distortion D (averaged over the last 10 epochs), and the δxδy performance index P; see Fig. 4. We observe that for high values of Ke and Ki, the stability condition of Corollary 2 is violated (the black solid line overpasses the black dashed line). The distortion (orange curve) closely follows the left-hand side of condition (15) (up to a scaling factor), suggesting that distortion can serve as a measure of stability of system (1)–(5). Furthermore, the distortion and the δxδy performance index P indicate that the learning process degrades for high values of (Ke,Ki), in line with the fact that condition (15) is violated. Figure 5 confirms this good alignment between the theoretical stability condition and the performance of the self-organizing map: for the first five cases, it properly maps the input space to the neural one, whereas the topology of the input space is not preserved in the last two cases, and a malformed topographic map is obtained.

Figure 4.

Figure 4

Numerical investigation of Corollary 2. Eight different pairs of the parameters (Ke,Ki) were used to investigate the conservativeness of the stability condition given by Corollary 2. We ran eight different simulations for 7000 epochs, keeping always the rest of the parameters as in Table 1 and the same PRNG seed as before (7659). The black curve indicates the numerical value of the left-hand side of (15): the stability is guaranteed if it is below the black dashed line. The green curve indicates the δxδy performance index P. The orange curve represents the distortion D averaged of the 10 last epochs. It is apparent that as the values of (Ke,Ki) increase the Corollary 2 becomes violated and the self-organizing map fails to map the input space to the neural one (see Fig. 5 for more detail)

Figure 5.

Figure 5

Numerical Investigation of Corollary 2. For the same eight experiments as in Fig. 4, the obtained self-organizing map is provided (first line), together with its δxδy representation (second line) and the evolution of the distortion (third line). The mean δx is represented as a red line, whereas the slope of the linear regression is given as a magenta line. (A): (Ke=0.30, Ki=0.25), (B): (Ke=0.4, Ki=0.35), (C): (Ke=0.5, Ki=0.45), (D): (Ke=0.7, Ki=0.63), (E): (Ke=0.9, Ki=0.86), (F): (Ke=1.0, Ki=0.92). In line with Fig. 4, a relevant map is obtained for the first five experiments (for which condition (15) is fulfilled), whereas for the two last self-organizing maps (G): (Ke=2, Ki=1.85) and (H): (Ke=3, Ki=2.85), the stability condition (15) is violated. This violation results in a nonstable neural field equation, and thus the self-organizing maps do not learn properly the representations

Conclusion

In this work, we have presented theoretical conditions for the stability of a neural field system coupled with an Oja-like learning rule [26]. Numerical assessments on a two-dimensional self-organizing map indicate that the theoretical condition is closely aligned with the capacity of the network to form a coherent topographic map.

Previous works have shown through simulations that the dynamical system described by Eqs. (1)–(5) can develop topographic maps through an unsupervised self-organization process [8, 9]. The model relies on the activity of a neural field to drive a learning process. This type of models are capable of developing topographic maps and reorganize them in face of several kinds of disturbances. Here we proceed to a rigorous theoretical analysis of such kind of models by employing neural field Lyapunov theory.

The obtained stability conditions are reminiscent of those obtained for general neural fields dynamics, in which the spatial L2-norm of synaptic weights plays an essential role [11, 22, 33]. In our setting, these conditions translate in a good balance between excitation and inhibition for the exponential stability of the model equilibrium, thus allowing the self-organizing process to develop topographic maps. It is worth stressing that the proof techniques employed here do not rely on a linearization of the system around the considered equilibrium; it thus allows us to cover nondifferentiable activation functions (such as classical saturation or rectification functions).

These stability conditions provide a means to identify the parameters set within which the unsupervised learning works efficiently and thus provides an indication on how to tune them in practice. In particular, they can be used to further investigate how the dynamics of an underlying system affects the learning process during an unsupervised training process and what is the effect of the parameters on the final topographic map: as Fig. 4 indicates, the parameters of the model directly affect the quality of the topographic map. However, a limitation of the present work is that it does not offer a way of choosing the parameters in an optimal way. Furthermore, although the conditions provided by Theorem 1 guarantee the stability of the neural field, they do not predict the quality of the obtained map: Stability ensures that the learning will converge to an equilibrium, but the quality of the obtained map strongly depends on the structure of this equilibrium and hence on the chosen initial values of the feed-forward weights. This is a well-known problem with self-organizing maps [19], which is generally solved using a decreasing neighborhood, starting from a very wide one. In our case the neighborhood function is directly correlated with the profile of the field activity and is fixed (stereotyped). We thus cannot always ensure the proper unfolding of the map. It is to be noted that when the neighborhood of a Kohonen is kept fixed, it suffers from similar problems. Nevertheless, the numerical assessment of the proposed theoretical stability conditions suggests that the stability condition accurately predicts the emergence of topographic maps through unsupervised learning: see Figs. 4 and 5.

Other works have studied stability conditions for Kohonen maps and vector quantization algorithms using methods from linear systems stability theory [32] or through energy functions [10]. However, these works focus on the learning rule for the Kohonen self-organizing maps [20], and the dynamics are not explicitly given by dynamical systems. Our work goes beyond by taking into account not only the learning dynamics, but also the neural dynamics that drives the self-organizing process.

Last but not least, it has been shown that neural adaptation is crucial in the development of the neocortex [23] and neurons tend to adapt their input/output relation according to the statistics of the input stimuli. Our theoretical results provide conditions under which this input/output adaptation successfully takes place at least at a computational level.

Proof of the theoretical results

Proof of Theorem 1

To place the equilibrium at the origin, we employ the following change of variables:

u˜(r,t)=u(r,t)u(r),w˜f(r,t)=wf(r,t)wf(r),

where u and wf denote the equilibrium patterns of Eq. (6a) and (6b), as defined in Eq. (7a) and (7b). Then system (6a) and (6b) can be written as

τu˜t(r,t)=u˜(r,t)+Ωwl(r,r)f˜l(r,u˜(r,t))dr+f˜s(w˜f(r,t)), 17a
w˜ft(r,t)=γw˜f(r,t)Ωwe(r,r)f˜e(u˜(r,t))dr, 17b

where for all xR and all rΩ,

f˜l(r,x)=fl(x+u(r))fl(u(r)),f˜s(x)=fs(x)fs(0),f˜e(r,x)=fe(x+u(r)).

With this notation, we have f˜l(r,0)=f˜s(0)=0 for all rΩ, meaning that (17a) and (17b) owns an equilibrium at zero. Thus the stability properties of the origin of (17a) and (17b) determine those of the equilibria of (6a) and (6b).

First, observe that since we is a bounded function and Ω is compact, there exists ω¯e>0 such that

Ωwe(r,r)2drw¯e2,rΩ. 18

To assess the stability of (17a) and (17b), we may be tempted to rely on linearization techniques. Nevertheless, the linearized system (17a) and (17b) around the origin would necessarily involve the derivative of fs at zero, which may be undefined if fs is not differentiable at zero (which is the case for the system of interest (1)–(5), where fs involves an absolute value). Consequently, the proof we propose here relies on Lyapunov methods [17], which were extended to neural fields in [11].

Consider the following Lyapunov functional:

V(t):=τ2Ωu˜(r,t)2dr+ρ2γΩw˜f(r,t)2dr, 19

where ρ>0 denotes a parameter whose value will be decided later. First, observe that the following bounds hold at all t0:

α_(u˜(,t)2+w˜f(,t)2)V(t)α(u˜(,t)2+w˜f(,t)2), 20

where α_:=12min{τ;ρ/γ}>0 and α:=12max{τ;ρ/γ}>0. The derivative of V along the solutions of (6a) and (6b) reads

V˙(t)=τΩu˜(r,t)u˜(r,t)tdr+ργΩw˜f(r,t)w˜f(r,t)tdr=Ωu˜(r,t)[u˜(r,t)+Ωwl(r,r)f˜l(r,u˜(r,t))dr+f˜s(w˜f(r,t))]drρΩ[w˜f(r,t)2Ωwe(r,r)f˜e(r,u˜(r,t))dr]dr. 21

Moreover, denoting by s, e, and l the Lipschitz constants of fs, fe, and fl respectively, we have that for all xR and all rΩ.

|f˜l(r,x)|l|x|, 22
|f˜s(x)|s|x|, 23
|f˜e(r,x)f˜e(r,0)|e|x|. 24

Applying the Cauchy–Schwarz inequality and using Eq. (22), it follows that

Ωwl(r,r)f˜l(r,u˜(r,t))drΩ|wl(r,r)||f˜l(r,u˜(r,t))|drlΩ|wl(r,r)||u˜(r,t)|drlΩwl(r,r)2drΩu˜(r,t)2dr.

Hence, using again the Cauchy–Schwarz inequality,

Ωu˜(r,t)[Ωwl(r,r)f˜l(r,u˜(r,t))dr]drlΩ|u˜(r,t)|[Ωwl(r,r)2drΩu˜(r,t)2dr]drlΩu˜(r,t)2drΩ[Ωwl(r,r)2drΩu˜(r,t)2dr]dr.

Observing that Ωu˜(r,t)2dr is independent of r and defining

w¯l:=ΩΩwl(r,r)2drdr, 25

it follows that

Ωu˜(r,t)[Ωwl(r,r)f˜l(r,u˜(r,t))dr]drlw¯lΩu˜(r,t)2dr. 26

Furthermore, using Eq. (23), we have that

Ωu˜(r,t)f˜s(wf(r,t))drsΩ|u˜(r,t)||wf(r,t)|drsΩu˜(r,t)2drΩwf(r,t)2dr.

Invoking the inequality 2ab(a2/λ+λb2) for all a,bR and λ>0, we obtain that

Ωu˜(r,t)f˜s(w˜f(r,t))drs2(λΩu˜(r,t)2dr+1λΩw˜f(r,t)2dr) 27

for any λ>0.

Now assumption (10) ensures that infrΩΩwe(r,r)f˜e(r,0)dr>0. It follows that there exists c>0 such that

Ωwe(r,r)f˜e(r,0)dr2c,rΩ.

Consequently, using (24) and the Cauchy–Schwarz inequality, we get that for any vL2(Ω,R),

Ωwe(r,r)f˜e(r,v(r))dr=Ωwe(r,r)f˜e(r,0)dr+Ωwe(r,r)(f˜e(r,v(r))f˜e(r,0))dr2cΩ|we(r,r)||f˜e(r,v(r))f˜e(r,0)|dr2ceΩ|we(r,r)||v(r)|dr2ceΩwe(r,r)2drΩv(r)2dr2cew¯ev,

where the last bound comes from (18). Let Bε denote the ball (in L2-norm) of radius ε>0, that is, Bε:={vL2(Ω,R):v<ε}. Letting ε:=c/ew¯e, we conclude from the above expression that

Ωwe(r,r)f˜e(r,v(r))drc,rΩ,vBε. 28

Consider an initial condition such that u˜(,0)Bε and let T[0,+] denote the time needed for u˜(,t) to leave Bε. Then, by definition, u˜(,t)Bε for all t[0,T), and u˜(,T)Bε if T is finite. Note that by the continuity of solutions, T>0. Moreover, in view of (28),

Ωwe(r,r)f˜e(r,u˜(r,t))drc,t[0,T),rΩ. 29

Combining Eqs. (21), (22), (23), and (29), we obtain that for all t[0,T),

V˙(t)(1lw¯lλs2)Ωu˜(r,t)2dr(ρcs2λ)Ωw˜f(r,t)2dr

Pick λ=(1lw¯l)/s. Note that λ>0 since lw¯l<1 by assumption (see Eq. (9)). Then the choice ρ=scλ=s2c(1lw¯l)>0 leads to

V˙(t)12u˜(,t)2ρc2w˜f(,t)212min{1;ρc}(u˜(,t)2+w˜f(,t)2).

Using (20) and letting α:=12αmin{1;ρc}>0, we finally obtain that

V˙(t)αV(t),t[0,T).

Integrating this gives V(t)V(0)eαt for all t[0,T), which yields, using (20),

u˜(,t)2+w˜f(,t)2αα_(u˜(,0)2+w˜f(,0)2)eαt,t[0,T). 30

Thus if initial conditions are picked within the L2-ball of radius α_εα, then u˜(,t)+w˜f(,t)<ε at all times t0. This means that for these initial conditions, solutions never leave the ball Bε, and hence T=+. Thus Eq. (30) ensures the exponential stability on this set of initial conditions.

Proof of Corollary 1

Assumption described by Eq. (13) is equivalent to requiring w¯l<1, with w¯l defined in Eq. (25). Since the Lipschitz constant of the rectification is l=1, this makes Eq. (9) fulfilled.

Moreover, we claim that the solution u of the implicit Eq. (12a) and (12b) is necessarily positive on some subset of Ω of nonzero measure. To see this, assume on the contrary that u(r)0 for almost all rΩ. Then rect(u(r))=0 for almost all rΩ, which implies that

Ωwl(r,r)rect(u(r))dr=0,rΩ.

In view of Eq. (12a) and (12b), this implies that u(r)=1 for all rΩ, thus leading to a contradiction. Consequently, as claimed, u is necessarily positive on some subset Ω+ of Ω of nonzero measure. Recalling that here Ω is assumed to be a compact set, it follows that

infrΩΩe|rr|2/2σe2rect(u(r))drinfrΩΩ+e|rr|2/2σe2u(r)dr>0,

which makes Eq. (10) satisfied. The conclusion then follows from Theorem 1.

Proof of Corollary 2

The following one-dimensional relation holds:

ababe|xy|22σ2dxdy=ξa,b(σ). 31

To compute its two-dimensional counterpart, let r=(r1,r2) and r=(r1,r2). Then, for Ω=[a,b]×[a,b],

ΩΩe|rr|22σ2drdr=ababababexp((r1r1)2+(r2r2)22σ2)dr1dr2dr1dr2.

Using Fubini’s theorem, it follows that

ΩΩe|rr|22σ2drdr=ababexp((r1r1)22σ2)(ababexp((r2r2)22σ2)dr2dr2)dr1dr1.

By Eq. (31) this gives

ΩΩe|rr|22σ2drdr=ababexp((r1r1)22σ2)ξa,b(σ)dr1dr1.=ξa,b(σ). 32

The left-hand term of Eq. (13) then reads

ΩΩ(Kee|rr|22σe2Kie|rr|22σi2)2drdr=ΩΩ(Ke2e|rr|2σe2+Ki2e|rr|2σi22KeKie|rr|2σeσi2σe2+σi2)drdr=Ke2ξa,b(σe/2)+Ki2ξa,b(σi/2)2KeKiξa,b(σeσiσe2+σi2),

which concludes the proof.

PRNG seed

We ran both stable and nonstable experiments ten times with different PRNG seeds. All the PRNG seeds we used are 10, 74, 433, 721, 977, 1330, 3433, 5677, 9127, 7659.

Acknowledgements

Not applicable.

Abbreviations

SOM

Self-organizing map

DSOM

Dynamic self-organizing map

FFT

Fast Fourier transform

PRNG

Pseudorandom number generator

Authors’ contributions

GID conceived the idea, wrote the code, designed, and ran the numerical experiments, AC performed the mathematical derivation and analysis. GID, AC, and NPR contributed to preparing, writing, and revising the manuscript. All authors read and approved the final manuscript.

Funding

This work was partially funded by grant ANR-17-CE24-0036.

Availability of data and materials

The source code used in this work for running the simulations, analyzing the results, and plotting the figures is freely distributed under the GPL-3 License and can be found at https://github.com/gdetor/som_stability.

Ethics approval and consent to participate

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Consent for publication

Not applicable.

Footnotes

Georgios Detorakis and Antoine Chaillet contributed equally to this work.

References

  • 1.Abbott LF, Nelson SB. Synaptic plasticity: taming the beast. Nat Neurosci. 2000;3:1178–1183. doi: 10.1038/81453. [DOI] [PubMed] [Google Scholar]
  • 2.Abramowitz M, Stegun IA. Handbook of mathematical functions with formula, graphs and mathematical tables. New York: Dover; 1983. [Google Scholar]
  • 3.Amari S. Dynamics of pattern formation in lateral-inhibition type neural fields. Biol Cybern. 1977;27(2):77–87. doi: 10.1007/BF00337259. [DOI] [PubMed] [Google Scholar]
  • 4.Bressloff PC. Spatiotemporal dynamics of continuum neural fields. J Phys A, Math Theor. 2012;45(3):033001. [Google Scholar]
  • 5.Coombes S. Waves, bumps, and patterns in neural field theories. Biol Cybern. 2005;93(2):91–108. doi: 10.1007/s00422-005-0574-y. [DOI] [PubMed] [Google Scholar]
  • 6.Cover TM, Thomas JA. Elements of information theory. New York: Wiley; 2012. [Google Scholar]
  • 7. Demartines P. Organization measures and representations of Kohonen maps. In: First IFIP Working Group. vol. 10. Citeseer; 1992.
  • 8.Detorakis GI, Rougier NP. A neural field model of the somatosensory cortex: formation, maintenance and reorganization of ordered topographic maps. PLoS ONE. 2012;7(7):e40257. doi: 10.1371/journal.pone.0040257. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Detorakis GI, Rougier NP. Structure of receptive fields in a computational model of area 3b of primary sensory cortex. Front Comput Neurosci. 2014;8:76. doi: 10.3389/fncom.2014.00076. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Erwin E, Obermayer K, Schulten K. Self-organizing maps: ordering, convergence properties and energy functions. Biol Cybern. 1992;67(1):47–55. doi: 10.1007/BF00201801. [DOI] [PubMed] [Google Scholar]
  • 11.Faugeras O, Grimbert F, Slotine J-J. Absolute stability and complete synchronization in a class of neural fields models. SIAM J Appl Math. 2008;69(1):205–250. [Google Scholar]
  • 12.Faugeras O, Veltz R, Grimbert F. Persistent neural states: stationary localized activity patterns in nonlinear continuous n-population, q-dimensional neural networks. Neural Comput. 2009;21(1):147–187. doi: 10.1162/neco.2008.12-07-660. [DOI] [PubMed] [Google Scholar]
  • 13.Grajski KA, Merzenich M. Advances in neural information processing systems. 1990. Neural network simulation of somatosensory representational plasticity; pp. 52–59. [Google Scholar]
  • 14. Grossberg S. Physiological interpretation of the self-organizing map algorithm. 1994.
  • 15. Hebb DO. The organization of behavior: a neuropsychological theory. Psychology Press; 2002.
  • 16.Kaski S, Kangas J, Kohonen T. Bibliography of self-organizing map (SOM) papers: 1981–1997. Neural Comput Surv. 1998;1(3–4):1–176. [Google Scholar]
  • 17.Khalil H. Nonlinear systems. 2. New York: Macmillan Publishing Co.; 1996. [Google Scholar]
  • 18.Knight RT, Staines WR, Swick D, Chao LL. Prefrontal cortex regulates inhibition and excitation in distributed neural networks. Acta Psychol. 1999;101(2):159–178. doi: 10.1016/s0001-6918(99)00004-9. [DOI] [PubMed] [Google Scholar]
  • 19.Kohonen T. Self-organized formation of topologically correct feature maps. Biol Cybern. 1982;43(1):59–69. [Google Scholar]
  • 20.Kohonen T. Self-organizing maps. Berlin: Springer; 2001. [Google Scholar]
  • 21.Krubitzer LA, Kaas JH. The organization and connections of somatosensory cortex in marmosets. J Neurosci. 1990;10(3):952–974. doi: 10.1523/JNEUROSCI.10-03-00952.1990. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Galtier MN, Faugeras OD, Bressloff PC. Hebbian learning of recurrent connections: a geometrical perspective. Neural Comput. 2012;24(9):2346–2383. doi: 10.1162/NECO_a_00322. [DOI] [PubMed] [Google Scholar]
  • 23.Mease RA, Famulare M, Gjorgjieva J, Moody WJ, Fairhall AL. Emergence of adaptive computation by single neurons in the developing cortex. J Neurosci. 2013;33(30):12154–12170. doi: 10.1523/JNEUROSCI.3263-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Miikkulainen R, Bednar JA, Choe Y, Sirosh J. Computational maps in the visual cortex. Berlin: Springer; 2006. [Google Scholar]
  • 25.Nasrabadi NM, Feng Y. Proc. IEEE int. conf. Neural networks. 1988. Vector quantization of images based upon the kohonen self-organizing feature maps; pp. 101–105. [Google Scholar]
  • 26.Oja E. Simplified neuron model as a principal component analyzer. J Math Biol. 1982;15(3):267–273. doi: 10.1007/BF00275687. [DOI] [PubMed] [Google Scholar]
  • 27.Oja M, Kaski S, Kohonen T. Bibliography of self-organizing map (som) papers: 1998-2001 addendum. Neural Comput Surv. 2003;3(1):1–156. [Google Scholar]
  • 28.Qi HX, Preuss TM, Kaas JH. Somatosensory areas of the cerebral cortex: architectonic characteristics and modular organization. Senses Compr Ref. 2008;6:143. [Google Scholar]
  • 29.Rougier N, Boniface Y. Dynamic self-organising map. Neurocomputing. 2011;74(11):1840–1847. [Google Scholar]
  • 30.Schaefer M, Heinze H-J, Rotte M. Task-relevant modulation of primary somatosensory cortex suggests a prefrontal–cortical sensory gating system. NeuroImage. 2005;27(1):130–135. doi: 10.1016/j.neuroimage.2005.04.005. [DOI] [PubMed] [Google Scholar]
  • 31.Sirosh J, Miikkulainen R. Advances in neural information processing systems. 1995. Ocular dominance and patterned lateral connections in a self-organizing model of the primary visual cortex; pp. 109–116. [Google Scholar]
  • 32.Tucci M, Raugi M. The 2010 international joint conference on neural networks (IJCNN) Los Alamitos: IEEE Comput. Soc.; 2010. Stability analysis of self-organizing maps and vector quantization algorithms; pp. 1–5. [Google Scholar]
  • 33.Veltz R, Faugeras O. Local/global analysis of the stationary solutions of some neural field equations. SIAM J Appl Dyn Syst. 2010;9(3):954–998. [Google Scholar]
  • 34.Wilson SP, Law JS, Mitchinson B, Prescott TJ, Bednar JA. Modeling the emergence of whisker direction maps in rat barrel cortex. PLoS ONE. 2010;5(1):e8778. doi: 10.1371/journal.pone.0008778. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Xing J, Gerstein GL. Networks with lateral connectivity. I. dynamic properties mediated by the balance of intrinsic excitation and inhibition. J Neurophysiol. 1996;75(1):184–199. doi: 10.1152/jn.1996.75.1.184. [DOI] [PubMed] [Google Scholar]
  • 36.Xing J, Gerstein GL. Networks with lateral connectivity. II. Development of neuronal grouping and corresponding receptive field changes. J Neurophysiol. 1996;75(1):200–216. doi: 10.1152/jn.1996.75.1.200. [DOI] [PubMed] [Google Scholar]
  • 37.Xing J, Gerstein GL. Networks with lateral connectivity. III. Plasticity and reorganization of somatosensory cortex. J Neurophysiol. 1996;75(1):217–232. doi: 10.1152/jn.1996.75.1.217. [DOI] [PubMed] [Google Scholar]
  • 38.Yin H. Principal manifolds for data visualization and dimension reduction. Berlin: Springer; 2008. Learning nonlinear principal manifolds by self-organising maps; pp. 68–95. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The source code used in this work for running the simulations, analyzing the results, and plotting the figures is freely distributed under the GPL-3 License and can be found at https://github.com/gdetor/som_stability.


Articles from Journal of Mathematical Neuroscience are provided here courtesy of Springer-Verlag

RESOURCES