Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2012 Jan 12;8(1):e1002348. doi: 10.1371/journal.pcbi.1002348

Maximization of Learning Speed in the Motor Cortex Due to Neuronal Redundancy

Ken Takiyama 1, Masato Okada 1,2,*
Editor: Jörn Diedrichsen3
PMCID: PMC3257280  PMID: 22253586

Abstract

Many redundancies play functional roles in motor control and motor learning. For example, kinematic and muscle redundancies contribute to stabilizing posture and impedance control, respectively. Another redundancy is the number of neurons themselves; there are overwhelmingly more neurons than muscles, and many combinations of neural activation can generate identical muscle activity. The functional roles of this neuronal redundancy remains unknown. Analysis of a redundant neural network model makes it possible to investigate these functional roles while varying the number of model neurons and holding constant the number of output units. Our analysis reveals that learning speed reaches its maximum value if and only if the model includes sufficient neuronal redundancy. This analytical result does not depend on whether the distribution of the preferred direction is uniform or a skewed bimodal, both of which have been reported in neurophysiological studies. Neuronal redundancy maximizes learning speed, even if the neural network model includes recurrent connections, a nonlinear activation function, or nonlinear muscle units. Furthermore, our results do not rely on the shape of the generalization function. The results of this study suggest that one of the functional roles of neuronal redundancy is to maximize learning speed.

Author Summary

There are overwhelmingly more neurons than muscles in the motor system. The functional roles of this neuronal redundancy remains unknown. Our analysis, which uses a redundant neural network model, reveals that learning speed reaches its maximum value if and only if the model includes sufficient neuronal redundancy. This result does not depend on whether the distribution of the preferred direction is uniform or a skewed bimodal, both of which have been reported in neurophysiological studies. We have confirmed that our results are consistent, regardless of whether the model includes recurrent connections, a nonlinear activation function, or nonlinear muscle units. Additionally, our results are the same when using either a broad or a narrow generalization function. These results suggest that one of the functional roles of neuronal redundancy is to maximize learning speed.

Introduction

In the human brain, numerous neurons encode information about external stimuli, e.g., visual or auditory stimuli, and internal stimuli, e.g., attention or motor planning. Each neuron exhibits different responses to stimuli, but neural encoding, especially in the visual and auditory cortices, can be explained by the maximization of stimulus information [1][3]. This maximization framework can also explain learning that occurs when the same stimuli are repeatedly presented; previous neurophysiological experiments have suggested that perceptual learning causes changes in neural encoding to enhance the Fisher information of a visual stimulus [4]. However, a recent study has suggested that information maximization alone is insufficient to explain neural encoding. Salinas has suggested that “how encoded information is used” needs to be taken into account: neural encoding is influenced by the downstream circuits and output units to which neurons project, and it is ultimately influenced by animal behavior [5]. In the motor cortex, neural encoding is influenced by the characteristics of muscles (output units) because motor cortex neurons send motor commands to muscles through the spinal cord. In adaptation experiments, some motor cortex neurons exhibit rotations in their preferred directions (PDs), and these rotations result in a population vector that is directed toward a planned target [6]. Neural encoding therefore changes to minimize errors between planning and behavior, suggesting that neural encoding is influenced by behavior and properties of output units.

A critical problem exists in the relationship between motor cortex neurons and output units: the neuronal redundancy problem, or overcompleteness, which refers to the fact that the number of motor cortex neurons far exceeds the number of output units. Many different combinations of neural activities can therefore generate identical outputs. Neurophysiological and computational studies have revealed that the motor cortex exhibits neuronal redundancy [7], [8]. However, it remains unknown how neuronal redundancy influences neural encoding. In other words, we do not yet understand the functional roles of neuronal redundancy in motor control and learning, though other types of redundancies are known to play various functional roles [9].

One of these types of redundancy is muscle redundancy: many combinations of muscle activities can generate identical movements. The functional roles of this muscle redundancy include impedance control to achieve accurate movements [10], reduction of motor variance by constructing muscle synergies [11], and learning internal models by changing muscle activities [12]. Another redundancy is kinematic redundancy: many combinations of joint angles result in identical hand positions. This redundancy ensures the stability of posture even if one joint is perturbed [13], and it facilitates of motor learning by increasing motor variance in a dimension irrelevant to the desired movements [14]. Redundancies therefore play important functional roles in motor control and learning.

Similar to the muscle and kinematic redundancies, neuronal redundancy likely has functional roles in motor control and learning. However, the functional roles of this redundancy are unclear. Here, using a redundant neural network, we investigate these functional roles by varying the number of model neurons while holding the number of output units constant. This manipulation allows us to control the degree of neuronal redundancy because, if a neural network includes a large number of neurons and a small number of output units, many different combinations of neural activities can generate identical outputs. It should be noted that we used a redundant neural network model that can explain neurophysiological motor cortex data [7]. The key conclusion arising from our study is that one of the functional roles of neuronal redundancy is the maximization of learning speed.

Initially, a linear model with a fixed decoder was used. Analytical calculations revealed that neuronal redundancy is a necessary and sufficient condition to maximize learning speed. This maximization is invariant whether the distribution of PDs is unimodal [6] or bimodal [15][17]; both distributions have been reported in neurophysiological investigations. Second, numerical simulations confirmed the invariance of our results, even when the neural network included an adaptable decoder, a nonlinear activation function, recurrent connections, or nonlinear muscle units. Third, we show that our results do not depend on learning rules by using weight and node perturbation, both of which are representative stochastic gradient methods [18]. Finally, we demonstrate that our hypothesis does not depend on the shape of the generalization function which shape depends on the task (broad or sharp in force field [19], [20] or visuomotor rotation adaptation [21], respectively). Our results strongly support our hypothesis that neuronal redundancy maximizes learning speed.

Results

Neuronal redundancy is defined as the dimensional gap between the number of neurons Inline graphic and the number of outputs Inline graphic. It is synonymous with overcompleteness [22]: many combinations of neural activities Inline graphic can generate identical outputs Inline graphic through a decoder Inline graphic (Inline graphic) because there are more neurons than necessary, i.e., Inline graphic (Figure 1). It should be noted that neuronal redundancy is defined not by Inline graphic but by the relationship between Inline graphic and Inline graphic. In most parts of this study, the number of constrained tasks Inline graphic is the same as Inline graphic and is set to two, i.e., Inline graphic, so there is neuronal redundancy if Inline graphic. Thus, throughout this paper, the extent of neuronal redundancy can be expressed simply using the number of neurons. In this study, we can change only the neuronal redundancy; Inline graphic can be increased while Inline graphic is held constant at two, enabling the investigation of the functional roles of neuronal redundancy. In the Importance of Neuronal Redundancy section, we distinguish the effects of neuronal redundancy from the effects of neuron number by varying both Inline graphic and Inline graphic.

Figure 1. Graphical model of a redundant neural network.

Figure 1

In this study, we discuss the relationship between neuronal redundancy and learning speed by assuming adaptation to either a visuomotor rotation or a force field. These tasks are simulated by using a rotational perturbation Inline graphic where Inline graphic is the rotational angle. Due to this perturbation, if an error occurs between target position Inline graphic and output (motor command) Inline graphic in the Inline graphicth trial, neural activities Inline graphic are modified to minimize the error, where Inline graphic is the angle of the Inline graphicth target which is radially and equally distributed (Inline graphic, Inline graphic, Inline graphic). To model the learning process in the motor cortex, we used a linear rate model, which can reproduce neurophysiological data [7] and be easily analyzed. In this model, Inline graphic is given by a weighted average of Inline graphic, and each component of Inline graphic is accordingly set to Inline graphic, i.e., Inline graphicth component of Inline graphic is defined as Inline graphic, where Inline graphic is a variable that is independent of Inline graphic. Because of this assumption, the learning rate is set to Inline graphic such that the trial-to-trial variation of Inline graphic do not depend on Inline graphic (Inline graphic), but the optimized learning rate Inline graphic is Inline graphic (see Text S1), i.e., Inline graphic, suggesting that we consider the quasi-optimal learning rate in this study. It should be noted that, because the following results do not depend on Inline graphic, our results hold when the optimal learning rate is used. Furthermore, even when each component of Inline graphic is Inline graphic, the following results are invariant if we set the learning rate to its optimal value (see Text S1). Our study shows that neuronal redundancy is necessary and sufficient to maximize learning speed.

Neuronal redundancy maximizes learning speed

Fixed homogeneous decoder

In the case of a fixed decoder, Inline graphic, the Inline graphicth neuron has uniform force amplitude (FA) (Inline graphic) and force direction (FD), Inline graphic, which is randomly sampled from a uniform distribution. Because of its uniformity, we refer to this decoder as a fixed homogeneous decoder. This model corresponds to the one proposed by Rokni et al. [7].

In this case, the squared error can be calculated recursively as

graphic file with name pcbi.1002348.e053.jpg (1)

where Inline graphic. Here, we assume that a single target is repeatedly presented for simplicity (general case is discussed in the Methods section), Inline graphic is the identity matrix, Inline graphic, Inline graphic is the learning rate, and neural activity Inline graphic is updated as

graphic file with name pcbi.1002348.e059.jpg (2)

for the Inline graphicth trial to minimize the squared error. Multiplication by Inline graphic in equation (2) is included for the purpose of scaling; it ensures that the amount of trial-to-trial variation in Inline graphic does not explicitly depend on Inline graphic. Equation (1) can thus be simplified as

graphic file with name pcbi.1002348.e064.jpg (3)

where the diagonal elements of Inline graphic, Inline graphic and Inline graphic, are eigenvalues of Inline graphic, Inline graphic is decomposed as Inline graphic (Inline graphic), Inline graphic, and learning speed is therefore determined based on the eigenvalues of

graphic file with name pcbi.1002348.e073.jpg (4)

each component of which is Inline graphic. The larger Inline graphic becomes, the faster learning becomes (Inline graphic). It should be noted that learning speed and Inline graphic do not explicitly depend on Inline graphic.

Analytical calculations can yield necessary and sufficient conditions to maximize learning speed (see the Methods section). The following self-averaging properties [23] maximize learning speed or maximize the minimum eigenvalue of Inline graphic:

graphic file with name pcbi.1002348.e080.jpg (5)
graphic file with name pcbi.1002348.e081.jpg (6)

and

graphic file with name pcbi.1002348.e082.jpg (7)

where Inline graphic is the probability distribution in which FDs are randomly sampled. It remains unknown what kind of conditions can satisfy the self-averaging properties. The self-averaging properties are satisfied if and only if the neural network model includes sufficient neuronal redundancy. In other words, learning speed is maximized if and only if Inline graphic. If the neural network includes neuronal redundancy, the self-averaging properties exist. Conversely, if the self-averaging properties exist, the neural network model should include sufficient neuronal redundancy because Monte Carlo integration shows a fluctuation of Inline graphic [24]. Thus, in the case of a fixed homogeneous decoder, neuronal redundancy plays a functional role in maximizing learning speed.

We numerically confirmed the above analytical results. Figures 2A and 2B show the learning speed and learning curves calculated using the results of 1,000 sets of randomly sampled Inline graphic values, an identical target sequence (Inline graphic), and Inline graphic. The more neuronal redundancy grows, the faster learning speed becomes. Figure 2C shows the relationship between learning speed and neuronal redundancy. The horizontal axis denotes the number of neurons, and the vertical axis denotes the increase in learning speed. Although a saturation of the increase can be seen, greater neuronal redundancy still yields faster learning speed. Therefore, these figures support our analytical results: in the case of a fixed homogeneous decoder, neuronal redundancy maximizes learning speed.

Figure 2. Relationship between learning speed and neuronal redundancy (Inline graphic).

Figure 2

(A): Learning speed when Inline graphic, or Inline graphic. The bar graph and error bars depict sample means and standard deviations, both of which are calculated using the results of randomly sampled sets of 1000 Inline graphic values. (B): Learning curves when Inline graphic, or Inline graphic. These curves and error bars show averaged values and standard deviations of errors. (C): Relationship between learning speed and the number of model neurons when Inline graphic, or Inline graphic. The horizontal axis represents the number of neurons Inline graphic and the vertical axis represents Inline graphic, where Inline graphic is the learning speed when the number of neurons is Inline graphic. Dotted and solid lines denote the average learning speed and power functions fitted to the values, respectively.

Fixed non-homogeneous decoder

The question remains whether it is necessary for FD and FA to be distributed uniformly, so we assume that the values Inline graphic are randomly sampled from the probability distribution Inline graphic to make FD and FA non-homogeneous, i.e., FDs are non-uniformly distributed, and FAs are different for each neuron. In the case of a non-homogeneous decoder, the necessary and sufficient conditions to maximize learning speed are also the following self-averaging properties:

graphic file with name pcbi.1002348.e103.jpg (8)

and

graphic file with name pcbi.1002348.e104.jpg (9)

where Inline graphic and Inline graphic are marginalized distributions. Figures 3A and 3D show distributions of Inline graphic that satisfy equations (8) and (9). Inline graphic is randomly sampled from unimodal and bimodal Gaussian distributions in Figures 3A and 3D, respectively. Because these figures show the non-uniformity in both FD and FA, neuronal redundancy maximizes learning speed regardless of these non-uniformities.

Figure 3. Network properties when Inline graphic satisfies equations (8) and (9).

Figure 3

(A): Scatter plot of Inline graphic when Inline graphic and Inline graphic are randomly sampled from a unimodal Gaussian distribution (Inline graphic). (B), (C): Histogram of preferred direction and modulation depth when Inline graphic is randomly sampled as shown in (A). (D): Scatter plot of Inline graphic when Inline graphic are randomly sampled from a bimodal Gaussian distribution. (E), (F): Histograms of preferred direction and modulation depth when Inline graphic is randomly sampled as shown in (D).

Distribution of preferred directions

Some neurophysiological studies have suggested that the distribution of PD is a skewed bimodal [15][17], but other neurophysiological studies have suggested that the distribution of PD is uniform [6]. We investigated whether our results were consistent with the results of these neurophysiological studies. Figures 3B and 3E depict the distribution of preferred directions (PDs) that results when Inline graphic is randomly sampled as shown in Figures 3A and 3D, respectively, with PDs calculated as Inline graphic (see the Methods section). Figures 3B and 3E show that both a skewed bimodal distribution and a uniform distribution can be observed when Inline graphic satisfies equations (8) and (9), suggesting that our hypothesis is consistent with the results of previous neurophysiological experiments.

Figures 3C and 3F show the distribution of modulation depth, which is calculated as Inline graphic (see the Methods section). Our results suggest that the distribution of modulation depth is skewed.

Adaptable decoder

We have analytically elucidated the relevance of neuronal redundancy to learning speed only when Inline graphic is fixed, but the question remains of whether neuronal redundancy can maximize learning speed even when Inline graphic is adaptable. In this case, it is analytically intractable to calculate learning speed, so we used numerical simulations. Figure 4A shows the learning speed when Inline graphic, or Inline graphic in the case of an adaptable decoder. Although there was no significant difference in learning speed between the cases in which Inline graphic and Inline graphic, neuronal redundancy maximized learning speed even if the decoder was adaptable. Figure 4B, which shows the learning curve when Inline graphic, or Inline graphic, also supports the maximization.

Figure 4. Relationship between learning speed and neuronal redundancy when the decoder is adaptable (Inline graphic).

Figure 4

(A): Bar graphs and error bars depict sample means and standard deviations both of which are calculated using the results from 1000 sets of Inline graphic values. (B): Learning curves when Inline graphic, or Inline graphic. These curves and error bars show averaged values and the standard deviations of the errors.

Importance of neuronal redundancy

Although we have revealed that neuronal redundancy maximizes learning speed when Inline graphic, it is important to verify that the effect is caused by the neuronal redundancy, i.e., the dimensional gap between Inline graphic and Inline graphic, and not simply the number of neurons Inline graphic. In this section, we investigate this question by varying both Inline graphic and Inline graphic while assuming that each component of Inline graphic is randomly sampled from a Gaussian distribution.

Figures 5A and 5B show the learning speed and the learning curve produced when Inline graphic, and Inline graphic with a fixed non-homogeneous decoder. If Inline graphic alone were important for maximizing learning speed, learning speed would be faster when Inline graphic than when Inline graphic or Inline graphic. However, the results shown in these figures support the opposite conclusion, i.e., learning speed becomes slower when Inline graphic compared to the other cases. This result suggests that the number of neurons alone is not important for maximizing learning speed.

Figure 5. Importance of neuronal redundancy (Inline graphic).

Figure 5

(A): Learning speed when Inline graphic, Inline graphic, or Inline graphic, where N and T are the number of neurons and constrained tasks, respectively. The bar graphs and error bars depict the sample means and standard deviations, both of which are calculated using the results of 1000 sets of Inline graphic values. (B): Learning curves when Inline graphic, Inline graphic, or Inline graphic. These curves and error bars show the average values and the standard deviations of the errors. (C): Learning speed when Inline graphic, or Inline graphic, and Inline graphic. The bar graphs and error bars depict the sample means and the standard deviations, both of which are calculated using the results of 1000 sets of Inline graphic values. (D): Learning curves when Inline graphic, and Inline graphic. These curves and error bars show the average values and the standard deviations of the errors. (E): Learning curves calculated when Inline graphic, Inline graphic, or Inline graphic and decoder Inline graphic is adaptable. (F): Learning curves calculated when Inline graphic, or Inline graphic; Inline graphic; and the decoder Inline graphic is adaptable.

Figures 5C and 5D show the learning speed and learning curve produced when Inline graphic, or Inline graphic with Inline graphic and a fixed non-homogeneous decoder. If neuronal redundancy were important, the learning speed would be faster when Inline graphic than when Inline graphic or Inline graphic. These figures support this hypothesis; learning speed increased when Inline graphic compared to the other cases. Taken together, these results indicate that the important factor for maximizing learning speed is in fact neuronal redundancy and not simply the number of neurons.

In addition, we investigated whether neuronal redundancy or neuron number is important when Inline graphic is adaptable. In this case, we only show learning curves because learning speed cannot be exponentially fitted, which makes it impossible to calculate learning speed. Figures 5E and 5F show the learning curves calculated when Inline graphic, or Inline graphic and Inline graphic, or Inline graphic with Inline graphic. These figures show the same results as the case when Inline graphic is fixed; even when Inline graphic is adaptable, the important factor for maximizing learning speed is neuronal redundancy, not simply the number of neurons.

Generality of our results

The generality of our results should be investigated because we analyzed only linear and feed-forward networks, but neurophysiological experiments have suggested the existence of recurrent connections [25] and nonlinear neural activation functions [26]. Also, only a linear rotational perturbation task was considered, so we need to investigate whether our results hold when the constrained tasks are nonlinear because, in fact, motor cortex neurons solve nonlinear tasks. The neurons send motor commands and control muscles whose activities are nonlinearly determined: muscles can pull but cannot push. Using numerical simulations, we show that neuronal redundancy maximizes learning speed, even when the neural network includes recurrent connections (Figure S1), when it includes nonlinear activation functions (Figure S2), and when the task is nonlinear (Figure S3).

In addition, we used only deterministic gradient descent, so the generality regarding the learning rule needs to be investigated. In fact, previous studies have suggested that stochastic gradient methods are more biologically relevant than deterministic ones [27], [28]. Analytical and numerical calculations confirm that our results are invariant even when the learning rule is stochastic (Figure S4). Our results therefore have strong generality.

Activity noise and plasticity noise

Although our results have strong generality, there is still an open question regarding the robustness of noise: does neuronal redundancy maximize learning speed even in the presence of neural noise? Actually, neural activities show trial-to-trial variation [29], and the neural plasticity mechanism also includes trial-to-trial fluctuations [7]. This section investigates the relationships between neuronal redundancy, learning speed, and neural noise.

Figures 6A and 6D show the variance of learning curves when Inline graphic and Inline graphic, respectively, with Inline graphic, or Inline graphic and Inline graphic and Inline graphic representing the standard deviations of activity noise and plasticity noise, respectively. The definition of the variance is Inline graphic, which is a measure of the stability of learning. Examples of learning curves are shown in Figures 6B, 6C, 6E, and 6F. These figures show that neuronal redundancy enhances the stability of learning by eliminating the influences of activity and plasticity noise. Neuronal redundancy therefore not only maximizes learning speed but also facilitates robustness in response to neural noise.

Figure 6. Relationship between neuronal redundancy and neural noise (Inline graphic).

Figure 6

(A): Variance of the learning curve when Inline graphic, or Inline graphic and Inline graphic. The bar graphs show the average values of randomly sampled sets of 1000 Inline graphic values. (B): Learning curves calculated when Inline graphic, or Inline graphic, and Inline graphic. These curves and error bars show the average values and the standard deviations of the errors. (C): Learning curves calculated when Inline graphic, or Inline graphic, and Inline graphic. (D): Variance of the learning curve when Inline graphic, or Inline graphic and Inline graphic. (E): Learning curves when Inline graphic, or Inline graphic, and Inline graphic. (F): Learning curves calculated when Inline graphic, or Inline graphic, and Inline graphic.

Shape of the generalization function

In many situations, learning in one context is generalized to different contexts, such as different postures [30], different arms [31], and different movement directions [19][21], with the degree of generalization depending on the task. In this study, we define the generalization function as the degree of generalization to different movement directions. The performance of reaching towards Inline graphic is generalized to that of reaching towards Inline graphic, and the degree of this generalization is determined by the generalization function Inline graphic. In visuomotor rotation adaptation, the generalization function is narrow in the direction metric [21]. In contrast, the generalization function is broad in force field adaptation [19], [20]. To investigate the generality of our results with respect to various kinds of tasks, it is necessary to investigate the relationships between neuronal redundancy, learning speed, and the shape of the generalization function.

Figure 7 shows the relationship between the shape of the generalization function and learning speed. Figures 7A and 7B show the learning speed and learning curve calculated when the generalization function is broad (Figure 7C). Figures 7D and 7E show the learning speed and learning curve calculated when the generalization function is narrow (Figure 7F). Although these figures show that narrower generalization results in a slower learning speed, neuronal redundancy maximizes learning speed independently of the shape of the generalization function.

Figure 7. Relationship between neuronal redundancy, learning speed, and the shape of the generalization function (Inline graphic).

Figure 7

(A): Learning speed when Inline graphic, or Inline graphic, and Inline graphic. The bar graphs and error bars depict sample means and standard deviations, both of which are calculated using the results of randomly sampled sets of 1000 Inline graphic values in the case of a broad generalization function. (B): Learning curves calculated when Inline graphic, and Inline graphic. These curves and error bars show the average values and standard deviations of the errors. (C): The generalization function with Inline graphic. (D): The learning speed when Inline graphic, or Inline graphic, and Inline graphic. Bar graphs and error bars depict the sample means and standard deviations when the generalization function is narrow (Inline graphic). (E): Learning curves calculated when Inline graphic, and Inline graphic. (F): The generalization function with Inline graphic.

Discussion

We have quantitatively demonstrated that neuronal redundancy maximizes learning speed. The larger the dimensional gap grows between the number of neurons and the number of constrained tasks, the faster learning speed becomes. This maximization does not depend on whether the PD distribution is unimodal or bimodal, the decoder is fixed or adaptable, the network is linear or nonlinear, the task is linear or nonlinear, or the learning rule is stochastic or non-stochastic. Additionally, we have shown that neuronal redundancy has another important functional role: it provides robustness in response to neural noise. Furthermore, neuronal redundancy maximizes learning speed in a manner independent of the shape of the generalization function. These results strongly support the generality of our results.

Neuronal redundancy maximizes learning speed because only Inline graphic equalities, Inline graphic, need to be satisfied, and Inline graphic-dimensional neural activity Inline graphic is adaptable (Inline graphic). This dimensional gap yields the large Inline graphic dimensional subspace of Inline graphic in which the Inline graphic equalities are satisfied. The more Inline graphic increases, the greater the fraction of the subspace becomes: Inline graphic. Neuronal redundancy, rather than the number of neurons, thus enables Inline graphic to rapidly reach a single point in the subspace. This interpretation likely applies even in the cases of an adaptable decoder, recurrent connections, a nonlinear network, a nonlinear task, and a stochastic learning rule. Furthermore, this interpretation is supported by the results shown in Figure 5; the bigger Inline graphic grows, the faster learning speed becomes.

At first glance, our results may seem inconsistent with the results of Werfel et al. [18], who concluded that learning speed is inversely proportional to Inline graphic. In their model, because they considered the single-layer linear model, Inline graphic is the same as the number of input units, which is defined as Inline graphic( = Inline graphic) in the present study. A similar tendency can be observed in Figure 5; the more Inline graphic increases, the slower learning speed becomes. We calculated the optimal learning rate and speed as shown in Text S1, and confirmed that learning speed is inversely proportional to Inline graphic. Thus, our results are consistent with Werfel's study and additionally suggest that neuronal redundancy maximizes learning speed.

Neuronal redundancy plays another important role: generating robustness in response to neural noise (Figure 6). Because neuronal redundancy has the same meaning as overcompleteness, its functional role is the same as the robustness of overcompleteness in the face of perturbations in signals [32]. This additional functional role further supports our hypothesis that neuronal redundancy is a special neural basis on which to maximize learning speed. For example, if we increase the learning rate Inline graphic in a non-redundant network, the learning speed approaches the maximal speed in a redundant network in which the learning rate is fixed to Inline graphic. As shown in Figure 6, however, a non-redundant network is not robust with respect to neural noise. Furthermore, neuronal redundancy minimizes residual errors when the neural network includes synaptic decay [7] (see the Methods section and Figure S5). Thus, neuronal redundancy represents a special neural basis for maximizing learning speed while minimizing residual error and maintaining robustness in response to neural noise.

Methods

Model definition

Our study assumed the following task: participants move their arms towards one of Inline graphic radially distributed targets. If the Inline graphicth target is presented in the Inline graphicth trial, the neural network model receives the input Inline graphic (Inline graphic, Inline graphic), where Inline graphic. The input units project to neurons (hidden units), the activities of which are determined by

graphic file with name pcbi.1002348.e257.jpg (10)

where Inline graphic is synaptic weight in the Inline graphicth trial, Inline graphic is the standard deviation of neural activity noise, Inline graphic denotes independent normal Gaussian random variables, and Inline graphic is the number of neurons (Figure 1). The Inline graphicth neuron has a PD given by Inline graphic and a modulation depth Inline graphic, where Inline graphic, this cosine tuning having been reported by many neurophysiological studies.

The neural population generates a force of Inline graphic through a decoder matrix Inline graphic:

graphic file with name pcbi.1002348.e269.jpg (11)

where Inline graphic is the number of outputs, which, in most cases, is set to 2. When Inline graphic is fixed and homogeneous, the Inline graphicth and Inline graphicth components of Inline graphic are defined as Inline graphic and Inline graphic, respectively, where division by Inline graphic is used for scaling and FD Inline graphic is randomly sampled from a uniform distribution (Inline graphic). When Inline graphic is fixed and non-homogeneous, Inline graphic is randomly sampled from a probability distribution Inline graphic and divided by Inline graphic. As a result, the neural network generates a final hand coordinate Inline graphic:

graphic file with name pcbi.1002348.e285.jpg (12)

which means that Inline graphic is perturbed by a rotation Inline graphic which assumes a visuomotor rotation or curl force field. Rotational perturbations are assumed because many behavioral studies have used them. Because we discuss only the endpoint of the movement, we refer to Inline graphic as the motor command. The constrained tasks are those that the neural network generates Inline graphic toward Inline graphic, i.e., Inline graphic, which means the number of constrained tasks Inline graphic is the same as Inline graphic. We used Inline graphic instead of Inline graphic in the following sections.

If the error occurs between Inline graphic and Inline graphic, synaptic weights Inline graphic are adapted to reduce the squared error, which is defined as Inline graphic, based on a gradient descent method

graphic file with name pcbi.1002348.e300.jpg (13)

where Inline graphic is the synaptic decay rate, Inline graphic is the learning rate (Inline graphic is set to 0.2 in most parts of the present study), Inline graphic is the strength of synaptic drift, and Inline graphic denotes normal Gaussian random variables. Since each component of Inline graphic is Inline graphic, multiplying Inline graphic by Inline graphic allows trial-by-trial variation of both Inline graphic and Inline graphic to be Inline graphic. As shown in Text S1, the optimal learning rate Inline graphic is Inline graphic (Inline graphic), suggesting that we consider a quasi-optimal learning rate. It should be noted that our results hold whether the learning rate is optimal or quasi-optimal because the results do not depend on Inline graphic. It should also be noted that the amount of variation in Inline graphic does not explicitly depend on Inline graphic.

Learning curve

Equation (13) yields the following update rule of squared error:

graphic file with name pcbi.1002348.e319.jpg (14)

where Inline graphic, and Inline graphic denotes the identity matrix. At first, we assume a case in which Inline graphic for simplicity. Because Inline graphic is symmetric, Inline graphic can be decomposed as Inline graphic, where each row of Inline graphic is one of the eigenvectors (Inline graphic) and each diagonal component of a diagonal matrix Inline graphic is one of the eigenvalues of Inline graphic. This decomposition transforms equation (14) into the simple form

graphic file with name pcbi.1002348.e330.jpg (15)

where Inline graphic and Inline graphic. This recurrence formula yields the analytical form of the learning curve:

graphic file with name pcbi.1002348.e333.jpg (16)

Equation (16) requires that the larger the eigenvalues become, the faster the learning speed becomes and the smaller the residual error becomes (Figure S5). Because

graphic file with name pcbi.1002348.e334.jpg (17)

whose component is Inline graphic, simple algebra gives the analytical form of the eigenvalues,

graphic file with name pcbi.1002348.e336.jpg (18)

which are also Inline graphic, suggesting that learning speed does not depend explicitly on Inline graphic. Because the learning speed is determined by the smaller eigenvalue, the necessary and sufficient conditions to maximize learning speed, or to maximize the smaller eigenvalue, are

graphic file with name pcbi.1002348.e339.jpg (19)

and

graphic file with name pcbi.1002348.e340.jpg (20)

What kind of conditions can simultaneously satisfy equations (19) and (20)? The only answer is sufficient neuronal redundancy, i.e., Inline graphic, because sufficient neuronal redundancy enables self-averaging properties to exist in a neural network, i.e.,

graphic file with name pcbi.1002348.e342.jpg (21)
graphic file with name pcbi.1002348.e343.jpg (22)

and

graphic file with name pcbi.1002348.e344.jpg (23)

where Inline graphic is the probability distribution in which FDs are randomly sampled. Conversely, if equations (21), (22), and (23) are satisfied in all of the sets of randomly sampled FDs, the number of neurons needs to satisfy Inline graphic because the fluctuation of Monte Carlo integrals is Inline graphic [24]. Therefore, to maximize learning speed, the necessary and sufficient condition is sufficient neuronal redundancy.

The above analytical calculations hold even when Inline graphic. Equation (13) yields the recurrence equation of the squared error:

graphic file with name pcbi.1002348.e349.jpg (24)

where Inline graphic is set to 1 for simplicity. Using Inline graphic, this equation can be written as

graphic file with name pcbi.1002348.e352.jpg (25)

The larger the eigenvalue becomes, the faster learning speed becomes if Inline graphic and Inline graphic have the same sign, or if Inline graphic. This inequality is appropriate if the equality Inline graphic can be proved, where Inline graphic is a positive constant. To prove this equality, let us assume that in the 1st trial after the rotational perturbation Inline graphic is applied, output can be written as Inline graphic because the neural network can generate accurate outputs if there is no perturbation. In this case,

graphic file with name pcbi.1002348.e360.jpg (26)

where Inline graphic is a positive constant. Thus, the larger Inline graphic becomes, the faster learning speed becomes even when Inline graphic; analytical calculations show that neuronal redundancy maximizes learning speed even when Inline graphic.

Fixed non-homogeneous decoder

When Inline graphic is fixed and non-homogeneous, i.e., Inline graphic, Inline graphic, Inline graphic, Inline graphic, and Inline graphic, the necessary and sufficient conditions for maximizing learning speed are given by the following equations:

graphic file with name pcbi.1002348.e371.jpg (27)
graphic file with name pcbi.1002348.e372.jpg (28)

with neuronal redundancy assumed. Equations (27) and (28) can be satisfied when, for example,

graphic file with name pcbi.1002348.e373.jpg (29)

(shown in Figure 3A with Inline graphic and Inline graphic), or

graphic file with name pcbi.1002348.e376.jpg (30)

(shown in Figure 3D with Inline graphic, Inline graphic, Inline graphic and Inline graphic), where Inline graphic. Because the learning rate of motor commands is determined by Inline graphic (see the following section), Inline graphic is determined based on the results of behavioral studies [33]. We cannot analytically calculate the general class of Inline graphic and the distributions of PDs satisfying equations (27) and (28), but broad classes of those distributions can satisfy these equations because the classes include even asymmetric distributions, e.g., when Inline graphic, Inline graphic.

Learning rule of decoder Inline graphic

When Inline graphic is adaptable, this is also adapted to minimize the squared error:

graphic file with name pcbi.1002348.e389.jpg (31)

where Inline graphic is set to Inline graphic, Inline graphic is a normal Gaussian random variable, and Inline graphic is set to 0.1 in the Adaptable Decoder section and 0.05 in the Importance of Neuronal Redundancy section. This learning rule corresponds to back-propagation [34].

High dimensional tasks

In the Importance of Neuronal Redundancy section, the neural network generates the output Inline graphic, which is determined by

graphic file with name pcbi.1002348.e395.jpg (32)

for the Inline graphicth trial. An initial value of Inline graphic is randomly sampled from the normal Gaussian distribution and divided by Inline graphic for scaling. The input Inline graphic is randomly sampled from the normal Gaussian distribution and is normalized to satisfy Inline graphic to avoid the effect of this value on learning speed. In addition, we used a fixed value of Inline graphic because the generalization function (see the following section) strongly depends on Inline graphic, i.e., Inline graphic. It should be noted that learning speed does not explicitly depend on Inline graphic because learning speed is determined only by the minimum eigenvalue of Inline graphic.

The generalization function and the update rule for motor commands

Equation (13) yields the following update rule for motor commands:

graphic file with name pcbi.1002348.e406.jpg (33)

If equations (27) and (28) (or (22) and (23)) are satisfied, equation (33) can be written as

graphic file with name pcbi.1002348.e407.jpg (34)

where the cross term of Inline graphic and Inline graphic determines the generalization function Inline graphic, e.g., Inline graphic, if we define Inline graphic. We set Inline graphic and Inline graphic to satisfy Inline graphic. It should be noted that equation (34) corresponds to a model for sensorimotor learning that can explain the results of behavioral experiments [35], suggesting that our hypothesis is consistent with the results of behavioral experiments.

Because the shape of the generalization function depends on the task, we need to confirm the generality of our results with regard to the shape of the generalization function. To simulate various shapes of generalization functions, we used the von-Mises function

graphic file with name pcbi.1002348.e416.jpg (35)

where Inline graphic, Inline graphic, and Inline graphic are the precision parameter, the preferred direction of the Inline graphicth input unit, and the number of input units, respectively. The normalization factor Inline graphic is determined to make Inline graphic to avoid the influence of this value on the learning speed, where Inline graphic. This normalization permits us to investigate the influence of the shape of the generalization function alone on learning speed. The larger the value of Inline graphic, the sharper the shape of the generalization function becomes. We set Inline graphic to 100 throughout this study.

Numerical simulation procedure

We conducted 100 baseline trials with Inline graphic and Inline graphic to identify the baseline values of Inline graphic. The initial value of Inline graphic, Inline graphic, was set to Inline graphic. After these trials, 100 learning trials were conducted using Inline graphic and Inline graphic. Learning speed Inline graphic was calculated by fitting the exponential function Inline graphic to Inline graphic. All the figures denote Inline graphic which was obtained only in learning trials. The present study calculated learning speed and learning curves by averaging the results of 1000 sets of baseline and learning trials, each set including an identical target sequence that was randomly sampled, and each set using different FD values.

For all of the statistical tests, we used the Wilcoxon sign rank test. It should be noted that the Inline graphic-value was indicated only if the value was significantly different from 0; no statistically significant differences were detected.

Supporting Information

Figure S1

Relationship between learning speed, neuronal redundancy, and adaptable recurrent connections ( Inline graphic ). (A): Learning speed when Inline graphic and Inline graphic. The whiter the color, the faster the learning speed. (B): Learning curves obtained when N = 10, 50, or 100 and Inline graphic. These curves show the average values of 1,000 randomly sampled sets of Inline graphic. Error bars represent the standard deviations of the errors. (C): Learning curves obtained when Inline graphic and Inline graphic. These curves and error bars show average values and standard deviations. (D): Variance of the learning curve when Inline graphic and Inline graphic (Inline graphic). These variances are average values from 1,000 randomly sampled sets of Inline graphic.

(EPS)

Figure S2

Relationship between learning speed and neuronal redundancy in the case of a nonlinear neural network ( Inline graphic ). (A): Learning speed when N = 10, 50, 100, and 1000. The bar graphs and error bars depict sample means and standard deviations, both of which are calculated using the results of 1,000 randomly sampled sets of Inline graphic values. (B): Learning curves obtained when Inline graphic, or Inline graphic. These curves and error bars show average values and the standard deviations of the errors.

(EPS)

Figure S3

Relationship between learning speed and neuronal redundancy when the neural network includes nonlinear muscle units ( Inline graphic ). (A): The bar graphs and error bars depict sample means and standard deviations, both of which were calculated using the results of 1,000 randomly sampled sets of Inline graphic values. (B): Learning curves obtained when Inline graphic or Inline graphic. These curves and error bars show average values and the standard deviations of the errors.

(EPS)

Figure S4

Relationship between learning speed and neuronal redundancy in the case of weight perturbation and node perturbation ( Inline graphic ). (A): Learning speed when Inline graphic, or Inline graphic, with weight perturbation as the learning rule. The bar graphs and error bars depict sample means and standard deviations, both of which are calculated using the results of 1,000 randomly sampled sets of Inline graphic. (B): Learning curves obtained when Inline graphic, or Inline graphic, with weight perturbation as the learning rule. These curves and error bars show the average values and the standard deviations of the errors. (C): Learning speed when Inline graphic, or Inline graphic, with node perturbation as the learning rule. The bar graphs and error bars depict sample means and standard deviations, both of which are calculated using the results of 1,000 randomly sampled sets of Inline graphic. (D): Learning curves obtained when Inline graphic, or Inline graphic, with node perturbation as the learning rule. These curves and error bars show average values and the standard deviations of the errors.

(EPS)

Figure S5

Relationship between residual error, learning speed, and neuronal redundancy with synaptic decay included ( Inline graphic ). (A): Residual error when Inline graphic. The bar graphs and error bars denote sample means and standard deviations, both of which are calculated using the results of 1,000 randomly sampled sets of Inline graphic values. (B): Learning speed when Inline graphic. The bar graphs and error bars depict sample means and standard deviations. (C): Learning curves obtained when Inline graphic, and Inline graphic and Inline graphic. These curves and error bars show average values and standard deviations. (D): Residual error when Inline graphic. (E): Learning speed when Inline graphic. (F): Learning curve when Inline graphic. (G): Residual error when Inline graphic. (H): Learning speed when Inline graphic. (I): Learning curve when Inline graphic.

(EPS)

Text S1

Generality of our results. This file contains the detailed descriptions of Generality of our results section.

(PDF)

Acknowledgments

We thank D. Nozaki, Y. Sakai, Y. Naruse, K. Katahira, T. Toyoizumi, and T. Omori for their helpful discussions.

Footnotes

The authors have declared that no competing interests exist.

This work was partially supported by a Grant-in-Aid for Scientific Research (A) (Grant No. 20240020) and a Grant-in-Aid for Special Purposes (Grant No. 10J04910) from the Ministry of Education, Culture, Sports, Science and Technology of Japan. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Barlow H. Redundancy reduction revisited. Network. 2001;12:241–253. [PubMed] [Google Scholar]
  • 2.Olshausen BA, Field DJ. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature. 1996;381:607–609. doi: 10.1038/381607a0. [DOI] [PubMed] [Google Scholar]
  • 3.Lewicki MS. Efficient coding of natural sounds. Nat Neurosci. 2002;5:356–363. doi: 10.1038/nn831. [DOI] [PubMed] [Google Scholar]
  • 4.Gutnisky D, Dragoi V. Adaptive coding of visual information in neural populations. Nature. 2008;452:220–224. doi: 10.1038/nature06563. [DOI] [PubMed] [Google Scholar]
  • 5.Salinas E. How behavioral constraints may determine optimal sensory representations. PLoS Biol. 2006;4:2383–2392. doi: 10.1371/journal.pbio.0040387. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Li CS, Padoa-Schioppa C, Bizzi E. Neuronal correlates of motor performance and motor learning in the primary motor cortex of monkeys adapting to an external force field. Neuron. 2001;30:593–607. doi: 10.1016/s0896-6273(01)00301-4. [DOI] [PubMed] [Google Scholar]
  • 7.Rokni U, Richardson AG, Bizzi E, Seung HS. Motor learning with unstable neural representations. Neuron. 2007;54:653–666. doi: 10.1016/j.neuron.2007.04.030. [DOI] [PubMed] [Google Scholar]
  • 8.Narayanan NS, Kimchi EY, Laubach M. Redundancy and synergy of neuronal ensembles in motor cortex. J Neurosci. 2005;25:4207–4216. doi: 10.1523/JNEUROSCI.4697-04.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Bernstein N. The coordination and regulation of movements. London: Pergamon; 1967. [Google Scholar]
  • 10.Gribble PL, Mullin LI, Cothros N, Mattar A. Role of cocontraction in arm movement accuracy. J Neurophysiol. 2003;89:2396–2405. doi: 10.1152/jn.01020.2002. [DOI] [PubMed] [Google Scholar]
  • 11.Latash ML, Scholz JP, Schoner G. Motor control strategies revealed in the structure of motor variability. Exerc Sport Sci Rev. 2002;30:26–31. doi: 10.1097/00003677-200201000-00006. [DOI] [PubMed] [Google Scholar]
  • 12.Thoroughman KA, Shadmehr R. Electromyographic correlates of learning an internal model of reaching movements. J Neurosci. 1999;19:8573–8588. doi: 10.1523/JNEUROSCI.19-19-08573.1999. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Latash ML. The organization of quick corrections within a two-joint synergy in conditions of unexpected blocking and release of a fast movement. Clin Neurophysiol. 2000;11:975–987. doi: 10.1016/s1388-2457(00)00263-7. [DOI] [PubMed] [Google Scholar]
  • 14.Yang JF, Scholz JP, Latash ML. The role of kinematic redundancy in adaptation of reaching. Exp Brain Res. 2007;176:54–69. doi: 10.1007/s00221-006-0602-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Scott SH, Gribble PL, Cabel DW. Dissociation between hand motion and population vectors from neural activity in motor cortex. Nature. 2001;413:161–165. doi: 10.1038/35093102. [DOI] [PubMed] [Google Scholar]
  • 16.Kurtzer I, Pruszynski JA, Herter TM, Scott SH. Nonuniform distribution of reach-related and torque-related activity in upper arm muscles and neurons of primary motor cortex. J Neurophysiol. 2006;96:3220–3230. doi: 10.1152/jn.00110.2006. [DOI] [PubMed] [Google Scholar]
  • 17.Naselaris T, Merchant H, Amirikian B, Georgopoulos AP. Large-scale organization of preferred directions in the motor cortex. I. motor cortical hyperacuity for forward reaching. J Neurophysiol. 2006;96:3231–3236. doi: 10.1152/jn.00487.2006. [DOI] [PubMed] [Google Scholar]
  • 18.Werfel J, Xie X, Seung S. Learning curves for stochastic gradient descent in linear feedforward networks. Neural Compt. 2005;17:2699–2718. doi: 10.1162/089976605774320539. [DOI] [PubMed] [Google Scholar]
  • 19.Thoroughman KA, Shadmehr R. Learning of action through adaptive combination of motor primitives. Nature. 2000;407:742–747. doi: 10.1038/35037588. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Donchin O, Francis JT, Shadmehr R. Quantifying generalization from trial-by-trial behavior of adaptive systems that learn with basis functions. J Neurosci. 2003;23:9032–9045. doi: 10.1523/JNEUROSCI.23-27-09032.2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Krakauer JW, Pine ZM, Ghilardi MF, Ghez C. Learning of visuomotor transformations for vectorial planning of reaching trajectories. J Neurosci. 2000;20:8916–892. doi: 10.1523/JNEUROSCI.20-23-08916.2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Lewicki MS, Sejnowski TJ. Learning Overcomplete Representations. Neural Comput. 2000;12:337–365. doi: 10.1162/089976600300015826. [DOI] [PubMed] [Google Scholar]
  • 23.Hidetoshi N. Statistical Physics of Spin Glasses and Information Processing: An Introduction. 2001. Oxford University Press.
  • 24.Bishop CM. Pattern Recognition and Machine Learning. 2006. Springer.
  • 25.Capaday C, Ethier C, Brizzi L, Sik A, van Vreewijk C. On the nature of the intrinsic connectivity of the cat motor cortex: evidence for a recurrent neural network topology. J Neurophysiol. 2009;102:2131–2141. doi: 10.1152/jn.91319.2008. [DOI] [PubMed] [Google Scholar]
  • 26.Tsodyks M, Markram H. The neural code between neocortical pyramidal neurons depends on neurotransmitter release probability. Proc Natl Acad Sci U S A. 1997;94:719–723. doi: 10.1073/pnas.94.2.719. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Seung HS. Learning in spiking neural networks by reinforcement of stochastic synaptic transmission. Neuron. 2003;40:1063–1073. doi: 10.1016/s0896-6273(03)00761-x. [DOI] [PubMed] [Google Scholar]
  • 28.Fiete IR, Fee MS, Seung HS. Model of birdsong learning based on gradient estimation by dynamic perturbation of neural conductances. J Neurophysiol. 2007;98:2038–2057. doi: 10.1152/jn.01311.2006. [DOI] [PubMed] [Google Scholar]
  • 29.Lee D, Port NL, Kruse W, Georgopoulos AP. Variability and correlated noise in the discharge of neurons in motor and parietal areas of the primate cortex. J Neurosci. 1998;18:1161–1170. doi: 10.1523/JNEUROSCI.18-03-01161.1998. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Shadmehr R, Mussa-ivaldi FA. Adaptive representation of dynamics during learning of a motor task. J Neurosci. 1994;14:3208–3224. doi: 10.1523/JNEUROSCI.14-05-03208.1994. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Criscimagna-Hemminger SE, Donchin O, Gazzaniga MS, Shadmehr R. Learned dynamics of reaching movements generalize from dominant to nondominant arm. J Neurophysiol. 2003;89:168–176. doi: 10.1152/jn.00622.2002. [DOI] [PubMed] [Google Scholar]
  • 32.Simoncelli EP, Freeman WT, Adelson EH, Heeger DJ. Shiftable multiscale transforms. IEEE Trans Info Theory. 1992;38:587–607. [Google Scholar]
  • 33.Cheng S, Sabes PN. Calibration of visually guided reaching is driven by error-corrective learning and internal dynamics. J Neurophysiol. 2007;97:3057–3069. doi: 10.1152/jn.00897.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Rumelhart DE, Hinton GE, Williams RJ. Learning representations by backpropagating errors. Nature. 1986;323:533–536. [Google Scholar]
  • 35.van Beers RJ. Motor learning is optimally tuned to the properties of motor noise. Neuron. 2009;63:406–417. doi: 10.1016/j.neuron.2009.06.025. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Figure S1

Relationship between learning speed, neuronal redundancy, and adaptable recurrent connections ( Inline graphic ). (A): Learning speed when Inline graphic and Inline graphic. The whiter the color, the faster the learning speed. (B): Learning curves obtained when N = 10, 50, or 100 and Inline graphic. These curves show the average values of 1,000 randomly sampled sets of Inline graphic. Error bars represent the standard deviations of the errors. (C): Learning curves obtained when Inline graphic and Inline graphic. These curves and error bars show average values and standard deviations. (D): Variance of the learning curve when Inline graphic and Inline graphic (Inline graphic). These variances are average values from 1,000 randomly sampled sets of Inline graphic.

(EPS)

Figure S2

Relationship between learning speed and neuronal redundancy in the case of a nonlinear neural network ( Inline graphic ). (A): Learning speed when N = 10, 50, 100, and 1000. The bar graphs and error bars depict sample means and standard deviations, both of which are calculated using the results of 1,000 randomly sampled sets of Inline graphic values. (B): Learning curves obtained when Inline graphic, or Inline graphic. These curves and error bars show average values and the standard deviations of the errors.

(EPS)

Figure S3

Relationship between learning speed and neuronal redundancy when the neural network includes nonlinear muscle units ( Inline graphic ). (A): The bar graphs and error bars depict sample means and standard deviations, both of which were calculated using the results of 1,000 randomly sampled sets of Inline graphic values. (B): Learning curves obtained when Inline graphic or Inline graphic. These curves and error bars show average values and the standard deviations of the errors.

(EPS)

Figure S4

Relationship between learning speed and neuronal redundancy in the case of weight perturbation and node perturbation ( Inline graphic ). (A): Learning speed when Inline graphic, or Inline graphic, with weight perturbation as the learning rule. The bar graphs and error bars depict sample means and standard deviations, both of which are calculated using the results of 1,000 randomly sampled sets of Inline graphic. (B): Learning curves obtained when Inline graphic, or Inline graphic, with weight perturbation as the learning rule. These curves and error bars show the average values and the standard deviations of the errors. (C): Learning speed when Inline graphic, or Inline graphic, with node perturbation as the learning rule. The bar graphs and error bars depict sample means and standard deviations, both of which are calculated using the results of 1,000 randomly sampled sets of Inline graphic. (D): Learning curves obtained when Inline graphic, or Inline graphic, with node perturbation as the learning rule. These curves and error bars show average values and the standard deviations of the errors.

(EPS)

Figure S5

Relationship between residual error, learning speed, and neuronal redundancy with synaptic decay included ( Inline graphic ). (A): Residual error when Inline graphic. The bar graphs and error bars denote sample means and standard deviations, both of which are calculated using the results of 1,000 randomly sampled sets of Inline graphic values. (B): Learning speed when Inline graphic. The bar graphs and error bars depict sample means and standard deviations. (C): Learning curves obtained when Inline graphic, and Inline graphic and Inline graphic. These curves and error bars show average values and standard deviations. (D): Residual error when Inline graphic. (E): Learning speed when Inline graphic. (F): Learning curve when Inline graphic. (G): Residual error when Inline graphic. (H): Learning speed when Inline graphic. (I): Learning curve when Inline graphic.

(EPS)

Text S1

Generality of our results. This file contains the detailed descriptions of Generality of our results section.

(PDF)


Articles from PLoS Computational Biology are provided here courtesy of PLOS

RESOURCES