Significance
Adolescence is a period during which there are important changes in behavior and the structure of the brain. In this manuscript, we use theoretical modeling to show how improvements in working memory and reinforcement learning that occur during adolescence can be explained by the reduction in synaptic connectivity in prefrontal cortex that occurs during a similar period. We train recurrent neural networks to solve working memory and reinforcement learning tasks and show that when we prune connectivity in these networks, they perform the tasks better. The improvement in task performance, however, can come at the cost of flexibility as the pruned networks are not able to learn some new tasks as well.
Keywords: neural network, reinforcement learning, working memory, pruning
Abstract
Adolescent development is characterized by an improvement in multiple cognitive processes. While performance on cognitive operations improves during this period, the ability to learn new skills quickly, for example, a new language, decreases. During this time, there is substantial pruning of excitatory synapses in cortex and specifically in prefrontal cortex. We have trained a series of recurrent neural networks to solve a working memory task and a reinforcement learning (RL) task. Performance on both of these tasks is known to improve during adolescence. After training, we pruned the networks by removing weak synapses. Pruning was done incrementally, and the networks were retrained during pruning. We found that pruned networks trained on the working memory task were more resistant to distraction. The pruned RL networks were able to produce more accurate value estimates and also make optimal choices more consistently. Both results are consistent with developmental improvements on these tasks. Pruned networks, however, learned some, but not all, new problems more slowly. Thus, improvements in task performance can come at the cost of flexibility. Our results show that overproduction and subsequent pruning of synapses is a computationally advantageous approach to building a competent brain.
In both artificial and biological neural networks, cognitive operations are defined by patterns of synaptic connectivity in the network (1–4). During development, neurons across multiple cortical and subcortical systems first establish an abundance of synaptic contacts with downstream areas (5–7). Following this period, weak synapses are pruned through activity-dependent mechanisms (8–11). This process happens early in development for some systems, including retinal–geniculate synapses (5) and climbing fiber to Purkinje cell synapses in the cerebellum (12). However, in human prefrontal cortex, substantial pruning takes place beginning in late childhood and continuing into adulthood (13–15). The number of excitatory synapses in prefrontal cortex peaks between the ages of 5 and 10. After this it decreases exponentially. Estimates suggest that up to 40% of excitatory synapses are pruned in prefrontal cortex, between the ages of 10 and 30 (13). Studies also suggest that most of the pruning occurs on recurrent layer III connections within prefrontal cortex, as opposed to inputs from other areas (16). During this time the number of neurons remains relatively constant, having reached adult levels by about 6 mo of age (14). Similar changes in the number of prefrontal synapses during development have also been seen in monkeys (17–20) and rats (21).
Cognition and learning also change during adolescence (22–26). For example, there are consistent and systematic changes in performance on reinforcement learning (RL) tasks, in which participants have to learn from feedback to choose rewarding options and avoid punishing options (27–30). These studies often characterize performance using two variables, learning rate and decision noise (31). Learning rate is the rate at which subjects learn to select the better option, and decision noise characterizes the asymptotic level of performance. Decision noise, therefore, reflects the consistency with which subjects select the better option, after their performance has stabilized. When decision noise is high, subjects tend to sample multiple options, even late in learning. Studies most consistently report that decision noise decreases during adolescence (27), although some studies also suggest that learning rate increases (32). Similar developmental changes have also been seen in rodents (33). Although performance on well-defined bandit RL problems improves during adolescence, children are better at learning problems requiring exploration (34–37). Additionally, children can often learn new skills more quickly than adults, including languages, and this ability decreases during adolescence (9, 38). Some forms of exploration, for example, directed exploration, which requires an estimate of what is unknown, improve during adolescence (39).
Working memory performance and the ability to avoid distraction also improve during adolescence (40–44). Improvements in working memory may also underlie improvements in goal-directed behavior (45). Developmental improvements in working memory performance have also been seen in primates (46). Comparison of single-cell neurophysiology responses between pubertal and adult monkeys shows that neural responses reflecting remembered items were stronger in adults and less sensitive to distraction during delay periods (46).
Pruning of weak synapses was early recognized as an important technique to improve performance in artificial neural networks (47–50). Large artificial networks with many hidden units and connections can solve complex problems. However, these large networks generalize poorly beyond their training examples and rely on massive computational resources. Several techniques have been introduced to reduce network complexity by shrinking connection weights (51), as well as pruning or eliminating weights (47, 48), and these approaches have been shown to improve generalization performance. Pruning techniques, therefore, optimize the trade-off between network complexity and generalization.
Our goal in the present study was to understand how developmental pruning can account for changes in performance on working memory and RL tasks. To examine this, we trained two sets of recurrent neural networks. One set of networks was trained to carry out a delayed match to sample (DMS) task (46, 52, 53). The DMS task requires subjects to remember items over a delay, and therefore, it tests working memory. The second network was trained to carry out a two-armed bandit RL task (54–56). After the networks were trained, we incrementally pruned networks by removing weak connections. During pruning we continued to train the networks. We also trained a set of unpruned networks in parallel, such that they saw the same number of training examples. We then compared the performance of the pruned networks to the unpruned networks. We found that pruning led to changes in network dynamics that improved performance on both tasks. In the DMS task, the pruned network was less susceptible to distractors, and this improved performance was related to a strengthening of the attractor basin. In the RL task, the pruned network produced more accurate value estimates in the presence of noise. We also found corresponding changes in the dynamics that accounted for these improvements. The pruned networks, however, learned some new tasks and new choice options more slowly than the unpruned networks. Therefore, the increased performance on the tasks for which the networks were trained came at the cost of decreased flexibility.
Results
DMS.
We fit recurrent neural networks to two cognitive tasks, performance on which is known to improve during adolescence: a DMS task that engages working memory (52) and a two-armed bandit RL task (54). DMS tasks come in many forms. Here we trained networks on a task inspired by a left/right auditory spatial match to sample task. In the task, a cue is presented either on the left or the right of the network agent (Fig. 1). Following a delay period during which the network has to remember the location of the first cue, a second cue is presented either on the left or the right. If the two cues are on the same side (i.e., if the sign of the input matches), it is a match trial and the network should give a match response. If the two cues are on opposite sides, it is a nonmatch trial, and the network should give a nonmatch response. There are four possible conditions identified by the side of the first and second cues. After training, a fully connected recurrent neural network (Fig. 1B) does well at the task (Fig. 1 C–F). Following cues presented on the same side, the network produces a match response (Fig. 1 C and D), and following cues presented on opposite sides the network produces a nonmatch response (Fig. 1 E and F).
We examined the effect of synaptic pruning on task performance and recurrent dynamics. To examine the effect of pruning, we first trained a large ensemble (400) of fully connected networks on the task. Each of these original 400 networks was then used to generate a sequence of pruned and unpruned networks (Fig. 1G). For the networks trained with pruning, we first pruned 10% of the weakest recurrent connections (after the initial training), either positive or negative, and then retrained the networks on a new set of training examples. Each new set of training examples, which was fixed for a training episode, was five trials of each condition, with additive noise. The noise was frozen for a training episode (see Methods for further details). We then pruned an additional 5% of the weakest connections and retrained on a new set of examples. We continued this iteratively until we had pruned 95% of the connections. For the unpruned networks, we did not prune the connectivity matrix. However, we introduced a new set of training examples and retrained to criterion. We repeated this, in parallel with the pruned networks, such that each unpruned network saw the same number of training examples as a corresponding pruned network. We found that with small amounts of pruning, the pruned networks tended to have higher final loss (Fig. 1H) and required more training iterations to converge (Fig. 1I). Over an intermediate range of pruning values, however, the pruned networks had lower final training loss and required a smaller number of training iterations. At the highest pruning levels, the pruned networks again had high loss and converged slowly. The unpruned networks had consistent training loss and iterations across the multiple training episodes. When we examined the appended series of training iterations (SI Appendix, Fig. S1), it could be seen that pruning led to a large training loss on the first iteration, which quickly decreased.
We next examined the performance of an example pruned (70%) and unpruned (trained for the same number of examples) network. The pruned network had a larger proportion of strong connections than the unpruned network (Fig. 2A). This is consistent with a biological process known as homeostatic synaptic scaling, which maintains the total input to single cells (57). To begin examining the effect of pruning on the neural dynamics we carried out singular value decomposition on the recurrent connectivity matrix, A, and found that the scaling of the singular values was similar between the pruned and unpruned networks (Fig. 2B). Thus, we had not reduced the rank of the recurrent dynamics, which suggests that all of the recurrent units were participating in the task after pruning.
To compare the performance of the example pruned and unpruned networks we examined the ability of the networks to avoid distraction. Distraction was provided by delivering a probe cue during the delay period between cue 1 and cue 2. Probe cues were always delivered on the side opposite of the first cue. Therefore, if the network output was the same on probe and nonprobe trials, it was resistant to the distractor. Note that neither network was trained to ignore the probes. Thus, if the network could not ignore the probe cue, it would give the wrong answer. Both the pruned and unpruned networks gave the correct answer in unprobed trials (Fig. 2 C and E). In this example probe trial, however, the unpruned network provided the wrong response (Fig. 2D). The pruned network, however, was insensitive to the same probe cue (Fig. 2F).
We explored the latent dynamics of the example networks to see the effects of the probe stimuli and the pruning (Fig. 3). For each network we calculated the principal components (PCs) of the activity in the recurrent layer and examined the time evolution of the activity in the first two PCs for probed and unprobed trials. When we examined activity in the first two PCs we found that the probe, delivered at time 20, drove the activity off the corresponding trajectory for the unprobed, condition 4 trial (Fig. 3 A and B). The activity diverged slowly over several time steps. However, later in the trial, after cue 2 was delivered, the activity was far off the trajectory of the unprobed trial, which led to the network producing an incorrect output at the response time (Fig. 3C). In the pruned network, however, the probed and unprobed trials had similar trajectories (Fig. 3 D and E), at least in the first two PC dimensions, and only diverged slightly later in the trial in PC2 (Fig. 3E). Thus, the network produced the correct output at the response time (Fig. 3F).
We next examined the network dynamics in two-dimensional (2D) PC space and overlaid the network’s vector field. The vector field shows how the network evolves from different points in phase space over one time step. We examined the vector field centered on the unprobed trajectory (condition 4) for the unpruned network (Fig. 4A) and the pruned network (Fig. 4B). The probe stimulus perturbs the network from the unprobed trajectory. If the vector field of the network is organized such that it pushes a perturbed trajectory back to the unperturbed trajectory, the network will be resistant to the probe and give the correct answer (46). Note that we only illustrate the 2D vector field that projects into low-dimensional PC space. The full vector field can diverge more or less in dimensions that we do not plot. We can see, however, that the vector field for the pruned network pushes the perturbed trajectories back to the unperturbed trajectories more effectively than the vector field for the unpruned network (Fig. 4 A and B). This is represented by the convergence of the perturbed points (i.e., open circles) in the pruned network, back to the unperturbed trajectory (i.e., end points of vectors from open circles). Thus, when a probe stimulus perturbs the network off the unperturbed trajectory, the network relaxes back to the unperturbed trajectory, in the pruned network (Fig. 4B) but not in the unpruned network (Fig. 4A). In the unpruned network the perturbed trajectories tend to diverge from the unperturbed trajectory, and the probe trajectory also diverges (Fig. 4A).
To characterize this, we calculated the ratio of the spread of the end points (circles in Fig. 4 A and B) to the spread of the probed points after one iteration of the network (Fig. 4C). When this ratio < 1, the spread of the points after one iteration of the network is smaller than the spread when they are perturbed. This characterizes the extent to which the network collapses perturbed trajectories back to the unperturbed trajectory, by comparing the volume of perturbed trajectories before and after one time step of the network. We found that throughout the delay period, this fraction was smaller, and consistently less than 1, in the pruned network. It was greater than 1 throughout the delay in the unpruned network. Thus, when the network was pushed off the mean trajectory by a distractor, it tended to return to a point closer to the unperturbed trajectory in one iteration for the pruned network. In the unpruned network, however, the probed trajectory tended to diverge from the unprobed trajectory.
We further formalized this by calculating the spectrum of Lyapunov exponents (LEs) around the unperturbed trajectory averaged over the delay period (Fig. 4D). The LEs characterize the extent to which two nearby points diverge or converge over time, as the network is iterated. Values greater than 0 indicate that the network will diverge in that dimension over time, if perturbed. Values less than 0 indicate that the network will converge over time. The maximal LE characterizes the overall resistance of the network to perturbations. Thus, the LE characterizes the contraction (values < 0) or expansion (values > 0) over time of perturbed trajectories, and larger positive or negative values indicate networks in which the expansion or contraction occurs more rapidly. The maximal LE in this example was −0.02 for the pruned network and 0.27 for the unpruned network. Therefore, activity tended to diverge following a perturbation for the unpruned network but converge for the pruned network. This convergence is why the pruned network is resistant to distraction.
We characterized the performance and dynamics for our population of trained pruned and unpruned networks (Fig. 5). We found that networks pruned to intermediate values were more resistant to distractors of different strength, delivered throughout the delay period, than unpruned networks (Fig. 5A). The difference in performance was most pronounced early in the delay (Fig. 5B) and was minimal near the end of the delay, when pruned and unpruned networks were more resistant to distractors. In the absence of probe cues (probe strength 0), the networks were almost perfect (Fig. 5C). Performance, however, degraded slowly with increasing distractor strength (Fig. 5C) but was consistently higher for pruned networks. Pruned networks had lower maximum LEs, and the LE decreased with increased pruning (Fig. 5D). Unpruned networks had higher LEs, despite being continuously trained in parallel with the pruned networks (Fig. 5D). We also found that the networks that performed on average best when probed also had lower LEs for both pruned and unpruned networks (Fig. 5E). The LE was negatively correlated with performance for both the unpruned [r(400) = −0.299, P < 0.001] and pruned [r(400) = −0.318, P < 0.001] networks. The distribution of connection strengths, at 70% pruning, averaged across networks, showed fewer small and more large connections in the pruned than the unpruned networks (Fig. 5F).
In a final analysis, we trained pruned (70%) and unpruned networks and examined the rate at which they were able to learn the DMS task. To examine task learning of pruned and unpruned networks we randomly initialized networks, pruned 70% of the weakest connections, and then trained them on the DMS task. We compared these networks to randomly initialized unpruned networks. We found that the unpruned networks learned more quickly than the pruned networks (SI Appendix, Fig. S2A). The average performance of the pruned network became similar to the unpruned network after about 20 training iterations. We also compared this to retraining pruned (70%) and unpruned networks with a new pair of cues, after they had previously been trained on the original task. When we examined learning to do DMS with a new set of cues, we found that the pruned network was overall more efficient, similar to the results on the original cues (SI Appendix, Fig. S2B). Thus, the recurrent dynamics of the pruned network could be reused to learn DMS for new cues that the network had not seen previously, and unpruned networks can initially learn more quickly than pruned networks, but the pruned networks learned new cues more effectively, if not more quickly.
RL.
We next trained a series of networks to carry out an RL task. RL is often thought to be a dopamine-dependent process, with dopamine driving plasticity on frontal-striatal synapses during learning (58). However, recurrent neural networks can be trained to solve bandit tasks and may better reflect the neural processes that drive learning in these tasks, under some conditions (55, 56). After training, the weights of the network can be fixed, and the network can make correct choices, essentially matching performance of biological agents, by storing information about past outcomes in patterns of recurrent activity. We trained a population of networks on a two-state RL problem. Specifically, they were trained to select a fixation point when it was presented (SI Appendix, Fig. S3) and to learn to select one of two targets in each block of trials, depending on which target was most frequently rewarded. To match the noisy choice behavior of young subjects, networks were trained to match noisy value functions (Methods). We followed the same procedure of incrementally pruning and retraining the RL network that was used to train the DMS network and also training, in parallel, unpruned networks. Thus, we initially trained 100 networks to carry out the RL task. We then used the original set of 100 networks to generate a series of pruned and unpruned networks. At each stage, we pruned a fraction of connections in the pruned networks and retrained the networks with a new set of training examples. For the unpruned networks, we copied the unpruned weight matrix forward but retrained them on an equivalent set of training examples.
When we examined the performance of an example pruned and unpruned network we found that the pruned RL network had more large weights (in absolute value) than the unpruned network (Fig. 6A). SVD on the recurrent A matrix showed that it was full rank, and the spectrum of singular values was similar between the pruned and the unpruned networks (Fig. 6B). We also found that the unpruned network produced noisier results (higher Q-value error) during selection of both the fixation point (Fig. 6C) and the best choice option (Fig. 6D). The pruned network, on the other hand, produced results that matched the values predicted by the Q-learning algorithm well (Fig. 6 E and F). To explore the network features that led to the improved performance of the pruned network, we again examined activity in a low-dimensional PC space. The dynamics of both networks were relatively low dimensional and lower dimensional for the pruned than the unpruned network (Fig. 6G).
When we examined the data plotted in the first two PCs, we found that the recurrent activity patterns, which represent action values, during the fixation and choice periods were separated in state space. They were separated more, however, for the pruned network than the unpruned network (Fig. 6 H and I). When we calculated the Mahalanobis distances (59), in the full 20-dimensional latent space, between the value clouds for the two states, we found that the distances were larger for the pruned than the unpruned network (average values pruned = 39,261, unpruned = 5,200). Thus, there was more separation in the recurrent activity representations for the two states for the pruned network. We also characterized the LEs across the whole trial and found that they were consistently smaller in the pruned networks (−0.020) than the unpruned networks (0.031).
When we compared performance of a population of pruned and unpruned networks for a range of pruning values, we found that pruned networks outperformed unpruned networks, by better approximating Q values, when pruning fractions were greater than about 0.3 (Fig. 7A). Correspondingly, pruned networks also more frequently chose the same option as the Q-learning algorithm, which we defined as correct decisions (Fig. 7B), and obtained more total rewards per block (Fig. 7C). Similar to the DMS networks, pruned RL networks had lower LEs (Fig. 7D). For the pruned RL networks the LEs were negative on average. The pruned RL networks also had lower dimensional latent dynamics (Fig. 7E) and a larger fraction of higher strength connections (Fig. 7F). The separation of latent activity related to values for the two states was also larger for pruned than unpruned networks (Fig. 7G). When we examined the relation between LE and the fraction of decisions that were consistent with the Q-learning algorithm, we found that pooled across both pruned and unpruned networks, there was a significant correlation between fraction correct (i.e., consistency between network and Q-learning algorithm) and the LE at low [r(100) = −0.158, P = 0.025), medium [Fig. 7H; r(100) = −0.873, P < 0.001], and high [Fig. 7I; r(100) = −0.593, P < 0.001] training noise levels. Thus, lower LEs, mostly in the pruned networks, lead to better performance. Overall, performance mostly increased with additional pruning for the RL network, up to 90%. For the DMS network, performance began to decrease at about 85% pruning. This difference may relate to the underlying complexity of the task.
In a final analysis we trained both pruned (70%) and corresponding unpruned networks to learn to select between a new option, which the network had not previously seen (choice 3; SI Appendix, Fig. S3), and one of the options with which the network was previously familiar (SI Appendix, Fig. S4). In this case both networks were first trained to select between options 1 and 2. After this, the pruned and unpruned networks were retrained to select between options 1 and 3. Although there was some advantage for the pruned network over the first training iteration, the unpruned network quickly surpassed the performance of the pruned network. With extensive training the pruned network approached the performance of the unpruned network, which suggests that the pruned network had enough residual capacity to learn this task. However, the unpruned network learned more quickly.
Discussion
We have examined the effect of synaptic pruning in recurrent neural network models of working memory and RL. The networks were able to capture many features of behavior that change during development. Specifically, in the delayed DMS task, the pruned networks were more resistant to distractors. When we examined the local dynamics, we found that the distractor resistance followed from a smaller LE around the unperturbed trajectory in the pruned network. Thus, when a distractor perturbed the pruned network, it returned to the unperturbed trajectory, whereas the unpruned network did not. We used an incremental training and pruning approach. When we directly trained a network, after randomly initializing the weights and then pruning 70% of them, we found that the network learned more slowly than the unpruned network. However, when we first trained networks on the DMS task, with and without pruning, the pruned network learned with lower loss to do DMS with new cues.
When we examined the performance of a network trained to perform a two-armed bandit RL task, we also found that pruning improved performance. Pruned networks trained in the presence of noise were able to more accurately approximate optimal Q values than unpruned networks. Pruning has long been used as an approach in the feed-forward neural network literature, to improve generalization performance (47, 48), and pruning can be computationally optimized (60). The RL results follow from improved generalization performance. Every block of trials in a two-armed bandit RL task is unique because choice outcomes are stochastic. Additionally, there was noise in the Q values we used for training as this noise approximates noisier learning in children. Therefore, the network has to learn to generalize. Although the pruned network was better able to perform the RL task on which it was trained, it learned to select a novel option more slowly than the unpruned network. Thus, pruned networks are better at the tasks for which they have been trained, but they learn some new tasks slowly. It is not clear why the pruned DMS network learned a new option more effectively, which was not the case for the pruned RL network. These differences may be due to whether learned weights in the pruned network can generalize better to new options within a task.
The brain, at least at a thalamocortical level, produces an overabundance of synapses during development (7). The reasons for this are unclear. It may be a constraint of developmental processes, which are able to guide axons to their target structures but not able to establish precise connections (61). It is unlikely that the precise wiring of the brain, even in systems like the early visual system where adult plasticity is only minimally important, could be specified by genetics (9). In general, the brain relies on learning during development to establish adult behavior and corresponding neural connectivity (62). Development uses a mechanism by which axons are steered toward their target structures, and then synapses are overproduced. The synapses are then pruned through activity-dependent mechanisms in which weak synapses are eliminated (5, 8). These mechanisms have been most thoroughly investigated in a few model systems, including the retinogeniculate pathway, thalamocortical synapses in barrel cortex, and climbing fibers in the cerebellum (8, 9).
In addition to these model systems, it is also known that substantial pruning of excitatory synapses happens throughout cortex, including in prefrontal cortex (13, 14). Studies in human postmortem tissue, monkeys (17, 19, 20), and rats (21) have consistently shown substantial pruning. Across these studies, synapses appear to peak before puberty (13). They then decline exponentially and continue to decline with age. The timing of pruning differs from the timing of cell elimination. Specifically, the number of neurons in prefrontal cortex, which is highest at birth, drops to adult levels by about 6 mo of age (14). Thus, in cortical systems that support cognition, there is substantial synaptic pruning, specific to excitatory synapses on both excitatory and inhibitory neurons, during adolescent development.
Pruning was developed in machine learning to improve the generalization and computational efficiency of artificial neural networks (47, 48, 63). Large networks can learn complex problems. However, they generalize to new data poorly, especially when training data are limited. This is often referred to as the bias/variance trade-off (64). Large networks can approximate complex functions with minimal bias. However, they produce highly variable fits when they are trained on different datasets that represent variations on the same learning problem. Simpler networks produce biased fits to complex functions as they may smooth over the complexity. However, they produce less variable fits when trained on different datasets and therefore generalize better to new data. Pruning reduces the complexity of a large neural network, and therefore, pruning can be used to optimize the trade-off between bias and variance, for a given learning problem. Early (47) and more recent work (65) on pruning was also intended to reduce the memory requirements and computational demands of large networks. One important point is that it is not clear whether the pruning used in the present work directly produces networks that are better at the task or whether pruning facilitates training to higher levels of performance. Additionally, regularization techniques in machine learning, like pruning, improve performance in the presence of limited training data. With very large training datasets, regularization is less important. Other regularization approaches, for example, weight shrinkage (66), in which weights are shrunk toward zero on each training iteration, or adding noise to connection weights during training (67), could also lead to improved performance. However, these approaches are less related to known developmental processes.
Consistent with the earlier work in artificial intelligence and aside from possible developmental constraints, we have shown that there can be computational advantages to overproducing synapses and then pruning them incrementally while tasks are learned, in artificial neural networks. The DMS network was able to learn more efficiently before pruning. When we pruned a DMS network before training, it learned more slowly than the unpruned network. However, trained pruned DMS networks were able to learn new cues as fast, and more effectively, than unpruned DMS networks. Additionally, when we explored the ability of the pruned and unpruned RL networks to learn to select a novel option, to which the networks had not been previously exposed, the pruned network learned more slowly. Thus, consistent with developmental processes, unpruned networks are able to learn some new tasks or problems more quickly than pruned networks.
The rate-based recurrent network is simplified and lacks biological features. Simplified, formal models have provided insight into neural population coding (68, 69) and RL (31, 70). Connectionist models, particularly as extended in deep networks, provide a powerful description of tuning functions in sensory cortex (71, 72). Therefore, the basic computational principles captured by our model can provide a reasonable description of the effects of neural pruning in cortex. One feature of biological pruning not captured by our model, which should be explored in the future, is that only excitatory synapses are pruned. Studying the effects of pruning in spiking neural networks that learn arbitrary cognitive tasks and that have excitatory and inhibitory units will be challenging. There are methods to map between rate models and spiking models (73), and there has been progress in training spiking networks (74). Therefore, future work may allow a more detailed exploration of the effects of pruning in more realistic networks.
Our network makes several predictions that can be explored in neurophysiology recordings. First, as already shown in previous neurophysiology data, pruned networks should be more resistant to distractors in working memory tasks (46). Second, the resistance to distractors follows from steeper or larger attractor basins and correspondingly less positive or negative Lyapunov coefficients in the model networks. Thus, when activity is perturbed, it should relax back to the unperturbed trajectory more effectively in a pruned (i.e., adult) network. It should be possible to estimate statistics that capture the size and steepness of attractor basins with large-scale neural recordings that are becoming more prevalent (75). Finally, we also found that the dimensionality of neural activity was lower in pruned networks in the RL task. Although dimensionality has been estimated in many neural systems (76, 77), future work could compare dimensionality across developmental time points.
During the period over which synapses are pruned in cortical areas that support cognition, there is also improvement in several behavioral processes (25). Two processes that improve are working memory (40) and RL (22, 27–29, 32). In working memory tasks, both the span and resistance to distraction improve with development (40, 41). Similar improvements have been seen in monkeys (46). Furthermore, in monkeys, physiology recordings have shown that resistance to distraction is reflected in neural activity in prefrontal cortex. In young animals, distractors affected neural activity during a memory interval. However, in adult animals, distractor effects quickly dissipated. These results are consistent with the dynamic analysis we carried out on our network. The pruned network was more resistant to perturbation during the delay interval. By several related measures, when the pruned network was perturbed by a distractor, the trajectory relaxed back to the unperturbed trajectory. The unpruned network, on the other hand, frequently diverged after perturbation, and this led to the network sometimes providing the wrong answer. The robustness of the pruned network, reflected in a deeper basin of attraction and more stable dynamics, is similar to results we have seen when similar networks were trained on RL tasks, and the evolution over learning was examined (55).
Performance on RL tasks also improves during adolescence (22). Some studies report improvements in learning rates (32). However, the most consistent result is a decrease in decision noise, such that older subjects more frequently choose the most rewarded options (27). We found that pruned networks outperformed unpruned networks, having lower decision noise, when both were trained in the presence of noise. The pruned networks produced more accurate approximations to the optimal Q values and more consistently picked the same options as the Q algorithm. Thus, pruning affected decision noise, consistent with most studies (27). The networks were trained to reproduce a specific learning rate, and therefore, it is not straightforward to assess effects of pruning on learning rates. When we examined the latent space, we found that the state-dependent value functions were better separated in the pruned network than in the unpruned network. The result in the pruned network is consistent with analysis of neural data from prefrontal cortex in monkeys, which showed that when new pairs of options were learned, they were represented in different subspaces in prefrontal cortex (76). Qualitatively, there appeared to be more interference or cross-talk between value representations in the unpruned network than the pruned network. Therefore, outcomes tended to affect value updates across multiple options. This is also reminiscent of results following lesions of orbitofrontal cortex or ventrolateral prefrontal cortex in moneys, which lead to problems with credit assignment (78, 79).
Pruning appears to be a closely regulated process, such that insufficient pruning or overpruning can lead to disorders (80, 81). For example, it has been suggested that excessive synapses and insufficient pruning may underlie deficits in autism (82). In the opposite direction, it has been suggested that excessive pruning may underlie deficits in schizophrenia (83–87), and subsequent genetic and molecular work supports this hypothesis (88–90). We found that pruning led to a strengthening of the attractor basin. Thus, pruned networks are less distractable, which also implies that they may be harder to move out of certain attractor basins into which they have settled. While this can be useful when the pruned network functions on the task for which it has been trained, it can be less useful when the pruned and trained network is producing pathological output. This may apply to psychiatric disorders. For example, in depression, brain networks that underlie motivational processes (91) do not appropriately motivate behavior. In addiction, motivational processes are highly active but motivate harmful behavior. Both depression and addiction can be resistant to therapy, which is an effort to modify the input–output mappings of a patient’s brain. Consistent with this, we found that pruned networks can be more difficult to train on new tasks. Ultimately, unpruning an overpruned network, and restoring plasticity, may be a highly effective therapeutic approach. Unpruning a network can lead to a weakening of pathological attractors, and restoring plasticity can allow retraining the network such that it no longer produces pathological output.
Recent work with both ketamine and psilocybin, both of which may be treatments for depression (92–95), suggests that these drugs may lead to production of new spines (96–98). Similar results may also follow from other therapeutic techniques like electroconvulsive therapy or deep brain stimulation. These techniques may lead to therapeutic effects via an increase in spines and synapses. This increase in spines may flatten the attractor landscape of the brain and allow plasticity. Both effects can lead to eliminating pathological attractors. This may be why these therapies can open a therapeutic window. It is also the case that early interventions for mental health issues, when the brain is more plastic, can be more effective than interventions targeted at the adult, after pruning has taken place. Although this remains highly speculative, our simulation results at least provide a theoretical framework for approaching these issues (99). The theory suggests that an overpruned network can effectively become stuck in a deep basin of attraction, and the best way out is to increase synapses in the network, making the network more amenable to learning new solutions.
Methods
We trained a series of rate-based, recurrent neural networks to solve two cognitive tasks. The first set of networks was trained to solve a DMS working memory task (52). The second set of networks was trained to solve a two-armed bandit RL task (54). We trained networks of various sizes, N. As these are low-dimensional problems, small networks (n = 10) could solve the tasks well. However, we present results using n = 30 for the DMS task and n = 20 for the RL task. Results at n = 30 were more consistent at the population level for the DMS task.
The recurrent dynamics of our network are given by Eqs. 1–3. The network was simulated as a nonlinear iterative map. The variable represents inputs to the network at time k; represents the latent dynamics, or activity in the recurrent layer, at time k; and is the output of the network. For consistency with tasks used in experimental work, k indexes ∼50-ms time bins.
[1] |
[2] |
[3] |
All networks were trained using the scaled conjugate gradient algorithm and back propagation through time (100, 101). The CG algorithm iteratively minimizes the loss using second-order information. One CG iteration is composed of weight updates over a full set of conjugate vectors, where the number of vectors is given by the dimensionality of the weight vector, W, for the network. The first update vector is given by the gradient calculated with back-propagation. Subsequent conjugate vectors are computed recursively, and step size for weight updates is calculated from the second-order approximation.
We trained matrices A, B, and C using a quadratic loss function on the output given by
[4] |
The variable W indicates the vector of all weights formed by vectorizing and concatenating matrices A, B, and C. The inputs and outputs were specific to each task. For additional details on network training, see SI Appendix, SI Methods.
LE.
We estimated a maximal LE for each network (102). The LE characterizes the rate at which trajectories, in the undriven network, diverge or converge in the latent space. Specifically, if we consider two nearby points in latent space, and , the distance between these points at time k is given by . After n iterations of the network, the distance between the subsequent points on each trajectory is given by . The LE is then defined by .
For characterizing our networks we calculated (a local estimate of) the maximal LE by first calculating
[5] |
and then calculated the eigen-decomposition of . The maximal LE is the log of the absolute value of the largest eigenvalue of Z divided by the number of time steps K. This is defined as where the values of are the eigenvalues of Z. The matrices are the Jacobians of the network update equation, , evaluated at each time step k.
Supplementary Material
Acknowledgments
This work was supported by the intramural research program of the National Institute of Mental Health (ZIA MH002928). Simulations were carried out on the NIH/High Performance Computing Biowulf cluster (https://hpc.nih.gov/).
Footnotes
The authors declare no competing interest.
This article is a PNAS Direct Submission.
This article contains supporting information online at https://www.pnas.org/lookup/suppl/doi:10.1073/pnas.2121331119/-/DCSupplemental.
Data Availability
Code has been deposited in GitHub (https://github.com/baverbeck/Pruning-RNNs.git).
References
- 1.Rosenblatt F., The perceptron: A probabilistic model for information storage and organization in the brain. Psychol. Rev. 65, 386–408 (1958). [DOI] [PubMed] [Google Scholar]
- 2.Widrow B., Hoff M. E., “Adaptive switching circuits” in IRE Wescon Convention Record (Institute of Electrical and Electronics Engineers, Western Electronic Show and Convention, 1960), pp. 96–104. [Google Scholar]
- 3.Hebb D. O., Organization of behavior. J. Clin. Psychol. 6, 307–307 (1950). [Google Scholar]
- 4.Rumelhart D. E., McClelland J. L., Parallel Distributed Processing: Explorations in the Microstructure of Cognition, Computational Models of Cognition and Perception (MIT Press, Cambridge, MA, 1986). [Google Scholar]
- 5.Lee H., et al. , Synapse elimination and learning rules co-regulated by MHC class I H2-Db. Nature 509, 195–200 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Huttenlocher P. R., Morphometric study of human cerebral cortex development. Neuropsychologia 28, 517–527 (1990). [DOI] [PubMed] [Google Scholar]
- 7.Rakic P., Bourgeois J. P., Eckenhoff M. F., Zecevic N., Goldman-Rakic P. S., Concurrent overproduction of synapses in diverse regions of the primate cerebral cortex. Science 232, 232–235 (1986). [DOI] [PubMed] [Google Scholar]
- 8.Faust T. E., Gunner G., Schafer D. P., Mechanisms governing activity-dependent synaptic pruning in the developing mammalian CNS. Nat. Rev. Neurosci. 22, 657–673 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Hensch T. K., Critical period regulation. Annu. Rev. Neurosci. 27, 549–579 (2004). [DOI] [PubMed] [Google Scholar]
- 10.Schafer D. P., et al. , Microglia sculpt postnatal neural circuits in an activity and complement-dependent manner. Neuron 74, 691–705 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Riccomagno M. M., Kolodkin A. L., Sculpting neural circuits by axon and dendrite pruning. Annu. Rev. Cell Dev. Biol. 31, 779–805 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Watanabe M., Kano M., Climbing fiber synapse elimination in cerebellar Purkinje cells. Eur. J. Neurosci. 34, 1697–1710 (2011). [DOI] [PubMed] [Google Scholar]
- 13.Petanjek Z., et al. , Extraordinary neoteny of synaptic spines in the human prefrontal cortex. Proc. Natl. Acad. Sci. U.S.A. 108, 13281–13286 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Huttenlocher P. R., Synaptic density in human frontal cortex—Developmental changes and effects of aging. Brain Res. 163, 195–205 (1979). [DOI] [PubMed] [Google Scholar]
- 15.Huttenlocher P. R., Dabholkar A. S., Regional differences in synaptogenesis in human cerebral cortex. J. Comp. Neurol. 387, 167–178 (1997). [DOI] [PubMed] [Google Scholar]
- 16.Woo T. U., Pucak M. L., Kye C. H., Matus C. V., Lewis D. A., Peripubertal refinement of the intrinsic and associational circuitry in monkey prefrontal cortex. Neuroscience 80, 1149–1158 (1997). [DOI] [PubMed] [Google Scholar]
- 17.Elston G. N., Oga T., Fujita I., Spinogenesis and pruning scales across functional hierarchies. J. Neurosci. 29, 3271–3275 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Elston G. N., Oga T., Okamoto T., Fujita I., Spinogenesis and pruning from early visual onset to adulthood: An intracellular injection study of layer III pyramidal cells in the ventral visual cortical pathway of the macaque monkey. Cereb. Cortex 20, 1398–1408 (2010). [DOI] [PubMed] [Google Scholar]
- 19.Chung D. W., Wills Z. P., Fish K. N., Lewis D. A., Developmental pruning of excitatory synaptic inputs to parvalbumin interneurons in monkey prefrontal cortex. Proc. Natl. Acad. Sci. U.S.A. 114, E629–E637 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Bourgeois J. P., Goldman-Rakic P. S., Rakic P., Synaptogenesis in the prefrontal cortex of rhesus monkeys. Cereb. Cortex 4, 78–96 (1994). [DOI] [PubMed] [Google Scholar]
- 21.Mallya A. P., Wang H. D., Lee H. N. R., Deutch A. Y., Microglial pruning of synapses in the prefrontal cortex during adolescence. Cereb. Cortex 29, 1634–1643 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Palminteri S., Kilford E. J., Coricelli G., Blakemore S. J., The computational development of reinforcement learning during adolescence. PLOS Comput. Biol. 12, e1004953 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Larsen B., Luna B., Adolescence as a neurobiological critical period for the development of higher-order cognition. Neurosci. Biobehav. Rev. 94, 179–195 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Gopnik A., et al. , Changes in cognitive flexibility and hypothesis search across human life history from childhood to adolescence to adulthood. Proc. Natl. Acad. Sci. U.S.A. 114, 7892–7899 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Hartley C. A. N., Cohen K. A. O., Interactive development of adaptive learning and memory. Annu. Rev. Dev. Psychol. 3, 59–85 (2021). [Google Scholar]
- 26.Steinberg L., Cognitive and affective development in adolescence. Trends Cogn. Sci. 9, 69–74 (2005). [DOI] [PubMed] [Google Scholar]
- 27.Nussenbaum K., Hartley C. A., Reinforcement learning across development: What insights can we draw from a decade of research? Dev. Cogn. Neurosci. 40, 100733 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Lange-Küttner C., Averbeck B. B., Hirsch S. V., Wießner I., Lamba N., Sequence learning under uncertainty in children: Self-reflection vs. self-assertion. Front. Psychol. 3, 127 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Lange-Kuttner C., Averbeck B. B., Hentschel M., Baumbach J., Intelligence matters for stochastic feedback processing during sequence learning in adolescents and young adults. Intelligence 86, 101542 (2021). [Google Scholar]
- 30.Cohen A. O., Nussenbaum K., Dorfman H. M., Gershman S. J., Hartley C. A., The rational use of causal inference to guide reinforcement learning strengthens with age. NPJ Sci. Learn. 5, 16 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Averbeck B. B., Costa V. D., Motivational neural circuits underlying reinforcement learning. Nat. Neurosci. 20, 505–512 (2017). [DOI] [PubMed] [Google Scholar]
- 32.Xia L., et al. , Modeling changes in probabilistic reinforcement learning during adolescence. PLOS Comput. Biol. 17, e1008524 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Moin Afshar N., Keip A. J., Taylor J. R., Lee D., Groman S. M., Reinforcement learning during adolescence in rats. J. Neurosci. 40, 5857–5870 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Blanco N. J., Sloutsky V. M., Systematic exploration and uncertainty dominate young children’s choices. Dev. Sci. 24, e13026 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Sumner E. S., et al. , The exploration advantage: Children’s instinct to explore leads them to find information that adults miss. psyarXiv [Preprint] (2019). https://psyarxiv.com/h437v/. Accessed 4 June 2021.
- 36.Gopnik A., Childhood as a solution to explore-exploit tensions. Philos. Trans. R. Soc. Lond. B Biol. Sci. 375, 20190502 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Gopnik A., Griffiths T. L., Lucas C. G., When younger learners can be better (or at least more open-minded) than older ones. Curr. Dir. Psychol. Sci. 24, 87–92 (2015). [Google Scholar]
- 38.Cohen L. G., et al. , Period of susceptibility for cross-modal plasticity in the blind. Ann. Neurol. 45, 451–460 (1999). [DOI] [PubMed] [Google Scholar]
- 39.Somerville L. H., et al. , Charting the expansion of strategic exploratory behavior during adolescence. J. Exp. Psychol. Gen. 146, 155–164 (2017). [DOI] [PubMed] [Google Scholar]
- 40.Gathercole S. E., Pickering S. J., Ambridge B., Wearing H., The structure of working memory from 4 to 15 years of age. Dev. Psychol. 40, 177–190 (2004). [DOI] [PubMed] [Google Scholar]
- 41.Fry A. F., Hale S., Relationships among processing speed, working memory, and fluid intelligence in children. Biol. Psychol. 54, 1–34 (2000). [DOI] [PubMed] [Google Scholar]
- 42.Ullman H., Almeida R., Klingberg T., Structural maturation and brain activity predict future working memory capacity during childhood development. J. Neurosci. 34, 1592–1598 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Luciana M., Nelson C. A., Assessment of neuropsychological function through use of the Cambridge Neuropsychological Testing Automated Battery: Performance in 4- to 12-year-old children. Dev. Neuropsychol. 22, 595–624 (2002). [DOI] [PubMed] [Google Scholar]
- 44.Vuontela V., et al. , Audiospatial and visuospatial working memory in 6-13 year old school children. Learn. Mem. 10, 74–81 (2003). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Unger K., Ackerman L., Chatham C. H., Amso D., Badre D., Working memory gating mechanisms explain developmental change in rule-guided behavior. Cognition 155, 8–22 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Zhou X., et al. , Neural correlates of working memory development in adolescent primates. Nat. Commun. 7, 13423 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Le Cun Y., Denker J. S., Solla S. A., “Optimal brain damage” in Advances in Neural Information Processing Systems, D. S. Touretzky, Ed. (Morgan Kaufmann, Vancouver, 1990), pp. 598–605. [Google Scholar]
- 48.Hassibi B., Stork D. G., “Second order derivatives for network pruning: Optimal brain surgeon” in Advances in Neural Information Processing Systems, Hanson S. J., Cowan J. D., Giles C. L., Eds. (Morgan Kaufmann, Vancouver, 1993), pp. 164–171. [Google Scholar]
- 49.Hassibi B., Stork D., Wolff G., Optimal brain surgeon: Extensions and performance comparisons. Adv. Neural Inf. Process. Syst. 6 (1993), pp. 263–270. [Google Scholar]
- 50.Bishop C. M., Neural Networks for Pattern Recognition (Oxford University Press, Oxford, ed. 1, 1995). [Google Scholar]
- 51.MacKay D. J., Probable networks and plausible predictions—A review of practical Bayesian methods for supervised neural networks. Network 6, 469 (1995). [Google Scholar]
- 52.Napoli J. L., et al. , Correlates of auditory decision-making in prefrontal, auditory, and basal lateral amygdala cortical areas. J. Neurosci. 41, 1301–1316 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Chaisangmongkon W., Swaminathan S. K., Freedman D. J., Wang X. J., Computing by Robust Transience: How the fronto-parietal network performs sequential, category-based decisions. Neuron 93, 1504–1517 e1504 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Costa V. D., Dal Monte O., Lucas D. R., Murray E. A., Averbeck B. B., Amygdala and ventral striatum make distinct contributions to reinforcement learning. Neuron 92, 505–517 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Márton C. D., Schultz S. R., Averbeck B. B., Learning to select actions shapes recurrent dynamics in the corticostriatal system. Neural Netw. 132, 375–393 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Wang J. X., et al. , Prefrontal cortex as a meta-reinforcement learning system. Nat. Neurosci. 21, 860–868 (2018). [DOI] [PubMed] [Google Scholar]
- 57.Pozo K., Goda Y., Unraveling mechanisms of homeostatic synaptic plasticity. Neuron 66, 337–351 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Frank M. J., O’Reilly R. C., A mechanistic account of striatal dopamine function in human cognition: Psychopharmacological studies with cabergoline and haloperidol. Behav. Neurosci. 120, 497–517 (2006). [DOI] [PubMed] [Google Scholar]
- 59.Johnson R. A., Wichern D. W., Applied Multivariate Statistical Analysis (Prentice Hall, Saddle River, NJ, 1998). [Google Scholar]
- 60.Scholl C., Rule M. E., Hennig M. H., The information theory of developmental pruning: Optimizing global network architectures using local synaptic rules. PLOS Comput. Biol. 17, e1009458 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.O’Leary D. D., Chou S. J., Sahara S., Area patterning of the mammalian cortex. Neuron 56, 252–269 (2007). [DOI] [PubMed] [Google Scholar]
- 62.Leopold D. A., Averbeck B. B., Self-tuition as an essential design feature of the brain. Philos. Trans. R. Soc. Lond. B Biol. Sci. 377, 20200530 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Hanson S., Pratt L., Comparing biases for minimal network construction with back-propagation. Adv. Neural Inf. Process. Syst. 1, 177–185 (1988). [Google Scholar]
- 64.Bishop C. M., Pattern Recognition and Machine Learning (Information Science and Statistics, Springer, New York, 2006). [Google Scholar]
- 65.Wang Z., Wohlwend J., Lei T., Structured pruning of large language models. arXiv [Preprint] (2019). https://arxiv.org/abs/1910.04732. Accessed 20 January 2022.
- 66.Hinton G. E., “Learning translation invariant recognition in a massively parallel networks” in International Conference on Parallel Architectures and Languages Europe, J. W. de Bakker, A. J. Nijman, P. C. Treleaven, Eds. (Springer, 1987), pp. 1–13.
- 67.Hinton G. E., Van Camp D., “Keeping the neural networks simple by minimizing the description length of the weights” in Proceedings of the Sixth Annual Conference on Computational Learning Theory (1993), pp. 5–13.
- 68.Averbeck B. B., Latham P. E., Pouget A., Neural correlations, population coding and computation. Nat. Rev. Neurosci. 7, 358–366 (2006). [DOI] [PubMed] [Google Scholar]
- 69.Zohary E., Shadlen M. N., Newsome W. T., Correlated neuronal discharge rate and its implications for psychophysical performance. Nature 370, 140–143 (1994). [DOI] [PubMed] [Google Scholar]
- 70.Rescorla R. A., Wagner A. R., “A theory of Pavlovian conditioning: Variations in the effectiveness of reinforcement and nonreinforcement” in Classical Conditioning II: Current Research and Theory, Black A. H., Prokasy W. F. (Appleton-Century-Crofts, New York, 1972), pp. 64–99. [Google Scholar]
- 71.Poggio T., A theory of how the brain might work. Cold Spring Harb. Symp. Quant. Biol. 55, 899–910 (1990). [DOI] [PubMed] [Google Scholar]
- 72.Yamins D. L., et al. , Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proc. Natl. Acad. Sci. U.S.A. 111, 8619–8624 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.Shriki O., Hansel D., Sompolinsky H., Rate models for conductance-based cortical neuronal networks. Neural Comput. 15, 1809–1841 (2003). [DOI] [PubMed] [Google Scholar]
- 74.Kim C. M., Chow C. C., Learning recurrent dynamics in spiking networks. eLife 7, e37124 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Mitz A. R., et al. , High channel count single-unit recordings from nonhuman primate frontal cortex. J. Neurosci. Methods 289, 39–47 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Bartolo R., Saunders R. C., Mitz A. R., Averbeck B. B., Dimensionality, information and learning in prefrontal cortex. PLOS Comput. Biol. 16, e1007514 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Cunningham J. P., Yu B. M., Dimensionality reduction for large-scale neural recordings. Nat. Neurosci. 17, 1500–1509 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78.Murray E. A., Rudebeck P. H., Specializations for reward-guided decision-making in the primate ventral prefrontal cortex. Nat. Rev. Neurosci. 19, 404–417 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Walton M. E., Behrens T. E., Buckley M. J., Rudebeck P. H., Rushworth M. F., Separable learning systems in the macaque brain and the role of orbitofrontal cortex in contingent learning. Neuron 65, 927–939 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80.Kasai H., Ziv N. E., Okazaki H., Yagishita S., Toyoizumi T., Spine dynamics in the brain, mental disorders and artificial neural networks. Nat. Rev. Neurosci. 22, 407–422 (2021). [DOI] [PubMed] [Google Scholar]
- 81.Cao P., et al. , Early-life inflammation promotes depressive symptoms in adolescence via microglial engulfment of dendritic spines. Neuron 109, 2573–2589.e9 (2021). [DOI] [PubMed] [Google Scholar]
- 82.Hutsler J. J., Zhang H., Increased dendritic spine densities on cortical projection neurons in autism spectrum disorders. Brain Res. 1309, 83–94 (2010). [DOI] [PubMed] [Google Scholar]
- 83.Feinberg I., Schizophrenia: Caused by a fault in programmed synaptic elimination during adolescence? J. Psychiatr. Res. 17, 319–334 (1982-1983). [DOI] [PubMed] [Google Scholar]
- 84.Hoffman R. E., McGlashan T. H., Parallel distributed processing and the emergence of schizophrenic symptoms. Schizophr. Bull. 19, 119–140 (1993). [DOI] [PubMed] [Google Scholar]
- 85.Hoffman R. E., Dobscha S. K., Cortical pruning and the development of schizophrenia: A computer model. Schizophr. Bull. 15, 477–490 (1989). [DOI] [PubMed] [Google Scholar]
- 86.Selemon L. D., Zecevic N., Schizophrenia: A tale of two critical periods for prefrontal cortical development. Transl. Psychiatry 5, e623 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Glantz L. A., Lewis D. A., Decreased dendritic spine density on prefrontal cortical pyramidal neurons in schizophrenia. Arch. Gen. Psychiatry 57, 65–73 (2000). [DOI] [PubMed] [Google Scholar]
- 88.Kim M., et al. , Brain gene co-expression networks link complement signaling with convergent synaptic pathology in schizophrenia. Nat. Neurosci. 24, 799–809 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 89.Sellgren C. M., et al. , Increased synapse elimination by microglia in schizophrenia patient-derived models of synaptic pruning. Nat. Neurosci. 22, 374–385 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 90.Sekar A., et al. ; Schizophrenia Working Group of the Psychiatric Genomics Consortium, Schizophrenia risk from complex variation of complement component 4. Nature 530, 177–183 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91.Averbeck B. B., Murray E. A., Hypothalamic interactions with large-scale neural circuits underlying reinforcement learning and motivated behavior. Trends Neurosci. 43, 681–694 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 92.Griffiths R. R., et al. , Psilocybin produces substantial and sustained decreases in depression and anxiety in patients with life-threatening cancer: A randomized double-blind trial. J. Psychopharmacol. 30, 1181–1197 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 93.Carhart-Harris R. L., et al. , Psilocybin with psychological support for treatment-resistant depression: Six-month follow-up. Psychopharmacology (Berl.) 235, 399–408 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 94.Krystal J. H., Abdallah C. G., Sanacora G., Charney D. S., Duman R. S., Ketamine: A paradigm shift for depression research and treatment. Neuron 101, 774–778 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95.Bahji A., Vazquez G. H., Zarate C. A. Jr., Comparative efficacy of racemic ketamine and esketamine for depression: A systematic review and meta-analysis. J. Affect. Disord. 278, 542–555 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.Moda-Sava R. N., et al. , Sustained rescue of prefrontal circuit dysfunction by antidepressant-induced spine formation. Science 364, eaat8078 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Shao L. X., et al. , Psilocybin induces rapid and persistent growth of dendritic spines in frontal cortex in vivo. Neuron 109, 2535–2544.e4 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98.Savalia N. K., Shao L. X., Kwan A. C., A dendrite-focused framework for understanding the actions of ketamine and psychedelics. Trends Neurosci. 44, 260–275 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Durstewitz D., Huys Q. J. M., Koppe G., Psychiatric illnesses as disorders of network dynamics. Biol. Psychiatry Cogn. Neurosci. Neuroimaging 6, 865–876 (2021). [DOI] [PubMed] [Google Scholar]
- 100.Moller M. F., A scaled conjugate-gradient algorithm for fast supervised learning. Neural Netw. 6, 525–533 (1993). [Google Scholar]
- 101.Herz J., Krogh A., Palmer R. G., Introduction to the Theory of Neural Computation (Perseus Books, Cambridge, MA, 1991). [Google Scholar]
- 102.Kantz H., Schreiber T., Nonlinear Time Series Analysis (Cambridge University Press, Cambridge, 2004). [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Code has been deposited in GitHub (https://github.com/baverbeck/Pruning-RNNs.git).