Version Changes
Revised. Amendments from Version 1
We have answered the questions our Reviewers have asked and also implemented changes to the text to increase its readability and clarity. We have added a glossary (Supplementary File 1) and a Supplementary Figure S1 about the analogy of standard Darwinism and Darwinian neurodynamics, to guide the reader.
Abstract
Background: The fact that surplus connections and neurons are pruned during development is well established. We complement this selectionist picture by a proof-of-principle model of evolutionary search in the brain, that accounts for new variations in theory space. We present a model for Darwinian evolutionary search for candidate solutions in the brain.
Methods: We combine known components of the brain – recurrent neural networks (acting as attractors), the action selection loop and implicit working memory – to provide the appropriate Darwinian architecture. We employ a population of attractor networks with palimpsest memory. The action selection loop is employed with winners-share-all dynamics to select for candidate solutions that are transiently stored in implicit working memory.
Results: We document two processes: selection of stored solutions and evolutionary search for novel solutions. During the replication of candidate solutions attractor networks occasionally produce recombinant patterns, increasing variation on which selection can act. Combinatorial search acts on multiplying units (activity patterns) with hereditary variation and novel variants appear due to (i) noisy recall of patterns from the attractor networks, (ii) noise during transmission of candidate solutions as messages between networks, and, (iii) spontaneously generated, untrained patterns in spurious attractors.
Conclusions: Attractor dynamics of recurrent neural networks can be used to model Darwinian search. The proposed architecture can be used for fast search among stored solutions (by selection) and for evolutionary search when novel candidate solutions are generated in successive iterations. Since all the suggested components are present in advanced nervous systems, we hypothesize that the brain could implement a truly evolutionary combinatorial search system, capable of generating novel variants.
Keywords: attractor network, autoassociative neural network, evolutionary search, Darwinian dynamics, neurodynamics, learning, problem solving
Introduction
The idea that functional selection on a large set of neurons and their connections takes place in the brain during development 1– 3 is now experimentally validated 4– 7. As originally portrayed, this process is only one round of variation generation and selection, even if it requires several years. Evolution by natural selection works differently: variants are generated and then selected in iterative rounds. The field of “Neural Darwinism” 1– 3 fails to include generation of variants and thus could justifiably be regarded as a misnomer because the process that it describes is not evolutionary in the strict sense 8. Evidence indicated that the development of the brain is more “constructivist” 9 than pictured by the original selectionist accounts: for example, repeated rounds of neuron addition and loss happen during development 10. Structural plasticity (synaptic remodelling) is now known to be a lifelong process with implications for memory and learning ( e.g. 11, 12). The addition and deletion of synapses and neurons takes several hours or days 13. Our main goal here is to present a proof of principle that bona fide evolutionary dynamics could happen in the brain on a much faster time scale.
Maynard Smith 14 identified multiplication, inheritance and variability as necessary features of evolution. In genetic evolution the variability operators are mutation and recombination. If there are hereditary traits that affect the survival and/or the fecundity of the units, then in a population of these units, evolution by natural selection can take place. While this characterization qualitatively outlines the algorithmic aspect of evolution 15, concrete realizations require quantitative conditions too: population size cannot be too small (if it is too small, neutral drift dominates over selection 16) and replication accuracy cannot be too low (if it is too low, hereditary information is lost 17). Note, that this description says nothing about the nature of the units: they could be genes, organisms, linguistic constructions or anything else.
The proper implementation of an evolutionary process within the nervous system could have major implications for neuroscience and cognition 8, 18– 25. A main benefit of neuro-evolutionary dynamics would be that it could harness the parallelism inherent in the nervous system and the redistribution of resources at the same time. The latter process means that hopeless variants are thrown away and are replaced in the “breeding space” by more promising ones 8. Another important aspect of the process is that it is generative: it could explain where new hypotheses and new policies come from in Bayesian approaches to cognition 26, 27 and reinforcement learning 28– 30, respectively. Bayesian inference and natural selection are analogous 31, 32 in that candidate hypotheses in the brain (the prior distribution) represent a population of evolutionary units, which are evaluated, or selected, based on the evidence. There is a mathematical isomorphism between the discrete-time replicator equation and Bayesian updates 31. The likelihood function is analogous to the fitness function and the posterior distribution to the selected population. Relations like this suggest that Bayesian update could be one of the pillars of “universal Darwinism” 33. We believe that convincing models for neuro-evolution could empower Bayesian approaches by providing a mechanism to generate candidate hypotheses.
Attractor networks have been used (among others) as models of long-term memory, which are able to complete partial input 34, 35. These networks consist of one layer of units that recurrently connect back to the same layer. The recurrent connections can learn (store) a set of patterns with a Hebbian learning rule. Later, if these patterns or their noisy versions are used to provoke the network, it settles on the original patterns after several rounds of activation updates on the recurrent weights (recall), thus stored patterns act as attractors. It is of high importance that the existence of such networks has been experimentally validated in the visual cortex of awake mice by optogenetic methods 36. Some versions of the learning rule allow for iterative learning without catastrophic forgetting and enable palimpsest memory. A network with palimpsest memory is able to learn new patterns one-by-one, while sequentially forgetting earlier patterns.
In this paper we describe a model that implements evolution of activation patterns in the brain with the help of attractor networks. We see it as a model of problem solving, which is able to generate new candidate solutions to a problem based on past experiences. Any cognitive problem of the brain is encoded by the activity pattern of neurons. We represent neurons as binary units, being able to continuously maintain or suppress firing. A group of neurons therefore has a binary activation pattern. In our model, the units of evolution are these activation patterns, represented as bitstrings. Attractor neural networks can store patterns stably in form of corresponding attractors and are able to recall them given the appropriate trigger ( Figure 1A). This memory allows for heredity, which is indispensable for Darwinian dynamics (in genetic populations memory is the genotype pool). Attractor neural networks can generate new pattern variants in different ways (corresponding to mutation in a genetic system), see below under Discussion. Owing to memory and pattern generation, the possibility of iterated selection over a population of activation patterns becomes feasible. Our approach thus offers a way to implement hereditary dynamics at a faster scale than what could be provided by structural plasticity (cf. 37). This fast-scale dynamics is missing from Edelmanian Neural Darwinism. See the glossary ( Supplementary File 1) and Supplementary Figure S1 for more clarification about the analogy between our model and genetic evolution.
The patterns represent candidate hypotheses or candidate solutions to a problem, which are evaluated based on a fitness function that measures their goodness as a solution. The best patterns (those with the highest fitness) are selected and copied (with variation) back to the networks, which in turn generate the next generation of patterns ( Figure 1B). Stored patterns (attractors) constitute the long-term memory; output patterns constitute the working memory ( Figure 1B). While pattern generation is a simple recall task, which is only able to reproduce previously learnt patterns, the whole system is able to generate new variants due to noisy recall, spurious patterns (see later), noisy copying of patterns, and iterative learning, thus enabling the evolution of novel solutions.
Methods
Recurrent attractor networks. The basic units in our model are attractor networks. Attractor networks are recurrent neural networks consisting of one layer of units that are potentially fully connected. An attractor neural network produces the same (or highly correlated) output whenever the same input is provided (omitting retraining). The pattern that was learned becomes the attractor point of a new basin of attraction, i.e. it is the prototype pattern that the attractor network should return. Consequently, an attractor with a non-zero sized basin should also return the same output to different input patterns. However, the amount and type of correlation of input patterns that retrieve the same prototype, i.e., the actual structure of the basin of attraction, is hard to assess, let alone visualize. Still, it is safe to assume that most input patterns correlated with the prototype, produce the same output – the prototype itself.
The Hopfield network is a recurrent artificial neural network with binary neurons at nodes and weighted connectivity between nodes, excluding self-connections. According to the usual convention, the two states of binary neurons are +1 and -1. In our model, a neuron fires (state +1) if the total sum of incoming collaterals is greater than 0. Accordingly, the update rule has the following form:
The original Hebbian (covariance) learning rule has the following form (where m is the index of the patterns):
The Hebb rule is both local and incremental. A rule is local if the update of a connection depends only on the information available on either side of the connection (including information coming from other neurons via weighted connections). A rule is incremental if the system does not need information from the previously learnt patterns when learning a new one, thus the update rule uses the present values of the weights and the new pattern. The above update rule performs immediate update of the connection weights (“one shot” process; not a limit process requiring multiple update rounds). The covariance rule has a capacity of 0.14 N 38. However, if the network is trained sequentially, and it reaches its capacity, catastrophic forgetting ensues and the network will be unable to retrieve any of the previously stored patterns, forgetting all it has learnt.
To overcome this problem and to preserve the favorable properties of the covariance rule (one-shot, local and incremental updating) Storkey has introduced a palimpsest learning scheme 39 as follows:
and
Using the above rule, the memory becomes palimpsest ( i.e. new patterns successively replace earlier ones during sequential learning) with a capacity of C = 0.25 N (for details and proper definition of palimpsest capacity, see 39). Palimpsest memory is essential for our training regime, where we trained the networks sequentially with one pattern at a time.
An interesting feature of some autoassociative neural networks is the appearance of spurious patterns. In some cases, the network converges to a pattern different from any other patterns learnt previously. These spurious patterns can be the linear combination of an odd number of stored patterns:
where S is the number of the stored patterns 38. This effect can be thought of as an effective implementation of a neuronal recombination operator.
In what follows, we will treat a pattern of activations over binary neurons as the unit of selection. In the general case, the adaptive fitness of this pattern may be some complicated function that is contextualized by the current inputs and may or may not be a function of the history of inputs and outputs. To keep things simple, we will just consider the fitness of a pattern in terms of its Hamming distance to some target pattern. This means, we are effectively using selectionist and evolutionary schemes to optimize the connections (and ensuing dynamics) to recover a target pattern.
We present a series of simulations graduating from purely selectionist to evolutionary schemes in the face of a changing environment. In the first set of simulations we preclude variation in transmission over generations to examine the sufficiency of dynamical instabilities in supporting a selective process. This involves taking the outputs of one neural network and using them (after selection) to train another network. In the second set of simulations, we consider evolution proper and the transmission of patterns from generation to generation. Here, the patterns that are transmitted are subject to mutations (and selection) to illustrate the efficiency with which optimal patterns (those with high fitness) emerge.
Selection. For the selection experiment, we used N A = 20 structurally identical attractor networks, each consisting of N =200 neurons. These networks were initially trained with random patterns plus a special pattern for each with Storkey’s palimpsest learning rule. The 20 special training patterns were as follows. The worst special pattern was the uniform -1, the best special pattern was the uniform +1. Intermediate special patterns had increasing number of +1-s from the left. Fitness was measured as the relative Hamming similarity from the globally best target O target ( i.e. the proportion of +1-s in the pattern). The worst special pattern was presented only to network #1, the second worst to #2, etc., while the best special pattern (which was the target pattern) was presented to network #20. In this scenario, no further training occurred ( i.e., the dashed arrows on Figure 1 are not there). Assuming that the attractor basins of these patterns overlap among networks ( Figure 2A) the output of one network will be the cue to trigger one or more close special patterns in other networks. The special patterns ensure that there exists a search trajectory leading from the worst to the best pattern, fitness-wise. Starting from any arbitrary initial pattern, if any of the special patterns gets triggered at any time, the system can quickly converge to the optimum – providing the output of each network is delivered to the appropriate network that successively converges on the global optimum.
After initial training, each network received the same random input and generated an output according to its internal attractor dynamics. The output population was evaluated and the best output O best was selected based on its fitness. Noisy copies (with μ I, where μ is the per-bit mutation probability) of O best were redistributed for each network as new input for the next generation. These steps were iterated until fitness reached the theoretical optimum ( i.e. the system found special pattern #20). The crucial assumption for selection to work is continuity, namely the possibility that the output of one attractor of one network could fall in a different attractor basin of another network returning an output that is closer to the global optimum than the input was (see Figure 1 and Figure 2).
Evolutionary optimization on a single-peak landscape. In contrast to purely selective dynamics, in the evolutionary experiment, networks could learn new patterns during the search process. At the start, each network was trained with a different set of random patterns. The fitness of a pattern is defined as the relative (per bit) Hamming similarity between the given pattern and an arbitrarily set globally best target pattern O target. The selection process for the actual best output O best and redistribution of its noisy copies (with μ I = 0.005) for input was the same as before. Most importantly, the mutated versions (with μ T = 0.01) of O best were also used for retraining N T different networks in each generation (see Figure 1): this forms the basis for the Darwinian evolutionary search over attractor networks, as it allows for replication with variation of (learnt) patterns over networks (thin lines in Figure 3).
We have also analyzed how mutation rates affect the speed of evolution. We measured the number of generations needed for the system to find the globally best target pattern. The previously introduced parameters were used and the mutation rates were varied between 10 -4 and 0.01. We ran each simulation 100 times with different random seeds and measured the average number of generations and standard deviation. We found that only μ l affects the speed of evolution with an almost constant minimum in a wide range between approx. 0.003-0.007. Mutation rates lower than 0.003 or higher than 0.007 result in rapidly growing number of generations necessary for the system to find the optimum. μ T does not affect the speed of the evolution considerably (data not shown).
We have compared the search behavior of our system of attractor networks with a simpler model. In this model networks were represented as abstract storage units, which could store exactly C fix patterns ( C fix was set to be close to the actual capacity of networks). When such a storage unit receives an input pattern it simply returns the closest (in Hamming distance) of its stored patterns as output, with additional noise ( μ O = 0.001). The units simulate the almost perfect recall property of attractor networks and effectively approximate attractor behavior. We compared evolution in this simple model with evolution in the system of attractor networks (thick and thin lines in Figure 3).
Optimization in a changing environment. In order to test the effect of memory on successive search, we have implemented a periodically changing selective environment, i.e., we periodically changed the fitness function. The environment alternated between E 1 and E 2, with a stable period length of T E = 2000. Each environmental change reset the global optimum: for this scenario, we assumed a uniform +1 sequence for E 1 and its inverse, uniform -1 for E 2 as global optima, and used the relative Hamming similarity as a fitness measure.
In the first phase of the simulation, networks were allowed to learn in each environment for a total of T nolearn = 12000 generations (three periods per environments). Afterwards, learning was turned off to test the effect of memory. To make sure that the optimal pattern was not simply carried over as an output pattern from the previous environment but was recalled from memory, the input patterns were set to random patterns (instead of inheriting the previous output population) at the start of each new environmental period after T nolearn. This ensures that the population could only maintain high fitness afterwards in an environment if the optimum was stored and could be successfully recalled (see Figure 4). In order to assess the memory of a network, we also measured the distance between the actual best output of the population and the closest one of the set of previously learnt patterns within the same network (as different networks have different training history). A small distance indicates that the network outputs a learnt pattern from memory ( i.e. recalls it) instead of a spurious pattern.
For this scenario, we introduced a different selection method (also used in the next section). Each network in the population produces an output according to its internal attractor dynamics and the input it received from the previous generation. From all output sequences one was randomly chosen and mutated ( µ R = 1/ N per bit mutation rate). If the mutant had a higher fitness than the worst of the output pool, the worst pattern was replaced by it ( elimination of the worst). Furthermore, the superior mutant was used to train N T number of different networks. Lastly, the resulting output population is shuffled and fed to the networks as input in the next generation (except when the environment changes and input is reset externally).
Optimization on a difficult landscape. To investigate the applicability of this optimization process, we adopted a complex, deceptive landscape with scalable correlation, and also modified the selection algorithm introduced above. We used the general building-block fitness (GBBF) function of Watson and Jansen 40. According to the GBBF function, each sequence of length N is partitioned into blocks of uniform length P, so that N = P B ( P, B ∈ Z +) where B is the number of blocks. For each block, L arbitrarily chosen subsequences are designated as local optima, with randomly chosen but higher-than-average subfitness values. The overall fitness F( G) of a pattern G (“genotype”) is as follows:
where f( g i) is the fitness contribution of the ith block in the pattern, t j is the j th local optimum of length P (all L different optima are the same for each block in our experiments) with subfitness value w j > 1, and d is the Hamming distance. Consequently, this landscape has many local optima, a single global optimum and a highly structured topology. Furthermore, since there are no nonlocal effects of blocks, each block can be optimized independently, favoring a metapopulation search.
Accordingly, in this experiment, we introduced multiple populations of attractor networks. Each population of N A attractor neural networks forms a deme and N D demes are arranged in a 2D square lattice of Moore neighborhood (all of the eight surrounding demes are considered neighbors). Demes might accept output sequences from neighboring demes with a low probability p migr per selection event; this slow exchange of patterns can provide the necessary extra variability for recombination. These demes correspond to the groups of columns in the brain.
Networks in the deme are the same as those used in previous experiments. However, selection is modeled in a different way, similar to the selective dynamics outlined in 40. In turn, we only give a brief description. Given a deme, each network produces an output according to its internal attractor dynamics and the input it received from the previous generation. Output sequences are pooled and either one or two is randomly chosen for mutation or recombination, respectively ( i.e. no elitist selection). With probability p rec, two-point recombination is performed in the two selected partners, with 1- p rec probability, a single selected sequence is mutated, with µ R = 1/ N per bit mutation rate. With p migr probability, the recombinant partner is chosen from another neighboring deme instead of the focal one. Next, the output(s) of recombination or mutation are calculated: if the resulting sequence (any of the two recombinants or the mutant) has a higher fitness than the worst of the output pool, it is replaced by the better one ( elimination of the worst). Furthermore, the superior mutant or recombinant was used to train N T number of different networks within the deme. Lastly, the resulting output population is shuffled and fed to the networks as input in the next generation. Each deme is updated in turn according to the outlined method; a full update of all networks in all demes constitutes a generation ( i.e. a single time step).
The GBBF landscape was set up identically to the test case in 40, as follows. For each block uniformly, two target sequences of length P, T 1 and T 2, were appointed. T 1 is the uniform plus-one sequence T 1 = {+1} P and T 2 is alternating between -1 and +1 ( T 2 = {-1, +1} P/2). According to the fitness rule (Equation 5–Equation 6 in 40 and Equation 1– Equation 3 above), the best subfitness of each block in a sequence can be calculated and the sum of all the subfitness values is the fitness of the global optimum sequence. Thus for sake of simplicity, we used relative fitness values with the global optimum (the uniform +1 sequence) having maximal fitness 1. The sequence(s) with lowest fitness always have a nonzero value.
The source code of all models and data presented in this paper is freely available as a supplement to this paper.
Results
Selection. We should distinguish between two processes: (i) search without learning among the stored patterns to find the best available solution ( i.e., selection without step 5 on Figure 1A), and (ii) search with learning: retraining one or more networks with the selected and mutated patterns ( Figure 1A with step 5). The first is a purely selectionist approach because it cannot generate heritable variants, while the second implements Darwinian evolution because learning changes the output behavior of the networks, thus they generate new patterns. First, we analyze the strictly selectionist version, and then the evolutionary version of the model.
In the selectionist version we pre-trained each network with a random set of patterns (excluding the target pattern) and started by provoking them with a different random input. Each network produced an output pattern according to its own attractors and then the best pattern was selected. This pattern was used in turn to provoke the networks in the next generation, and so on. This search has found among all the available stored (pre-trained) patterns the one with the highest fitness; it could not find the global optimum, as the networks were not pre-trained with it and there was no way for new variants to survive in this simulation.
Next, we specifically composed the sets of pre-training patterns: each network was pre-trained with random patterns as before but also with one special pattern. This set of special patterns (in which individual patterns can be ordered according to gradually increasing fitness values) delineate a route to the optimum through overlapping basins of attractors in different networks (see Figure 2A) so that we can test whether in this simplified case the algorithm converges quickly to the optimum. The first population was initiated with the special pattern that was farthest from the optimum. We have found that the selected output gets closer to the optimum in each generation, but the optimization process is saltatory: it skips over many intermediate neighboring special patterns (and thus networks). This is due to the fact that attractor basins of neighboring special patterns were highly overlapping. For example, in Figure 2A, the stored special pattern of network #3 is in the basins of stored special patterns of networks #4–#11, and since the stored pattern of network #11 is closest to the optimum, networks #4–#10 were skipped. A typical sequence of networks generating the actual best output is: #3, #11, #17 and #20 (of 20 networks; for actual parameters, see Figure 2A).
Evolution. Learning new patterns as attractors ( Figure 2B) allows networks to adapt to the problem and perform evolutionary search. The results of the evolutionary experiments clearly prove that a population of attractor networks can implement evolutionary search in problem spaces of different complexity ( i.e. different levels of correlation and deceptiveness).
Evolution on a simple fitness landscape. In this scenario, neither the global optimum nor a route toward it is assumed to pre-exist in the system as in the selectionist experiments: networks are pre-trained only with random patterns. Even under these stringent conditions, we have found that the system can converge to the global optimum, and this convergence is robust against a wide range of mutation rates. Our simplified abstract model, which always returns the stored prototype that is closest to the input, behaves qualitatively the same way (see Figure 3). The speed of convergence to the optimum is mainly affected by the number of retrained networks ( Figure 3): as we increase the number of networks that are retrained we find a faster fitness increase, albeit with diminishing returns. Mutation has an optimal range in terms of the speed of evolution. On one hand, if mutation rate is too low evolution slows down, because there is not enough variation among patterns. On the other hand, if mutation rate is too high it hinders evolution as the offspring is too dissimilar to the parent and cannot exploit the attractor property of the system. When mutation rate is zero, the source of variation is only the probabilistic input-output behavior of the networks due to their asynchronous update and the appearance of spurious patterns when the input is too far from the stored patterns.
While the attractor networks have memory, due to the monotonic, single-peak nature of the fitness landscape there is no need to use it: the system works almost equally well if the networks only store the last training pattern ( i.e., weights are deleted before each learning event). Next, we present experiments where both the attractor property and the palimpsest memory of the networks are used.
Evolution in a changing environment. In this experiment we alternated two environments: in every 2000th generation the target pattern (the optimum), against which fitness was measured, was changed. From an evolutionary point of view, this can be perceived as a changing environment, whereas from a cognitive point of view, this procedure simulates changing task demands. Figure 4 shows that the system found and learnt the optima of each of the two environments separately. Then, after generation 12000, we switched off learning. The fact that networks are nevertheless able to recall the target pattern right after the environmental change proves that they use previously stored memories. After we switched off learning, we used random patterns to provoke networks at the first generation of each new environment. A single network that can recall the optimum from the random input is enough to produce a correct output that is amplified by selection for the next generational input, ultimately saturating the population with optimal patterns. This experiment effectively proves that a system of attractor networks can reliably recall earlier stored solution patterns, therefore solving the problem faster in an alternating environment than a system without long-term memory.
Evolution on a difficult fitness landscape. The previous evolutionary experiment (where search was on a single-peak fitness landscape with a single population of networks) is a proof of principle of the effectiveness of our evolutionary algorithm. In order to assess the capacity of population search of attractor networks, we introduce a considerably harder fitness landscape with higher dimensionality, where the deceptiveness of the problem can be tuned. The GBBF fitness landscape of 40 provides a method to easily generate scalable and complex landscapes with many deceptive local optima. The complexity of the problem requires the introduction of multiple interacting populations of networks. Though explicit spatial arrangement of the networks is not required to solve the problem, we have nevertheless included it in our implementation to imitate real spatial arrangement of neurons in the brain. Locality allows the exchange of information among neighboring populations ( i.e. recombination) that is essential to solve the GBBF problem (or similar deceptive problems) in a reasonable time.
We have investigated the performance of search in a metapopulation with different problem sizes (pattern lengths; see Figure 5). Results indicate that despite the vastness of the search space, the metapopulation is always able to converge to the global optimum, given enough time. The most complex landscape of 100-bit patterns is of size 2 100 with one global optimum and a huge number of local optima. The metapopulation consists of 10 5 neurons (100 populations of 10 networks each with 100 neurons per network) and can find the single global optimum in ~10 4 time steps. The limit of further increasing the problem size is in the computational capacity of our resources.
Discussion
Summary of results. Attractor networks can be used to implement selection, replication and evolutionary search. As a proof of principle, we showed that attractor networks find the global optimum in a purely selectionist model ( i.e. without learning) if they are pre-trained with attractors that overlap in their basins and lead to the optimum. The population can effectively select over the standing variation of all stored patterns and find the trajectory to the (single) peak of the fitness landscape (see Figure 2B). Furthermore, if learning is allowed during search, the relative frequency of good patterns (those closer to the optimum) can be increased by re-training networks with such patterns, so that they are stored in the long-term memory in more copies ( Figure 3). Overwriting an older memory trace with a new pattern corresponds to a copying operation, potentially with mutations. A particularly interesting aspect of a population of attractor networks under the given coupling is that even if learning is not erroneous, the Lamarckian nature of inheritance of patterns (as output → memory → output; see Figure 1B) means that there is room for heritable variation to emerge at other stages of the cycle, thus implementing Darwinian dynamics.
The explicit benefit of memory manifests in a periodically changing environment. In a single, stable environment, memory is not very useful because the attractor property acts against exploring variation, and networks might be even slower than gradient hill-climbers ( i.e. searchers who generate mutant output blindly and only take a step on the landscape if the output is better than the input). However, in a periodically changing environment, attractor networks with memory are able to recall the actual global optimum if they have already encountered the given environment in the past and stored its optimum; hill-climbers or naïve attractor networks lacking the necessary memory would have to perform the search all over again. Attractor network can recall the appropriate pattern even if there is no initial cue for the population to know which environment it is in. A network with memory can complete a partial cue and thus can recall the global optimum. After learning has ceased, it is enough to only have a few networks in the population that can recall the optimum from random cues that are accidentally close to the actual optimum (maximal fitness is immediately 1 in the population, see green curve in Figure 4). Selection then simply amplifies these optimal output patterns in the population (as output is fed back as input in the next generation) until all networks receive optimal patterns as input. At that point, average fitness also reaches the maximum ( Figure 4, blue curve). A network without memory would have to search for the new optima in each environment over and over again, finding the whole uphill path on the landscape.
We also proved that a metapopulation of attractor networks can successfully optimize on a complex, correlated landscape of higher dimensionality. This is a notoriously hard optimization problem (cf. 40) as hill-climbers can easily get stuck in deceptive local optima. The spatial organization of attractor networks resembles the spatial organization of neural structures of the cortex and it allows parallel optimization of subproblems. By this independent optimization of solution parts, local populations can exchange successful partial solutions with each other and form recombinants that are necessary to solve such complex problems in reasonable time.
Evolutionary context. We have chosen attractor networks to demonstrate Darwinian neurodynamics because (i) the search for existing solutions uses the same architecture for generating, testing and storing novel ones; (ii) stored patterns help evolutionary search by readily employing related past (learnt) experience, and (iii) the particular choice of Storkey’s model naturally results in some new recombinant patterns. This is an important point because, as we know from population genetics 41, recombination speeds up the response to selection by creating new variants 42. For a non-exhaustive mapping between population genetics ad Darwinian neurodynamics (see Supplementary File 1 and Supplementary Figure S1).
Our choice of implementation of attractor networks, following Storkey’s model 39, is based on three important aspects: (i) it has palimpsest memory, so that it can learn new and forget old patterns without catastrophic forgetting, as happens in Hopfield and other networks; (ii) its attractors are roughly of equal size and are well-structured according to a Hamming distance measure, and (iii) unlike most other neural networks, it is able to store correlated patterns. The downside is that these networks require full connectivity, which is neuronally unrealistic. However, its functional properties reflect well what we know of long-term memory in the brain, which is enough for a proof or principle of an evolutionary implementation of neuronal function. To our knowledge no model exists in the literature that could satisfy all requirements above and that, at the same time, work with diluted connectivity 43.
It is important to clarify that the units of evolution in our proposed Darwinian neurodynamics system are the activation patterns: they are copied potentially with errors, selected and amplified. However, patterns live in two stages: in the working memory and in the long-term memory ( Figure 1). This implies different inheritance methods (routes to pass on information) from what is expected in a purely genetic system. Changed attractor prototypes imply changed outputs, just like a mutated genotype implies a different phenotype. However, in our proposed system, changes made to output patterns (by mutation) can also be “inherited” by the stored attractor prototypes via learning. Furthermore, there is another difference in the dynamics, explained in turn.
Darwinian evolution is often described as a parallel, distributed search on a fitness landscape 8. The population, as an amoeba, climbs the peaks of the landscape via iterative replication and selection in successive generations. The attractors, however, impose a different mode of evolution, because they simply return the prototype pattern closest to the input, even if it is less fit than the input pattern itself. Consequently, attractor networks work against variability and slow down hill-climbing. However, attractor networks resist fitness increase only half of the time on average; the other half of the time they effectively push inferior patterns uphill in the fitness landscape at a speed much higher than that expected for the same (reduced) amount of genetic variation. Consequently, attractor networks can facilitate evolution 21.
We stress the importance of evolutionary combinatorial search. In cases where ab initio calculation of molecular functionality is impossible, artificial selection on genetic variants has proven to be an extremely useful method to generate molecular solution to functional problems, as experiments on the generation of catalytic RNA molecules (ribozymes) illustrate (see 44 for a recent review). By the same token when a brute force numerical calculation of a combinatorial functionality problem is impossible for the brain, given the adequate architecture it could (and we suggest it does) use an evolutionary search mechanism, as shown in this paper.
Implementation in the brain. It is of primary importance that all the components in our ‘algorithmic diagram’ ( Figure 1) can be implemented by mechanisms known in the brain. It is likely that the cortical hypercolumn 45 behaves like an (at least one) attractor network. The reciprocal connections between the long-term and working memory networks are assumed to be like those in Figure 3 in 46. We propose that, first, the reinforcing signal from the basal ganglia via the thalamus keeps active the good candidate pattern solutions in the rewarded auto-associative network and, second, that the latter sends a copy of the active pattern to (unconscious) working memory for eventual redistribution. When there is an increase in the quality of a solution (fitness increase) or when a local or global optimum is reached, the central executive elicits the transmission of a copy of the solution to the conscious memory 47.
Our proposed mechanism relies on information transmission of patterns between cortical groups and relevant subcortical structures with variation, and in this way it differs from all previous models. A discussion of timescales is in order. Without learning newly generated variants, the selective mode would require about the same time as suggested by 48 as the “cognitive cycle” (based on data), but without perception at the beginning of the cycle, i.e. it would be in the 160–310 ms range. In the evolutionary mode, learning of new variants is required which would take more time. A conservative estimate for the reliable expression of changed synaptic weights is between seconds to minutes 49, 50. The second scale would allow several dozen cycles/generations per minute—a very good number in artificial selection experiments. A more accurate estimate of timing will require a fully neuronal model of our proposed mechanism.
Copying of information is necessary for evolutionary processes: this is where previous approaches 22– 24, 51 have been too artificial. There are four well known instances where scholars invoke copying of information in the brain: (i) the transfer of information from short to long-term memory 35, 52, 53, (ii) the transfer of information from specialized cortical areas to working memory 54 or to the global workspace 55, 56 and, finally (iii) the possible copying of patterns from working memory to conscious processing 57. Undoubtedly, all these approaches require accurate information transfer between different brain components 58.
In the context of Darwinian neurodynamics, inheritance is replaced by learning. De Vladar and Szathmáry 21 showed that the rate of “mutation” in spiking diminishes as there is convergence to the optimum. In other words, as the population of attractors approaches the solution, the rates of mutation decrease, since the need to generate new hypotheses decreases. This type of adaptive mutation rate results in what is known as “evolvability” 59 and in Darwinian neurodynamics it comes for free: while selection ensures that more accurate hypotheses are implemented, learning through the strengthening of connections between simultaneously active neurons decreases the rate of variability. Hence, the crucial effect in neurodynamics comes from the built-in mechanism that effectively implements second-order selection by adapting the mutation rate according to the quality of the standing hypotheses.
There are three sources of variation in our system: (i) due to the finite size of our networks and asynchronous update of the neurons, the output patterns show some variation even if a network is repeatedly provoked by the same input pattern (noisy recall), (ii) acknowledging the noisiness of neuronal transmission we introduce variation akin to mutations when patterns are transmitted among the blocks of the model (we mutate the patterns with mutation rate μ), and finally (iii) we have realized that “spurious patterns” 38 emerge as by-products of previous training patterns and they might act as (new) recombinants of learnt patterns that facilitate the evolutionary search.
The non-conscious or implicit working memory, which has received considerable attention lately 47, 57 is crucial for our proposed mechanism. Irrespective of whether working memory overlaps with the conscious domain 60 or not 57 (in the latter case a ‘conscious copy’ must be sent from working memory to conscious access), the important factor is that the bound on the number of patterns that can be held in the unconscious part of the working memory is larger than that of the conscious working memory 60. In other words, our mechanism suggests that the total storage capacity of the unconscious network population is much higher than that of the conscious one. Crucially, there is support for this requirement: there is evidence that the central executive function of working memory is not restricted to the conscious domain either 47. The relatively large capacity of (unconscious) working memory can hold not one, but several patterns selected by the cortex-striatum-basal ganglia loop. This type of selection can be realized by a winner-share-all (WSA) mechanism 61.
The latter point requires special attention. The reader is referred to the recent review by 62 on models of action selection and reinforcement learning. We wish to make a few critical points in this regard. First, as we are considering problem solving that unfolds its capacity online, there is no reason to select one pattern, since the interim patterns are not alternative actions but only candidate solutions. They can be turned into actions sometimes during, or only at the very end, of the evolutionary search. Weak lateral inhibition within the evaluation mechanism enhances value differences in selection, but a single winner is not selected 61, 62. Second, parallelism of the evolutionary approach loses considerable power if the evaluations are not done in parallel, and if poor solutions cannot be replaced by better solutions in the storage. (In a subsequent study we shall show that the number of parallel evaluations are allowed to be considerably smaller than population size but also that purely serial evaluation of candidates is a killer). Third, it is perfectly possible that the WSA part is implemented by the cortex rather than the striatum (cf. 62): we are agnostic on this point for the time being. (Admittedly that option would require a different version of the full model). Fourth, we maintain that parallel survival of a number of candidates should happen, and the mechanism for this might have evolved with selection for complex (offline) problem solving. Fifth, it is well possible that WSA is gradually reduced towards WTA (one winner takes all) during evolutionary search in the brain: this would also guard against premature convergence early and fast convergence towards the end of the search.
To sum up, we have seen that a process analogous to natural selection can be rendered into a neuronal model employing known neurophysiological mechanisms. Now we discuss relations to some other publications and outline future work.
Related work. Several examples show that evolution with neurodynamics can be more powerful than either of the components alone. Fernando et al. 25 proved that the path evolution algorithm – which includes both elements of structural plasticity 63, 64 and iterative generation of variation – is more powerful in several regards than classical genetic algorithms. Fernando et al. have also shown that replication combined with Hebbian learning is more powerful than classical natural selection in a model of mechanistic copying of binary activity 22. De Vladar and Szathmáry provided proof that the synergy between selection and learning results in increased evolvability; also they pointed out that synaptic plasticity helps escaping impasses and build circuits that are tailored to the problem 21. Finally, in a recent model Fernando and his colleagues have used autoencoders for the generation of “neuronal genotypes” 65. Since autoencoders produce compressed representations of the input, we expect them to successfully replace the identity function ( i.e. bit by bit copying, as in DNA replication). Indeed, applying this neural component within the context of a genetic algorithm turned out to be rewarding.
Unless the envisaged information transfer is accurate enough in space and time, the evolutionary dynamics breaks down. Similar to genetic evolution, where the error threshold 17 had to be raised before long informative genomes could arise by evolving adaptations for more accurate copying 66, in the neuronal context the element of accuracy was raised by Adams 18. In his “Hebb and Darwin” paper Adams talks about synaptic replication and synaptic mutation as important ingredients for a Darwinian view of the brain. Synaptic replication means either the strengthening of an existing synapse, or the making of a new synapse between two neurons that already have one synapse between them. Adams’ is an important insight: evolutionary dynamics does not need copying for selection (scoring or strengthening is enough), but it needs copying with errors to test the new variants against the others. Synaptic mutation happens when a neuron grows a synapse towards a neighboring neuron with which previously it had no contact. Interestingly, these thoughts preceded the burst of interest in structural synaptic plasticity (SSP, 63, 64). Following his expansion-renormalization model for SSP 67 Kilgard observes that when SSP is used for learning something new, this could be regarded as a Darwinian mechanism 37, as it generates and tests variations in successive rounds, based on what is already there (unlike the original models of “neural Darwinism”). Kilgard’s mechanism has not been formalized yet (although see 21), but the path evolution model 25 bears some relationship to it.
We share the view of Eliasmith 54 that the cortex/basal ganglia/thalamus/cortex loop plays a crucial role not only in elementary action selection but also in symbolic reasoning. We conjecture that non-primate animals (in particular mammals and birds) employ the same (or at least an analogous) loop in order to retrieve old and to innovate new solutions, in a similar way as we have shown using our elementary model.
Another view to which we feel strongly related to is that of Bayesian models that advocate “theory learning as stochastic search in a language of thought” 27. We are reasonably confident that we have found a candidate mechanism for the search process. If true, the rugged learning landscape in Figure 3 of 27 can be directly interpreted as the fitness landscape of our neuro-evolutionary model. A task for the future is to work out the explicit relations in detail. We note again the formal link between Bayesian inference and evolutionary selection 31, 32 mentioned in the Introduction. This has further implications for the Bayesian brain, which indicate that our neuro-evolutionary framework is not only a metaphor, but works in a way that is at least equivalent to an evolutionary search. From the Bayesian perspective, the brain weights a series of candidate hypotheses according to their likelihood. From the evolutionary perspective, this update is interpreted precisely as the action of selection on the standing round of hypotheses, with the next round or generation given by the posterior probability of these hypotheses. Moreover, the selective likelihood has been shown in previous models of population genetics to be analogous to a free energy function that is minimized during evolution 68– 71. These models incorporate mechanisms for generating variability, a component that seems to be absent in previous theories of Bayesian brains. But as evolutionary models show, it is utterly essential to account for mechanisms that create new variation, or, in the neurodynamical context, new hypotheses.
Our mechanism ( Figure 1) could in principle implement, with appropriate modifications, an estimation of distribution algorithm (EDA). The population-based incremental learning (PBIL) algorithm consists of the following steps 72: (i) Generate a population from the probability vector; (ii) Evaluate and rank the fitness of each member; (iii) Update the probability vector based on the elite individual; (iv) Mutate the probability vector; (v) Repeat steps (i–iv) until a finish criterion is met. EDA can work better than copying algorithms making it an interesting line to pursue.
Future work will be to link our recurrent model with the feedforward autoencoder model of Fernando et al. 65, since the latter can generate interesting genotypes (better substrates for selection) due to the emerging compressed representations of the inputs.
As two experts aptly remark: “The Bayesian brain falls short in explaining how the brain creates new knowledge” ( 73 p. 9). We suggest that neuronal evolutionary dynamics might serve as a remedy.
Data and software availability
Zenodo: István Zachar at doi: 10.5281/zenodo.154113 74; András Szilágyi at doi: 10.5281/zenodo.803989 75.
The algorithm described in the paper is also available on GitHub at https://github.com/IstvanZachar/Neurodynamics and at https://github.com/andszilagyi/darwinianneurodyn.
Acknowledgements
We are thankful to Luc Steels, Chrisantha Fernando, Mauro Santos, and Thomas Filk for useful comments and discussions. We are also grateful for our reviewers, László Acsády, Stuart J. Edelstein and Karl Friston, for their invaluable comments and suggestions.
Funding Statement
The research leading to these results has received funding from the European Union Seventh Framework Program (FP7/2007-2013) under grant agreement numbers 308943 (INSIGHT project) and 294332 (EvoEvo project). It was also supported by the project GINOP-2.3.2-15-2016-00057 (In the light of evolution).
The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
[version 2; referees: 3 approved]
Supplementary material
Supplementary File 1. Glossary of terms, summarizing the important evolutionary and neurobiological concepts and explaining the analogy between genetic and neuronal Darwinism.Supplementary Figure S1. Analogy between genetic evolution and the attractor network-based model of Darwinian Neurodynamics. A) genetic evolution; B) attractor network-based model of Darwinian Neurodynamics. The encoded information (genotypes or attractor networks) express a trait (a phenotypic component in case of genetics; an output pattern in case of the model). Selection acts over these expressed traits, and the new generation’s distribution of encoded information depends on this selection process. In genetic evolution, the genotype is inherited at generation t+1 with some mutations. In our neurodynamical model, some attractor networks (and attractors) are left unaffected, and some are altered by training them with one of the selected and mutated patterns. For more information about the analogy of Darwinian inheritance systems, see 76.
References
- 1. Changeux JP: Neuronal man: The biology of mind. Princeton, NJ: Princeton University Press; Translated by Garey, L. 1985. Reference Source [Google Scholar]
- 2. Changeux JP, Courrége P, Danchin A: A theory of the epigenesis of neuronal networks by selective stabilization of synapses. Proc Natl Acad Sci U S A. 1973;70(10):2974–2978. 10.1073/pnas.70.10.2974 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Edelman GM: Neural Darwinism. The theory of neuronal group selection. New York: Basic Books;1987. Reference Source [DOI] [PubMed] [Google Scholar]
- 4. Williams RW, Rakic P: Elimination of neurons from the rhesus monkey’s lateral geniculate nucleus during development. J Comp Neurol. 1988;272(3):424–436. 10.1002/cne.902720310 [DOI] [PubMed] [Google Scholar]
- 5. O’Leary DD: Development of connectional diversity and specificity in the mammalian brain by the pruning of collateral projections. Curr Opin Neurobiol. 1992;2(1):70–77. 10.1016/0959-4388(92)90165-H [DOI] [PubMed] [Google Scholar]
- 6. Rabinowicz T, de Courten-Myers GM, Petetot JM, et al. : Human cortex development: estimates of neuronal numbers indicate major loss late during gestation. J Neuropathol Exp Neurol. 1996;55(3):320–328. 10.1097/00005072-199603000-00007 [DOI] [PubMed] [Google Scholar]
- 7. Miller-Fleming TW, Petersen SC, Manning L, et al. : The DEG/ENaC cation channel protein UNC-8 drives activity-dependent synapse removal in remodeling GABAergic neurons. eLife. 2016;5: pii: e14599. 10.7554/eLife.14599 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Fernando C, Szathmáry E, Husbands P: Selectionist and evolutionary approaches to brain function: a critical appraisal. Front Comput Neurosci. 2012;6:24. 10.3389/fncom.2012.00024 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Quartz SR, Sejnowski TJ: The neural basis of cognitive development: a constructivist manifesto. Behav Brain Sci. 1997;20(4):537–556; discussion 556-96. 10.1017/S0140525X97001581 [DOI] [PubMed] [Google Scholar]
- 10. Bandeira F, Lent R, Herculano-Houzel S: Changing numbers of neuronal and non-neuronal cells underlie postnatal brain growth in the rat. Proc Natl Acad Sci U S A. 2009;106(33):14108–14113. 10.1073/pnas.0804650106 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Caroni P, Donato F, Muller D: Structural plasticity upon learning: regulation and functions. Nat Rev Neurosci. 2012;13(7):478–490. 10.1038/nrn3258 [DOI] [PubMed] [Google Scholar]
- 12. Bernardinelli Y, Nikonenko I, Muller D: Structural plasticity: mechanisms and contribution to developmental psychiatric disorders. Front Neuroanat. 2014;8:123. 10.3389/fnana.2014.00123 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Tetzlaff C, Kolodziejski C, Markelic I, et al. : Time scales of memory, learning, and plasticity. Biol Cybern. 2012;106(11–12):715–726. 10.1007/s00422-012-0529-z [DOI] [PubMed] [Google Scholar]
- 14. Maynard Smith J: The problems of biology. USA: Oxford University Press;1986. Reference Source [Google Scholar]
- 15. Maynard Smith J: Genes, memes, and minds. The New York Review of Books.1995;42(19):46–48. Reference Source [Google Scholar]
- 16. Kimura M: The Neutral Theory of Molecular Evolution. Cambridge: Cambridge University Press;1983. 10.1017/CBO9780511623486 [DOI] [Google Scholar]
- 17. Eigen M: Selforganization of matter and the evolution of biological macromolecules. Naturwissenschaften. 1971;58(10):465–523. 10.1007/BF00623322 [DOI] [PubMed] [Google Scholar]
- 18. Adams P: Hebb and Darwin. J Theor Biol. 1998;195(4):419–438. 10.1006/jtbi.1997.0620 [DOI] [PubMed] [Google Scholar]
- 19. Calvin WH: The brain as a Darwin Machine. Nature. 1987;330(6143):33–34. 10.1038/330033a0 [DOI] [PubMed] [Google Scholar]
- 20. Calvin WH: The cerebral code: thinking a thought in the mosaics of the mind.Cambridge, MA: MIT Press.1996. Reference Source [Google Scholar]
- 21. de Vladar HP, Szathmáry E: Neuronal boost to evolutionary dynamics. Interface Focus. 2015;5(6):20150074. 10.1098/rsfs.2015.0074 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Fernando C, Goldstein R, Szathmáry E: The neuronal replicator hypothesis. Neural Comput. 2010;22(11):2809–2857. 10.1162/NECO_a_00031 [DOI] [PubMed] [Google Scholar]
- 23. Fernando C, Karishma KK, Szathmáry E: Copying and evolution of neuronal topology. PLoS One. 2008;3(11):e3775. 10.1371/journal.pone.0003775 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Fernando C, Szathmáry E: Natural selection in the brain. In: Glatzeder B, Goel V, von Müller A, editors. Towards a theory of thinking. vol. 5 of On thinking.Berlin/Heidelberg: Springer-Verlag;2010;291–322. 10.1007/978-3-642-03129-8_20 [DOI] [Google Scholar]
- 25. Fernando C, Vasas V, Szathmáry E, et al. : Evolvable neuronal paths: a novel basis for information and search in the brain. PLoS One. 2011;6(8):e23534. 10.1371/journal.pone.0023534 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Kemp C, Tenenbaum JB: The discovery of structural form. Proc Natl Acad Sci U S A. 2008;105(31):10687–10692. 10.1073/pnas.0802631105 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Ullman TD, Goodman ND, Tenenbaum JB: Theory learning as stochastic search in the language of thought. Cognitive Dev.The Potential Contribution of Computational Modeling to the Study of Cognitive Development: When, and for What Topics?2012;27(4):455–480. 10.1016/j.cogdev.2012.07.005 [DOI] [Google Scholar]
- 28. Börgers T, Sarin R: Learning through reinforcement and replicator dynamics. J Econ Theory. 1997;77(1):1–14. 10.1006/jeth.1997.2319 [DOI] [Google Scholar]
- 29. Niekum S, Barto AG, Spector L: Genetic programming for reward function search. IEEE Trans Auton Ment Dev. 2010;2(2):83–90. 10.1109/TAMD.2010.2051436 [DOI] [Google Scholar]
- 30. Sutton RS, Barto AG: Introduction to reinforcement learning. 1st ed. Cambridge, MA, USA: MIT Press;1998. Reference Source [Google Scholar]
- 31. Harper M: The replicator equation as an inference dynamic. ArXiv e-prints.2009. Reference Source [Google Scholar]
- 32. Shalizi CR: Dynamics of Bayesian updating with dependent data and misspecified models. Electron J Statist. 2009;3:1039–1074. 10.1214/09-EJS485 [DOI] [Google Scholar]
- 33. Campbell JO: Universal Darwinism As a Process of Bayesian Inference. Front Syst Neurosci. 2016;10:49. 10.3389/fnsys.2016.00049 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Hopfield JJ: Neural networks and physical systems with emergent collective computational abilities. Proc Natl Acad Sci U S A. 1982;79(8):2554–2558. 10.1073/pnas.79.8.2554 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Rolls ET, Treves A: Neural networks and brain function. Oxford, New York: Oxford University Press;1998. 10.1093/acprof:oso/9780198524328.001.0001 [DOI] [Google Scholar]
- 36. Carrillo-Reid L, Yang W, Bando Y, et al. : Imprinting and recalling cortical ensembles. Science. 2016;353(6300):691–694. 10.1126/science.aaf7560 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Kilgard MP: Harnessing plasticity to understand learning and treat disease. Trends Neurosci. 2012;35(12):715–722. 10.1016/j.tins.2012.09.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38. Hertz J, Palmer RG, Krogh AS: Introduction to the Theory of Neural Computation.1st ed. Perseus Publishing;1991. 10.1063/1.2810360 [DOI] [Google Scholar]
- 39. Storkey AJ: Efficient covariance matrix methods for Bayesian gaussian processes and Hopfield neural networks.Imperial College, Department of Electrical Engineering, Neural System Group;1999. Reference Source [Google Scholar]
- 40. Watson RA, Jansen T: A building-block royal road where crossover is provably essential. In: Proceedings of the 9th Annual Conference on Genetic and Evolutionary Computation GECCO ’ 07. New York, NY, USA: ACM;2007;1452–1459. 10.1145/1276958.1277224 [DOI] [Google Scholar]
- 41. Maynard Smith J: The Evolution of Sex. Cambridge University Press;1978. Reference Source [Google Scholar]
- 42. Maynard Smith J: The units of selection. Novartis Found Symp. 1998;213:203–11; discussion 211–7. [PubMed] [Google Scholar]
- 43. Sompolinsky H: Computational neuroscience: beyond the local circuit. Curr Opin Neurobiol.Theoretical and computational neuroscience.2014;25:xiii–xviii. 10.1016/j.conb.2014.02.002 [DOI] [PubMed] [Google Scholar]
- 44. Müller S, Appel B, Balke D, et al. : Thirty-five years of research into ribozymes and nucleic acid catalysis: where do we stand today? [version 1; referees: 2 approved]. F1000Res. 2016;5: pii: F1000 Faculty Rev-1511. 10.12688/f1000research.8601.1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45. Mountcastle VB: Modality and topographic properties of single neurons of cat’s somatic sensory cortex. J Neurophysiol. 1957;20(4):408–434. [DOI] [PubMed] [Google Scholar]
- 46. Rolls ET: Attractor networks. Wiley Interdiscip Rev Cogn Sci. 2010;1(1):119–134. 10.1002/wcs.1 [DOI] [PubMed] [Google Scholar]
- 47. Soto D, Silvanto J: Reappraising the relationship between working memory and conscious awareness. Trends Cogn Sci. 2014;18(10):520–525. 10.1016/j.tics.2014.06.005 [DOI] [PubMed] [Google Scholar]
- 48. Madl T, Baars BJ, Franklin S: The timing of the cognitive cycle. PLoS One. 2011;6(4):e14803. 10.1371/journal.pone.0014803 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49. Gustafsson B, Asztely F, Hanse E, et al. : Onset Characteristics of Long-Term Potentiation in the Guinea-Pig Hippocampal CA1 Region in Vitro. Eur J Neurosci. 1989;1(4):382–394. 10.1111/j.1460-9568.1989.tb00803.x [DOI] [PubMed] [Google Scholar]
- 50. Hirsch JC, Crepel F: Use-dependent changes in synaptic efficacy in rat prefrontal neurons in vitro. J Physiol. 1990;427(1):31–49. 10.1113/jphysiol.1990.sp018159 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51. Fernando C, Szathmáry E: Chemical, neuronal, and linguistic replicators.In Pigliucci M, Müller G, editors. Cambridge, MA: MIT Press;2010;209–250. 10.7551/mitpress/9780262513678.003.0009 [DOI] [Google Scholar]
- 52. Nadel L, Land C: Commentary-Reconsolidation: Memory traces revisited. Nat Rev Neurosci. 2000;1(3):209–212. 10.1038/35044572 [DOI] [PubMed] [Google Scholar]
- 53. Nadel L, Moscovitch M: Memory consolidation, retrograde amnesia and the hippocampal complex. Curr Opin Neurobiol. 1997;7(2):217–227. 10.1016/S0959-4388(97)80010-4 [DOI] [PubMed] [Google Scholar]
- 54. Stewart TC, Choo X, Eliasmith C: Symbolic Reasoning in Spiking Neurons: A Model of the Cortex/Basal Ganglia/Thalamus Loop. In: 32nd Annual Meeting of the Cognitive Science Society;2010. Reference Source [Google Scholar]
- 55. Dehaene S, Kerszberg M, Changeux JP: A neuronal model of a global workspace in effortful cognitive tasks. Proc Natl Acad Sci U S A. 1998;95(24):14529–14534. 10.1073/pnas.95.24.14529 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56. Shanahan M: A spiking neuron model of cortical broadcast and competition. Conscious Cogn. 2008;17(1):288–303. 10.1016/j.concog.2006.12.005 [DOI] [PubMed] [Google Scholar]
- 57. Jacobs C, Silvanto J: How is working memory content consciously experienced? The ‘conscious copy’ model of WM introspection. Neurosci Biobehav Rev. 2015;55:510–519. 10.1016/j.neubiorev.2015.06.003 [DOI] [PubMed] [Google Scholar]
- 58. Kumar A, Rotter S, Aertsen A: Spiking activity propagation in neuronal networks: reconciling different perspectives on neural coding. Nat Rev Neurosci. 2010;11(9):615–627. 10.1038/nrn2886 [DOI] [PubMed] [Google Scholar]
- 59. Dawkins R: The evolution of evolvability.In: On growth, form and computers(eds. Kumar S, Bentley PJ): Elsevier Academic Press, London, chapter 13:2003;239–255. 10.1016/B978-012428765-5/50046-3 [DOI] [Google Scholar]
- 60. Oberauer K: Access to information in working memory: exploring the focus of attention. J Exp Psychol Learn Mem Cogn. 2002;28(3):411–421. 10.1037/0278-7393.28.3.411 [DOI] [PubMed] [Google Scholar]
- 61. Fukai T, Tanaka S: A simple neural network exhibiting selective activation of neuronal ensembles: from winner-take-all to winners-share-all. Neural Comput. 1997;9(1):77–97. 10.1162/neco.1997.9.1.77 [DOI] [PubMed] [Google Scholar]
- 62. Morita K, Jitsev J, Morrison A: Corticostriatal circuit mechanisms of value-based action selection: Implementation of reinforcement learning algorithms and beyond. Behav Brain Res. 2016;311:110–121. 10.1016/j.bbr.2016.05.017 [DOI] [PubMed] [Google Scholar]
- 63. Chklovskii DB, Mel BW, Svoboda K: Cortical rewiring and information storage. Nature. 2004;431(7010):782–788. 10.1038/nature03012 [DOI] [PubMed] [Google Scholar]
- 64. Holtmaat A, Svoboda K: Experience-dependent structural synaptic plasticity in the mammalian brain. Nat Rev Neurosci. 2009;10(9):647–658. 10.1038/nrn2699 [DOI] [PubMed] [Google Scholar]
- 65. Churchill AW, Sigtia S, Fernando C: Learning to generate genotypes with neural networks.Evolutionary Computation.2015. Reference Source [Google Scholar]
- 66. Maynard Smith J, Szathmáry E: The major transitions in evolution.Oxford: Freeman & Co.1995. Reference Source [Google Scholar]
- 67. Reed A, Riley J, Carraway R, et al. : Cortical map plasticity improves learning but is not necessary for improved performance. Neuron. 2011;70(1):121–131. 10.1016/j.neuron.2011.02.038 [DOI] [PubMed] [Google Scholar]
- 68. Iwasa Y: Free fitness that always increases in evolution. J Theor Biol. 1988;135;265–281. 10.1016/S0022-5193(88)80243-1 [DOI] [PubMed] [Google Scholar]
- 69. Sella G, Hirsh AE: The application of statistical physics to evolutionary biology. Proc Natl Acad Sci U S A. 2005;102(27):9541–9546. 10.1073/pnas.0501865102 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70. de Vladar HP, Barton NH: The statistical mechanics of a polygenic character under stabilizing selection, mutation and drift. J Roy Soc Interface. 2011;8(58):720–739. 10.1098/rsif.2010.0438 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71. Friston K: The free-energy principle: a unified brain theory? Nat Rev Neurosc. 2010;11(2):127–138. 10.1038/nrn2787 [DOI] [PubMed] [Google Scholar]
- 72. Baluja S, Caruana R: Removing the genetics from the standard genetic algorithm.Morgan Kaufmann Publishers;199538–46. 10.1016/B978-1-55860-377-6.50014-1 [DOI] [Google Scholar]
- 73. Friston K, Buzsáki G: The Functional Anatomy of Time: What and When in the Brain. Trends Cogn Sci. 2016;20(7):500–511. 10.1016/j.tics.2016.05.001 [DOI] [PubMed] [Google Scholar]
- 74. Zachar I: Code for “Breeding novel solutions in the brain: A model of Darwinian neurodynamics”: Publication release.2016. Data Source [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75. Szilágyi A: Code for "Breeding novel solutions in the brain: A model of Darwinian neurodynamics": Publication release.2017. Data Source [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76. Zachar I: Replicator Formalism. A general account of replication. PhD thesis. Eötvös Loránd University, Department of Plant Systematics, Ecology and Theoretical Biology. Budapest.2011. 10.13140/RG.2.1.3981.5209 [DOI] [Google Scholar]