Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2020 Nov 30;16(11):e1008425. doi: 10.1371/journal.pcbi.1008425

Phenotypes to remember: Evolutionary developmental memory capacity and robustness

András Szilágyi 1,2,3, Péter Szabó 1,4, Mauro Santos 1,5, Eörs Szathmáry 1,2,3,*
Editor: Stuart A Newman6
PMCID: PMC7703877  PMID: 33253184

Abstract

There is increased awareness of the possibility of developmental memories resulting from evolutionary learning. Genetic regulatory and neural networks can be modelled by analogous formalism raising the important question of productive analogies in principles, processes and performance. We investigate the formation and persistence of various developmental memories of past phenotypes asking how the number of remembered past phenotypes scales with network size, to what extent memories stored form by Hebbian-like rules, and how robust these developmental “devo-engrams” are against networks perturbations (graceful degradation). The analogy between neural and genetic regulatory networks is not superficial in that it allows knowledge transfer between fields that used to be developed separately from each other. Known examples of spectacular phenotypic radiations could partly be accounted for in such terms.

Author summary

The development of individual organisms from embryo to adult state is under the control of many genes. During development the initially active genes activate other genes, which in turn change the composition of regulatory elements. The behavior of genetic regulatory systems shows similarities to that of neural networks, of which the most remarkable one is developmental memory, the ability to quickly adapt to environments that have occurred in the past, occasionally several generations earlier. This is because each previously evolved developmental pathway leaves an “imprint” in the gene regulatory network. We investigated the properties of this system; the number of different developmental pathways that can be “memorized”, how this number depends on the number of expressed genes, how fast the system can switch between these pathways, and its robustness against various disturbances affecting either the embryo state or the gene interaction networks. Our results suggest that developmental memory may also provide the mechanism behind some rapid speciation processes.

Introduction

Alan Turing, the father of machine learning, also formulated one of the most important mathematical models in developmental biology: the reaction-diffusion model for pattern generation [1]. This is striking because, although both gene regulatory networks [25] and associative neural networks [68] have extensive literature, only recently a conceptual analogy between evolutionary developmental processes and artificial neural network-based learning models has been articulated. [914]. Since development is the process whereby the phenotype is specified by the evolving genotype, late-evolved morphologies or functional capacities retain aspects of earlier stages (“memory”) that were likely shaped by natural selection. These earlier stages might become reactivated if they are again useful in a different or a changing environment [15]. In this formulation evolutionary changes provide no novel structures that are non-homologous to an ancestral or existing one [16,17], but allow for recursion. For instance, mimetic color patterns of an extinct morph of the butterfly Heliconius cydno, presumably as a result of human disturbance, can be reconstructed from wild-caught butterflies [18], meaning that the morph could recur in nature if the former conditions reappear. Also surprising is the repeatability of evolution among closely related lineages [19,20]. An iconic textbook example is the extraordinary morphological convergence associated with adaptation to distinct ecological niches in cichlid fishes [21], with a large taxonomic diversity in the African Great Lakes Tanganyika (the oldest radiation, around 9–12 Myr ago with about 250 species), Malawi (less than 0.8 Myr ago and over 700 species) and Victoria (about 700 species evolved within the past 15,000 years) [22].

The idea that developmental processes can retain a memory of past selected phenotypes [13], together with the exceptional ability of genomes to find adaptive solutions that quickly converge upon remarkably similar states (“attractors” [23]) in closely related lineages, clearly suggests a non-linear genotype-phenotype mapping capable of producing multiple distinct phenotypes [13,24]. Non-linearity is also the hallmark of reaction-diffusion (Turing) and signaling systems involved in patterning processes [25], and developmental evolutionary biology (evo-devo) views the genotype-phenotype mapping as highly non-linear [26,27]. Furthermore, it might not be farfetched to think of some sort of developmental memory in the cichlid adaptive radiation. The explosive diversification in Lake Victoria was predated by an ancient admixture between two distantly related riverine lineages, one from the Upper Congo and one from the Upper Nile drainage [28]. Many phenotypic traits known to contribute to the adaptation of different ecological niches in the Lake Victoria radiation are also divergent between the riverine species [29,30]. Thus, when referring to the anatomical and morphological variation of Haplochromine cichlids, which are at the origin of the Lake Victoria radiation [28], Greenwood writes [30, p. 266]: “It is amongst the species of these various lacustrine flocks that one encounters the great range of anatomical, dental and morphological differentiation usually associated with the genus. The fluviatile species appear to be less diversified, but even here there is more diversity than is realized at first.” If the high diversity in the Haplochromine cichlids of Lake Victoria is, to some extent, the result of re-evolved (similar) phenotypes in the ancestral fluviatile lineages, then the enduring question of why such an explosive diversification happened within a short time interval might have a simpler solution than previously thought. We aim here to sketch what the solution could be.

The genomic program for development operates primarily by the regulatory inputs and functional outputs of control genes that constitute network-like architectures [31], which are mathematically equivalent to artificial neural networks [10,11]. Although the insights of Vohradsky [10,11] and Watson et al. [13] shed light on an important analogy between neural and genetic regulatory networks, the conclusion of the theory of autoassociative networks cannot yet be readily extended to developmental systems. This is because of the different state space representations, as well as the nature of the task to be solved. Models of autoassociative networks tend to work with positive/negative state variables (inherited from ferromagnetic systems, but see [6]). In contrast to this, in ontogenetic systems the relevant space is that of nonnegative real numbers, corresponding to concentrations of different molecules, see e.g. [32]. Due to the nonlinear activation function features of models working with the above mentioned, alternative state representations can markedly differ. Another important consideration is that autoassociative networks (as their name indicates) solve the problem of the recovery of a particular state (attractor property). During ontogeny we require something more: not only should the adult stage be stable, but the system should reach this state from a particular embryo state (referred to as heteroassociative property). Moreover the system has to find transitions from different embryo states to corresponding, different adults states. In short, we require networks that solve problems of auto/heteroassociativity in one.

In the present investigation we extend the theory of gene regulatory networks to involve both auto- and heteroassociativity. We derive a heuristic formula for regulatory weights to obtain a functional system with the desired properties and we follow how Darwinian dynamics shapes the regulatory networks to acquire these properties. We compare the resulting regulatory matrices and analyze their robustness against different kinds of perturbation. Evolution of developmental pathways is interpreted within the context of ontogenetic dynamics.

Methods

Developmental model

Our model is a formal description of ontogenetic development operating primarily by the regulatory inputs and functional outputs of control genes. Consider an organism with N genes. Its developmental state at time t, expressed by its gene product composition (e.g., proteins), can be represented by the vector p(t) = (p1,p2,…,pN)T with each element being the quantity of the product of a gene. These quantities are assumed to change due to protein decay and gene expression processes. Following [13], the ontogenetic dynamics of the developmental state can be described by the difference equation

pi(t+1)=(1δ)pi(t)+τf([Mp(t)]i), (1)

where τ denotes the decay rate, τ denotes the maximal gene expression rate, f(.) is the activation function, and the matrix M stands for the regulatory network. An mij entry of the matrix gives the regulatory effect of the product of gene j on the expression level of gene i; positive and negative elements imply activation and inhibition, respectively. The cumulative regulatory effects on any single gene i, i.e. the ith element of the product Mp, determine the gene expressions via a sigmoid activation function modelled here as f(x) = (1+tanh (ωx))/2, where ω is the slope parameter.

From an ontogenetic viewpoint, the role of the gene regulatory network is to guide the individual along a developmental pathway from an initial embryonic state p(0) = e to a specific adult state p(T)→a. In real systems, an ensemble of different developmental pathways is desired, each responsible for achieving some environment-specific adult state from a particular embryonic state. We used T = 150 iterations to reach the steady state, when the output does not change in time.

Evolutionary model

In the evolutionary model we considered a population of K individuals, with each member of the population represented by its regulatory matrix. All the interaction matrix elements were zero initially, representing an undeveloped regulation. Every individual shared the same environment, but the environment can change in time. We assumed Q = 3 number of different selective environments, each defining an embryonic state e(q) and a corresponding adapted adult state a(q). The selective environments alternated randomly; if the average fitness of the population approached the optimum (w¯>0.95 for at least 20 consecutive generations), or after 10000 generations, a new environment was chosen at random. In each generation the individuals underwent mutation, development and selection steps as follows.

Mutation: The mutation of the regulation network was implemented by adding a normally distributed random value, with zero mean and μW variance, to a randomly selected matrix element. Matrix elements were clipped into the range [−1,1].

Development: The equilibrium, adult state of each member of the population was obtained by iterating Eq (1).

Selection: The fitness of individual k was expressed by a similarity index derived from the Euclidean distance between the actual adult state p(T) and the environment-specific optimal adult phenotype a(q) as

wk=1n=1N[pn(T)an(q)τ/δ]2 (2)

Then the regulatory matrix of a randomly selected individual was replaced by that of the individual with the highest fitness (elitist selection). Although a stochastic Moran process [33] would be a more realistic selection scheme, for computational reasons simulations were performed using a relatively small population size (K = 100) that would result in too much genetic drift.

Embryonic and (optimally adapted) adult vectors: The number of genes was N = 100 with a low average expression level of σ = 0.1, where 40% of the expressed genes were common, 20% were partially common (appear in two pairs only), and 40% were unique in all the embryonic and all the adult vectors. Specifically, the expression sites of the employed state vectors were e1={13_,19_,32_,36¯,39,49,55,72_,81,87¯}, e2={13_,19_,31,32_,40¯,60,62,72_,87¯,100}, e3={5,13_,19_,32_,36¯,40¯,47,67,72_,94}, a1={6_,12¯,20,24,46_,65,84_,86¯,88,92_}, a2={6_,11,28,46_,79,84_,86¯,91¯,92_,96},a3={6_,12¯,46_,56,61,66,80,84_,91¯,92_};

where underlines and overlines denote the common and partially common elements, respectively. E.g. gene 13 is expressed in all embryo states (common element), gene 12 is expressed in two adult states (partially common element). The initial state was always a perturbed embryonic state. The perturbation was performed, similar to the mutations, by adding a normally distributed random value, with zero mean and μe variance, to a randomly selected element of the environment-specific embryo vector. Vector elements were clipped into the range [0,τ/δ].

Perturbation analysis

To investigate the robustness of the resulting gene regulatory networks we evaluated their performance against three different kind of perturbations. The embryo states were perturbed by flipping the vector elements from low to high, or vice versa, with the given probability. The interaction matrices were perturbed by either adding random values to all matrix elements, drawn from a normal distribution with the given standard deviation, or by nullifying a proportion of the elements. Note that, in the evolutionary algorithm we perturbed only single elements of the embryonic states. In contrast, in the analytical matrix construction we perturbed all elements of the embryonic vectors to incorporate the accumulating effects of many consecutive perturbations on the interaction matrix.

Results

To perform the developmental task, the network must guarantee that (i) each adult state is a stable equilibrium point of the dynamics (stability condition), and (ii) each embryonic state is within the basin of attraction of its corresponding adult state (attraction condition); these two conditions correspond to the auto- and heteroassociative properties in a neural network [34]. Note that this is a more difficult task than a simple pattern recovery problem, which is known to be achievable by a neural network with the standard Hebbian learning rule that fulfils only the stability condition [7]. Not only must all the adult states have a basin of attraction, but these basins must include the corresponding embryonic states.

We found that the task-optimized structure of the regulatory network can be inferred from the embryo-adult state vector pairs in the form of an interaction matrix M (Fig 1). Consider the simplest case with one embryo-adult pair (i.e. one developmental pathway). Depending on whether a gene is expressed in the adult state or not, all the other expressed gene products, in either the embryonic or the adult state, must enhance or block its expression, respectively. This would provide, on the one hand, stability for the adult state and, on the other hand, attraction from the embryonic state. Note, however, that if a gene is expressed in neither the embryonic nor the adult state, then its regulatory effect is irrelevant, therefore the corresponding matrix elements are undetermined. In summary, an mij element of the regulatory matrix M should be positive or negative, depending on whether the ith gene is expressed in the adult state or not, except when the jth gene is expressed in neither the embryonic nor the adult state. The above line of thought can be generalized for arbitrary Q number of embryo-adult state pairs. Denoting the zero-one normalized embryonic and adult state vectors by e and a, such a matrix can be obtained by averaging two dyadic products for all developmental pathways as

M=12Qq=1Q(2a^(q)1)a^(q)+(2a^(q)1)e^(q), (3)

where Q stands for the number of embryo-adult state pairs and q denotes the different pairs. The first and second dyadic products are responsible for the stability and attraction conditions, respectively. Note the similarity of our treatment and Kurikava and Kaneko 2012 and 2013 approach [35,36]. Within each dyadic product the right term determines whether an entry is relevant from the viewpoint of the state vector, whereas the left dyadic term determines its sign. The resulting matrix contains positive values, negative values and zeros for activator, inhibitory and undetermined elements, respectively. Notice that the developmental pathways can be in conflict with each other as to whether a gene should be up- or downregulated by another gene. It is instructive to compare this formula with the standard Hebbian learning rule H = aa for ai∈{−1,+1}. Its modification for ai∈{0,+1} vectors that preserves that stability condition is H = (2a−1)∘a, which is identical to the first term in Eq (3), c.f. Table 1, and see [6]. (This is the standard procedure in ANNs with unsigned states as it converts the training patterns to signed values even though the state vector is unsigned. Natural selection will not operate like this–we are just showing how to hand-construct a solution that works.)

Fig 1. Illustration of the construction rules of interaction matrices based on theoretical considerations on the optimal pairwise interaction types between genes. e(1) and a(1) are the first embryo-adult pair, e(2) and a(2).

Fig 1

the second pair. Depending on the combination of gene expressions ei(n) and ai(n) in an embryo-adult vector pair (n = 1,2), an mij element of the interaction matrix can be positive (′+′, activation), negative (′−′, inhibition), or undefined (′U′). To ensure correct development (Me(n)e(n)a(n)) the Me(n) matrices must have the structure indicated in the figure. (If ej(n)=1 and ai(n)=1, then mij(n) = ′+′; if ej(n)=1 and ai(n)=0, then mij(n) = ′−′; if ej(n)=0, then mij(n) = ′U′; irrespective of the value of ai(n).) A similar argument holds for the stability criteria (Ma(n)a(n)a(n)) and results in the Ma(n) matrices. By combining Me(n) and Ma(n) the resulting M(n) fulfills both the attractivity and stability criteria. The combination rules are the following: (+,+)→+; (−,−)→−; (±,U)→±; and (±,∓)→C, which can be done practically by taking the element-wise average of the two matrices. The ultimate combination of all M(n)s results in a matrix that fulfills the attraction and stability criteria for all different embryo-adult pairs.

Table 1. Comparison of the resulting analytic interaction matrices for an autoassociative task with the two representations.

{-1,+1} representation {0,+1} representation
learning rule Hij=qai(q)aj(q)(Hebbrule) Mij=q(2ai(q)1)aj(q)
symmetry Hij = H non-symmetric
neutralities in weight matrix no neutral elements can be neutral elements (“opposite to” zero vector elements)
main diagonal always positive (if allowed) can be negative or positive
structure has a unique structure many different realizations

We investigated the parameter dependence of the analytic model. As for the regulatory matrix we used a slightly modified version of Eq (3). Treves [8] claims that the interaction terms should be modified by the average expression σ, i.e. the proportion of expressed genes. This is because if a larger proportion of genes is expressed, then proportionally smaller interaction strengths are needed for the same regulatory effect on any single gene. Incorporating this consideration into Eq (3) gives

M=12Qq=1Q(2a^(q)1)(a^(q)σ+e^(q)σ) (4)

The performance of a regulatory network constructed by the above rule changes with the number of developmental pathways and gene expression levels (Fig 2). With increasing number of embryo-adult pairs, the accumulating conflicts between them inevitably corrupt the regulatory ability of the network; some adult states will be unreachable from their embryonic states. Nevertheless, the network is able to tolerate a fair number of conflicts, related to its structural stability. Since conflicts can occur only between non-orthogonal state vectors, the performance of the network also depends on the amount of overlap in the expression patterns of states belonging to different pairs. This highlights the importance of the proportion of expressed genes; i.e., the sparseness of the state vectors. If these vectors are very sparse, then they are likely to be orthogonal, therefore the number of learnable embryo-adult pairs does not reduce. The number of learnable pairs is a decreasing function of sparseness, because the numerous non-orthogonal state vectors and the largely different gene expressions in the adult states lead to several conflicts among them. Regarding the effect of system size on functionality, the results are in line with the expectations; the higher the number of genes, the higher is the number of “error free” developmental pathways (Fig 3).

Fig 2. Performance of the analytic developmental networks.

Fig 2

We assumed different sparseness (proportion on non-zero entries in the state vectors) values and different number of embryo-adult pairs. Embryo and desired adult vectors were generated by independently setting each vector element to high or low randomly according to the sparseness value. The performance was measured by the averaged (over 400 realizations) Pearson correlation(s) between the desired and the experienced adult state(s) for all developmental pathways (panel A). Panels B and C show a more detailed view for the two cross-sections of the parameter space (indicated by dashed lines in panel A). Orange horizontal lines show the maximum number of orthogonal state vectors for the given sparseness values. Parameters: N = 100, δ = 0.2, τ = 1, ω = 25).

Fig 3. Performance of the analytic developmental networks with different number of genes and developmental pathways.

Fig 3

Relevant parameters as in Fig 2 and σ = 0.1.

We have also analyzed the “memory capacity”, i.e. the number of learnable developmental pathways as a function of the system size (N). There are several studies on the memory capacity of Hopfield networks [e.g. 37,38], but these results are not applicable to our hetero/autoassociative system, and the capacity definitions themselves are valid only in the N→∞ limit, while our systems are relatively small (N≤500). Therefore, we define the memory capacity of a given system as the average number of developmental pathways which can be reconstructed at r = 0.95 performance (denoted by Q0.95, c.f. Fig 2C). Fig 4 shows that Q0.95 depends linearly on N: Q0.95 = ϑ(σ)∙N+b(σ).

Fig 4. The memory capacity (Q0.95) as a function of the number of neurons (N) at different sparseness (σ).

Fig 4

Similar to Fig 2, each point is an average of 400 realizations. The lines are the linear regression fits (R2>0.99 in all cases).

Since decreasing sparseness increases the probability of controversial interactions (cf.. 1) ϑ(σ) is a decreasing function of σ, in line with the expectations. The ϑ values, which are in the range [0.0588, 0.0236] depending on σ, can be treated as a proportionality factor of memory capacity. These are significantly lower values than the proportionality value of 0.138 in the widespread used linear formula for autoassociative networks, see e.g. [39]. Notice, however, that these values should be compared with caution on one hand due to the different interpretation of memory capacity, and on the other hand owing to the “double task” nature (hetero- and autoassociative tasks) of our system.

A key question is whether a functional network is attainable by Darwinian selection via a series of mutation-selection steps. In our evolutionary model we used a more realistic Darwinian dynamics than the solitary stochastic hill climbing [13]. From the viewpoint of the theory of artificial neural networks this process can be regarded as a Darwinian dynamics-driven learning process. The evolutionary algorithm yields interaction matrices that contain positive and negative values where the heuristic formulation predicts them (Fig 5). While the individual interaction matrices vary, their average is in line with the heuristically derived matrix. The values are arranged into a characteristic structure; positive and negative entries form horizontal stripes, intermitted with vertical stripes of near-zero values (c.f. Fig 1). Those genes have the largest effect on the developmental process, which are expressed in any embryonic or adult states (c.f. marked columns in Fig 5). Depending on whether the affected gene is expressed in any of the adult states, they have a strong positive or negative effect (c.f. marked rows in Fig 5). The rest of the genes drift freely in individual realizations due to a lack of selective pressure. Consequently, the average values in these positions are approximately zero (c.f. grey columns in left panel of Fig 5). The corresponding values in the analytic treatment (undefined elements) are zero by definition. The only major difference from the heuristic matrix is that the main diagonal elements of the evolutionary matrix are mainly negative, which means that the expression of every gene is under negative feedback by its own inhibitory product. A possible explanation is that without a strong negative feedback a gene could be easily overexpressed due to the perturbations of the interaction elements. This is more probable if the sparseness of the expression vectors is low, as it was in our case. This picture is likely to change with hierarchical developmental regulation, the evolution of which takes longer time and should be investigated in the future.

Fig 5. Structure of the interaction matrix obtained with the evolutionary algorithm as compared to the analytically derived one for three developmental pathways.

Fig 5

(A) The evolutionary interaction matrix was obtained by averaging the output of 300 independent runs of the evolutionary algorithm. The three applied environment-specific embryonic and adult state vectors are shown along the sides. Orange guidelines highlight those rows where the corresponding genes are expressed in at least one adult state, whereas green guidelines highlight those columns where at least one gene is expressed in any of the embryonic or adult states. (B) The theoretically predicted interaction matrix was constructed from the embryonic and adult state vectors using Eq (3). Parameters as in Fig 2 and Q = 3, K = 100, μW = 0.05, μe = 0.1, σ = 0.1.

A detailed view of the evolutionary process is shown in Fig 6. During the early generations, where the gene regulation is undeveloped, it takes many generations (i.e., mutation-selection steps) to approach the environment-specific optimum. In addition, selection for one environment can have adverse effects on performance in another environment if the basin of attraction of the actually selected adult state engulfs the neighborhood of the embryonic states of other adult states. But those interactions which are not beneficial in any of the environments are eliminated. Deleterious mutations may arise any time also in a well-functioning system, but selection eliminates them over the timescale of a few environmental changes.

Fig 6. Learning of three different developmental pathways in the evolutionary model.

Fig 6

(A) Average fitness and the mutation-selection steps needed to achieve a well-functioning developmental network during random environmental changes. The three environments are denoted by red, green and blue. Parameters as in Fig 5. (B) Schematic illustration of the changes in the state-space dynamics during the evolutionary process with three developmental pathways (indicated by red, blue and green colors). The panels show the basins of attraction of an initial, random regulation system with two embryo-adult pairs (left), a well-functioning one (top right) and a bad one, where the basins of attraction of the adult states (filled dots encircled by dotted lines indicating variation around the target phenotypes) include not only their corresponding embryo states (bottom right).

A developmental process must be sufficiently robust against stochastic perturbations of both the embryonic state and the gene interaction matrix [5,12]. It requires that the neighborhood (according to a given metric) of the embryonic states must also be in the basin of their corresponding adult states. Therefore, some inputs of variation should produce little or no phenotypic variation at all, a phenomenon that has received a lot of attention under the labels of canalization, robustness or buffering [31,32,4042]. The recovery performance of the network changes with increasing amount of perturbations (Fig 7). The system is very robust against perturbations regarding the embryonic state, it is moderately robust against additive perturbations and has limited robustness against eliminated interactions regarding the interaction network. The sharp difference between the robustness of evolutionary and analytic models in the latter case is the consequence of the peculiarities of the evolutionary process. In the evolutionary matrix the regulatory weights are less evenly distributed as compared to the analytic one, due to the stochastic nature of the evolutionary process. It makes the system more sensitive to eliminating perturbations. Resilience understandably decreases with the number of developmental pathways in all cases, but conforming to “graceful degradation” in artificial neural networks; i.e., performance first decreases mildly and drops fast only beyond a critical strength of perturbation [7]. To sum up, variation is apportioned into discontinuous (basins of attraction) and continuous (small perturbations around the target) phenotypes (Fig 6B). Evo-devo mainly focuses on the first kind of variation whereas standard evolutionary genetics focuses on the second [26,43].

Fig 7. Robustness of the developmental dynamics against perturbations.

Fig 7

The interaction matrices were constructed from the given number of embryo-adult vector pairs according to Eq (4). The performance was expressed by the Pearson correlation(s) between the desired and the experienced adult state(s) for all developmental pathways after T = 150 iterations averaged over 300 matrices and 100 perturbations for each parameter combination. (A) Performance against the proportion of the flipped embryonic vector elements. (B). Performance against the standard deviation (SD) of the perturbation of the interaction matrix. All elements of the matrix were perturbed additively by an N(0,SD) random number. (C) Performance against the proportion of nullified elements of the interaction matrix. Each element of the interaction matrix was set to zero with the given probability. Relevant parameters are as in Fig 5.

Discussion

Treating gene regulatory networks as formally analogous to artificial neural networks [10,11] allows translating the well-known dynamics of the latter [44] to model genomic programs for development. There is widespread natural variation in morphogenic pathways [45], and the developmental memory of past selected phenotypes [13] is akin to the memory capacity of neural networks. This developmental memory allows populations to re-evolve phenotypes much faster than it would be possible if they had to evolve de novo. Previous speculation on the effect of the heat-shock protein Hsp90 as a capacitor for releasing hidden morphogenetic variation that could allow fast morphological radiations [45] has been criticized on the grounds that the function of Hsp90 is to prevent morphological aberrations. Furthermore, some sense of purposive evolution, fully incompatible with the lack of foresight of natural selection, lays behind this sort of interpretations [46].

These criticisms do not apply here because in our developmental model past selected states can recur in the population if they appear useful again in a different environment or body context. As any theoretical model, ours obviously has inherent limitations and highly simplifies the representation of biological systems. However, to the extent that it captures sufficient conditions to generate the phenomenon of morphological radiations, more complex explanations are not required. Thus, the assumption that structural novelties (or “key innovations”) are associated with adaptive radiations into new ecological niches (e.g. [47, p. 159]) might be unwarranted. There is a noteworthy implication in the foregoing consideration for the understanding of atavism. Crocodilian teeth can grow in mutant birds, which suggests the reactivation of the associated developmental machinery [48], that required the resurrection [49] of a key aspect of regulation. The same neurons participate in the storage of different engrams in neural networks. The same holds for the storage of devo-engrams in genetic regulatory networks. Resurrection leading to atavism requires only limited reactivation of a few connections in a network that is maintained by the current selective forces. An exciting question is how evo-devo learning can generalize from the “training set” (previously selected target phenotypes) to novel ones [13,50]. The prediction described by these authors is that generalization potential works within a set that can be characterized by the same formal grammar.

While the theory of neural networks can (and does) infer the same conclusions based on different representations, in the case of modelling real biological situations the adequacy of the representation can be crucial (the same holds for neuronal networks). Our results show that a linear change to the representation has profound impact on the essential features of the system. While in the customary (neural) {-1,+1} representation there are no neutral elements in the interaction matrix, the biologically adequate {0,+1} representation of genetic regulatory networks allows for the free choice of interaction elements being opposite to “0”. This feature turns out to increase the robustness of the system against the disturbance of interaction coefficients only if the system is very sparse, which guarantees the commonly zero elements in the embryo and adult states. Another feature of our representation is the large number of different interaction matrices entailing the same developmental process, thus evolution “from scratch” does not face so many constraints. In other words, starting with a single ancestor an extraordinarily rapid morphological diversification could be attained, which is the hallmark of adaptive radiations.

Acknowledgments

The authors thank Ferenc Jordán and Harold. P. de Vladar for comments on an earlier draft.

Data Availability

All relevant data are within the manuscript and its Supporting Information files. The source code of the implementation of the model is available on GitHub (https://github.com/peterszabo77/developmental_memory_2020).

Funding Statement

This work was supported by the National Research, Development and Innovation Office (NKFIH, https://nkfih.gov.hu) under OTKA grant numbers K124438 (A.S.), K119347 (E.S. and A.S.) and GINOP-2.3.2-15-2016-00057 (E.S., P.S. and A.S.) research grants; the ATTRACT Project EmLife (AS, ES and MS) (https://attract-eu.com); The Volkswagen Foundation (https://www.volkswagenstiftung.de) (initiative “Leben? –Ein neuer Blick der Naturwissenschaften auf die grundlegenden Prinzipien des Lebens”, project “A unified model of recombination in life”) (A.S., E.S. and M.S.). A.S. was supported by the Bolyai János Research Fellowship of the Hungarian Academy of Sciences (https://mta.hu/) and the ÚNKP-19-4 New National Excellence Program of the Ministry of Human Capacities. M.S. was supported by the research grant CGL2017-89160-P from the Ministerio de Economía, Industria y Competitividad, research grant 2017SGR 01379 from Generalitat de Catalunya, and the MTA Distinguished Guest Fellowship Programme in Hungary. All authors were supported by Templeton Foundation under grant number TWCF0268 (“Learning in evolution, evolution in learning”). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Turing AM. The chemical basis of morphogenesis. Philos Trans R Soc B. 1952;237:37–72. [Google Scholar]
  • 2.Kauffman SA. Metabolic stability and epigenesis in randomly constructed genetic nets. J Theor Biol. 1969;22(3):437–467. 10.1016/0022-5193(69)90015-0 [DOI] [PubMed] [Google Scholar]
  • 3.Glass L, Kauffman SA. The logical analysis of continuous, non-linear biochemical control networks. J Theor Biol. 1973;39(1):103–129. 10.1016/0022-5193(73)90208-7 [DOI] [PubMed] [Google Scholar]
  • 4.Salazar-Ciudad I, Newman SA, Solé RV. Phenotypic and dynamical transitions in model genetic networks I. Emergence of patterns and genotype-phenotype relationships. Evol Dev. 2001;3:84–94. 10.1046/j.1525-142x.2001.003002084.x [DOI] [PubMed] [Google Scholar]
  • 5.Kaneko K. Evolution of robustness to noise and mutation in gene expression dynamics. PLoS ONE. 2007;2:e434 10.1371/journal.pone.0000434 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Hopfield JJ. Neural networks and physical systems with emergent collective computational abilities. Proc Natl Acad Sci USA. 1982;79(8):2554–2558. 10.1073/pnas.79.8.2554 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Rolls ET. Memory, attention and decision-making Oxford university press, 2008. [Google Scholar]
  • 8.Treves A. Graded-response neurons and information encodings in autoassociative memories. Phys Rev A. 1990;42:2418–2430. 10.1103/physreva.42.2418 [DOI] [PubMed] [Google Scholar]
  • 9.Mjolsness E, Sharp DH, Reinitz J. A connectionist model of development. J Theor Biol. 1991;152(4): 429–453 10.1016/s0022-5193(05)80391-1 [DOI] [PubMed] [Google Scholar]
  • 10.Vohradsky J. Neural model of the genetic network. J Biol Chem. 2001;276:36168–36173. 10.1074/jbc.M104391200 [DOI] [PubMed] [Google Scholar]
  • 11.Vohradsky J. Neural network model of gene expression. FASEB J. 2001;15:846–854. 10.1096/fj.00-0361com [DOI] [PubMed] [Google Scholar]
  • 12.Ciliberti S, Martin OC, Wagner A. Innovation and robustness in complex regulatory gene networks. Proc Natl Acad Sci USA. 2007;104(34): 13591–13596. 10.1073/pnas.0705396104 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Watson RA, Wagner GP, Pavlicev M, Weinreich DM, Mills R. The evolution of phenotypic correlations and “developmental memory”. Evolution. 2014;68:1124–1138 (2014). 10.1111/evo.12337 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Watson RA, Szathmáry E. How Can Evolution Learn? Trends Ecol Evol. 2016;31:147–157. 10.1016/j.tree.2015.11.009 [DOI] [PubMed] [Google Scholar]
  • 15.Sharov AA. Evolutionary constraints or opportunities? Biosystems. 2014;123:9–18. 10.1016/j.biosystems.2014.06.004 [DOI] [PubMed] [Google Scholar]
  • 16.Müller GB, Wagner GP. Novelty in evolution: Restructuring the concept. Annu Rev Ecol Syst. 1991;22:229–256. [Google Scholar]
  • 17.Peterson T, Müller GB. What is evolutionary novelty? Process versus character based definitions. J Exp Zool (Mol Dev Evol). 2013;320B:345–350. 10.1002/jez.b.22508 [DOI] [PubMed] [Google Scholar]
  • 18.Linares M. The ghost of mimicry past: laboratory reconstitution of an extinct butterfly 'race'. Heredity. 1997;78:628–635. [Google Scholar]
  • 19.Orgogozo V. Replaying the tape of life in the twenty-first century. Interface Focus. 2015;5:20150057 10.1098/rsfs.2015.0057 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Blount ZD, Lenski RE, Lossos JB. Contingency and determinism in evolution: Replying life’s tape. Science. 2018;362: eaam5979 10.1126/science.aam5979 [DOI] [PubMed] [Google Scholar]
  • 21.Muschick M, Indermaur A, Salzburger W. Convergent evolution within an adaptive radiation of cichlid fishes. Curr Biol. 2012;22:2362–2368. 10.1016/j.cub.2012.10.048 [DOI] [PubMed] [Google Scholar]
  • 22.Salzburger W. Understanding explosive diversification through cichlid fish genomics. Nat Rev Genet. 2018;19:705–717. 10.1038/s41576-018-0043-9 [DOI] [PubMed] [Google Scholar]
  • 23.Conway Morris S. Life’s solution. Inevitable humans in a lonely universe Cambridge university press, 2003. [Google Scholar]
  • 24.Kouvaris K, Clune J, Kounios L, Brede M, Watson RA. How evolution learns to generalise: Using the principles of learning theory to understand the evolution of developmental organisation. PLoS Comput Biol. 2017;13:e1005358 10.1371/journal.pcbi.1005358 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Murray JD. Mathematical Biology, 2nd ed Springer, 1993. [Google Scholar]
  • 26.Alberch P. Developmental constraints in evolutionary processes In: Evolution and development Dahlem Konferenzen 1982, Bonner JT, ed. Springer-Verlag, 1982, pp 313–332. 10.1111/j.1558-5646.1982.tb05066.x [DOI] [Google Scholar]
  • 27.Gjuvsland AB, Vik JO, Beard DA, Hunter PJ, Omholt SW. Bridging the genotype–phenotype gap: what does it take? J Physiol. 2013;591:2055–2066. 10.1113/jphysiol.2012.248864 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Meier JI, Marques DA, Mwaiko S, Wagner CE, Excoffier L, Seehausen O. Ancient hybridization fuels rapid cichlid fish adaptive radiations. Nat Commun. 2017;8:14363 10.1038/ncomms14363 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Seehausen O, Koetsier E, Schneider MV, Chapman LJ, Chapman CA, et al. Nuclear markers reveal unexpected genetic variation and a Congolese-Nilotic origin of the Lake Victoria cichlid species flock. Proc R Soc Lond B Biol Sci. 2003;270:129–137. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Greenwood PH. Towards a phyletic classification of the ‘genus’ Haplochromis (Pisces, Cichlidae) and related taxa. Part 1. Bull Br Mus Nat Hist Zool. 1979;35:265–322. [Google Scholar]
  • 31.Levine M, Davidson EH. Gene regulatory networks for development. Proc Natl Acad Sci USA. 2005;102:4936–4942. 10.1073/pnas.0408031102 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Brun-Usan M, Thies C, Watson RA. How to fit in: The learning principles of cell differentiation. PLoS Comput Biol. 2020;16(4), e1006811 10.1371/journal.pcbi.1006811 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Moran PAP. Random processes in genetics. Proc Camb Phil Soc. 1958;54(1):60–71. [Google Scholar]
  • 34.Hirahara M. "Associative memory" in Encyclopedia of Neuroscience Binder MD, Hirokawa N, Windhorst U, eds. Springer, 2009, pp 195–201. [Google Scholar]
  • 35.Kurikawa T, Kaneko K. Associative memory model with spontaneous neural activity. Europhys Lett. 2012;98:48002. [Google Scholar]
  • 36.Kurikawa T, Kaneko K. Embedding responses in spontaneous neural activity shaped through sequential learning. PLoS Comput Biol. 2013;9:e1002943 10.1371/journal.pcbi.1002943 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Amit DJ, Gutfreund H, Sompolinsky H. Storing infinite numbers of patterns in a spin-glass model of neural networks. Phys Rev Lett. 1985;55:1530 10.1103/PhysRevLett.55.1530 [DOI] [PubMed] [Google Scholar]
  • 38.Folli V, Leonetti M, Ruocco G. On the maximum storage capacity of the Hopfield model. Front Comput Neurosci. 2017;10:144 10.3389/fncom.2016.00144 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Hertz J,Krogh A, Palmer RG. Introduction to the theory of neural computation Redwood City, CA, Addison-Wesley, 1991. [Google Scholar]
  • 40.Waddington CH. Canalization of development and the inheritance of acquired characters. Nature. 1942;150:563–565. [DOI] [PubMed] [Google Scholar]
  • 41.Nijhout HF. The nature of robustness in development. Bioessays. 2002;24:553–563. 10.1002/bies.10093 [DOI] [PubMed] [Google Scholar]
  • 42.Hallgrimsson B, Greena RM, Katz DC, Fish JL, Bernier FP, et al. The developmental-genetics of canalization. Sem Cell Dev Biol. 2019;88:67–79. 10.1016/j.semcdb.2018.05.019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Müller GB. Evo-devo: extending the evolutionary synthesis. Nat Rev Genet. 2007;8:943–949. 10.1038/nrg2219 [DOI] [PubMed] [Google Scholar]
  • 44.Haykin S. Neural networks: a comprehensive foundation, 2nd ed Pearson Education, 1999. [Google Scholar]
  • 45.Rutherford SL, Lindquist S. Hsp90 as a capacitor for morphological evolution. Nature. 1998;396:336–342. 10.1038/24550 [DOI] [PubMed] [Google Scholar]
  • 46.Dickinson WJ, Seger J. Cause and effect in evolution. Nature. 1999;399:30 10.1038/19894 [DOI] [PubMed] [Google Scholar]
  • 47.Carroll SB, Grenier JK, Weatherbee SD. From DNA to diversity. Molecular genetics and the evolution of animal design, 2nd ed Blackwell, 2005. [Google Scholar]
  • 48.Harris MP, Hasso SM, Ferguson MWJ, Fallon JF. The development of Archosaurian first-generation teeth in a chicken mutant. Curr Biol. 2006;16:371–377. 10.1016/j.cub.2005.12.047 [DOI] [PubMed] [Google Scholar]
  • 49.Marshall CR, Raff EC, Raff RA. Dollo’s law and the death and resurrection of genes. Proc Natl Acad Sci USA. 1994;91:12283–12287. 10.1073/pnas.91.25.12283 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Parter M, Kashtan N, Alon U. Facilitated variation: how evolution learns from past environments to generalize to new environments. PLoS Comput Biol. 2008;4:e1000206 10.1371/journal.pcbi.1000206 [DOI] [PMC free article] [PubMed] [Google Scholar]
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1008425.r001

Decision Letter 0

Stuart A Newman, Natalia L Komarova

7 Sep 2020

Dear Professor Szathmáry,

Thank you very much for submitting your manuscript "Phenotypes to remember: Evolutionary developmental memory capacity and robustness" for consideration at PLOS Computational Biology.

As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.

Please note that apart from Reviewer #1's narrative comments below he also provided an annotated copy of your manuscript (see attachment). Reviewer #2 asks you to cite earlier work by Kauffman, Kaneko, and Newman. I would elaborate on this by noting that Kauffman and Kaneko have each presented gene regulatory network systems, which like the one in the present manuscript, generates systems of hierarchically nested attractors that serve as models of development. Newman, in contrast, has presented not a dynamical network treatment, but a perspective in which inherent physical properties of developmental systems provide a basis for understanding recurrent features in evolution. This is also a precedent for the present work. Newman has also recently presented a critical view of the sufficiency of dynamical systems models for evolutionary simulations (doi: 10.1016/j.jtbi.2019.110031). The authors may or may not wish to address this critique.      

We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.

Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Stuart A Newman

Guest Editor

PLOS Computational Biology

Natalia Komarova

Deputy Editor

PLOS Computational Biology

***********************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: This is a very clear and well-developed paper. It develops work on the functional equivalence of learning and evolution - specifically, in the context of GRNs and correlation learning, and the ability to evolve GRNs that store multiple phenotypic memories. The work is carried out carefully and presented clearly - I see no technical problems. The results are clear and not overstated. The finding - that the functional equivalence of evolved GRNs and correlation learning models can be deepened and extended - is significant. Previous work in this area used simplifying assumptions, and assessing whether the effects depend on these assumptions is important. The work also extends the scope of evolved behaviours from autoassociative to heteroassociative - and this is a significant extension. The following comments are therefore minor/suggestions for improvements.

1. I like an introduction to be clear about the aims and claims. As it is, the intro is excellent at setting the scene but it could be more clear about the hypotheses that will be tested or the questions that will be answered. This has a knock-on effect later in the paper.

2. It seems like the aim is primarily to figure out whether/how the simplifying assumptions of prior work, compared to a more realistic modelling set-up for the GRN, matters to the idea of evolution learning multiple developmental memories. It turns out, by the end of the paper, that the analogy remains sound, but there are some interesting differences along the way. The results also show that the task can be extended to heteroassociative tasks which hasn’t been done before. Clarity on the balance of these two findings could improve the paper. Specifically,

- If the aim of the paper was really to detail the impact of these different implementational assumptions, it would have been valuable to compare the two models on the same task. I don’t think that the non-negative state models work on the tasks of earlier work. And you’ve showed that the non-negative state models do work on your task. But we don’t have a side-by-side comparison of both models on both tasks.

- Likewise, you suggest that the non-negative state model results in more neutrality and that this increases robustness (compared to the model with signed states). I don’t think its clear that the signed state model has a problem with robustness despite having no neutrals. And we don’t have a side-by-side comparison.

- However, if the claim of the paper is that these differences don’t matter much – ie that the analogy is basically sound, and the task space can be extended, then a side-by-side comparison is not needed.

- I think that the implementational details do matter to some tasks – and your paper is very useful in helping us think that through more carefully. And the implementational details sometimes don’t matter - and you’ve shown a case where they don’t. So, both points are worth making. Yet, I think the balance in the writing at present is a little bit towards the former, whereas the results better support the latter. So, perhaps the paper could be improved if you stated your aims/claims/questions/hypotheses more clearly at the beginning in the intro.

3. A technical point relating to contrast of signed and unsigned models. You say (Table 1) that signed states will result in Hebbian learning and symmetric interaction terms (as observed in prior work). But the evolved interactions in that work are symmetric because there is no reason for them to be asymmetric – not because they have to be symmetric or because Hebb’s rule is applied. Hebb’s rule isn’t applied, of course – and they aren’t forced to be symmetric - but when there is no reason to be asymmetric, symmetric Hebbian changes are observed. When evolving weights to a single pattern with unsigned states, the weights are not symmetric and not Hebbian – as you say. But I happen to know that in other conditions (e.g. when connections are costly) asymmetric interactions evolve even when the states are signed. So, question: Is there a reason for selection to create asymmetric weights in the heteroassociative task, does it require asymmetric weights (your Fig 1 suggests yes), and would asymmetric weights then evolve even in the signed state model? If we don’t know the answers to these questions, then Im not sure that all the rows of your Table 1 follow from the choice of signed/unsigned state models. As per 2 above, Im not sure that such questions/issues regarding the comparison of the modelling choices are the main contribution of this paper – in which case this doesn’t matter so much.

4. Since capacity and robustness are in the title….

a. There is early (1980s) and recent work on Hop Net capacity (e.g. Folli, V., Leonetti, M., & Ruocco, G. (2017). “On the maximum storage capacity of the Hopfield model”. Frontiers in computational neuroscience, 10, 144. and refs therein)

b. Fig 6 top and middle supports your claims nicely wrt robustness under large perturbations. Im not sure that the bottom fig does though – If Im reading it right. It does have a larger tail in the limit of large perturbations, but the initial drop off is large – I don’t see how this supports the claim that neutrality increases robustness.

5. some other (optional) suggestions and thoughts for consideration are provided in the commented pdf

In sum, this paper makes a very strong contribution to the topic, and very nicely written and easy to read - with strong technical analysis. I hope these suggestions above are helpful.

Reviewer #2: (attached)

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: None

Reviewer #2: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Richard A. Watson

Reviewer #2: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, PLOS recommends that you deposit laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions, please see http://journals.plos.org/compbiol/s/submission-guidelines#loc-materials-and-methods

Attachment

Submitted filename: PCOMPBIOL-D-20-00996_review.pdf

Attachment

Submitted filename: renamed_6c9c8.pdf

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1008425.r003

Decision Letter 1

Stuart A Newman, Natalia L Komarova

6 Oct 2020

Dear Professor Szathmáry,

We are pleased to inform you that your manuscript 'Phenotypes to remember: Evolutionary developmental memory capacity and robustness' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Stuart A Newman

Guest Editor

PLOS Computational Biology

Natalia Komarova

Deputy Editor

PLOS Computational Biology

***********************************************************

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1008425.r004

Acceptance letter

Stuart A Newman, Natalia L Komarova

28 Oct 2020

PCOMPBIOL-D-20-00996R1

Phenotypes to remember: Evolutionary developmental memory capacity and robustness

Dear Dr Szathmáry,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Melanie Wincott

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: PCOMPBIOL-D-20-00996_review.pdf

    Attachment

    Submitted filename: renamed_6c9c8.pdf

    Attachment

    Submitted filename: Answers.docx

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting Information files. The source code of the implementation of the model is available on GitHub (https://github.com/peterszabo77/developmental_memory_2020).


    Articles from PLoS Computational Biology are provided here courtesy of PLOS

    RESOURCES