Abstract
Performing more tasks in parallel is a typical feature of complex brains. These are characterized by the coexistence of excitatory and inhibitory synapses, whose percentage in mammals is measured to have a typical value of 20-30%. Here we investigate parallel learning of more Boolean rules in neuronal networks. We find that multi-task learning results from the alternation of learning and forgetting of the individual rules. Interestingly, a fraction of 30% inhibitory synapses optimizes the overall performance, carving a complex backbone supporting information transmission with a minimal shortest path length. We show that 30% inhibitory synapses is the percentage maximizing the learning performance since it guarantees, at the same time, the network excitability necessary to express the response and the variability required to confine the employment of resources.
Brain functions, such as learning and memory, operate through coordinated neuronal activations in which highly connected neurons (hubs) have an orchestrating role. Synaptic plasticity1 regulates the balance of excitation and inhibition shaping cortical networks into a complex scale-free functional structure, where hubs are found to be inhibitory neurons2. Experimental observations suggest that in mammalian brains the fraction of inhibitory synapses is close to 20–30%, however this value is not justified by any theoretical argument. Evoked activity in the mammalian cortex exhibits a large variability to a repeated stimulus, which is expression of the fluctuations in ongoing activity when the stimulus is applied3. Spontaneous activity therefore represents the fundamental support on which neuronal systems develop learning skills. In recent years, a novel mode of spontaneous activity has been detected, neuronal avalanches. These are bursts of firing neurons measured both in vitro and in vivo, whose size and duration distributions show a robust power law behaviour4,5,6. In this framework, learning can be interpreted as a phenomenon occurring in a critical state, where the ongoing activity does not exhibit a characteristic size. Recently, learning of single Boolean rules has been investigated7 in a neuronal network model able to fully reproduce the scaling properties and the temporal organization of neuronal avalanches8,9,10. The model, without any reinforcement learning11 or error back propagation12, is able to learn even complex rules: The learning performance increases monotonically with the number of times a rule is applied and all rules can be learned (and remembered) provided that the plastic adaptation is sufficiently slow. The performance rate and the learning time exhibit universal scaling features, independent of the particular rule and, surprisingly, the percentage of inhibitory synapses in the network. Multi-task learning requires a more complex organization of neuronal activity where inhibitory synapses may indeed play an important role since competing rules have to establish a synaptic backbone for information transmission under interference effects, not present in single rule learning. Synaptic plasticity is the key process sculpting this backbone. Recently, interesting homeostatic features have been detected in plasticity consisting in balanced depression and potentiation of excitatory and inhibitory synapses aiming at the conservation of the total synaptic strength1.
Here we study the parallel learning of Boolean rules with two inputs (AND, OR, XOR and RAN, a rule associating to each input configuration a random output) on a neuronal network able to reproduce the statistical behaviour of spontaneous activity. A firing (non-firing) neuron corresponds to the Boolean variable 1 (0) (see Methods). The response of the network to the successive applications of each rule is monitored for several trials, where plastic adaptation is performed according to a non-uniform negative feedback algorithm if the system gives a wrong answer (see Methods). We demonstrate that the 30% fraction of inhibitory synapses, measured in mammalian brains, optimizes the network's performance. We provide an understanding of this behaviour in terms of the structural features of the cluster of paths supporting information transmission and the activity dynamical properties.
In the following we will investigate the learning performance of the model as function of the parameter pin. The role of other parameters has been addressed in previous works13,14 and found to be irrelevant for the scaling behaviour of avalanche activity. We first verify that for each rule the choice of hubs inhibitory neurons and homeostatic plastic adaptation optimize the performance in both single and parallel learning (see Suppl. Fig. 1, 2). On this basis we adopt these two ingredients in our procedure. Moreover, we verify that also for parallel learning the performance improves for increasing average connectivity in the network and for decreasing distance between input and output neurons. Most importantly, slow plastic adaptation improves the performance also in multi-task learning and curves corresponding to different plastic responses collapse onto a universal function by appropriately rescaling the axes in terms of the plastic adaptation strength (Suppl. Fig. 3).
Results
We first analyse the evolution of the performance in learning more rules in parallel. Surprisingly, this exhibits a percentage of success for each rule that does not increase monotonically with the number of trials, as for single rule learning7 (Fig. 1). In the case of learning in parallel the OR and AND rules the non-monotonicity is dramatic: Already after few applications of both rules in parallel the network recovers the right answer to OR (blue circles) with a performance of about 100%, however it cannot provide the right answer to AND (red triangles) and is therefore unable to learn both rules (black line). After about 500 trials, the network starts recovering the right answer to AND, however this leads to a sharp decrease in the 100% OR performance, down to about 50% performance for both rules, corresponding to a random 0/1 answer. After further applications, the network improves again its performance for each individual rule, which implies that it has to partially forget the right answer to OR in order to learn the AND rule. A good performance of 90% in learning both rules is reached only after a longer training time. The above observation suggests that some rules (OR in the previous case) are “easier” to learn and will drive the initial performance. However the network needs a longer training time to organize the synaptic connections able to provide the right answer to both rules. This effect is also observed for three-rule-learning, where oscillations are less striking but well beyond statistical fluctuations. In this case the easiest rule to learn is the random rule (non-monotonicity shown in the inset), whereas AND and XOR learning exhibit similar performances. In both two and three-rule learning, the performance in the whole task is always monotonic. This suggests that the final response in complex learning results from the alternation of learning and forgetting, where partially “unlearning” is functional to the improvement in the overall performance.
We then question the dependence of the multi-task learning performance on the percentage of inhibitory synapses pin in the network. Surprisingly, the success rate does not show a monotonic trend with pin (Fig. 2). In particular the asymptotic value of the performance strongly depends on pin and exhibits a maximum for . Increasing the network size improves the asymptotic performance, as in single rule learning7, but does not affect the non-monotonic behaviour. Interestingly, the initial configuration of synaptic strengths does not affect the performance results. In particular, even starting from a uniform distribution of strengths for all neurons, at the end of the learning routine inhibitory neurons are on average stronger than excitatory ones and, for , the average strength ratio is close to typical balanced networks15. Therefore, the plastic adaptation regulates the relative strength and hinders the creation of excessively strong inhibitory synapses.
To better understand this result, we first analyse the structure of the union of all paths of active synapses connecting input and output neurons which provide the right answer to all rules. This constitutes the backbone supporting information transmission. Each backbone in Fig. 3 shows the temporal sequence of firing neurons, where neurons active simultaneously lie on the same horizontal line and the temporal evolution describes how firings propagate in time from input neurons (green) at t = 0 to the output neuron (black). The lines connect the pre-synaptic firing neuron with the post-synaptic stimulated neuron drawn at its own firing time. Therefore, a one-time-step line indicates a strong temporal correlation between successive firings, whereas long temporal connections evidence a temporal delay between the pre and post-synaptic neuron firings. The last time step contains always the output neuron, that can be also reached at earlier times, namely more than once during an avalanche. Fig. 3 shows that for a purely excitatory network and for pin = 50%, the backbone has a simpler structure involving few neurons, whose firings are mostly successive in time. Conversely, for pin = 30% the structure becomes larger and with an intricate network of firings. Indeed the size of the backbone averaged over many configurations exhibits a maximum value for this fraction of inhibitory synapses (Suppl. Fig. 4). Moreover, the average shortest path connecting input and output neurons within the backbone exhibits an opposite trend for different pin, namely a minimum value for (Suppl. Fig. 4). Furthermore, for these values of pin the multiplicity of independent paths through each neuron (Suppl. Fig. 5) varies over a wider range, indicating the emergence of alternative paths, enhancing cooperative effects under the combined action of excitatory and inhibitory synapses. At the same time, within this complex backbone a preferential path for information transmission is identified under the combined action of excitatory and inhibitory synapses. The fraction thus optimizes the structural features.
Recent experimental results have suggested that neuronal systems in a resting state can be viewed as systems acting close to a critical state where activity does not have a characteristic size. We question if the same behaviour is observed for the response to an external stimulus. We measure the size of neuronal avalanches as the number of neurons active when the system provides the right answer to each step of all rules. For purely excitatory networks learning is achieved only if all neurons are involved in the avalanche, i.e., the distribution is peaked at s* = N (vertical line Fig. 4). However, this does not imply that all neurons belong to the backbone, because many neurons are activated in an inefficient attempt to reach the output neuron. Conversely, for increasing pin, activity is progressively confined, namely the peak height decreases and the distribution extends towards smaller sizes. This implies that the system can recover the right answer by involving a limited number of neurons (Fig. 4) and therefore by a more efficient employment of resources. Results suggest that purely excitatory networks, able to exhibit avalanches without a characteristic size in spontaneous activity, cannot provide the right answer to a stimulus unless the entire system is involved. This observation is particularly striking in the case of a right answer “zero”. In the following, we will provide an understanding of these optimal learning conditions.
An avalanche of size s* represents a configuration able to learn the rule where s* neurons are active and N − s* are inactive. In the framework of spin models, this avalanche is therefore analogous to a configuration with s* up spins and N − s* down spins. The probability P(s*) to obtain an s*-avalanche, regardless of the individual firing neurons, is the probability to observe such configuration. Here pin tunes the degree of structural disorder in the system and therefore plays the role of temperature in thermal systems. We then define the entropy associated to the learning dynamics as
which quantifies the variability in the response provided by the system. The entropy S is equal to zero for pin = 0%, as for thermal systems at zero temperature, and progressively increases with pin (Fig. 5). In a learning routine the network must be able to trigger and sustain the activity in response to the external stimulus, which is, for instance, impossible is a purely inhibitory network. As a consequence, the level of excitability of the whole network must be adequate to tackle the task. We quantify the level of excitability of the network by evaluating the average synaptic strength E = <gij>, where gij is positive (negative) for excitatory (inhibitory) synapses. The excitability E is a decreasing function of pin and tends to zero for (Fig 5). In order to combine the above ingredients, we propose a novel functional
which quantifies the balance between variability, both in the structure and in the response, and the excitability of the system. This definition is reminiscent of the free energy in statistical physics, with the additional requirement that the system cannot learn in a state with zero excitability. As the free energy, F is composed of an energetic and an entropic term and vanishes in the extreme cases of absence of variability or excitability, assuring that both features are required for learning. This functional is a non-monotonic function of pin and exhibits a maximum for (Fig. 5), percentage that leads to optimal learning by the ability of the system to react to an external stimulus combined to the possibility of tuning the response saving resources.
Discussion
Real brains are able to perform more tasks in parallel. We investigate the multiple task learning performance of a neuronal network able to reproduce the critical behaviour of spontaneous activity. Networks undergo the teaching procedure of two or more rules in parallel, via a numerical procedure (different inputs and the same output for all rules) which allows, at the same time, to introduce neuronal firing interference and to monitor separately the performance to each rule. The network starts by learning the “easiest” rule, the one which requires less plastic adaptation. Under this point of view, OR is the easiest rule to learn, whereas XOR requires a longer training. Even if the first rule is learned, further applications stimulate the system to accomplish the entire task. The network is then obliged to redefine the connectivity network, even if it was successful in providing the right answer to the first rule. This operation implies that the system partially forgets the previous right answer, down to a level of performance where the answer is compatible to a random outcome. Only after this partial forgetting the system is able to identify a synaptic structure providing the right answer to all rules. Therefore, learning and forgetting appear to be the two, apparently opposite, mechanisms that must coexist to realize multi-task-learning.
Current evidence from functional magnetic resonance imaging (fMRI) experiments16,17 and EEG data18,19 shows that a greater brain signal variability indicates a more sophisticated neural system, able to explore multiple functional states. Signal variability also reflects a greater coherence between regions and a natural balance between excitation and inhibition originating the inherently variable response in neural functions. Furthermore, the observation that older adults exhibit less variability reflecting less network complexity and integration, suggests that variability can be a proxy for optimal systems. In our study, we quantify variability in the response by the Shannon entropy associated to neuronal activity in learning, combined to the structural disorder measured by the percentage of inhibitory synapses in the system. The quantity pinS is therefore an entropic term measuring the level of variability, far from being just noise. At the same time, the cognitive performance must rely on the capability of the system to react to an external stimulation. This feature can be interpreted as an energetic term, which is maximum for purely excitatory networks, where learning always triggers an all-encompassing activity, involving also unnecessary resources. Therefore, it is beneficial for brains performing complex tasks to realize a balance between the entropic and the energetic features.
Our study shows that multi-task learning is optimized not for a structurally balanced network with 50% inhibitory synapses, but for 30% inhibitory synapses, value measured in mammalian brains. This percentage allows the desired balance between excitability and variability in the response. To better understand why this particular value complies with the two main requirements for learning, excitability and variability, it is useful to recall the structure of the backbone of paths carved by the plastic adaptation process. Indeed, the backbone results in a more complex structure of firing connections for 30% inhibitory synapses, than for a structurally balanced network: The number of neurons involved usefully in the process increases. Indeed, a purely excitatory network involves all neurons in successfully performing a task but the majority of them does not operate to convey the information from input to output, i.e. the backbone is quite bare. At the same time the multiplicity of paths going through each neuron varies over a range which is maximum for pin = 30%. This does not simply imply longer paths, since the average length of the shortest path connecting input and output is minimum for pin = 30%. Also this result is striking since it would be more reasonable to expect shorter paths for a higher percentage of inhibitory synapses, corresponding to a multiplicity range an order of magnitude smaller. Results suggest that a complex backbone, with a wide range of path multiplicity and therefore the coexistence of a larger number of paths, including very short ones, is the optimal firing structure for learning. This observation confirms that learning is a truly collective process, where a large number of units with intricate connections participate to the dynamics. An excess of excitatory or inhibitory synapses, with respect to this optimal value, would hinder the emergence of this complex structure and therefore limit the learning performance of the system. It would be interesting to extend the present study to other networks, as for instance modular networks, and verify, as found for spontaneous activity, if the particular network topology affects the multi-task learning behaviour.
Methods
Neuronal network
We consider N neurons placed at random in a two-dimensional space and characterized by a potential vi. Each neuron can be either excitatory or inhibitory, according to Dale's law, with a random out-going connectivity degree, kout, assigned according to the experimental distribution of functionality networks20, with . This distribution implies that the network does not exhibit a characteristic connectivity degree. On the contrary, few neurons are highly connected and act as hubs with respect to information transmission, whereas the majority of neurons are connected to few other neurons. Connections are established according to a distance dependent probability, , where r is their spatial distance and the connectivity spatial range21. Once the outgoing connections are chosen, we evaluate the in-degree of each neuron kin. The initial synaptic strengths are randomly drawn from a uniform distribution, , where . Since the neuronal connectivity degree is power law distributed, the level of inhibition is expressed in terms of pin, the fraction of inhibitory synapses in the network, with inhibitory neurons highly connected (kout > 10)2.
A neuron i fires as soon as its potential is above a fixed threshold, , changing the membrane potential of post-synaptic neurons proportionally to gij8,9
where is the in-degree of neuron j, the sum is extended to all out-going connections of i and the plus or minus sign is for excitatory or inhibitory synapses, respectively. The factor makes the potential variation of neuron j, induced by neuron i, independent of the connectivity of both neurons, whereas the factor normalizes the synaptic strength values for each neuron. We note that the unit time step represents the time unit for propagation from one neuron to the connected ones, which in real systems could be of the order of 10ms. After firing, a neuron is set to a zero resting potential and in a refractory state lasting tref = 1 time step. The initial neuron potentials are uniformly distributed between vmax− 1 and vmax. A small random fraction 10% of neurons is fixed at zero potential and act as boundary sites or sinks for the charge.
Learning routine
For each rule we choose at random two input neurons and a unique output neuron, under the condition that they are not boundary sites and they are mutually separated on the network by kd nodes. kd represents the minimum distance between input and output neurons, which can also be joined by much longer paths. In the standard multi-layer perceptron framework, kd would play the role of the number of hidden layers. Assigning the same output neuron to all rules enables, at the same time, to discriminate between different rules and to have interference among the different paths carrying information. We test the ability of the network to learn in parallel different Boolean rules: AND, OR, XOR and a rule RAN, which associates a random output to all possible binary states at two inputs. For each rule the binary value 1 is identified with the neuron firing, i.e., at any time during the avalanche propagation. Conversely, the binary state 0 corresponds to the neuron which has been depolarized but fails to reach the firing threshold during the entire avalanche.
Once the input sites are stimulated, their activity may bring to threshold other neurons and therefore lead to an avalanche of firings. We impose no restriction on the number of firing neurons in the propagation and let the avalanche evolve to its end. If the avalanche during the propagation did not reach the output neuron, we consider that the system was in a state unable to respond to the given stimulus, and as a consequence to learn. We therefore increase uniformly the potential of all neurons by a small quantity, β = 0.001, until the configuration reaches a state where the output neuron is first perturbed. We then compare the state of the output neuron with the desired output. A single learning step requires the application of the entire sequence of states for each rule, letting the activity propagate till the end and then monitoring the state of the output neuron. If the answer is right for all three entries of a rule, this has been learned. The routine proceeds until the system learns all rules.
Plastic adaptation
Plastic adaptation obeys a non-uniform generalization7 of the negative feedback algorithm22: If the output neuron is in the correct state according to the rule, we keep the value of synaptic strengths. Conversely, if the response is wrong, we modify the strengths of those synapses active during the avalanche23 by , where dk is the chemical distance of each presynaptic neuron from the output neuron and α represents the strength of synaptic adaptation. Here α represents the ensemble of all possible physiological factors influencing synaptic plasticity. Therefore synapses can be either strengthened or weakened depending on the mistake: If the output neuron fails to be in a firing state, we strengthen the synapses, conversely strengths are weakened if the right answer 0 is missed. Once the strength becomes smaller than a threshold, gij <10−4, the synapse is pruned. This ingredient is very important as since decades the crucial role of selective weakening and elimination of unneeded connections in adult learning has been recognized24,25.
The synapses involved in the signal propagation and responsible for the wrong answer, are therefore not adapted uniformly but inversely proportional to the chemical distance from the output site. Namely, synapses directly connected to the output neuron receive the strongest adaptation ±α. This rule dependence on dk models the feedback to the wrong answer triggered in the region of the output neuron and propagating backward towards the input neurons. In the brain this mechanism is regulated by the release of messenger molecules, as some hormones (dopamine suppressing LTD26 or adrenal hormones enhancing LTD27) or nitric oxide28. Three different procedures for plastic adaptations are tested: Homeostatic, uniform and restricted. In homeostatic plasticity the adaptation of excitatory and inhibitory synapses have opposite signs to realize conservation of the average strength1. In the uniform case all active synapses undergo the same adaptation, whereas in the restricted case only excitatory synapses are modified.
Author Contributions
L.d.A. and H.J.H. were involved in all the phases of this study. V.C. performed the numerical simulations and prepared the figures. All authors reviewed the manuscript.
Supplementary Material
Acknowledgments
We thank the European Research Council (ERC) Advanced Grant 319968-FlowCCS for financial support.
References
- Royer S. & Paré D. Conservation of total synaptic weight through balanced synaptic depression and potentiation. Nature 422, 518 (2003). [DOI] [PubMed] [Google Scholar]
- Bonifazi P. et al. GABAergic hub neurons orchestrate synchrony in developing hippocampal networks. Science 326, 1419 (2009). [DOI] [PubMed] [Google Scholar]
- Arieli A., Sterkin A., Grinvald A., Aertsen A. Dynamics of ongoing activity: Explanation of the large variability in evoked cortical response. Science 273, 1868 (1996). [DOI] [PubMed] [Google Scholar]
- Beggs J. M. & D. Plenz D. Neuronal avalanches in neocortical circuits. J. Neurosci. 23, 11167 (2003). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gireesh E. D. & Plenz D. Neuronal avalanches organize as nested theta- and beta/gamma-oscillations during development of cortical layer 2/3. Proc. Nat. Acad. Sci. USA 105, 7576 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Petermann T. et al. Spontaneous cortical activity in awake monkeys composed of neuronal avalanches Proc. Nat. Acad. Sci. USA 106, 15921 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- de Arcangelis L. & Herrmann H. J. Learning as a phenomenon occurring in a critical state. Proc. Natl. Acad. Sci. USA 107, 3977 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- de Arcangelis L., Perrone Capano C. & H J. Herrmann H. J. Self-organized criticality model for brain plasticity. Phys. Rev. Lett. 96, 028107 (2006). [DOI] [PubMed] [Google Scholar]
- Pellegrini G. L., de Arcangelis L. & Herrmann H. J. Perrone Capano, C. Modelling the brain as an Apollonian network. Phys. Rev. E 76, 016107 (2007). [DOI] [PubMed] [Google Scholar]
- F. Lombardi F., Herrmann H. J., Perrone-Capano C., Plenz C. &. d. e. Arcangelis L. Balance between excitation and inhibition controls the temporal organization of neuronal avalanches. Phys. Rev. Lett. 108, 228703 (2012). [DOI] [PubMed] [Google Scholar]
- Barto A. G., Sutton R. S. & Anderson C. W. Neuron-like adaptive elements that can solve difficult learning control problems. IEEE Trans. Syst. Man. Cybern. 15, 835 (1983). [Google Scholar]
- Rumelhart D. E. Hinton G. E. & Williams R. J Learning representations by back-propagating errors. Nature 323, 533 (1986). [Google Scholar]
- de Arcangelis L. & Herrmann H. J. Activity-dependent neuronal model on complex networks. Frontiers in Phisiol. 3, 62 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- de Arcangelis L., Lombardi F. & Herrmann H. J. Criticality in the brain. J. Stat. Mech. 3, P03026 (2014). [Google Scholar]
- Brunel N. Dynamics of Sparsely Connected Networks of Excitatory and Inhibitory Spiking Neurons. J. Comput. Neurosci. 8, 183 (2000). [DOI] [PubMed] [Google Scholar]
- Garrett D. D., Kovacevic N., McIntosh A. R. & Grady C. L. Blood Oxigen Level-Dependent Signal Variability Is More than Just Noise J. Neurosci. 30, 4914 (2010). [DOI] [PMC free article] [PubMed]
- Garrett D. D., Kovacevic N., McIntosh A. R. & Grady C. L. The Importance of Being Variable J. Neurosci. 23, 4496 (2011). [DOI] [PMC free article] [PubMed]
- Ghosh A., Rho Y., McIntosh A. R., Kötter R. & Jirsa V. K. Noise during rest enables the exploration of the brain's dynamic repertoire. PLoS Comput. Biol. 4,e1000196 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
- McIntosh A. R., Kovacevic N. & Itier R. J. Increased brain signal variability accompanies lower behavioral variability in development. PLoS Comput. Biol. 4,e1000106 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eguiluz V. M., Chialvo D. R., Cecchi G. A., Baliki M. & Apkarian A. V. Scale-free brain functional networks. Phys. Rev. Lett. 94, 018102 (2005). [DOI] [PubMed] [Google Scholar]
- Roerig B. & Chen B. Relationships of local inhibitory and excitatory circuits to orientation preference maps in ferret visual cortex. Cerebral Cortex 12, 187 (2002). [DOI] [PubMed] [Google Scholar]
- Bak P. &. Chialvo D. R. Adaptive learning by extremal dynamics and negative feedback. Physical Review E 63, 0319121 (2001). [DOI] [PubMed] [Google Scholar]
- Frey U. & Morris R. G. M. Synaptic tagging and long-term potentiation. Nature 385, 533 (1997). [DOI] [PubMed] [Google Scholar]
- Young J. Z. A model of the Brain (Clarendon, Oxford., 1964). [Google Scholar]
- Changeux J. P. Neuronal Man: The Biology of Mind (Oxford University Press, New York., 1985). [Google Scholar]
- Otmakhova N. A. & Lisman J. E. D1/D5 Dopamine receptors inhibit depotentiation at CA1 synapses via cAMP-dependent mechanism. J. Neurosci. 18, 1270 (1998). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Coussens C. M., Kerr D. S. & Abraham W. C. Glucocorticoid receptor activation lowers the threshold for NMDA-receptor-dependent homosynaptic long-term depression in the hippocampus through activation of voltage-dependent Calcium channels. J. Neurophysiol. 78, 1 (1997). [DOI] [PubMed] [Google Scholar]
- Reyes-Harde M., Empson R., Potter B. V. L., Galione A. & Stanton P. K. Evidence of a role for cyclic ADP-ribose in long-term synaptic depression in hippocampus. Proc. Natl. Acad. Sci. USA 96, 4061 (1999). [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.