Skip to main content
Springer logoLink to Springer
. 2026 Feb 23;120(2):5. doi: 10.1007/s00422-026-01038-4

Brain-inspired energy efficient technologies for next-generation artificial intelligence

Hillel J Chiel 1,2,3, Jay S Coggan 4, Gourav Datta 5, Jean-Marc Fellous 6, William R P Nourse 7, Roger D Quinn 7, Peter J Thomas 1,5,8,9,
PMCID: PMC12929283  PMID: 41731107

Abstract

Since the advent of widely accessible AI tools, AI technology has been in high demand by businesses, academic researchers and individuals. Technology companies are building AI infrastructure at a rapid pace, and these facilities consume vast and growing resources, particularly electricity and water, with significant real and projected climate impacts. There is a need for new research initiatives to support long time horizon efforts to develop energy efficient computing capabilities to support the continued growth of AI infrastructure in a sustainable fashion. Such efficiency is required at both the hardware and software levels. Where can industry turn for examples of ultra-low power, energy efficient computing? We argue here that neurobiological principles offer rich and under-exploited sources of inspiration for energy efficient NeuroAI, and that new partnerships between industry and academia should be developed in this direction.

Keywords: Neuroscience, Artificial intelligence, Energy efficiency, BioFlop

Introduction

One way to achieve energy efficient artificial intelligence (AI) is to study how nature generates naturally intelligent behavior that minimizes energy expenditure. All biological organisms are shaped by developmental and evolutionary processes that require energy for their survival and reproduction. An organism whose energetic costs chronically exceed its energy intake will suffer dire consequences, and on the evolutionary time scale, will be eliminated. We will briefly review energy usage in biological brains and bodies and then suggest potential new directions for artificial intelligence in general and NeuroAI in particular, a new type of artificial intelligence designs inspired by brain mechanisms. The ideas presented could also apply to robotics and autonomous agents.

Biological brains and bodies

Biological brains and energy

Biological brains (Fig. 1) are by far the most energy efficient computing devices that we know, using only about 20 W of power in the case of the human brain, which is more than 1 million-fold less than the world’s largest supercomputer, a machine that does not yet approximate human intelligence, in spite of the impressive feats of the best LLMs (large language models) today. As Moravec articulated, tasks that are difficult for humans, such as playing chess or learning languages, are easy for AI, while many tasks that are easy for humans are extremely difficult for machines (Moravec 1988). This points to many important issues that are beyond the scope of this perspective. But as a matter of example, self-driving car enthusiasts were premature and underestimated what the human brain needs to do to drive to work without wreaking havoc. Some of what humans do best, and easily, are among the hardest and most important things they do for survival. The dynamical thinking at which humans excel, and that is at the core of biological intelligence, is not even attempted in LLM or any other AI model.

Fig. 1.

Fig. 1

Energy usage in the brain. A. Glucose crosses the blood-brain barrier, and is metabolized by neurons, astrocytes, and other glial cells. B. Glucose utilization can be measured, creating an intensity map of activity throughout the brain. Images (used with permission) from (Jamadar et al. 2025)

It is possible that we will not know exactly how efficient the brain actually is in units of Watts/Flops until we establish a biological equivalent, such as the BioFlop, as coined by (Stiefel and Coggan 2023). This hypothetical value may be difficult to pin down precisely due to the analog nature of biological computing. In this same paper, the authors introduced a measure of AI efficiency; the ERASI equation. This calculation has initially shown that the cost of mimicking the human brain (based on what we currently know about its computations) would be orders of magnitude higher than the entire annual US energy output. As a benchmark, the authors used estimates of energy use from the now disbanded Blue Brain Project of EPFL (https://bluebrain.epfl.ch/bbp/research/domains/bluebrain/), a Swiss government initiative to simulate an entire mouse brain based on a biologically realistic reconstruction of brain circuits. This initiative has birthed a very different approach to AI than LLMs, one that attempts to discover new biological learning rules that could perhaps be implemented in future Machine Learning (ML) systems (see also https://www.openbraininstitute.org; https://www.inait.ai). Currently, these efforts and other similar AI work inspired by the brain are captured by the generic term “NeuroAI” (Zador et al. 2023; Sadeh and Clopath 2025; Arbib 2025).

Artificial neural networks were initially developed based on theories in neuroscience from the 1950s, but neuroscience has made a lot of progress since then, which has not been incorporated into mainstream AI and ML. For example, one of the many limitations of LLMs and other Artificial Neural Network approaches is underestimating how single cells process information. While LLMs use a “neural” architecture, their low-level components have little in common with the physiology of nerve cells and biological synapses, which are themselves capable of processing and transforming information. Artificial neural networks are abstracted from synaptic integration and all-or-none action potential generation present in biological networks, to a simplified linear summation followed by a nonlinear response such as a logistic, tanh or ReLu operator1 (Haykin 1999). There is a growing theoretical base for conceptualizing this level of computation (Rieke et al. 1997; Poirazi and Mel 2001; Arcas et al. 2003; Coggan et al. 2020; Lillicrap et al. 2020; Coggan et al. 2022). There may be more fundamental and yet unexplored units of computations than the neuron (e.g. dendrite, synapse, ion channels or metabolic states), which could also point in new directions for energy efficiency (Boahen 2022). In short, there may be great value in synergistically coupling brain research and new initiatives in AI research. Beyond biomimicry, we have much more to learn from biological solutions and computational neuroscience about the fundamental mechanisms of intelligence. It cannot be overemphasized that basic computational and experimental neurobiology research will lead to more powerful and efficient AI.

Here we review concepts from biology and use them as a springboard to suggest examples of areas where AI technologists might benefit from deeper understanding of neurobiology and behavior.

Biological neural dynamics

Biological neurons use energy to maintain their ability to process and generate both electrical and chemical signals. Across the membranes of nerve cells, energy requiring pumps (e.g. the sodium/potassium ATPase pump) ensure that, inside the cell, potassium ion concentrations are higher and sodium ion concentrations are lower than outside the cell. The differential concentrations of ions across the cell membrane are the basis for the resting potential, a nonzero membrane potential difference (often about -60 mV) that allows neurons to rapidly respond to and integrate electrical inputs (Berndt and Holzhutter 2013). Neurons also maintain a complex panoply of ion channels that can be gated (i.e., can be opened or closed) by voltage, specific chemicals that bind to specific receptors, the calcium ion, second messengers, or by different forms of energy (e.g., photons, temperature gradients, mechanical deformation, applied force, or vibration) (Bhattacharjee 2023). The specific complement of ion channels in a neuron’s membrane endows it with complex dynamical properties, so that neurons can be silent, or spontaneously fire action potentials. Some neurons can generate spontaneous rhythmic bursting (pacemaker neurons) (Ramirez et al. 2004), and neurons can have multiple stable dynamical properties (e.g., they can be switched from being silent to bursting regularly) (Connors and Gutnick 1990). Unlike most current NeuroAI systems where all neuronal elements are identical, biological neurons exhibit a high degree of individual differences that may be at the origin of high computational capacity and efficiency. Ion channels responsive to energy and chemicals in the environment allow neurons to sense environmental conditions and are the basis for the sensations of touch (Jin et al. 2020), taste (Spector and Travers 2021), smell (Fulton et al. 2024), sight (Barret et al. 2022), hearing (Zheng and Holt 2021), and the ability to sense the location, position and overall internal state of the body itself (proprioception; (Moon et al. 2021)). In the cognitive areas of the brain (involved in learning, memory, and decision making), neurons are mostly silent and communicate with each other either using bursts of action potentials (packets of information, temporal summation) or synchronized action potentials (simultaneous production of tightly time-locked action potentials to a target neuron, spatial summation) (Tiesinga et al. 2008; Lisman 1997). Again, these rich intrinsic firing properties and dependences of firing on the environments are usually not present in NeuroAI systems.

Biological neural synapses

Neurons connect to one another via synapses, which may be either electrical or chemical, and the energetic requirements of synaptic activity are one of the largest uses of energy by neurons (Li and Sheng 2022; Harris et al. 2012). Electrical synapses provide a way of rapidly linking activity between neurons, whereas chemical synapses provide somewhat slower but highly modifiable means of exciting or inhibiting another neuron, potentially greatly amplifying or dampening the strength of the signal. Chemical synapses can connect neurons over multiple temporal and spatial scales: the rapid binding of neurotransmitters released within the synaptic cleft can induce rapid changes in other neurons, with some receptors responding more slowly than others. Moreover, some neurotransmitters can diffuse out of the specialized synaptic endings and bind distant receptors, affecting neurons that are not directly connected over a much larger spatial volume. Neuromodulatory transmitters have been shown to lead to higher energy consumption during more complex cognitive processing in the cerebral cortex (Castrillon et al. 2023). Maintaining chemical synaptic transmission requires significant energy, and so does synaptic plasticity, the change in synaptic strength as a function of conjunctive activity of a neuron and other connected neurons. There are two kinds of plasticity: the first is short term (depression, facilitation) in the order of 10-100ms and controls the way sequences of signals are transferred from neuron to neuron (e.g. frequency filtering) (Asopa and Bhalla 2023; Yu et al. 2025b). The second is long term (LTP - long-term potentiation, LTD - long-term depression, minutes to days) and controls the “relevance” of the synapse to general neural computation (Murai and Goto 2025; Stanton 1996; Bear and Malenka 1994). This type of plasticity has inspired “changes in synaptic weight” rules in artificial neural networks, while the first short-term type is generally ignored. Another important energy requirement is the remodeling of synaptic connections, as neurons may grow additional connections to strengthen or remove connections to weaken interactions with other neurons. In cognitive areas, a given neuron receives potential inputs from about 10,000 other neurons, and a fraction of these inputs are almost always active, keeping the neurons in a constant state of subthreshold fluctuations, ready to fire in response to small but significant input signals. This type of background subthreshold activity is not yet accounted for and used by NeuroAI systems.

Biological neural shapes and efficient computation

Neurons have complex shapes (morphology) that allow them to process inputs both spatially and temporally, and studies suggest that energy is crucial for determining and maintaining their shapes (Wen and Chklovskii 2008). Early anatomists saw the similarity between neuronal branching and those of trees, referring to the fine branches of neurons on which connections are made as dendritic arborizations (Yuste and Tank 1996). Neuronal shapes determine whether sub-threshold electrical inputs travel to the spike-initiating zone and generate outputs from the entire neuron. Furthermore, localized interactions can allow neurons to process inputs at many different locations in parallel. Energy requirements influence the maintenance and changes in shape that neurons undergo over an animal’s lifetime (Rumpf et al. 2023). Artificial Neural Networks usually consider neurons as shapeless ‘point-like’ units.

Maintaining biological neural networks

Another major use of energy in the brain is the maintenance of the neuronal support cells, the glial cells, which play significant roles in providing energy to neurons, and for generating the wrapping material (myelin, which gives white matter its name) that allows rapid transmission along the long axons of neurons. White matter is responsible for long distance communication between regions of gray matter, which contain the dendrites, cell bodies, and synapses of the neurons. White matter consumes energy at a much lower rate than does gray matter (Yu et al. 2018). Glial cells also regulate the amount of energy available to nerve cells (Shoenhard and Sehgal 2025) and play a critical role in clearing debris and responding to attacks from bacteria, viruses, and other sources of inflammation, which also requires significant energy (Jamadar et al. 2025). There are no implementations of glial cells in NeuroAI system,

Biological neural architectures and sparse connectivity

Across phylogeny, different regions of the nervous system are organized into different complex architectures. The central complex of arthropods, the vertical lobe system of the octopus, the hippocampus, the cerebral and cerebellar cortices in primates and humans are all striking examples of highly organized anatomical organizations of input, output and interneuronal circuitry (Luo 2021). Neural architectures clearly subserve specific functions. In addition to highly organized local patterns of connectivity, there are long-range connection tracts, many of which constitute interconnection “hubs”, which have the highest energy requirements (Ceballos et al. 2025; Jamadar et al. 2025). Neuronal architectures are generated through developmental processes, which also require and are regulated by energy availability (Ghosh et al. 2023). Figure 2 illustrates some consequences of sparsity and structure in natural versus artificial neural networks. More needs to be done in current NeuroAI system to allow for and exploit energy-dependent ‘developmental-like’ changes in architectures.

Fig. 2.

Fig. 2

Sparsity and structure in natural and artificial neural networks. A. In modern AI systems, most computation is performed on hardware (GPU, NPU) which uses dense vector and matrix multiplication via multiply-accumulate (MAC) modules. As a toy example, multiple input vectors are multiplied by a weight matrix, then the resulting vectors are processed by some activation function. These operations can be carried out in parallel, but this dense MAC operation is at the core. B. In contrast, natural networks of neurons are interconnected in 3D space, with sparse connectivity and activity patterns. Image (used with permission) from (Gamlin et al. 2025). C. Energy use in dense 2D accelerators such as GPUs/NPUs scales quadratically with increasing numbers of neurons. Biological systems, on the other hand, scale linearly in energy use. Image (used with permission) from (Boahen 2022)

Transmission rates versus information transfer in biological nervous systems

Patterns of neural activity achieve a tradeoff between frequency of transmission and information transfer because of energetic constraints. Although neurons are capable of high tonic firing rates of firing, the energetic costs of operating action potentials (and especially synapses) favor lower rates of transmission. Ongoing spontaneous activity within the brain provides a floor to the minimum rates of transmission that are above the background noise, but also may provide the brain with the ability to mobilize long range connectivity much more rapidly and with relatively small energy expenditures (Harris et al. 2012). Interestingly, it has been suggested that transient patterns of high firing rate (bursts) could carry more information than single action potentials (Lisman 1997), and that bursting may be key to implementing classic artificial neural network mechanisms such as back-propagation (Sun et al. 2021; Payeur et al. 2021)

Biological neural network preprocessing

A major role of nervous systems is to filter out irrelevant stimuli and enhance stimuli that are most likely to be immediately relevant. Processes such as habituation act to reduce activity in sensory neurons that would otherwise be activated by irrelevant stimuli (Ramaswami 2014). Feedforward activation makes it easier to begin to excite neurons that are most likely to be encountering sensory inputs or generating motor outputs next (Briggs 2020). These type of preprocessing increases the effectiveness and efficiency of information processing.

Biological brains and behavior

Unless an organism can photosynthesize and reproduce asexually, it must be able to move through and interact with its surrounding environment to obtain food and mates, and to avoid predators. Within an organism, internal organs, such as the heart, lungs, and digestive system, also need to be controlled. Thus, neurons must interact, through the neuromuscular junction, with smooth, cardiac and skeletal muscle. Once again, energy plays critical roles in both the function of muscle and in the behavior that it generates. Depending on the size and speed of a behavior, energy may be partitioned among inertial energy, viscous (dissipative) energy, elastic energy, and responses to gravity (Sutton et al. 2023). These different forms of energy imply different stabilities in response to perturbations, which in turn demand different controls strategies for behavior. Energy obtained through feeding must also be partitioned between the needs of the body and of the brain, and complex regulatory mechanisms are responsible for maintaining the appropriate levels of energy for each (Myers et al. 2021). Such energy-expenditure specializations would likely benefit autonomous mobile or actuating artificial systems such as NeuroAI-enabled robots.

Unsupervised learning in biological nervous systems

Over longer periods of time, behavior is regulated by positive and negative reinforcement from the environment (Sutton and Barto 2018). One component of these regulators is the physical effort and energy that must be expended to achieve goals (Jiang et al. 2024). Another is the experience of pain or pleasure in response to an organism’s actions, and emotions that reflect an organisms overall internal state and response to past, present and future events (Fellous and Arbib 2005; Arbib and Fellous 2004). Artificial neural networks are trained and modified using datasets in which elements are considered intrinsically equally important. Their relative importance is determined by how often they occur, rather than by how relevant they might be to learning a particular body of knowledge. AI systems typically do not learn in “one shot”, as humans and animals often do (Yu et al. 2025a). More generally, living organisms are autopoietic, networks of processes that maintain and renew themselves through their own activity (Maturana and Varela 1980). Reducing the amount of training data, using considerations of their intrinsic value to learning could yield significant improvement in the explainability of AI algorithms (Wells and Bednarz 2021) and energy consumption savings in future NeuroAI learning systems.

Biologically-inspired energetically efficient approaches to AI

Given these general insights from biological systems, we next list some ideas that could be explored in the near-term to improve the energy efficiency of AI systems.

Neuromorphic computation

New approaches to computing that are inspired by biological neurons may lead to significant savings in energy expenditure. Hardware approaches subdivide into attempts to capture the complex analog processing that occurs in biological neurons using very low energy levels; others attempts to use standard silicon technology to emulate neurons, and some commercially available neuromorphic chips can be used for rapid preprocessing of sensory information on chip (e.g., neuromorphic cameras); others use sparse connectivity and computation for high energy efficiency (e.g., BrainChip, SpikeCore, Brain Corp). Intel has developed a neuromorphic chip (Loihi2) that uses a spiking neural network for computation, and that communicates and computes using “spike events”, reducing power consumption. In contrast to this chip, which ensures that all processing is completed at each step, a more distributed architecture has been pioneered by SpiNNcloud, which has multiple cores that simulate neurons communication asynchronously. By using denser local connectivity (but not full connectivity) and sparser global connectivity, these chips can also be more energy efficient. Many neuromorphic chips also incorporate plastic synapses that allow for learning.

Sparse representations may be energy efficient

There is a large body of neurobiological and computational evidence pointing to sparse coding in the nervous system (Beyeler et al. 2019). Sparse representations that improve reconstruction from noisy data have been intensively studied in applied mathematics (Calvetti et al. 2019; Calvetti et al. 2020; Calvetti and Somersalo 2024). The connection between sparse representation and energy efficiency has been investigated in neurobiology (Hu et al. 2012; Sacramento et al. 2015; Moosavi et al. 2024). In addition to saving energy, sparse representations could also render AI systems more robust to adversarial attacks and catastrophic failures.

Sparse bursting activity is common in motor control/central pattern generator systems

In neuro-motor systems, it is inefficient to simultaneously tense opposing muscles, or to activate muscles that are not needed for a given movement. Thus, it is no surprise that patterns of activity in biological motor control systems exhibit sparse bursting patterns, in which a small number of neurons are active at any one time. Such sparse activity patterns are mediated by inhibition, in which the activity of one cell suppresses the activities of others. Similarly, inhibition plays a key role in winner-take-all (WTA) algorithms that select one out of many possible answers (Maass 2000). If a unit consumes more energy when it is “active” than when it is “silent” then WTA dynamics lends itself to energy-efficient implementation. Inhibition also underlies stable heteroclinic channels (SHCs: Fig. 3) that have been proposed as a dynamical architecture supporting sparse, functionally effective activation patterns in motor systems (Shaw et al. 2015; Horchler et al. 2015; Rouse and Daltorio 2021; Mengers et al. 2025), sensory systems (Laurent et al. 2001; Rabinovich et al. 2010) and cognitive processing (Afraimovich et al. 2011; Rabinovich and Varona 2018). Research initiatives to study SHCs, WTA architectures, and other mechanisms for inhibition-dominated connectivity in AI systems may lead to novel energy-efficient solutions.

Fig. 3.

Fig. 3

Diagram illustrating a heteroclinic channel. A heteroclinic cycle in a dynamical system is a sequence of trajectories connecting the flow out of and into successive saddle point equilibria. A stable heteroclinic channel attracts all nearby trajectories. See (Rouse and Daltorio 2021) for a detailed description. Image (used with permission) from (https://en.wikipedia.org/wiki/Heteroclinic_channels)

Neuromodulation suppresses the activity of subnetworks that are not needed for a given task

In motor control systems, metabolic resources are redirected to muscle systems that are actively being used. Switching muscles “on” and “off” is regulated by neuromodulation (Sillar et al. 2014). Similarly, activity in neural subsystems is also regulated by neuromodulation. Specializing circuits for specific computational tasks and then powering down those circuits when those tasks are not needed may recapitulate biological-like (biomorphic) efficiencies. More generally, the extent to which neuromodulation in general (e.g. via serotonin, norepinephrine, dopamine, and others) has evolved to improve the efficiency of neural computation (rather than perform neural computation per se) is understudied and deserves attention (Castrillon et al. 2023; Yu et al. 2025c). Implementing neuromodulatory principles in “classic” AI or more modern NeuroAI models might significantly improve their performance, and consequently, their energy expenditure.

Plasticity modulation for adaptive intelligence

In addition to gating circuit activity for energy efficiency, neuromodulation in biological systems plays a central role in regulating when and where learning occurs. Neuromodulators such as dopamine, acetylcholine, and norepinephrine do not directly trigger but instead modulate synaptic plasticity, guiding long-term changes in neural circuits based on behavioral relevance, reward prediction, or uncertainty (Marder 2012). This dynamic plasticity control enables animals to learn selectively, preserving stable functions while adapting rapidly to new situations—a capability that current AI systems, including LLMs, largely lack. Instead, these models rely on static, global learning schedules and costly retraining for adaptation (Bommasani et al. 2021). By integrating neuromodulatory principles, such as plasticity gating conditioned on internal goals or novelty signals, future AI systems could enable targeted, context-aware updates to their internal states. This approach offers a path to learning-to-learn (meta-learning) in large-scale models, improving data efficiency, stability, and the ability to generalize across tasks, while avoiding the overhead of continual full-network retraining (Wang et al. 2016; Wang et al. 2018; Hattori et al. 2023; Goudar et al. 2023).

Memory changes with time and use, yielding more efficient representations

Unlike computers, we (humans and animals alike) store information neither “forever” nor exactly as it was acquired. We do not store pixel-level images, or second-by-second sequences of speech or episodic memories. We have an ability to abstract, simplify or ignore our inputs as they come in, and to re-shape our memories as a function of how or how often we use them, or as a function of their intrinsic “importance”. These features have undoubtedly evolved (at a cost for reliability) to address our limited capacities to perceive and memorize, saving energy and time. AI systems do not usually implement the (useful) mechanisms of forgetting, or memory consolidation (as during sleep (Rouast and Schonauer 2023)). While industry considers these features as deleterious and artifactual, to be avoided and corrected to achieve reliable and precise recall, research into LSTM networks with “forgetting states” has shown promise for continuous learning and other applications (Gers et al. 2000; Wang et al. 2025). New research should explore the extent to which a trade-off can or should be achieved between full recall and precision, and efficiency of representations.

Neuroscience-inspired memory efficiency

Biological brains achieve remarkable memory efficiency through mechanisms such as sparse coding, synaptic consolidation, and multi-scale memory hierarchies. Unlike large-scale AI models that rely on dense parameter storage and exhaustive training data exposure, the brain encodes information using compact, context-dependent neural activations, often reusing the same networks across tasks via dynamic routing or population coding. Furthermore, systems consolidation, involving transfer from fast-learning hippocampal circuits to slower cortical storage, enables long-term retention without continuous memory access. These strategies contrast with the monolithic memory structures of LLMs, which grow in size and cost with increasing data. Incorporating such biological insights—e.g., sparsity-inducing priors, gated memory consolidation, or attention-modulated storage and retrieval mechanisms—could significantly reduce the memory footprint of large models without sacrificing performance. This approach not only improves energy and storage efficiency, but also supports more flexible, lifelong learning paradigms where memory is treated as a dynamic, structured resource rather than a static archive (Olshausen and Field 1996; Zenke and Gerstner 2017; McClelland et al. 1995).

Multi-resolution representations may be energy efficient

Receptive fields such as those found in visual cortex have inspired significant contributions in Artificial Learning systems, such as convolutional neural networks (Celeghin et al. 2023). Typically, in these systems, all convolutional kernels are similar. Not all information, however, needs to be represented at the same resolution. The visual system has evolved multi-scale coding (different receptive field sizes within cortical areas, e.g. cortical area V1, and across areas, e.g. cortical areas V1-V4-IT). Spatial navigation in large environments uses multi-scale hippocampal place fields (Harland et al. 2021; Eliav et al. 2021). This multi-scale representation has been shown to have many computational advantages, including efficiency and fast attentional shifting. Most AI systems do not make use of multi-scale representations.

Cortical traveling waves and structured dynamics in large-scale neural models

A growing body of neuroscience research points to the central role of spatiotemporal cortical waves, such as alpha, beta, and theta oscillations, in orchestrating perception, attention, and memory in the brain. These waves enable multiplexed signaling and efficient integration across distributed cortical areas by rhythmically modulating neuronal excitability (Muller et al. 2018). Recent work suggests that oscillatory activity in the brain may be a form of low-energy analog computing that supports working memory (Lundqvist et al. 2023). Yet modern large-scale AI models, including LLMs, operate with fundamentally static or token-synchronous dynamics. Bridging this gap presents a compelling research frontier: introducing traveling-wave-like dynamics and oscillatory gating into LLMs and other large-scale architectures (Muller et al. 2024). This approach could involve rhythm-based attention windows (Tiesinga et al. 2004; Tiesinga and Sejnowski 2010; Fries 2023), phase-coupled memory activation (Roehri et al. 2022), and dynamic routing paths that emulate the selective coherence seen in cortical circuits (Fries 2023; Banaie Boroujeni and Womelsdorf 2023). Such structured dynamics would not only reduce inference cost and improve temporal coherence but also pave the way for closed-loop agentic systems, where internal wave states regulate external actions in response to changing sensory or goal contexts. Simulating these wave phenomena in neuromorphic or FPGA-based edge architectures can further support real-time, biologically grounded inference under strict energy constraints.

Synthetic nervous systems

A new neural network architecture, Synthetic Nervous Systems (SNS), has begun to be developed that is more directly inspired by biological nervous systems, requires much smaller training sets, and because of the sparsity of connections, is likely to be far more energy efficient. In the common instantiation, model neurons do not spike, but they can contain multiple ionic conductances, endowing them with complex temporal dynamics similar to those of biological neurons, and their synaptic connections are also dynamic and can change with experience (Szczecinski et al. 2017a). The SNS framework has been extended to simple spiking systems as well (Szczecinski et al. 2020). By using the functional sub-network approach, it is possible to reliably and automatically map a particular function to a small network of these model neurons (Szczecinski et al. 2017b). Once the dynamics of the network have been established, improving parameters using standard optimization techniques is very effective. A software tool allows SNS networks with thousands of neuron to be designed and simulated in real-time (Nourse et al. 2023). These networks have been used for modeling neuromechanical systems, and for the control of biologically-inspired robots.

Leaner training algorithms should save energy

One of the major differences between natural (human/biological) and artificial intelligence is the ability to (1) learn with few examples, (2) learn continuously with minimal teaching or supervision, and (3) use common-sense reasoning (Choi 2022). These well-documented features save time and energy. Due to availability and relatively low cost of computing resources, current AI systems do not typically attempt to address this need for enormous training datasets, and using data generated by AI systems to train themselves has proven dangerous (Shumailov et al. 2024). The industry has not yet shown a motivation to develop methods to minimize the training sets. New research specifically targeted at continuous learning and making the most of small training datasets techniques and theories (Perera-Lago et al. 2024) could be a significant step forward for energy efficient AI, especially at the “edge” (i.e. close to the user). Alternative algorithms to those commonly used in LLMs today could also be tested for efficiency, as well as superior outcomes (Grossberg 2020; Wang et al. 2024).

Organoid computation

Investigating the limits of single neuron information processing capabilities has taken many forms over decades in living models from protists such as slime molds to mammalian neurons (e.g., (Zhu et al. 2018; Fitch 2021)). More recently, the organoid, or brain-in-a-dish, approach is showing some promise where small groups of neurons are harnessed to compute input-output relationships, even to the point of playing video games (Kagan et al. 2022). These efforts, and the ability to “design” neurons using technologies such as CRISPR/Cas9, will likely also lead to more synthetic biological approaches to discover multi-scale computation principles (Kagan et al. 2023).

Offline and offsite processing save time and energy

Our brain processes information during sleep, with minimal energy expenditure. It is also capable of dynamically allocating resources (e.g. attention) to specific tasks or sensory pathways in a context-dependent manner (Murai and Goto 2025). AI is essentially stimulus driven (prompt, get an answer), and is active when humans are active (peak energy expenditure). One could imagine world-wide AI architectures that allocate AI tasks to regions of the world where energy is the cheapest/demand is lowest (e.g. prompt during the day in the USA, get an immediate AI answer/processing at night, in India). One could also imagine AI systems that compile answers, restructure knowledge, and improve their database during “off-line” hours (when humans sleep). Not all AI computations need be human-stimulus driven.

Discussion

Longer term concerns

It is time to clarify what we want from AI systems and what AI actually means. Large Language Models (LLMs) are remarkable tools in their early stages but arguably not intelligent systems by any useful definition and their use to achieve Artificial General Intelligence (AGI) is increasingly in doubt (Hofkirchner 2023; Madabushi et al. 2025; Mumuni and Mumuni 2025). LLMs are structurally very limited, being simply rapid statistical sampling and prediction programs that operate on input data. In addition, the efficacy of representing higher-level human cognition with probabilistic models has been questioned (e.g. (Marcus and Davis 2013)). The well-known “hallucinations” (or confabulations) LLMs can produce result from intrinsic design choices that could be conceived as core to finding creative solutions to hard problems, but could also be the result of the processing of bad or incomplete data and inadequate algorithms (Farquhar et al. 2024; Ji et al. 2023). Furthermore, most current AI approaches are difficult to scale. AI technology is already hitting the wall of exponential energy use for very incremental gains in function. The costs of AI are already “obscene” as commented by the New Yorker2, elaborated by the MIT Tech Review3, and data centers are taking ever more from the grid (de Vries 2023). It costs about $700,000 per day in energy to run ChatGPT 3.54. This seems to us a questionable return on investment. The AI progress envisioned by the “Agent 4” superintelligence in the popular sci-fi speculation based on AI-2027 (https://ai-2027.com/) is very likely physically impossible with current chip architectures, on energy consumption grounds alone.

Novel mechanisms for research partnerships with industry and/or academia

AI can mimic and eventually improve natural, biological intelligence, the functions and efficiencies of which are far from understood. It therefore stands to reason that we need to understand more about biological intelligence, otherwise AI efforts will be searching for solutions in the dark, forced to implement untested strategies at potentially high financial, energy, and natural resource costs. One problem with the current scientific approach to AI is the limited cross-pollination of ideas between the architects of AI and neuroscientists. On the one hand, most computer scientists and physicists, including those enjoying notoriety in the AI field today, have limited appreciation of neurobiology or cognitive neuroscience. On the other hand, most neurobiologists lack sufficient understanding of computer algorithms, coding and the physics of information processing to contribute to AI technology development. Although there have been some efforts to integrate this knowledge (Hassabis et al. 2017; Botvinick et al. 2020), communication between these research communities remains hamstrung, partially due to siloed training experiences.

Improving and evolving AI will require establishing new research ecosystems where these two approaches to intelligence are encouraged to flourish synergistically, with a new generation of experts who are well-versed in biology, psychology, cognitive science, physics, engineering and computer science. Such integrated cross-disciplinary training should be reflected in new departments at universities, in new funding for multi-disciplinary training at multiple career stages, as well as in initiatives to bridge academic and industry priorities. In contrast to today, 60 years ago, there were no undergraduate neuroscience departments or programs; interested students had to choose between psychology or biology or computer science. It is not unprecedented therefore that the organization of academia should adapt to the new requirements of the society. This process has already started at a few institutions, such as at Rice University5, which now offer AI majors, but these curricula still fall short, particularly in biology. Given the challenges facing the research budget of the US government, the integrated involvement of foundations, non-profits and private industry would allow more efficient use of taxpayer allocated resources by reducing overhead costs, enabling administrative agility and creating a profit-sharing environment that fosters innovation.

Biologically-inspired intelligent devices

In the longer-term, it may be very valuable to consider the contrast between engineered systems and evolution-driven biological systems for generating novel efficient intelligent devices. In general, engineered devices are carefully designed, are usually functionally decomposable for ease of design, maintenance and repair, are manufactured to be identical, and are often controlled using techniques that minimize nonlinearities and maximize ease of predictability. In contrast, biologically systems are subject to evolutionary processes, in which individuals vary, and that variation may be essential for survival in an unpredictable and changing environment. The process of development organizes an exponentially increasing number of elements in polynomial time into organs and a functioning organism. During the lifespan of an organism, local plasticity rules allow for continual adjustments to a changing environment, allowing organisms to find regularities and “common sense” rules about how the world works through experience. All of these processes are constrained by the need to be as energy efficient as possible, and thus all may provide valuable lessons for creating novel energy efficient autonomous intelligent devices in the future.

Efficiency and hardware

An important consideration for incorporating biological features into engineered systems is that energy efficiency is controlled by both software and hardware. Many of the mechanisms described in this work, such as event-based communication and sparse computation, would have a minimal benefit on current AI hardware (GPUs/TPUs) which is not designed with those paradigms in mind. Current AI and ML systems have been optimized for these power-hungry platforms as they are the current leaders in the “hardware lottery” (Hooker 2020). Truly energy-efficient AI may require significant investment into developing hardware which is designed from the beginning to harness these biologically-grounded mechanisms of energy-efficiency and intelligence.

Acknowledgments

The authors thank Dr. M. Anthony Lewis for useful comments on an earlier version of the manuscript.

Author contributions

All authors wrote and reviewed the manuscript.

Funding

The funding was provided by National Science Foundation, Grant Nos.: DBI 2015317, 2342866, DMS 2052109 and National Institutes of Health, Grant No.: RO1 NS118606, NSF Grant IIS 2342866.

Data availability

No datasets were generated or analysed during the current study.

Declarations

Competing interest

The authors declare no competing interests.

Footnotes

This Perspective is adapted from a white paper submitted to the US Federal Register in response to a request for comments related to the US Government’s development of a 2025 National Artificial Intelligence Research and Development Strategic Plan, Docket ID No. NSF-2025-OGC-0001 (https://www.federalregister.gov/d/2025-07332/p-12).

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  1. Afraimovich V, Young T, Muezzinoglu MK, Rabinovich MI (2011) Nonlinear dynamics of emotion-cognition interaction: when emotion does not destroy cognition? B Math Biol 73(2):266–284. 10.1007/s11538-010-9572-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Arbib MA (2025) Artificial intelligence meets brain theory (again). Biol Cybern 119(4–6):16. 10.1007/s00422-025-01013-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Arbib MA, Fellous JM (2004) Emotions: from brain to robot. Trends Cogn Sci 8(12):554–561. 10.1016/j.tics.2004.10.004 [DOI] [PubMed] [Google Scholar]
  4. Arcas BAY, Fairhall AL, Bialek W (2003) Computation in a single neuron: Hodgkin and Huxley revisited. Neural Comput 15(8):1715–1749. 10.1162/08997660360675017 [DOI] [PubMed] [Google Scholar]
  5. Asopa A, Bhalla US (2023) A computational view of short-term plasticity and its implications for E-I balance. Curr Opin Neurobiol 81:102729. 10.1016/j.conb.2023.102729 [DOI] [PubMed] [Google Scholar]
  6. Banaie Boroujeni K, Womelsdorf T (2023) Routing states transition during oscillatory bursts and attentional selection. Neuron 111(18):2929-2944 e2911. 10.1016/j.neuron.2023.06.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Barret DCA, Kaupp UB, Marino J (2022) The structure of cyclic nucleotide-gated channels in rod and cone photoreceptors. Trends Neurosci 45(10):763–776. 10.1016/j.tins.2022.07.001 [DOI] [PubMed] [Google Scholar]
  8. Bear MF, Malenka RC (1994) Synaptic plasticity: LTP and LTD. Curr Opin Neurobiol 4(3):389–399. 10.1016/0959-4388(94)90101-5 [DOI] [PubMed] [Google Scholar]
  9. Berndt N, Holzhutter HG (2013) The high energy demand of neuronal cells caused by passive leak currents is not a waste of energy. Cell Biochem Biophys 67(2):527–535. 10.1007/s12013-013-9538-3 [DOI] [PubMed] [Google Scholar]
  10. Beyeler M, Rounds EL, Carlson KD, Dutt N, Krichmar JL (2019) Neural correlates of sparse coding and dimensionality reduction. PLoS Comput Biol 15(6):e1006908. 10.1371/journal.pcbi.1006908 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bhattacharjee A (2023) The Oxford handbook of neuronal ion channels. Oxford University Press, New York [Google Scholar]
  12. Boahen K (2022) Dendrocentric learning for synthetic intelligence. Nature 612(7938):43–50. 10.1038/s41586-022-05340-6 [DOI] [PubMed] [Google Scholar]
  13. Bommasani R, Hudson DA, Adeli E, Altman R, Arora S, von Arx S, Bernstein MS, Bohg J, Bosselut A, Brunskill E (2021) On the opportunities and risks of foundation models. arXiv preprint arXiv:210807258
  14. Botvinick M, Wang JX, Dabney W, Miller KJ, Kurth-Nelson Z (2020) Deep reinforcement learning and its neuroscientific implications. Neuron 107(4):603–616. 10.1016/j.neuron.2020.06.014 [DOI] [PubMed] [Google Scholar]
  15. Briggs F (2020) Role of feedback connections in central visual processing. Annu Rev Vis Sci 6:313–334. 10.1146/annurev-vision-121219-081716 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Calvetti D, Pragliola M, Somersalo E, Strang A (2020) Sparse reconstructions from few noisy data: analysis of hierarchical Bayesian models with generalized gamma hyperpriors. Inverse Probl. 10.1088/1361-6420/ab4d92 [Google Scholar]
  17. Calvetti D, Somersalo E (2024) Computationally efficient sampling methods for sparsity promoting hierarchical Bayesian models\ast. Siam-Asa J Uncert 12(2):524–548. 10.1137/23m1564043 [Google Scholar]
  18. Calvetti D, Somersalo E, Strang A (2019) Hierachical Bayesian models and sparsity: l(2)-magic. Inverse Probl. 10.1088/1361-6420/aaf5ab [Google Scholar]
  19. Castrillon G, Epp S, Bose A, Fraticelli L, Hechler A, Belenya R, Ranft A, Yakushev I, Utz L, Sundar L, Rauschecker JP, Preibisch C, Kurcyus K, Riedl V (2023) An energy costly architecture of neuromodulators for human brain evolution and cognition. Sci Adv. 10.1126/sciadv.adi7632 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Ceballos EG, Luppi AI, Castrillon G, Saggar M, Misic B, Riedl V (2025) The control costs of human brain dynamics. Netw Neurosci 9(1):77–99. 10.1162/netn_a_00425 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Celeghin A, Borriero A, Orsenigo D, Diano M, Mendez Guerrero CA, Perotti A, Petri G, Tamietto M (2023) Convolutional neural networks for vision neuroscience: significance, developments, and outstanding issues. Front Comput Neurosci 17:1153572. 10.3389/fncom.2023.1153572 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Choi Y (2022) The curious case of commonsense intelligence. Daedalus 151(2):139–155. 10.1162/daed_a_01906 [Google Scholar]
  23. Coggan JS, Keller D, Markram H, Schürmann F, Magistretti PJ (2020) Excitation states of metabolic networks predict dose-response fingerprinting and ligand pulse phase signalling. J Theor Biol. 10.1016/j.jtbi.2019.110123 [DOI] [PubMed] [Google Scholar]
  24. Coggan JS, Keller D, Markram H, Schürmann F, Magistretti PJ (2022) Representing stimulus information in an energy metabolism pathway. J Theor Biol. 10.1016/j.jtbi.2022.111090 [DOI] [PubMed] [Google Scholar]
  25. Connors BW, Gutnick MJ (1990) Intrinsic firing patterns of diverse neocortical neurons. Trends Neurosci 13(3):99–104. 10.1016/0166-2236(90)90185-d [DOI] [PubMed] [Google Scholar]
  26. de Vries A (2023) The growing energy footprint of artificial intelligence. Joule 7(10):2191–2194. 10.1016/j.joule.2023.09.004 [Google Scholar]
  27. Eliav T, Maimon SR, Aljadeff J, Tsodyks M, Ginosar G, Las L, Ulanovsky N (2021) Multiscale representation of very large environments in the hippocampus of flying bats. Science. 10.1126/science.abg4020 [DOI] [PubMed] [Google Scholar]
  28. Farquhar S, Kossen J, Kuhn L, Gal Y (2024) Detecting hallucinations in large language models using semantic entropy. Nature. 10.1038/s41586-024-07421-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Fellous JM, Arbib MA (2005) Who needs emotions?: The brain meets the robot. Affective science. Oxford University Press, New York [Google Scholar]
  30. Fitch WT (2021) Information and the single cell. Curr Opin Neurobiol 71:150–157. 10.1016/j.conb.2021.10.004 [DOI] [PubMed] [Google Scholar]
  31. Fries P (2023) Rhythmic attentional scanning. Neuron 111(7):954–970. 10.1016/j.neuron.2023.02.015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Fulton KA, Zimmerman D, Samuel A, Vogt K, Datta SR (2024) Common principles for odour coding across vertebrates and invertebrates. Nat Rev Neurosci 25(7):453–472. 10.1038/s41583-024-00822-0 [DOI] [PubMed] [Google Scholar]
  33. Gamlin CR, Schneider-Mizell CM, Mallory M, Elabbady L, Gouwens N, Williams G, Mukora A, Dalley R, Bodor AL, Brittain D, Buchanan J, Bumbarger DJ, Joyce E, Kapner D, Kinn S, Mahalingam G, Seshamani S, Takeno M, Torres R, Yin W, Nicovich PR, Bae JA, Castro MA, Dorkenwald S, Halageri A, Jia Z, Jordan C, Kemnitz N, Lee K, Li K, Lu R, Macrina T, Mitchell E, Mondal SS, Mu S, Nehoran B, Popovych S, Silversmith W, Turner NL, Wong W, Wu J, Yu SC, Berg J, Jarsky T, Lee B, Seung HS, Zeng H, Reid RC, Collman F, da Costa NM, Sorensen SA (2025) Connectomics of predicted Sst transcriptomic types in mouse visual cortex. Nature 640(8058):497–505. 10.1038/s41586-025-08805-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Gers FA, Schmidhuber J, Cummins F (2000) Learning to forget: continual prediction with LSTM. Neural Comput 12(10):2451–2471. 10.1162/089976600300015015 [DOI] [PubMed] [Google Scholar]
  35. Ghosh S, Korte A, Serafini G, Yadav V, Rodenfels J (2023) Developmental energetics: energy expenditure, budgets and metabolism during animal embryogenesis. Semin Cell Dev Biol 138:83–93. 10.1016/j.semcdb.2022.03.009 [DOI] [PubMed] [Google Scholar]
  36. Goudar V, Peysakhovich B, Freedman DJ, Buffalo EA, Wang XJ (2023) Schema formation in a neural population subspace underlies learning-to-learn in flexible sensorimotor problem-solving. Nat Neurosci 26(5):879. 10.1038/s41593-023-01293-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Grossberg S (2020) A path toward explainable AI and autonomous adaptive intelligence: deep learning, adaptive resonance, and models of perception, emotion, and action. Front Neurorobot. 10.3389/fnbot.2020.00036 [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Harland B, Contreras M, Souder M, Fellous JM (2021) Dorsal CA1 hippocampal place cells form a multi-scale representation of megaspace. Curr Biol 31(10):2178-2190 e2176. 10.1016/j.cub.2021.03.003 [DOI] [PubMed] [Google Scholar]
  39. Harris JJ, Jolivet R, Attwell D (2012) Synaptic energy use and supply. Neuron 75(5):762–777. 10.1016/j.neuron.2012.08.019 [DOI] [PubMed] [Google Scholar]
  40. Hassabis D, Kumaran D, Summerfield C, Botvinick M (2017) Neuroscience-inspired artificial intelligence. Neuron 95(2):245–258. 10.1016/j.neuron.2017.06.011 [DOI] [PubMed] [Google Scholar]
  41. Hattori R, Hedrick NG, Jain A, Chen SQ, You HJ, Hattori M, Choi JH, Lim BK, Yasuda R, Komiyama T (2023) Meta-reinforcement learning via orbitofrontal cortex. Nat Neurosci. 10.1038/s41593-023-01485-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Haykin SS (1999) Neural networks : a comprehensive foundation, 2nd edn. Prentice Hall, Upper Saddle River [Google Scholar]
  43. Hofkirchner W (2023) Large language models cannot meet artificial general intelligence expectations. Comput Sci Math Forum 8(1):67 [Google Scholar]
  44. Hooker S (2020) The Hardware lottery. arXiv:2009.06489
  45. Horchler AD, Daltorio KA, Chiel HJ, Quinn RD (2015) Designing responsive pattern generators: stable heteroclinic channel cycles for modeling and control. Bioinspir Biomim. 10.1088/1748-3190/10/2/026001 [DOI] [PubMed] [Google Scholar]
  46. Hu T, Genkin A, Chklovskii DB (2012) A network of spiking neurons for computing sparse representations in an energy-efficient way. Neural Comput 24(11):2852–2872. 10.1162/NECO_a_00353 [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Jamadar SD, Behler A, Deery H, Breakspear M (2025) The metabolic costs of cognition. Trends Cogn Sci 29(6):541–555. 10.1016/j.tics.2024.11.010 [DOI] [PubMed] [Google Scholar]
  48. Ji ZW, Lee N, Frieske R, Yu TZ, Su D, Xu Y, Ishii E, Bang YJ, Madotto A, Fung P (2023) Survey of hallucination in natural language generation. ACM Comput Surv. 10.1145/3571730 [Google Scholar]
  49. Jiang J, Foyard E, van Rossum MCW (2024) Reinforcement learning when your life depends on it: a neuro-economic theory of learning. PLoS Comput Biol 20(10):e1012554. 10.1371/journal.pcbi.1012554 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Jin P, Jan LY, Jan YN (2020) Mechanosensitive ion channels: structural features relevant to mechanotransduction mechanisms. Annu Rev Neurosci 43:207–229. 10.1146/annurev-neuro-070918-050509 [DOI] [PubMed] [Google Scholar]
  51. Kagan BJ, Gyngell C, Lysaght T, Cole VM, Sawai T, Savulescu J (2023) The technology, opportunities, and challenges of synthetic biological intelligence. Biotechnol Adv 68:108233. 10.1016/j.biotechadv.2023.108233 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Kagan BJ, Kitchen AC, Tran NT, Habibollahi F, Khajehnejad M, Parker BJ, Bhat A, Rollo B, Razi A, Friston KJ (2022) In vitro neurons learn and exhibit sentience when embodied in a simulated game-world. Neuron 110(23):3952-3969 e3958. 10.1016/j.neuron.2022.09.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Laurent G, Stopfer M, Friedrich RW, Rabinovich MI, Volkovskii A, Abarbanel HD (2001) Odor encoding as an active, dynamical process: experiments, computation, and theory. Annu Rev Neurosci 24:263–297. 10.1146/annurev.neuro.24.1.263 [DOI] [PubMed] [Google Scholar]
  54. Li S, Sheng ZH (2022) Energy matters: presynaptic metabolism and the maintenance of synaptic transmission. Nat Rev Neurosci 23(1):4–22. 10.1038/s41583-021-00535-8 [DOI] [PubMed] [Google Scholar]
  55. Lillicrap TP, Santoro A, Marris L, Akerman CJ, Hinton G (2020) Backpropagation and the brain. Nat Rev Neurosci 21(6):335–346. 10.1038/s41583-020-0277-3 [DOI] [PubMed] [Google Scholar]
  56. Lisman JE (1997) Bursts as a unit of neural information: making unreliable synapses reliable. Trends Neurosci 20(1):38–43. 10.1016/S0166-2236(96)10070-9 [DOI] [PubMed] [Google Scholar]
  57. Lundqvist M, Brincat SL, Rose J, Warden MR, Buschman TJ, Miller EK, Herman P (2023) Working memory control dynamics follow principles of spatial computing. Nat Commun 14(1):1429. 10.1038/s41467-023-36555-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Luo L (2021) Architectures of neuronal circuits. Science 373(6559):eabg7285. 10.1126/science.abg7285 [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Maass W (2000) On the computational power of winner-take-all. Neural Comput 12(11):2519–2535. 10.1162/089976600300014827 [DOI] [PubMed] [Google Scholar]
  60. Madabushi HT, Torgbi M, Bonial C (2025) Neither Stochastic Parroting nor AGI: LLMs Solve Tasks through Context-Directed Extrapolation from Training Data Priors. arXiv:abs/2505.23323
  61. Marcus GF, Davis E (2013) How robust are probabilistic models of higher-level cognition? Psychol Sci 24(12):2351–2360. 10.1177/0956797613495418 [DOI] [PubMed] [Google Scholar]
  62. Marder E (2012) Neuromodulation of neuronal circuits: back to the future. Neuron 76(1):1–11. 10.1016/j.neuron.2012.09.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Maturana HR, Varela FJ (1980) Autopoiesis and cognition : the realization of the living. Boston studies in the philosophy of science, vol 42. D. Reidel Pub. Co., Dordrecht, Holland, Boston
  64. McClelland JL, McNaughton BL, O’Reilly RC (1995) Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory. Psychol Rev 102(3):419–457. 10.1037/0033-295X.102.3.419 [DOI] [PubMed] [Google Scholar]
  65. Mengers N, Rouse N, Daltorio KA (2025) Stable heteroclinic channels for controlling a simulated aquatic serpentine robot in narrow crevices. Front Electron. 10.3389/felec.2025.1507644 [Google Scholar]
  66. Moon KM, Kim J, Seong Y, Suh BC, Kang K, Choe HK, Kim K (2021) Proprioception, the regulator of motor function. BMB Rep 54(8):393–402. 10.5483/BMBRep.2021.54.8.052 [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Moosavi SA, Pastor A, Ornelas AG, Tring E, Ringach DL (2024) Temporal dynamics of energy-efficient coding in mouse primary visual cortex. bioRxiv. 10.1101/2024.12.17.628997 [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Moravec HP (1988) Mind children : the future of robot and human intelligence. Harvard University Press, Cambridge [Google Scholar]
  69. Muller L, Chavane F, Reynolds J, Sejnowski TJ (2018) Cortical travelling waves: mechanisms and computational principles. Nat Rev Neurosci 19(5):255–268. 10.1038/nrn.2018.20 [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Muller L, Churchland PS, Sejnowski TJ (2024) Transformers and cortical waves: encoders for pulling in context across time. Trends Neurosci 47(10):788–802. 10.1016/j.tins.2024.08.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Mumuni AG, Mumuni F (2025) Large language models for artificial general intelligence (AGI): a survey of foundational principles and approaches. arXiv:abs/2501.03151
  72. Murai Y, Goto A (2025) Diverse synaptic mechanisms underlying learning and memory consolidation. Curr Opin Neurobiol 92:102996. 10.1016/j.conb.2025.102996 [DOI] [PubMed] [Google Scholar]
  73. Myers MG Jr., Affinati AH, Richardson N, Schwartz MW (2021) Central nervous system regulation of organismal energy and glucose homeostasis. Nat Metab 3(6):737–750. 10.1038/s42255-021-00408-5 [DOI] [PubMed] [Google Scholar]
  74. Nourse WRP, Jackson C, Szczecinski NS, Quinn RD (2023) SNS-Toolbox: an open source tool for designing synthetic nervous systems and interfacing them with cyber-physical systems. Biomimetics. 10.3390/biomimetics8020247 [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Olshausen BA, Field DJ (1996) Learning efficient linear codes for natural images: the roles of sparseness, overcompleteness, and statistical independence. P Soc Photo-Opt Ins 2657:132–138. 10.1117/12.238708 [Google Scholar]
  76. Payeur A, Guerguiev J, Zenke F, Richards BA, Naud R (2021) Burst-dependent synaptic plasticity can coordinate learning in hierarchical circuits. Nat Neurosci 24(7):1010–1019. 10.1038/s41593-021-00857-x [DOI] [PubMed] [Google Scholar]
  77. Perera-Lago J, Toscano-Duran V, Paluzo-Hidalgo E, Gonzalez-Diaz R, Gutierrez-Naranjo MA, Rucco M (2024) An in-depth analysis of data reduction methods for sustainable deep learning. Open Res Eur 4:101. 10.12688/openreseurope.17554.2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Poirazi P, Mel BW (2001) Impact of active dendrites and structural plasticity on the memory capacity of neural tissue. Neuron 29(3):779–796. 10.1016/s0896-6273(01)00252-5 [DOI] [PubMed] [Google Scholar]
  79. Rabinovich MI, Afraimovich VS, Varona P (2010) Heteroclinic binding. Dynam Syst 25(3):433–442. 10.1080/14689367.2010.515396 [Google Scholar]
  80. Rabinovich MI, Varona P (2018) Discrete sequential information coding: heteroclinic cognitive dynamics. Front Comput Neurosci. 10.3389/fncom.2018.00073 [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Ramaswami M (2014) Network plasticity in adaptive filtering and behavioral habituation. Neuron 82(6):1216–1229. 10.1016/j.neuron.2014.04.035 [DOI] [PubMed] [Google Scholar]
  82. Ramirez JM, Tryba AK, Pena F (2004) Pacemaker neurons and neuronal networks: an integrative view. Curr Opin Neurobiol 14(6):665–674. 10.1016/j.conb.2004.10.011 [DOI] [PubMed] [Google Scholar]
  83. Rieke F, Warland D, de Ruyter van Steveninck R, Bialek W (1997) Spikes: exploring the neural code. A Bradford book. The MIT Press, Cambridge
  84. Roehri N, Brechet L, Seeber M, Pascual-Leone A, Michel CM (2022) Phase-amplitude coupling and phase synchronization between medial temporal, frontal and posterior brain regions support episodic autobiographical memory recall. Brain Topogr 35(2):191–206. 10.1007/s10548-022-00890-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Rouast NM, Schonauer M (2023) Continuously changing memories: a framework for proactive and non-linear consolidation. Trends Neurosci 46(1):8–19. 10.1016/j.tins.2022.10.013 [DOI] [PubMed] [Google Scholar]
  86. Rouse NA, Daltorio KA (2021) Visualization of stable heteroclinic channel-based movement primitives. IEEE Robot Autom Lett 6(2):2343–2348. 10.1109/Lra.2021.3061382 [Google Scholar]
  87. Rumpf S, Sanal N, Marzano M (2023) Energy metabolic pathways in neuronal development and function. Oxf Open Neurosci 2:kvad004. 10.1093/oons/kvad004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  88. Sacramento J, Wichert A, van Rossum MCW (2015) Energy efficient sparse connectivity from imbalanced synaptic plasticity rules. Plos Comput Biol. 10.1371/journal.pcbi.1004265 [DOI] [PMC free article] [PubMed] [Google Scholar]
  89. Sadeh S, Clopath C (2025) The emergence of NeuroAI: bridging neuroscience and artificial intelligence. Nat Rev Neurosci. 10.1038/s41583-025-00954-x [DOI] [PubMed] [Google Scholar]
  90. Shaw KM, Lyttle DN, Gill JP, Cullins MJ, McManus JM, Lu H, Thomas PJ, Chiel HJ (2015) The significance of dynamical architecture for adaptive responses to mechanical loads during rhythmic behavior. J Comput Neurosci 38(1):25–51. 10.1007/s10827-014-0519-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Shoenhard H, Sehgal A (2025) Coordinating the energetic strategy of glia and neurons for memory. Trends Neurosci 48(2):93–95. 10.1016/j.tins.2025.01.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Shumailov I, Shumaylov Z, Zhao YR, Papernot N, Anderson R, Gal Y (2024) AI models collapse when trained on recursively generated data. Nature. 10.1038/s41586-024-07566-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  93. Sillar KT, Combes D, Simmers J (2014) Neuromodulation in developing motor microcircuits. Curr Opin Neurobiol 29:73–81. 10.1016/j.conb.2014.05.009 [DOI] [PubMed] [Google Scholar]
  94. Spector AC, Travers SP (2021) Taste buds and gustatory transduction: a functional perspective. Oxf Univ Press. 10.1093/acrefore/9780190264086.013.100 [Google Scholar]
  95. Stanton PK (1996) LTD, LTP, and the sliding threshold for long-term synaptic plasticity. Hippocampus 6(1):35–42. 10.1002/(SICI)1098-1063(1996)6:1%3c35::AID-HIPO7%3e3.0.CO;2-6 [DOI] [PubMed] [Google Scholar]
  96. Stiefel KM, Coggan JS (2023) The energy challenges of artificial superintelligence. Front Artif Intell. 10.3389/frai.2023.1240653 [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Sun W, Zhao X, Spruston N (2021) Bursting potentiates the neuro-AI connection. Nat Neurosci 24(7):905–906. 10.1038/s41593-021-00844-2 [DOI] [PubMed] [Google Scholar]
  98. Sutton GP, Szczecinski NS, Quinn RD, Chiel HJ (2023) Phase shift between joint rotation and actuation reflects dominant forces and predicts muscle activation patterns. PNAS Nexus 2(10):pgad298. 10.1093/pnasnexus/pgad298 [DOI] [PMC free article] [PubMed] [Google Scholar]
  99. Sutton RS, Barto AG (2018) Reinforcement learning : an introduction. Adaptive computation and machine learning series, 2nd edn, The MIT Press, Cambridge
  100. Szczecinski NS, Hunt AJ, Quinn RD (2017) Design process and tools for dynamic neuromechanical models and robot controllers. Biol Cybern 111(1):105–127. 10.1007/s00422-017-0711-4 [DOI] [PubMed] [Google Scholar]
  101. Szczecinski NS, Hunt AJ, Quinn RD (2017) A functional subnetwork approach to designing synthetic nervous systems that control legged robot locomotion. Front Neurorobot 11:37. 10.3389/fnbot.2017.00037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  102. Szczecinski NS, Quinn RD, Hunt AJ (2020) Extending the functional subnetwork approach to a generalized linear integrate-and-fire neuron model. Front Neurorobot 14:577804. 10.3389/fnbot.2020.577804 [DOI] [PMC free article] [PubMed] [Google Scholar]
  103. Tiesinga P, Fellous JM, Sejnowski TJ (2008) Regulation of spike timing in visual cortical circuits. Nat Rev Neurosci 9(2):97–107. 10.1038/nrn2315 [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Tiesinga PH, Fellous JM, Salinas E, Jose JV, Sejnowski TJ (2004) Inhibitory synchrony as a mechanism for attentional gain modulation. J Physiol Paris 98(4–6):296–314. 10.1016/j.jphysparis.2005.09.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Tiesinga PH, Sejnowski TJ (2010) Mechanisms for phase shifting in cortical networks and their role in communication through coherence. Front Hum Neurosci 4:196. 10.3389/fnhum.2010.00196 [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Wang JX, Kurth-Nelson Z, Kumaran D, Tirumala D, Soyer H, Leibo JZ, Hassabis D, Botvinick M (2018) Prefrontal cortex as a meta-reinforcement learning system. Nat Neurosci 21(6):860. 10.1038/s41593-018-0147-8 [DOI] [PubMed] [Google Scholar]
  107. Wang JX, Kurth-Nelson Z, Tirumala D, Soyer H, Leibo JZ, Munos R, Blundell C, Kumaran D, Botvinick M (2016) Learning to reinforcement learn. arXiv preprint arXiv:161105763
  108. Wang X, Wang S, Ding Y, Li Y, Wu W, Rong Y, Kong W, Huang J, Li S, Yang H (2024) State space model for new-generation network alternative to transformers: A survey. arXiv preprint arXiv:240409516
  109. Wang ZY, Yang EN, Shen L, Huang H (2025) A comprehensive survey of forgetting in deep learning beyond continual learning. IEEE Trans Pattern Anal Mach Intell 47(3):1464–1483. 10.1109/Tpami.2024.3498346 [DOI] [PubMed] [Google Scholar]
  110. Wells L, Bednarz T (2021) Explainable AI and reinforcement learning—a systematic review of current approaches and trends. Front Artif Intell. 10.3389/frai.2021.550030 [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Wen Q, Chklovskii DB (2008) A cost-benefit analysis of neuronal morphology. J Neurophysiol 99(5):2320–2328. 10.1152/jn.00280.2007 [DOI] [PubMed] [Google Scholar]
  112. Yu H, Mineyev I, Varshney LR, Evans JA (2025) Learning from one and only one shot. NPJ Artif Intell 1(1):13. 10.1038/s44387-025-00017-7 [Google Scholar]
  113. Yu Q, Tsodyks M, Sompolinsky H, Schmitz D, Gutig R (2025) Interactions between long- and short-term synaptic plasticity transform temporal neural representations into spatial. Proc Natl Acad Sci U S A 122(47):e2426290122. 10.1073/pnas.2426290122 [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Yu Y, Herman P, Rothman DL, Agarwal D, Hyder F (2018) Evaluating the gray and white matter energy budgets of human brain function. J Cereb Blood Flow Metab 38(8):1339–1353. 10.1177/0271678X17708691 [DOI] [PMC free article] [PubMed] [Google Scholar]
  115. Yu ZJ, Wang YY, Thomas PJ, Chiel HJ (2025) Tradeoffs in the energetic value of neuromodulation in a closed-loop neuromechanical system. J Theor Biol. 10.1016/j.jtbi.2025.112050 [DOI] [PMC free article] [PubMed] [Google Scholar]
  116. Yuste R, Tank DW (1996) Dendritic integration in mammalian neurons, a century after Cajal. Neuron 16(4):701–716. 10.1016/s0896-6273(00)80091-4 [DOI] [PubMed] [Google Scholar]
  117. Zador A, Escola S, Richards B, Olveczky B, Bengio Y, Boahen K, Botvinick M, Chklovskii D, Churchland A, Clopath C, DiCarlo J, Ganguli S, Hawkins J, Kording K, Koulakov A, LeCun Y, Lillicrap T, Marblestone A, Olshausen B, Pouget A, Savin C, Sejnowski T, Simoncelli E, Solla S, Sussillo D, Tolias AS, Tsao D (2023) Catalyzing next-generation Artificial Intelligence through NeuroAI. Nat Commun 14(1):1597. 10.1038/s41467-023-37180-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  118. Zenke F, Gerstner W (2017) Hebbian plasticity requires compensatory processes on multiple timescales. Philos Trans R Soc Lond B Biol Sci. 10.1098/rstb.2016.0259 [DOI] [PMC free article] [PubMed] [Google Scholar]
  119. Zheng W, Holt JR (2021) The mechanosensory transduction machinery in inner ear hair cells. Annu Rev Biophys 50:31–51. 10.1146/annurev-biophys-062420-081842 [DOI] [PMC free article] [PubMed] [Google Scholar]
  120. Zhu L, Kim SJ, Hara M, Aono M (2018) Remarkable problem-solving ability of unicellular amoeboid organism and its mechanism. R Soc Open Sci 5(12):180396. 10.1098/rsos.180396 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

No datasets were generated or analysed during the current study.


Articles from Biological Cybernetics are provided here courtesy of Springer

RESOURCES