Abstract
The article reviews the history development of artificial neural networks (ANNs), then compares the differences between ANNs and brain networks in their constituent unit, network architecture, and dynamic principle. The authors offer five points of suggestion for ANNs development and ten questions to be investigated further for the interdisciplinary field of brain simulation. Even though brain is a super-complex system with 1011 neurons, its intelligence does depend rather on the neuronal type and their energy supply mode than the number of neurons. It might be possible for ANN development to follow a new direction that is a combination of multiple modules with different architecture principle and multiple computation, rather than very large scale of neural networks with much more uniformed units and hidden layers.
Keywords: dual neural node, neuron type, energy source mode, hierarchical architecture, spike-time encoding, emergent computation, executive control to socially cognitive behavior
Introduction
Studies in ANNs have been the focus of contemporary society since the Image Net competition in visual object recognition was won by a deep neural network(Kietzmann et al., 2019; Kriegeskorte and Golan, 2019). Engineers dream of pursuing a class of brain-inspired machines. The situation of robots superseding humans in many functions may soon be a reality. Meanwhile, neurobiologists and psychologists wonder about progress in neural networks due to the great differences between ANNs and biological brains(Crick, 1989; Carlson et al., 2018).
This article reviews the contemporary theories and technical advances in three related fields: ANNs, neuroscience and psychology. ANNs were born mainly from more than two thousand years of mathematical theories and algorithms; over the past two hundred years, neuroscience has revealed more truths about the mysteries of the brain; and psychology has just passed its 143th anniversary of accumulating conceptions about conscious and unconscious cognitive processes. Cognitive sciences, encapsulating the disciplines of both natural and artificial intelligence, were founded in the mid-1970s to inquire into the mystery of intelligent substitutes, including humans, animals and machines. Although progress in these related fields is promising, a completed brain-inspired intelligent machine is far from the human brain. How can we reduce the distance between engineers’ dreams and reality? Some suggestions for brain simulation will be offered in the paper.
Three generations of ANNs
Artificial neural networks simulate brain intelligence by mathematical equations, software or electronic circuits. As the first constituent sub-discipline of cognitive sciences, ANNs already have a 79-year history since the conception of the “neural unit” and “Hebbian synapse” as well as the first model of the perceptron neural network in the 1940s–1950s (Mcculloch and Pitts, 1943; Hebb, 1949; Rosenblatt, 1958). We review the developmental course of the discipline and divide its history into three stages that represent the view of the human brain and intelligence from the perspective of mathematics. The first generation of ANNs is linear logic networks; the second generation is connectionist networks, including parallel distributed processing (PDP) and deep neural networks (DNNs); and the third generation is spike neural networks (SNNs).
The first generation of ANNs: Linear logic networks
The first sentence in the first ANN paper was “Because of the ‘all-or-none’ character nervous activity, neural events and the relations among them can be treated by means of propositional logic.”(Mcculloch and Pitts, 1943). Then, the authors described ten theorems and emphasized the calculus principle and academic significance in the following sentence in the last part of the paper: “Specification for any one time of afferent stimulation and of the activity of all constituent neurons, each an ‘all-or-none’ affair, determines the state.” It is worth paying attention to the keywords in the cited sentence, all-or-none and determine, which mean that perception is associated with the statistic separation of two-state affairs as a deterministic system. The second classical ANN paper stated, “When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficiency, as one of the cells firing B, is increased” (Hebb, 1949).
The first ANN perceptron model was a linear network model that comprised A units and R units. Eleven equations were used to analyze the model’s predominant phase through six parameters clearly defining physical variables that were independent of the behavioral and perceptual phenomena. As a result, the model was able to predict learning curves from neurological variables and likewise to predict neurological variables from learning curves. The author assertively wrote, “By the study of systems such as the perceptron, it is hoped that those fundamental laws of organization which are common to all information handling systems, machines and men included, may eventually be understood” (Rosenblatt, 1958). It is obvious for the classical writer to have too optimistic foresee for his linear network model. In fact, the linear model could not solve mathematic puzzles such as XOR partition. Therefore, a monograph (Minsky and Papert, 1969) claimed assertively that the investigation of linear networks must be only a game on paper. As a result, much foundation grant money was withdrawn, no longer supporting the projects.
The second generation of ANNs: Connectionist networks
2-G ANNs began to be developed in the mid-1980s and can be divided into two periods: parallel distributed processing (PDP) and deep learning. In addition to the demand for ANN development, the demand for debugging in artificial intelligence (AI) also promoted the renaissance of ANNs in the 1980s.
A few dedicated scientists persisted in their neural network programs after 1969, their achievements provided a prelude for an ANN renaissance. For example, two neural network models, the discrete two-state model(Hopfield, 1982) and the continuous-response model (Hopfield, 1984), were published. A two-volume book (Rumelhart and McClelland, 1986) about PDP written by 16 experts as coauthors was spread quickly over the world. Many neural models and learning algorithms, such as the massively parallel network with hidden unit-Boltzmann machine (Ackley et al., 1985), error backpropagation (Rumelhart et al., 1986), competitive learning, neural Darwinism, adaptive systems, and self-organized learning, were created quickly during the 1980s–1990s. Meanwhile, connectionist modern supercomputers were developed, and very large-scale analog integrated circuits, analog neural computers, and applied electronic chips, such as electronic retina chips and see-hear chips, emerged.
It is worth discussing backpropagation algorithm (BP) due to its broad applications in the field of ANNs. Rumelhart et al. (1986) described the learning procedure and abstracted a universal δ-learning rule. The results in implementation of the learning algorithm demonstrate that there is error gradient descent in weight-space, so that network’s output reaches at its targeted value in a little by a little, meanwhile the error derivatives (δ) between the output and the desired result propagates back along the gradient descent. In fact, the efficiency of learning procedure usually is very low. For example, it is necessary more than several hundred even thousand turns of training for a very simple network with a hidden unit to solve a task of XOR function partition. BP is not a nonlinear model, error back-propagation is just an implicit nonlinear mapping without a really feedback loop. The number of hidden nodes has to be make sure by experience or testing without theoretical guideline. The authors wrote in the end of their paper that “the learning procedure, in its current form, is not a plausible model of learning in brains. However, …it is worth looking for more biologically plausible ways of doing gradient descent in neural networks.”
The studies in DNNs were initiated by the heuristic ideas: parallel distributed processing (PDP), and stochastic computing, for example, Boltzmann machine. Although its initial step began early in 1980s, it did not shoot a flash in scientific society until Alex-Net won image competition in 2012. Alex-Net was composed of 650,000 artificial neural units, consisted five convolutional layers, some of which were followed by max-pooling layers and three fully connected layers with a final 1,000 ways to output the results. It successfully recognized and classified 1.2 million high-resolution images by 60 million parameters in the ILSVRC-2012 competition (Krizhevsky et al., 2012). Recently, a deep CNN based Kaggle network won the contest to recognize real-time facial expression in the 5th ICEEICT (Sultana et al., 2021).
Beside the multiple hidden layers of network, a number of new algorithm has given DNNs a strong support, for example, convolutional algorithm, wavelet, and principal component analysis (PCA), increased high accuracy of feature extracting and pattern classification. The roles of traditional shallow models in pattern recognition had been changed in recent years, because of deep CNNs with strong learning ability—the deep learning-based object detection. The situation is based on some premises, such as the database of large scale networks had been accumulated during Image Net twice competitions in LSVRC-2010 and LSVRC-2012 (Deng et al., 2009); CNNs for speech recognition had been reported (Hinton et al., 2012); a design of regions with CNN features was successively applied (Girshick et al., 2014). In short, CNNs can learn object representations without the need to design features manually and can straightforwardly process visual or auditory complex objects without object dividing or component extracting. The deep learning-based object detection models and algorithms, covering different application domains, had been reviewed in detail (LeCun et al., 2015; Zhao et al., 2019; Naveen and Sparvathi, 2022). It is possible for the progress in studies of machine deep learning to promote ANNs serving social life and economic development. But there are many mysteries opening with regard to how does the CNNs learn the ability. The answer is that BP learning procedure trains CNNs approximating the desired output by the mechanism of error-backpropagation (Lillicrap et al., 2020). But the article did not make sure that the same mechanism exists also in the biological brain. They cited a number of references which belong to the studies in either neuroanatomy or neurophysiology, missing the dada in studies of neurobiology or oscillatory encoding. For example, the theory of oscillatory encoding claims that information communication between brain areas might be implemented by the dynamic synchronization of different rhythmic activities among neural networks. GABA inter-neurons by their inhibition lock up pyramidal neurons coupling with a network oscillation, then Neurogliaform cells (NGFCs) dynamically decouple neuronal synchrony between brain areas (Sakalar et al., 2022). It is interesting that action potentials of NGFCs decoupled pyramidal cell activity from cortical gamma oscillations but did not reduce their firing nor affect local oscillations. Thus, NGFCs regulate information transfer by temporarily disengaging the synchrony without decreasing the activity of communicating networks. Such attribute of NGFCs in biological brain seems to be alike of backpropagation in ANNs or CNNs, it regulates error gradient in weight-space by implicit feedback fashion, but does not disturb feedforward information transfer of ANNs. The result in comparison of the functional fashion between cortical NGFCs and backpropagation in ANNs supports the idea that mechanism of implicit backpropagation exists in both biological brain and ANNs.
Despite the great progress in ANNs, DNNs, and CNNs, as well as their rapid spread throughout the world, their weaknesses sometimes spoiled their reputation by spontaneously generating adverse effects, leading to some strange results. Fortunately, this phenomenon has already been solved. In addition, the many neural units and large parameters required by a deep neural network created a somewhat complicated problem because they not only wasted energy but also prevented the application of DNNs to pragmatic problems. Spiking neural networks have an advantage in decreasing both the unit numbers and energy needed (Figure 1).
The third generation of ANNs: Spiking neural networks
1,2-G networks are composed of a deterministic system in which information transmission is determined by presynaptic and postsynaptic factors. Neurobiological achievements in three research fields, studies on the ion channel conductance of the neuronal membrane (Williams, 2004), studies on the probability of transmitter release from the presynaptic membrane (Tsodyks and Markram, 1997; Branco and Staras, 2009; Borst, 2010), and studies on spike timing-dependent plasticity (STDP; Song et al., 2000; Nimchinsky et al., 2002; Zucker and Regehr, 2002; Caporale and Dan, 2008; Losonczy et al., 2008; Spruston, 2008; Jia et al., 2010; Mark et al., 2017), produced a new conception of synaptic transmission and have attracted many experts to explore temporal coincidence (Montague and Sejnowski, 1994; Sejnowski, 1995; Markram et al., 1997; Roelfsema et al., 1997), noise sources (Maass, 1997; Roberts, 1999; Hopfield and Brody, 2004; Buesing and Maass, 2010; McDonnell and Ward, 2011), spiking neural networks (SNNs) and stochastic computing (Silver, 2010; Hamilton et al., 2014; Maass, 2014; Maass, 2015; Tavanaei and Maida, 2017) since the 1990s.
Such as showing in Figure 2, there are three categories of neuronal encoding: rate encoding, paired pulses ratio (PPR) encoding and spike-time encoding. Rate encoding has been adopted to present neuronal exciting level since 1930s, PPR encoding and spike-time encoding are discussed in the field of neurophysiology during 1990s. PPR encoding has been usually used to classify the neurons, Spike-time encoding employs the lengths of inter-spike intervals (Δt) to encode and transmit information (Sejnowski, 1995; Song et al., 2000; Nimchinsky et al., 2002; Zucker and Regehr, 2002; Caporale and Dan, 2008; Losonczy et al., 2008; Spruston, 2008; Jia et al., 2010; Mark et al., 2017). SNNs comprise neuromorphic devices in which information transmitted by more than three factors is constrained. Dendrites, as the third factor, are added to pre- and post-synaptic components, and any synaptic state is constrained by the locally surrounding patch of postsynaptic membrane, which contains hundreds of synapses. The pulses with the shortest Δt are identified as getting qualify to play a role in information transmission.
Spike neural networks are closer to the biological brain than 1,2-G networks are. SNNs do not need many units in the network architecture and save energy in contrast to DNNs. However, training SNNs remains a challenge because spike training requires discrete signals without differentiability, and the backpropagation algorithm cannot be directly used, as in DNNs. Recently, deep learning in SNNs has been reviewed in detail (Tavanaei and Maida, 2017), and a hybrid platform has been implemented by integrating DNNs with SNNs (Pei et al., 2019).
Theoretical weakness: From the energy function to the cost function
For brain-inspired ANNs, what is the dynamic source that drives state changes? Hopfield already answered this question with the energy function of the spin glass model, a solid-state physics model (Hopfield, 1982; Hopfield, 1984). Every unit of the system is a uniform attractor, and each unit’s orientation and position are statistically determined by the attractor’s surface energy and the interaction competition among attractors in the system state space. The energy is the dynamic source of the system and will trend toward global or local energy minimization. The system will reach the stationary state or convergent point when the energy function reaches its minimum. The spin-glass model is too far from the energy metabolism of the biological brain. Brain neurons are polymorphic, and their morphologic appearance and location in the brain are determined by phylogenetic evolution and ontogenetic processes rather than being uniform and stochastically allocated. Brain energy metabolism usually occurs at the basic metabolic level during an organism’s calm state, but it changes to global imbalance and the highest local level with cognitive activity. Therefore, the theoretical attractiveness of the energy function has decayed since the 1990s.
The concept of the cost function or loss function has recently been regarded as the dynamic basis of ANNs (Marblestone et al., 2016; Scholte et al., 2018; Richards and Lillicrap, 2019), but the definition differs among different references, for example, “The idea of cost functions means that neurons in a brain area can somehow change their properties, e.g., the properties of their synapses, so that they get better at doing whatever the cost function defines as their role”; “A cost function maps a set of observable variables to a real value representing the ‘loss’ of the system”; “A cost function is defined as the composition of mathematical operations, e.g., mean squared error”; “A loss function provides a metric for the performance of an agent on some learning task. In a neural circuit, loss functions are functions of synaptic strength”; and “Cost is the partial derivative of the error with respect to the weight.”
All of these definitions and terms are from mathematics or ANN, with very few neuroscientific or psychological implications. How can the cost function be understood by neuroscientists and psychologists? The brain-inspired ANN is an interdisciplinary field, and the main theoretical conceptions should include neuroscientific and psychological implications, at least in terms of brain physiology. Cost, error, weight, partial derivative, mathematical operations, the credit assignment problem, etc.; how many of these come from system neuroscience, integrative neurophysiology or cognitive psychology? As a primitive and basic conception of ANNs is far from the implications of the biological or human brain, it is difficult to establish an interdisciplinary field. Let us review the corresponding conceptions in neuroscience and psychology.
There are different models of energy supply for different brain structure. Large amount of parallel neural fiber in cerebellar did not put on its myelin coat, its energy supply comes only from its cell body; but the neural fiber from cerebrum can get additional energy supply from myelin (neuroglia cell). BOLD signals represent hemoglobin-responsive changes in local brain blood microcircuits or neuroglia assemblies, because hemoglobin cannot directly reach any neuron. Each neuron in the neocortex has 3.72 glial cells, and each Purkinje cell (PC) in the cerebellum has only 0.23 glia cells (Azevedo et al., 2009). In fact, energy supply in brain is, respectively, implemented by area, lamina, column. We suggest that the energy supply in ANNs needs to be improved, so that the energy consumption will be saved.
Information processing in the brain
Although the conception of information processing did not appear until the 1940s, brain anatomy and neurophysiology had already become important by the end of the 19th century. Through study of the anatomy of the nervous system, scientists had mastered the knowledge of sensory-motor pathways and the visceral autonomic nervous system as well as brain functional localization. The classical neurophysiological theory, in which the brain was regarded as a reflex organ, was founded at the beginning of the 20th century. By the mid-20th century, the brain was regarded as an information-processing organ since electrophysiological techniques provided scientific evidence of neuron firing and postsynaptic potentials (PSPs) in the 1930s–1960s. The brain is now considered an organ in which both neural and genetic information are processed; that is, the brain’s long-term memory is the result of a dialog between synapse and gene (Kandel, 2001). In recent years, transcriptomic expression has been used to classify the cell types of neuron distribution in the neocortex (Kim et al., 2017; Boldog et al., 2018). The concept of neural information processing is understood from four perspectives: the principle of the simultaneous existence of digital encoding and analog encoding, the principle of the simultaneous existence of multiple processing processes and multiple information streams, the principle of circular permutation and coupling between electrical transmission and biochemical transmission in the processing of neural information, and the principle of relevance between neural and genetic information (Shen, 2017). Therefore, neural information processing is more complicated than any communication information or industry information. Of course, the brain-simulated parameters must be simplified to build an ANN model, but the primitive unit, the network architecture, and the operational dynamics should approximate the biological brain.
Two types of primitive unit vs. an uniform excitatory unit
A series of discoveries in studies of the spinal reflex at the beginning of the 20th century, such as spatial and temporal summation, convergence, divergence, fraction, synaptic retardation, final common path, and reciprocal innervation, accounted for the interaction between excitatory and inhibitory processes (Sherrington, 1906). In particular, the conditioned reflex theory emphasized internal inhibition as a function of the cerebral hemispheres (Pavlov, 1927). Inhibitory neurons were not found until electrophysiological techniques and electro-microscopy were used. A model of the soma and large dendritic fields of a cat’s spinal motor neurons was published, showing that any nerve cell is encrusted with hundreds of thousands of synapses, with a mean diameter of approximately 1.0 μm (Haggar and Barr, 1950). Even when six different synaptic appearances were demonstrated, synaptic features could not credibly be used to discriminate whether a synapse is excitatory or inhibitory (Whittaker and Gray, 1962). The features of inhibitory synapses were judged by inhibitory postsynaptic potentials (IPSPs), and many more feedforward inhibitions, such as presynaptic inhibition and collateral inhibition, were found in brain networks, including the cerebral cortex, cerebellar cortex, and hippocampus (Eccles et al., 1954; Eccles, 1964).
Studies in brain chemical pathways in the 1960s–1970s found several categories of inhibitory neurons on the basis of their released transmitter molecules, such as gamma aminobutyric acid (GABA), glycine, and serotonin, and found that the inhibitory neurons in the neocortex are composed mainly of GABAergic neurons. The neurons in the neocortex can be divided into two categories: interneurons, which make local connections, and projection neurons, which extend axons to distant intra-cortical, sub-cortical and sub-cerebral targets. Projection neurons are excitatory, synthesizing glutamatergic transmitters with a typical pyramidal morphology that transmit information between different regions of the neocortex and to other regions of the brain (Paredes et al., 2016; Cadwell et al., 2019). Molecular neurobiology using transcriptomic techniques investigates the stereotyped distributions of mainly inhibitory cells in the brain and classifies three classes of inhibitory neurons: parvalbumin-positive interneurons (PV+ neurons), somatostatin-positive interneurons (SST+ neurons) and vasoactive intestinal polypeptide-positive interneurons (VIP+ neurons; Kim et al., 2017). Recently, the biological marker for GABAergic neurons in immunocytochemistry and molecular neurobiology has enabled direct classification into three categories: excitatory, inhibitory and non-neuronal cells, such as glial cells. By this method, a group of human interneurons with anatomical features in neocortical layer 1, with large ‘rosehip’-like axonal button and compact arborization (Boldog et al., 2018), was discriminated that had never been described in rodents.
Studies in GABAergic neurons found that the proportion of GABAergic neurons is approximately 15% of the population of all cortical areas in rats, whereas in the primates, the proportion reaches 20% in the visual cortex and 25% in the other cortex. The numbers of inhibitory interneurons have increased during phylogenetic evolution along with the appearance of unique interneuron subtypes in humans (Wonders and Anderson, 2006; Petanjek et al., 2008). In contrast to the prenatal development of excitatory neurons in the human cortex, interneuron production, at least in the frontal lobe, extends through 7 months after birth. GABA concentrations in the occipital cortex of adult subjects, as measured by magnetic resonance spectroscopy, are relatively stable over periods as long as 7 months(Near et al., 2014). In evolutionary history, the cortical projection neurons derive from the dorsal (pallial) telencephalon and migrate radially into the cortical mantle zone; the interneurons generated in the sub-pallium migrate tangentially across areal boundaries of the developing cerebral cortex into the overlying cortex (Mrzljak et al., 1992; Petanjek et al., 2009). The inhibition shapes cortical activity, and inhibitory rather than excitatory connectivity maintains brain functions(Isaacson and Scanziani, 2011; Mongillo et al., 2018; Fossati et al., 2019). There are large numbers of interneuron types with different morphological, electrophysiological, and transcriptomic properties in the human neocortex (Wonders and Anderson, 2006; Petanjek et al., 2008; Kim et al., 2017; Boldog et al., 2018).
Artificial neural networks have oversimplified the inhibitory unit as a supplementary variant of the excitatory process or a dependent variant of the activation function. ANN experts claim that “inhibition need not be relayed through a separate set of inhibitory units” (Kietzmann et al., 2019). Actually, “in the absence of inhibition, any external input, weak or strong, would generate more or less the same one-way pattern, an avalanche of excitation involving the whole population” (Buzsaki, 2006).
ANNs might be comprised of different categories of units with different connecting strength with their target units by different weight shift range 0–1 or 0–(−1). As a result, there are two types of primitive unit: excitatory and inhibitory unit. Each categories of units will be further divided into different types by their threshold activated in different encoding, for example, sensitivity to spike timing: the lower sensitivity(Δt < 10 ms) or the higher sensitivity(Δt < 6 ms). Also the unit types can be classified by pulse frequency to be evoked or by PPR standard.
Hierarchical small-world architecture with a circular vs. a uniform feedforward network
The structural basis of the brain and spinal reflex is the reflex arc that comprises afferent, efferent and neural centers. The dual neural processes—the excitatory and inhibitory processes—run in two directions in the reflex arc: along a centripetal pathway from the sensory organs, reaching the corresponding brain sensory center, and along a centrifugal pathway returning from the brain to the peripheral effectors. The small-world network comprises a few components, and each component contains a few constituent parts. The biological reflex arc comprises three components, and each component contains 3–5 neurons, for example, a monkey’s food-fetch reflex pathway (Thorpe and Fabre-Thorpe, 2001). The centripetal pathway is an input or afferent component that comprises three neurons from the retina, lateral geniculate body to primary visual cortex (V1), producing visual sense about the stimuli; the bottom-up stream is a perceptual mechanism from V1 to the anterior inferior temporal gyrus, producing visual perception by the ventral visual pathway; the prefrontal cortex to the premotor area of the cortex formats the decision-making mechanism; and the centrifugal pathway is the efferent or output component of the motor cortex (MC) to the spinal cord and finger muscles. In short word, the food-fetch reflex is not only implemented by 4 hierarchical networks with small-world architecture, but also there are many inhibitory modulating mechanisms: a self-feedback loop, the extrapyramidal system, the descending nonspecific reticular system and the cerebellum. The classical specific brain pathways are small-world architectures that have both the attributes of random rewiring and regular local organization; their synaptic path lengths are both the shortest for random networks and the longest for local networks (Watts and Strogatz, 1998; Buzsaki, 2006). Therefore, the efficiency of brain networks is excellent, with both high-speed information processing and saving of network sources.
On the contrary, DNNs are one-way feedforward networks and usually comprise an input layer, hidden layers, and an output layer in which the highest layer does not to be discriminated, and all units are uniform at the initial state or before the training period. The information stream moves in a feedforward or downward manner from the input layer to the output layer. Sometimes, the convolutional algorithm is implemented in all input sets, the hidden layer is comprised more than tens of layers, and whole connections among all units are implemented in layers.
Therefore, we suggest that the network architecture in ANNs need to be improved. As a commenter on an early version of this manuscript noted: “Current ANNs largely ignore anatomical organizations: topographic mapping, precise wiring between brain areas, layers, and cell types. This specific wiring bestows huge computational power with minimal request of energy consumption. For example, the wiring of the olfactory system is very different from that of the visual system. Similarly, the wiring of the sensory system differs drastically from that of the motor system.”
Biofeedback everywhere vs. only error backpropagation
The nervous system is a typical servo system with many biofeedback mechanisms that exist everywhere in brain networks but are not like error backpropagation in ANNs. Biofeedback can be found in reflex arcs, synapses, dendrites, axons, presynaptic membranes, biochemical transmission, etc. In the centrifugal pathway, recurrent inhibition in spinal networks is usually mediated by another inhibitory neuron, for example, Renshaw cells or their gamma loop between alpha and gamma motor neurons in the spinal reflex arc (Eccles, 1973, 1989). In the sensory-perceptual system, neural information is transformed along the centripetal pathway to the primary area of the sensory cortex, and the bottom-up information stream reaches the highest area of the perceptual cortex. In addition, there are many feedbacks or top-down streams from the highest perceptual or memory areas, and there are concurrent thalamus-cortical connections among the thalamus-cortical layer IV-cortical layer VI-thalamus (Kamigaki, 2019; Egger et al., 2020). This phenomenon is not consistent with the cognitive model rule that a unit can be connected to a unit in a higher layer or to a unit in the same layer but not to a unit in a lower layer (Feldman and Ballard, 1982; Cummins and Cummins, 2000). Recently, some new facts have been reported, such as dual whole-cell recordings from pyramidal cells in layer V of the rat somatosensory cortex, revealing an important mechanism for controlling cortical inhibition and mediating slow recurrent inhibition by SST+ neurons (Deng et al., 2020). In sensory experience and perceptual learning, input from the higher-order thalamus increases the activity of VP+ neurons and VIP+ neurons and decreases the activity of SST+ neurons, resulting in the disinhibition of pyramidal cells in the sensory cortex and the LTP effect. Contextual feedback from the higher-order thalamus is helpful in processing first-order sensory information (Williams and Holtmaat, 2019).
In the mechanism of biochemical transmission of neural information, the re-take-up and auto-receptor protein in the presynaptic membrane receives or reabsorbs its released transmitter. Reverse messengers such as NO and CO molecules released by postsynaptic neurons can quickly suppress the transmitter release of presynaptic neurons.
Sum up in a word, we hope that the feedback or recurrent inhibition should be set in everywhere of the whole networks, instead of only error backpropagation.
The multiple learning mechanisms vs. a uniform weight switch
The fixed reflex arc in the biological brain was established by phylogenetic evolution and provides a basis of unconditioned reflex; temporary connections between neural centers are the basis of the conditioned reflex. How can the temporary connection be established by training or experience? The unconditioned stimulus (US), for example, food, water or a sexual partner, induces a stronger excitatory process in the corresponding brain center; the conditioning stimulus (CS) at first is an unrelated neutral stimulus that usually induces a weaker excitatory process in the corresponding brain center. Temporary connection is the result of the attraction of a stronger center to a weaker center (Pavlov, 1927). Therefore, the CS must be presented first, and the US must follow in an interval of less than a couple of seconds. The conditioned reflex has been called classical conditioning as the physiological basis of unsupervised learning, in which increased connectionist strength takes place mainly among the neurons in the cerebral cortex. For supervised learning, a goal or a standard as the supervisor must be set in advance. The response should be aimed at an exact goal with a quick reaction time or with the refined skill that demands the inhibited actions. The neural connection between the cerebrum and cerebellum is the neural basis of supervised learning. The short−/long-term plasticity of the synapse in the sensory-perceptual neocortex is the basis of perceptual learning.
The causal role of learning is the coincidence in the tempo-spatial space between stimuli, inducing changes in synaptic plasticity. The temporal-difference learning algorithm has been tested, and an essential process in learning and memory is the transformation between spatiotemporal information in the brain (Fitzsimonds and Poo, 1998; Bil and Poo, 2001).
Two kinds of synapses were identified as the cellular mechanism of animal learning in the 1980s, while the biochemical mechanism of learning has been recognized as the molecular configuration of protein: NMDA-receptor protein and adenylate cyclase (Abrams and Kandel, 1988). Two kinds of synapses infer an activity-dependent synapse and a reinforcement synapse between pre- and post-synaptic components. The activity-dependent synapse takes place between two presynaptic components, and the facilitated component then acts on the third postsynaptic neuron, providing the conditioned response (CR). The reinforcement/rewarding synapse brings about the excitatory process in postsynaptic neuron, while the CS and US signals are transferred on the postsynaptic neuron that is facilitated by the two presynaptic components.
Artificial neural networks does not change their methodological or simulating strategy for machine learning, although a part of researchers in ANNs field has accepted the hypothesis on three types of learning (Doya, 2000). There is difficult for them to processing the differences among different types of learning. We suggest to regulate, respectively, the mode of energy supply for different learning network. The key neurons of supervised learning are Purkinje cells in cerebellar, each cell with a rich dendritic tree in a plane, and receive information from a lot of parallel nerve fibers uncovered myelin coat. The nerve fibers do not require additional energy supply from neuroglia cell (myelin coat). The electric source should be a stable lower power for supervised leaning circuit; on the contrary to the ANN circuit for the unsupervised learning. The electric source of the circuit should be a little higher power and can be quickly changed in a range, because pyramidal neuron in neocortex with a long axon and axonal tree in 3-diminssion space. Each pyramidal neuron has about 4 glial cells who work as both a myelin and additional energy supplier during the process of transmitting neural pulses.
Uniquely executive control mechanism vs. programing action command
So fa, there are 3–4 theories on the executive control mechanism in psychology: the unity and diversity model (Miyake et al., 2000), functional network of prefrontal cortex (Miller and Cohen, 2001) and the distributed executive control (van den Heuvell and Sporns, 2013). The unity and diversity model is also called “three factor model of update, inhibition and shifting or cognitive flexibility,” because the executive control is defined as the cognitive behavior containing updating of working memory, inhibition of unnecessary actions and frequently changing situating demand.
The executive control behavior is required to have mid-grade of inter-correlation among three experimental indices, so that it presents both the independent and common property of the behavior (Miyake et al., 2000). According to updating of working memory, subject is demanded to say the picture name which is demonstrated in the last one of a sequence of 60 pictures, but the sequence will be stopped randomly from time to time. So, the subject has to update his or her working memory without stopping during the process of the sequence running. Inhibition infers to eliminate the interfere for the performance running or inhibits undesired responses. Stroop color words interfere paradigm is usually adopted. A “blue” word written by red color and a “red” word written by blue color are inconsistent interfere items, the subjects are demanded to press a button according semantic blue, inhibiting the response to blue color. Shifting or cognitive flexibility confers that the experimental task will be randomly changed. For example, 2 digits are successively displayed on a screen, then the computational demand is shown either plus or minus.
The experimental models may not completely present human social intelligence. For example, Chinese proverb “show resourcefulness in an emergency” or “turned on by danger” means that working memory can reproduce a creative idea, during dangerous situation.
The small-world architecture may not be available for the social intelligence of the prefrontal cortex, even though it is important for brain network of basically cognitive process. A single cell axon tree of pyramidal neuron has been shown with thick interlaces and even with 60 thousand branches (Buzsaki, 2006; Boldog et al., 2018). The prefrontal cortex is the core structure to collect different information from other brain areas and efficiently processing the complicate information as well as controlling the related behavior. A lot of evidence from monkey and patients with brain injury demonstrates that prefrontal cortex contains different switches to control proactive, retroactive behavior (Hikosaka and Isoda, 2010), many local microcircuits to play a role in cognitive flexibility or shifting, and salience network to regulate interchanging between default mode network and central executive network (Tang et al., 2012; Goulden et al., 2014; Satpute and Lindquist, 2019). The network hubs in the human brain has been reported by the distributed executive control theory(van den Heuvell and Sporns, 2013), the network hubs are different from the other brain structures, there are the compact white matter in macroscopy and finely structured nerve fiber with high density spinous process in microscopy. It is the structure with the qualify to implement integrating information and controlling behavior. To date, the axon trees and their microcircuits as well as long-range output connectionist pathways in the prefrontal cortex has been believed to be the prerequisite of executive control ability. Especially, medial prefrontal cortex (mPFC) contains much more GABA-inhibitory neuron that collect a lot of information from the other neurons, such as Ach neuron in anterior brain, 5-serotonin neuron in thalamus, sensory cortex, and limbic-hippocampus system as well as memory system (Jayachandran et al., 2019; Nakajima et al., 2019; Ren et al., 2019; Sun et al., 2019; Weitz et al., 2019; Onorato et al., 2020). There are uniquely two kinds of neuron in PFC and mPFC: GABA-ergic chandelier cells (ChaCs) and von Economo neurons(VENs). ChCs are the only interneuron subtype that selectively innervates the axon initial segment of pyramidal neurons in frontal cortex (Tai et al., 2019); there are the higher density of VENs in the elderly above 80 year old with higher intelligence (DeFilepe et al., 2002; Evrard, 2018; Gefen et al., 2018).
The programing behavior sequences written by program editor cooperate robots, even though a guiding robot can performances the act of etiquette and answers question from customer. There are not creative idea or contingency approaches. That an auto drive bicycle runs on the street and avoids obstacles presents only human’s sensory-perception-motor ability but not general intelligence.
Intelligence is a neuro-computational system
Intelligence was regarded as the ability to compute discrete physical symbols when artificial intelligence was founded in 1956; PDP was regarded as neural computation when the second generation of ANNs was propagated in 1986. Here, intelligence might be a compound of multiple computing processes rather than any single computing process. Intelligence is a neuro-computational system, containing the following computations: deterministic computation, stochastic computation, chaotic computation, and the other computation can be implemented by the human brain.
Deterministic computation
Neuronal excitation is represented by the rate encoding of its discharge, while the postsynaptic potentials implement analog computation by amplitude modulation. The computation involved in combining digital and analog computing composes a deterministic system. The general intelligence is implemented by the classical specific nervous system, mainly through deterministic computation.
Population sparse encoding
The human brain comprises 1011 cells (approximately 160 billion), and the cerebral cortex comprises 82% of the total brain mass; however, only 19% of all neurons are located in the cerebral cortex, which is mainly responsible for intelligence (Azevedo et al., 2009). Each brain function is implemented only by a part of the neurons; therefore, the human brain implements most intelligent activities by population sparse encoding.
Oscillatory encoding
Spontaneous oscillatory encoding is a prerequisite of normal brain function because it provides optimal arousal or vigilance under the mechanism of the nonspecific reticular system.
Infra-slow oscillation
The spontaneous fluctuation in brain BOLD signals with infra-slow oscillation <0.1 Hz concerns not only the default mode networks(DMN) but also the cerebellum and salience network (Raichle et al., 2001; Hikosaka and Isoda, 2010; Raichle, 2010, 2015), which remain to be investigated further. ISA is a distinct neurophysiological process that reflects BOLD signals in fMRI, and its spatiotemporal trajectories are state dependent (wake versus anesthesia) and distinct from trajectories in delta (1–4 Hz) activity with the electrical property (Tang et al., 2012; Goulden et al., 2014; Satpute and Lindquist, 2019).
Stochastic computation
Spike time encoding
Even though the principal neurons in the classical specific nervous system have many dendritic branches and an obvious laminar distribution, their axon has only a few buttons, usually fewer than 15–20 (Branco and Staras, 2009; Borst, 2010; Silver, 2010). The neurons seem to meet the spike time encoding by their dendrites as the third factor in synaptic transmission (Borst, 2010; Scholte et al., 2018).
Stochastic encoding
Apart from dendritic algorism as the third factor of synaptic transmission, there are more factors that make up pyramidal neurons as a completely stochastic computational component: dendritic tree, axonal tree, laminar distributing effect, interlaced microcircuit interaction, stochastic noise and short-term synaptic plasticity. The multiple factors form hundreds of micrometers of interneuron space to assemble microcircuits (Bil and Poo, 2001; Fossati et al., 2019).
Chaotic computation
Chaotic computation is the field of “the qualitative study of unstable aperiodic behavior in deterministic dynamical systems.” A chaos system has three essential properties: it must be dramatically sensitive to initial conditions, with highly disordered behavior, and obey some laws that completely describe its motion. There are three computational indices (Faure and Korn, 2001; Tsuda, 2015; Chao et al., 2018; Sathiyadevi et al., 2019): Lyapunov exponent (λ), Kolmogorov-Sinai entropy (K), and strange attractors. λ is positive if the motion is chaotic; λ equals zero if the two trajectories are separate; λ is a constant amount if the trajectories are periodic. K equals zero if the system is periodic, whereas if it increases without interruption, the system is stochastic, or it increases to a constant value if the system is chaotic. Strange attractors are fractal objects, and their geometry is invariant against changes in scale or size.
Adaptive computation
Adaptive computation is a category of biological computation formatted in biological evolution. An individual changes him- or her-self to meet context-dependent multitask demands.
Adaptive encoding
The prefrontal cortex coordinates cognition, emotion, interpersonal communication, coping with situations, and other complicated mental activities. Adaptive encoding is an algorithmic principle that meets multiple demands for making decisions, planning, monitoring, error finding and revising in a wide range of tasks and successfully implements goal-directed behavior. The encoding feature is expressed mainly in context-dependent computation with chronometric and dynamic adaptation. Its algorithm is multivariate pattern analysis and multi-voxel pattern analysis via principal component analysis.
The emergent computation: A unique basis of human intelligence
Emergent computation is not completely new but has new implications. “Emergent collective computational abilities” appeared in a paper title in Hopfield (1982). In the article, the emergent computation refers to a functional model of creative intelligence in the human brain that has higher energy efficiency or uses less energy to obtain a stronger effect in nervous information processing.
Although dynamic coding (Stokes et al., 2013), abstract quantitative rules (Eiselt and Nieder, 2013), context-dependent computation (Mante et al., 2013), mixed selectivity (Rigotti et al., 2013), content-specific codes (Uluç et al., 2018) reconfiguration of responses (Jackson and Woolgar, 2018), situational processing (Lieberman et al., 2019), and so on have been reported, adaptive encoding cannot cover the crucial property of human social intelligence—creativity. Human beings not only adapt passively to situations but also actively create human society and themselves (Eccles, 1989; MacLean et al., 2014; MacLean, 2016). Therefore, a new mathematical computation that represents actively unique creativity in natural and social environments has existed in the human brain for more than ten thousand years. It might represent the emergent computation.
The efficiency of brain energy cost interchange for the amount of information processing, the partial derivative of the amount of information with respect to energy cost, might be a basic parameter of brain evolution and brain function. A new concept of emergent computing based on the efficiency of brain information processing is suggested as an approach to interdisciplinary theory.
The first principle of emergent computation is the active or initial desire that drives a human to create. The second is higher energy efficiency, meaning trade of more information by a lower energy cost. The third principle implicates the premise of creative intelligence, and the fourth principle encompasses the chronometric effect and spatiotemporal conversion effect. Because energy metabolism inside the human brain is a slow process (second scale), while nervous information changes or external environmental changes are usually rapid processes (millisecond scale), chronometric mechanisms and spatiotemporal conversion are necessary. The fifth principle is recruiting or reusing neurons or physical sources. It not only contains multiple encodings but also may enable more advanced computation that has not yet been identified by the best mathematical authorities. The algorithm of emergent computation might be based on answer of 10 questions as follows.
Suggestion and the questions to be investigated further
As mentioned above, ANNs, DNNs and CNNs have already made much progresses accompanied by some bugs. Except big energy consumption and larger number of hardware parts, it is very difficult for DNNs to simulate the social intelligence of the human brain, even by increasing the number of hidden layers or designing complicated input sets. The bug might be in the basically theoretical conceptions about the primitive unit, network architecture and dynamic principle as well as the basic attributes of the human brain.
We have given suggestion to ANNs in the article text, the list of suggestion is as follows:
the primitive neural node will be substituted by two types of unit: excitatory and inhibitory with different connectionist weight and threshold. For example, sensitivity to spike timing: the lower sensitivity (Δt < 10 ms) or the higher sensitivity (Δt < 6 ms).
The feedback or recurrent inhibition should be set in everywhere of the whole networks, ave. instead of only error backpropagation.
topographic mapping, precise wiring between brain areas, layers, and cell types. This specific wiring bestows huge computational power with minimal request of energy consumption. Small world network principle should be absorbed into ANN architecture.
Energy supply in brain is, respectively, implemented by area, lamina, column. The energy supply in ANNs needs to be improved, so that the energy consumption will be saved. Energy supply modes are different in different learning types, for example between supervised learning with lower energy consumption mainly in cerebellar and unsupervised learning with higher energy consumption in cerebral cortex.
the emergent computation as a basis of cognitive science substitutes “PDP-neuro-computation.” PDP is just one of neuro-computation.
Besides ANNs, the studies in brain simulation along biomedical line have already made much progress in recent decades. For example, the Blue Brain project simulated large-scale networks of neocortical micro circuit based on biological brain data, including morphological and electrophysiological properties as well as cell types of bioactive molecule expression (Markram et al., 2015). A reconstructed virtual brain slice was made up. The meso-circuit was 230 mm thick, containing a total of 139,834 neurons. The virtual slice reproduced oscillatory bursts (1 Hz) and displayed a spectrum of activity states. In addition, the project provided a neocortical model with 0.29 ± 0.01 mm3 containing 31,000 neurons, 55 layer-specific morphological neurons. But, the results are unsatisfactory, because there is only 1 Hz bioelectric activity in the simulated brain slice showing its alive state. The problem might be that how do the much neurons organize together? Apart from laminar and column-like architecture, it is necessary to consider the principle of small-world organization (Watt and Strogatz1998) that means a few neurons compose a microcircuit with a definite basically cognitive function, such as sensory-motor circuits. The questions are worth to inquire further into:
Except three neuronal encodings: rate, PPR, and spike time encoding described in Figures 2, a specifically dendritic action potentials (dAPs) has been reported (Gidon et al., 2020). The pyramidal neuron in II/III layer of neocortex has an ability of parting off XOR, because it can reproduce dendtric action potentials(dAPs). The dAPs mechanism is worth investigating further. How does such a neuron implement the computational ability to separate XOR function? The task should be finished by a deep neural network. So, it might be new neural encoding principle.
Neurogliaform cells (NGFCs) dynamically decouple neuronal synchrony between brain areas (Sakalar et al., 2022), but does neither decrease their own firing nor affect local oscillation. Is its function alike with backpropagation mechanism in BP learning. In the other words, Does NGFCs take a part in implicit information propagation in biological brain?
What is the relationship between brain information and brain energy during intelligent activity?
What is the relation of energy with computational power? Is there any difference in computational principle between the neocortex and cerebellum?
How does neuroglia cell perform in cognitive function and neural energy supply?
What is the meaning of mental capacity-limited or affordable processes for conscious cognitive function from the viewpoint of brain energy metabolism?
Is there any unique computation in the human brain beyond modern computational mathematics or postmodern mathematics?
Is there any other computing principle of the prefrontal cortex beyond adaptive encoding?
Is there any other core computing resource of human brain for social intelligence in addition to the axon tree with long-range networks and local microcircuit-to-microcircuit communication in the prefrontal cortex?
Is it possible to build a common theoretical basis of cognitive science to promote the interdisciplinary development of brain simulation?
Author contributions
ZS designed the study. FS performed the literature search. ZS and FS drafted, revised, and wrote the paper. All authors contributed to the article and approved the submitted version.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
We wish to thank five professors for reviewing and commenting on an early version of our manuscript. They are Lou Minmin at the National Institute of Biological Science, Beijing, China; Wei Chaocun at Shanghai Jiao Tong University; and Fang Fang and Jiong JiongYang and Yan Bao at Peking University, Beijing, China.
References
- Abrams T. W., Kandel E. R. (1988). Is contiguity detection in classical conditioning a system or a cellular property? Learning in Aplysia suggests a possible molecular site. Trends Neurosci. 11, 128–135. doi: 10.1016/0166-2236(88)90137-3, PMID: [DOI] [PubMed] [Google Scholar]
- Ackley D., Hinton G., Sejnowski T. (1985). A learning algorithm for Boltzmann machines. Cogn. Sci. 9, 147–169. doi: 10.1207/s15516709cog0901_7 [DOI] [Google Scholar]
- Azevedo F. A., Carvalho L. R., Grinberg L. T., Farfel J. M., Ferretti R. E., Leite R. E., et al. (2009). Equal numbers of neuronal and nonneuronal cells make the human brain an isometrically scaled-up primate brain. J. Comp. Neurol. 513, 532–541. doi: 10.1002/cne.21974, PMID: [DOI] [PubMed] [Google Scholar]
- Bil G., Poo M. (2001). Synaptic modification by correlated activity: Hebb's postulate revisited. Annu. Rev. Neurosci. 24, 139–166. [DOI] [PubMed] [Google Scholar]
- Boldog E., Bakken T. E., Hodge R. D., Novotny M., Aevermann B. D., Baka J., et al. (2018). Transcriptomic and morphophysiological evidence for a specialized human cortical GABAergic cell type. Nat. Neurosci. 21, 1185–1195. doi: 10.1038/s41593-018-0205-2, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Borst J. G. (2010). The low synaptic release probability in vivo. Trends Neurosci. 33, 259–266. doi: 10.1016/j.tins.2010.03.003, PMID: [DOI] [PubMed] [Google Scholar]
- Branco T., Staras K. (2009). The probability of neurotransmitter release: variability and feedback control at single synapses. Nat. Rev. Neurosci. 10, 373–383. doi: 10.1038/nrn2634, PMID: [DOI] [PubMed] [Google Scholar]
- Buesing L., Maass W. (2010). A spiking neuron as information bottleneck. Neural Comput. 22, 1961–1992. doi: 10.1162/neco.2010.08-09-1084, PMID: [DOI] [PubMed] [Google Scholar]
- Buzsaki G. (2006). Rhythms of the Brain. P61, P287. New York, NY: Oxford University Press. [Google Scholar]
- Cadwell C. R., Bhaduri A., Mostajo-Radji M. A., Keefe M. G., Nowakowski T. J. (2019). Development and arealization of the cerebral cortex. Neuron 103, 980–1004. doi: 10.1016/j.neuron.2019.07.009, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Caporale N., Dan Y. (2008). Spike timing-dependent plasticity: a Hebbian learning rule. Annu. Rev. Neurosci. 31, 25–46. doi: 10.1146/annurev.neuro.31.060407.125639 [DOI] [PubMed] [Google Scholar]
- Carlson T., Goddard E., Kaplan D. M., Klein C., Ritchie J. B. (2018). Ghosts in machine learning for cognitive neuroscience: moving from data to theory. NeuroImage 180, 88–100. doi: 10.1016/j.neuroimage.2017.08.019, PMID: [DOI] [PubMed] [Google Scholar]
- Chao Z. C., Takaura K., Wang L., Fujii N., Dehaene S. (2018). Large-scale cortical networks for hierarchical prediction and prediction error in the primate brain. Neuron 100, 1252–1266.e3. doi: 10.1016/j.neuron.2018.10.004, PMID: [DOI] [PubMed] [Google Scholar]
- Crick F. (1989). The recent excitement about neural networks. Nature 337, 129–132. doi: 10.1038/337129a0 [DOI] [PubMed] [Google Scholar]
- Cummins R., Cummins D. (2000). Minds, Brains, and Computers: An Historical Introduction to the Foundations of Cognitive Science. Oxford: Blackwell Publishers Ltd. [Google Scholar]
- DeFilepe J., Alonso-Nanclares L., Arellano J. I. (2002). Microstructure of the neocortex: comparative aspects. J. Neurocytol. 31, 299–316. doi: 10.1023/A:1024130211265 [DOI] [PubMed] [Google Scholar]
- Deng J., Dong W., Socher R., Li L. J., Li K., Li F. F. (2009). ImageNet: a large-scale hierarchical image database. Proc. CVPR 14067, 248–255. doi: 10.1109/CVPR.2009.5206848 [DOI] [Google Scholar]
- Deng S., Li J., He Q., Zhang X., Zhu J., Li L., et al. (2020). Regulation of recurrent inhibition by asynchronous glutamate release in neocortex. Neuron 105, 522–533.e4. doi: 10.1016/j.neuron.2019.10.038, PMID: [DOI] [PubMed] [Google Scholar]
- Doya K. (2000). Complementary roles of basal ganglia and cerebellum in learning and motor control. Curr. Opin. Neurobiol. 10, 732–739. doi: 10.1016/S0959-4388(00)00153-7, PMID: [DOI] [PubMed] [Google Scholar]
- Eccles J. C. (1953). The Neurophysiological Basis of Mind, London: Oxford University Press. [Google Scholar]
- Eccles J. C. (1964). The Physiology of Synapses. Berlin: Springer-Verlag. 201–215. [Google Scholar]
- Eccles J. C. (1973). "The Understanding of the Brain." New York, NY: McGraw-Hill book Company. 88–92. [Google Scholar]
- Eccles J. C. (1989). Evolution of the Brain: Creation of the Self. London; New York, NY: Routledge. 217–236. [Google Scholar]
- Eccles J. C., Fatt P., Koketsu K. (1954). Cholinergic and inhibitory synapses in a pathway from motor-axon collaterals to motoneurones. J. Physiol. 126, 524–562. doi: 10.1113/jphysiol.1954.sp005226, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Egger R., Narayanan R. T., Guest J. M., Bast A., Udvary D., Messore L. F., et al. (2020). Cortical output is gated by horizontally projecting neurons in the deep layers. Neuron 105, 122–137.e8. doi: 10.1016/j.neuron.2019.10.011, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eiselt A. K., Nieder A. (2013). Representation of abstract quantitative rules applied to spatial and numerical magnitudes in primate prefrontal cortex. J. Neurosci. 33, 7526–7534. doi: 10.1523/JNEUROSCI.5827-12.2013, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Evrard H. C. (2018). Von Economo and fork neurons in the monkey insula, implications for evolution of cognition. Curr. Opin. Behav. Sci. 21, 182–190. doi: 10.1016/j.cobeha.2018.05.006 [DOI] [Google Scholar]
- Faure P., Korn H. (2001). Is there chaos in the brain? I. concepts of nonlinear dynamics and methods of investigation. C. R. Acad. Sci. III 324, 773–793. doi: 10.1016/S0764-4469(01)01377-4, PMID: [DOI] [PubMed] [Google Scholar]
- Feldman J., Ballard D. (1982). Connectionist models and their properties. Cogn. Sci. 6, 205–254. doi: 10.1207/s15516709cog0603_1 [DOI] [Google Scholar]
- Fitzsimonds R. M., Poo M. M. (1998). Retrograde signaling in the development and modification of synapses. Physiol. Rev. 78, 143–170. doi: 10.1152/physrev.1998.78.1.143, PMID: [DOI] [PubMed] [Google Scholar]
- Fossati M., Assendorp N., Gemin O., Colasse S., Dingli F., Arras G., et al. (2019). Trans-synaptic signaling through the glutamate receptor delta-1 mediates inhibitory synapse formation in cortical pyramidal neurons. Neuron 104, 1081–1094.e7. doi: 10.1016/j.neuron.2019.09.027, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gefen T., Papastefan S. T., Rezvanian A., Bigio E. H., Weintraub S., Rogalski E., et al. (2018). Von Economo neurons of the anterior cingulate across the lifespan and in Alzheimer's disease. Cortex 99, 69–77. doi: 10.1016/j.cortex.2017.10.015, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gidon A., Zolnik T. A., Fidzinski P., Bolduan F., Papoutsi A., Poirazi P., et al. (2020). Dendritic action potentials and computation in human layer 2/3 cortical neurons. Science 367, 83–87. doi: 10.1126/science.aax6239, PMID: [DOI] [PubMed] [Google Scholar]
- Girshick R., Donahue J., Darrell T., Malik J. (2014). Rich feature hierarchies for accurate object detection and semantic segmentation. IEEE Conf. Comput. Vis. Pattern Recognit. 2014, 580–587. doi: 10.1109/CVPR.2014.81 [DOI] [Google Scholar]
- Goulden N., Khusnulina A., Davis N. J., Bracewell R. M., Bokde A. L., McNulty J. P., et al. (2014). The salience network is responsible for switching between the default mode network and the central executive network: replication from DCM. NeuroImage 99, 180–190. doi: 10.1016/j.neuroimage.2014.05.052, PMID: [DOI] [PubMed] [Google Scholar]
- Haggar R. A., Barr M. L. (1950). Quantitative data on the size of synaptic end-bulbs in the cat's spinal cord. J. Comp. Neurol. 93, 17–35. doi: 10.1002/cne.900930103, PMID: [DOI] [PubMed] [Google Scholar]
- Hamilton T., Afshar S., van Schaik A., Tapson J. (2014). Stochastic electronics: A neuro-inspired design paradigm for integrated circuits. Proc. IEEE 102, 843–859. doi: 10.1109/JPROC.2014.2310713 [DOI] [Google Scholar]
- Hebb D. O. (1949). The Organization of Behavior. New York, NY: John Willy. [Google Scholar]
- Hikosaka O., Isoda M. (2010). Switching from automatic to controlled behavior: cortico-basal ganglia mechanisms. Trends Cogn. Sci. 14, 154–161. doi: 10.1016/j.tics.2010.01.006, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hinton G., Deng L., Yu D., Dahl G. E., Mohamed A., Jaitly N., et al. (2012). Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Process. Mag. 29, 82–97. doi: 10.1109/MSP.2012.2205597 [DOI] [Google Scholar]
- Hopfield J. J. (1982). Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. U. S. A. 79, 2554–2558. doi: 10.1073/pnas.79.8.2554, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hopfield J. J. (1984). Neurons with graded response have collective computational properties like those of two-state neurons. Proc. Natl. Acad. Sci. U. S. A. 81, 3088–3092. doi: 10.1073/pnas.81.10.3088, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hopfield J. J., Brody C. D. (2004). Learning rules and network repair in spike-timing-based computation networks. Proc. Natl. Acad. Sci. U. S. A. 101, 337–342. doi: 10.1073/pnas.2536316100, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Isaacson J. S., Scanziani M. (2011). How inhibition shapes cortical activity. Neuron 72, 231–243. doi: 10.1016/j.neuron.2011.09.027, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jackson J. B., Woolgar A. (2018). Adaptive coding in the human brain: distinct object features are encoded by overlapping voxels in frontoparietal cortex. Cortex 108, 25–34. doi: 10.1016/j.cortex.2018.07.006, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jayachandran M., Linley S. B., Schlecht M., Mahler S. V., Vertes R. P., Allen T. A. (2019). Prefrontal pathways provide top-down control of memory for sequences of events. Cell Rep. 28, 640–654.e6. doi: 10.1016/j.celrep.2019.06.053, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jia H., Rochefort N. L., Chen X., Konnerth A. (2010). Dendritic organization of sensory input to cortical neurons in vivo. Nature 464, 1307–1312. doi: 10.1038/nature08947, PMID: [DOI] [PubMed] [Google Scholar]
- Kamigaki T. (2019). Prefrontal circuit organization for executive control. Neurosci. Res. 140, 23–36. doi: 10.1016/j.neures.2018.08.017, PMID: [DOI] [PubMed] [Google Scholar]
- Kandel E. R. (2001). The molecular biology of memory storage: a dialogue between genes and synapses. Science 294, 1030–1038. doi: 10.1126/science.1067020, PMID: [DOI] [PubMed] [Google Scholar]
- Kietzmann T., McClure P., Kriegeskorte N. (2019). "Deep neural networks in computational neuroscience." bioRxiv preprint [Epub ahead of preprint] doi: 10.1101/133504 [DOI] [PMC free article] [PubMed]
- Kim Y., Yang G. R., Pradhan K., Venkataraju K. U., Bota M., García Del Molino L. C., et al. (2017). Brain-wide maps reveal stereotyped cell-type-based cortical architecture and subcortical sexual dimorphism. Cells 171, 456–469.e22. doi: 10.1016/j.cell.2017.09.020, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kriegeskorte N., Golan T. (2019). Neural network models and deep learning. Curr. Biol. 29, R231–r236. doi: 10.1016/j.cub.2019.02.034 [DOI] [PubMed] [Google Scholar]
- Krizhevsky A., Sutskever I., Hinton G. E. (2012). ImageNet classification with deep convolutional neural networks., PMID: 18368112 [Google Scholar]
- LeCun Y., Bengio Y., Hinton G. (2015). Deep learning. Nature 521, 436–444. doi: 10.1038/nature14539 [DOI] [PubMed] [Google Scholar]
- Lieberman M. D., Straccia M. A., Meyer M. L., du M., Tan K. M. (2019). Social, self, (situational), and affective processes in medial prefrontal cortex (MPFC): causal, multivariate, and reverse inference evidence. Neurosci. Biobehav. Rev. 99, 311–328. doi: 10.1016/j.neubiorev.2018.12.021, PMID: [DOI] [PubMed] [Google Scholar]
- Lillicrap T. P., Santoro A., Marris L., Akerman C. J., Hinton G. (2020). Backpropagation and the brain. Nat. Rev. Neurosci. 21, 335–346. doi: 10.1038/s41583-020-0277-3 [DOI] [PubMed] [Google Scholar]
- Losonczy A., Makara J. K., Magee J. C. (2008). Compartmentalized dendritic plasticity and input feature storage in neurons. Nature 452, 436–441. doi: 10.1038/nature06725, PMID: [DOI] [PubMed] [Google Scholar]
- Maass W. (1997). Networks of spiking neurons: the third generation of neural network models. Neural Netw. 10, 1659–1671. doi: 10.1016/S0893-6080(97)00011-7 [DOI] [Google Scholar]
- Maass W. (2014). Noise as a resource for computation and learning in networks of spiking neurons. Proc. IEEE 102, 860–880. doi: 10.1109/JPROC.2014.2310593 [DOI] [Google Scholar]
- Maass W. (2015). To spike or not to spike: that is the question. Proc. IEEE 103, 2219–2224. doi: 10.1109/JPROC.2015.2496679 [DOI] [Google Scholar]
- MacLean E. L. (2016). Unraveling the evolution of uniquely human cognition. Proc. Natl. Acad. Sci. U. S. A. 113, 6348–6354. doi: 10.1073/pnas.1521270113, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- MacLean E. L., Hare B., Nunn C. L., Addessi E., Amici F., Anderson R. C., et al. (2014). The evolution of self-control. Proc. Natl. Acad. Sci. U. S. A. 111, E2140–E2148. doi: 10.1073/pnas.1323533111, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mante V., Sussillo D., Shenoy K. V., Newsome W. T. (2013). Context-dependent computation by recurrent dynamics in prefrontal cortex. Nature 503, 78–84. doi: 10.1038/nature12742, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Marblestone A. H., Wayne G., Kording K. P. (2016). Toward an integration of deep learning and neuroscience. Front. Comput. Neurosci. 10:94. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mark S., Romani S., Jezek K., Tsodyks M. (2017). Theta-paced flickering between place-cell maps in the hippocampus: a model based on short-term synaptic plasticity. Hippocampus 27, 959–970. doi: 10.1002/hipo.22743, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Markram H., Lübke J., Frotscher M., Sakmann B. (1997). Regulation of synaptic efficacy by coincidence of postsynaptic APs and EPSPs. Science 275, 213–215. doi: 10.1126/science.275.5297.213, PMID: [DOI] [PubMed] [Google Scholar]
- Markram H., Muller E., Ramaswamy S., Reimann M. W., Abdellah M., Sanchez C. A., et al. (2015). Reconstruction and simulation of neocortical microcircuitry. Cells 163, 456–492. doi: 10.1016/j.cell.2015.09.029, PMID: [DOI] [PubMed] [Google Scholar]
- Mcculloch W. S., Pitts W. (1943). A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 5, 115–133. doi: 10.1007/BF02478259 [DOI] [PubMed] [Google Scholar]
- McDonnell M. D., Ward L. M. (2011). The benefits of noise in neural systems: bridging theory and experiment. Nat. Rev. Neurosci. 12, 415–425. doi: 10.1038/nrn3061, PMID: [DOI] [PubMed] [Google Scholar]
- Miller E. K., Cohen J. D. (2001). An integrative theory of prefrontal cortex function. Annu. Rev. Neurosci. 24, 167–202. doi: 10.1146/annurev.neuro.24.1.167, PMID: [DOI] [PubMed] [Google Scholar]
- Minsky M., Papert S. (1969). "Perceptrons." Cambridge, MA: MIT Press. [Google Scholar]
- Miyake A., Friedman N. P., Emerson J. M., Witzki A. H., Howerter A., Wager T. D. (2000). The unity and diversity of executive functions and their contributions to complex “frontal lobe” tasks: a latent variable analysis. Cogn. Psychol. 41, 49–100. doi: 10.1006/cogp.1999.0734 [DOI] [PubMed] [Google Scholar]
- Mongillo G., Rumpel S., Loewenstein Y. (2018). Inhibitory connectivity defines the realm of excitatory plasticity. Nat. Neurosci. 21, 1463–1470. doi: 10.1038/s41593-018-0226-x, PMID: [DOI] [PubMed] [Google Scholar]
- Montague P. R., Sejnowski T. J. (1994). The predictive brain: temporal coincidence and temporal order in synaptic learning mechanisms. Learn. Mem. 1, 1–33. doi: 10.1101/lm.1.1.1, PMID: [DOI] [PubMed] [Google Scholar]
- Mrzljak L., Uylings H. B., Kostovic I., van Eden C. G. (1992). Prenatal development of neurons in the human prefrontal cortex. II. A quantitative Golgi study. J. Comp. Neurol. 316, 485–496. doi: 10.1002/cne.903160408, PMID: [DOI] [PubMed] [Google Scholar]
- Nakajima M., Schmitt L. I., Halassa M. M. (2019). Prefrontal cortex regulates sensory filtering through a basal ganglia to thalamus pathway. Neuron 103, 445–458.e10. doi: 10.1016/j.neuron.2019.05.026, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Naveen K., Sparvathi R. M. (2022). “Analysis of medical image by using machine learning applications of convolutional neural networks” In Proceedings of the Second International Conference on Artificial Intelligence and Smart Energy (ICAIS-2022) IEEE Xplore Part Number: CFP22OAB-ART, pp. 115–123.
- Near J., Ho Y. C., Sandberg K., Kumaragamage C., Blicher J. U. (2014). Long-term reproducibility of GABA magnetic reso nance spectroscopy. NeuroImage 99, 191–196. doi: 10.1016/j.neuroimage.2014.05.059, PMID: [DOI] [PubMed] [Google Scholar]
- Nimchinsky E. A., Sabatini B. L., Svoboda K. (2002). Structure and function of dendritic spines. Annu. Rev. Physiol. 64, 313–353. doi: 10.1146/annurev.physiol.64.081501.160008 [DOI] [PubMed] [Google Scholar]
- Onorato I., Neuenschwander S., Hoy J., Lima B., Rocha K.-S., Broggini A. C., et al. (2020). A distinct class of bursting neurons with strong gamma synchronization and stimulus selectivity in monkey V1. Neuron 105, 180–197.e5. doi: 10.1016/j.neuron.2019.09.039, PMID: [DOI] [PubMed] [Google Scholar]
- Paredes M. F., James D., Gil-Perotin S., Kim H., Cotter J. A., Ng C., et al. (2016). Extensive migration of young neurons into the infant human frontal lobe. Science 354, 70–73. doi: 10.1126/science.aaf7073 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pavlov I. (1927) Conditioned Reflexes: An Investigation of the Physiological Activity of the Cerebral Cortex. trans. ed. G. V. Anrep. Oxford: Oxford University Press. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pei J., Deng L., Song S., Zhao M., Zhang Y., Wu S., et al. (2019). Towards artificial general intelligence with hybrid Tianjic chip architecture. Nature 572, 106–111. doi: 10.1038/s41586-019-1424-8, PMID: [DOI] [PubMed] [Google Scholar]
- Petanjek Z., Berger B., Esclapez M. (2009). Origins of cortical GABAergic neurons in the cynomolgus monkey. Cereb. Cortex 19, 249–262. doi: 10.1093/cercor/bhn078, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Petanjek Z., Judas M., Kostović I., Uylings H. B. (2008). Lifespan alterations of basal dendritic trees of pyramidal neurons in the human prefrontal cortex: a layer-specific pattern. Cereb. Cortex 18, 915–929. doi: 10.1093/cercor/bhm124, PMID: [DOI] [PubMed] [Google Scholar]
- Raichle M. E. (2010). The brain's dark energy. Sci. Am. 302, 44–49. doi: 10.1038/scientificamerican0310-44 [DOI] [PubMed] [Google Scholar]
- Raichle M. E. (2015). The brain's default mode network. Annu. Rev. Neurosci. 38, 433–447. doi: 10.1146/annurev-neuro-071013-014030 [DOI] [PubMed] [Google Scholar]
- Raichle M. E., MacLeod A. M., Snyder A. Z., Powers W. J., Gusnard D. A., Shulman G. L. (2001). A default mode of brain function. Proc. Natl. Acad. Sci. U. S. A. 98, 676–682. doi: 10.1073/pnas.98.2.676, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ren S. Q., Li Z. Z.,Lin S., Bergami M., Shi S. H. (2019). “Precise long-range microcircuit-to-microcircuit communication connects the frontal and sensory cortices in the mammalian brain.” Neuron, 104: 385–401, doi: 10.1016/j.neuron.2019.06.028, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Richards B. A., Lillicrap T. P. (2019). Dendritic solutions to the credit assignment problem. Curr. Opin. Neurobiol. 54, 28–36. doi: 10.1016/j.conb.2018.08.003, PMID: [DOI] [PubMed] [Google Scholar]
- Rigotti M., Barak O., Warden M. R., Wang X. J., Daw N. D., Miller E. K., et al. (2013). The importance of mixed selectivity in complex cognitive tasks. Nature 497, 585–590. doi: 10.1038/nature12160, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Roberts P. D. (1999). Computational consequences of temporally asymmetric learning rules: I. differential hebbian learning. J. Comput. Neurosci. 7, 235–246. doi: 10.1023/A:1008910918445 [DOI] [PubMed] [Google Scholar]
- Roelfsema P. R., Engel A. K., König P., Singer W. (1997). Visuomotor integration is associated with zero time-lag synchronization among cortical areas. Nature 385, 157–161. doi: 10.1038/385157a0, PMID: [DOI] [PubMed] [Google Scholar]
- Rosenblatt F. (1958). The perceptron: a probabilistic model for information storage and organization in the brain. Psychol. Rev. 65, 386–408. doi: 10.1037/h0042519, PMID: [DOI] [PubMed] [Google Scholar]
- Rumelhart D., Hinton G., Williams R. (1986). Learning representations by Back propagating errors. Nature 323, 533–536. doi: 10.1038/323533a0 [DOI] [Google Scholar]
- Rumelhart D., McClelland J. (1986). Parallel Distributed Processing: Explorations in the Microstructure of Cognition: Foundations. Cambridge, MA: MIT Press. [Google Scholar]
- Sakalar E., Klausberger T., Lasztóczi B. (2022). Neurogliaform cells dynamically decouple neuronal synchrony between brain areas. Science, 377–324–328. doi: 10.1126/science.abo3355 [DOI] [PubMed] [Google Scholar]
- Sathiyadevi K., Karthiga S., Chandrasekar V. K., Senthilkumar D. V., Lakshmanan M. (2019). Frustration induced transient chaos, fractal and riddled basins in coupled limit cycle oscillators. Commun. Nonlinear Sci. Numer. Simul. 72, 586–599. doi: 10.1016/j.cnsns.2019.01.024 [DOI] [Google Scholar]
- Satpute A. B., Lindquist K. A. (2019). The default mode Network's role in discrete emotion. Trends Cogn. Sci. 23, 851–864. doi: 10.1016/j.tics.2019.07.003, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Scholte H. S., Losch M. M., Ramakrishnan K., de Haan E. H. F., Bohte S. M. (2018). Visual pathways from the perspective of cost functions and multi-task deep neural networks. Cortex 98, 249–261. doi: 10.1016/j.cortex.2017.09.019, PMID: [DOI] [PubMed] [Google Scholar]
- Sejnowski T. J. (1995). Pattern recognition. Time for a new neural code? Nature 376, 21–22. doi: 10.1038/376021a0 [DOI] [PubMed] [Google Scholar]
- Shen Z. (2017). Development of brain function theory and the frontier brain projects. Chin. Sci. Bull. 62, 3429–3439. doi: 10.1360/N972017-00426 [DOI] [Google Scholar]
- Sherrington C. (1906). "The Integrative Action of the Nervous System." New York, NY: Scribners. [Google Scholar]
- Silver R. A. (2010). Neuronal arithmetic. Nat. Rev. Neurosci. 11, 474–489. doi: 10.1038/nrn2864, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Song S., Miller K. D., Abbott L. F. (2000). Competitive Hebbian learning through spike-timing-dependent synaptic plasticity. Nat. Neurosci. 3, 919–926. doi: 10.1038/78829, PMID: [DOI] [PubMed] [Google Scholar]
- Spruston N. (2008). Pyramidal neurons: dendritic structure and synaptic integration. Nat. Rev. Neurosci. 9, 206–221. doi: 10.1038/nrn2286 [DOI] [PubMed] [Google Scholar]
- Stokes M. G., Kusunoki M., Sigala N., Nili H., Gaffan D., Duncan J. (2013). Dynamic coding for cognitive control in prefrontal cortex. Neuron 78, 364–375. doi: 10.1016/j.neuron.2013.01.039, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sultana A., Dey S., Rahma M. A. (2021). “A deep CNN based Kaggle contest winning model to recognize real-time facial exppression.” In 2021 5th International Conference on Electrical Engineering and Information and Communication Technology (ICEEICT) Military Institute of Science and Technology (MIST), Dhaka-1216, Bangladesh.
- Sun Q. X., Li M., Ren M., Zhao Q., Zhong Y., Ren P., et al. (2019). A whole-brain map of long-range inputs to GABAergic interneurons in the mouse medial prefrontal cortex. Nat. Neurosci. 22, 1357–1370. doi: 10.1038/s41593-019-0429-9, PMID: [DOI] [PubMed] [Google Scholar]
- Tai Y., Gallo N. B., Wang M., Yu J.-R., Aelst L. V. (2019). Axo-axonic innervation of neocortical pyramidal neurons by GABAergic chandelier cells requires AnkyrinG-associated L1CAM. Neuron 102, 358–372.e9. doi: 10.1016/j.neuron.2019.02.009, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tang Y. Y., Rothbart M. K., Posner M. I. (2012). Neural correlates of establishing, maintaining, and switching brain states. Trends Cogn. Sci. 16, 330–337. doi: 10.1016/j.tics.2012.05.001, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tavanaei A., Maida A. (2017). "Bio-inspired spiking convolutional neural network using layer-wise sparse coding and STDP learning." doi: 10.48550/arXiv.1611.03000. [Epub ahead of preprint]. [DOI]
- Thorpe S. J., Fabre-Thorpe M. (2001). Seeking categories in the brain. Science 291, 260–263. doi: 10.1126/science.1058249 [DOI] [PubMed] [Google Scholar]
- Tsodyks M. V., Markram H. (1997). The neural code between neocortical pyramidal neurons depends on neurotransmitter release probability. Proc. Natl. Acad. Sci. U. S. A. 94, 719–723. doi: 10.1073/pnas.94.2.719, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tsuda I. (2015). Chaotic itinerancy and its roles in cognitive neurodynamics. Curr. Opin. Neurobiol. 31, 67–71. doi: 10.1016/j.conb.2014.08.011, PMID: [DOI] [PubMed] [Google Scholar]
- Uluç I., Schmidt T. T., Wu Y. H., Blankenburg F. (2018). Content-specific codes of parametric auditory working memory in humans. NeuroImage 183, 254–262. doi: 10.1016/j.neuroimage.2018.08.024, PMID: [DOI] [PubMed] [Google Scholar]
- van den Heuvell M. P., Sporns O. (2013). Network hubs in the human brain. TICS 17, 683–696. doi: 10.1016/j.tics.2013.09.012 [DOI] [PubMed] [Google Scholar]
- Watts D. J., Strogatz S. H. (1998). Collective dynamics of 'small-world' networks. Nature 393, 440–442. doi: 10.1038/30918 [DOI] [PubMed] [Google Scholar]
- Weitz A. J., Lee H. J., ManK C., Lee J. H. (2019). Thalamic input to orbitofrontal cortex drives brain-wide, frequency-dependent inhibition mediated by GABA and zona incerta. Neuron 104, 1153–1167.e4. doi: 10.1016/j.neuron.2019.09.023, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
- Whittaker V. P., Gray E. G. (1962). The synapse: biology and morphology. Br. Med. Bull. 18, 223–228. doi: 10.1093/oxfordjournals.bmb.a069983 [DOI] [PubMed] [Google Scholar]
- Williams S. R. (2004). Spatial compartmentalization and functional impact of conductance in pyramidal neurons. Nat. Neurosci. 7, 961–967. doi: 10.1038/nn1305, PMID: [DOI] [PubMed] [Google Scholar]
- Williams L. E., Holtmaat A. (2019). Higher-order Thalamocortical inputs gate synaptic long-term potentiation via disinhibition. Neuron 101, 91–102.e4. doi: 10.1016/j.neuron.2018.10.049, PMID: [DOI] [PubMed] [Google Scholar]
- Wonders C. P., Anderson S. A. (2006). The origin and specification of cortical interneurons. Nat. Rev. Neurosci. 7, 687–696. doi: 10.1038/nrn1954 [DOI] [PubMed] [Google Scholar]
- Zhao Q.-Z., Zheng P., Xu S. T., Wu X. (2019). Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst. 30, 3212–3232. doi: 10.1109/TNNLS.2018.2876865 [DOI] [PubMed] [Google Scholar]
- Zucker R. S., Regehr W. G. (2002). Short-term synaptic plasticity. Annu. Rev. Physiol. 64, 355–405. doi: 10.1146/annurev.physiol.64.092501.114547 [DOI] [PubMed] [Google Scholar]