Abstract
Conjunctive encoding of inputs has been hypothesized to be a key feature in the computational capabilities of the brain. This has been inferred based on behavioral studies and electrophysiological recording from animals. In this report, we show that random neuronal ensembles grown on multi-electrode array perform a coarse-conjunctive encoding for a sequence of inputs with the first input setting the context. Such an encoding scheme creates similar yet unique population codes at the output of the ensemble, for related input sequences, which can then be decoded via a simple perceptron and hence a single STDP neuron layer. The random neuronal ensembles allow for pattern generalization and novel sequence classification without needing any specific learning or training of the ensemble. Such a representation of the inputs as population codes of neuronal ensemble outputs, has inherent redundancy and is suitable for further decoding via even probabilistic/random connections to subsequent neuronal layers. We reproduce this behavior in a mathematical model to show that a random neuronal network with a mix of excitatory and inhibitory neurons and sufficient connectivity creates similar coarse-conjunctive encoding of input sequences.
Introduction
Pattern or sequence recognition and classification is a well-studied problem in engineering that uses biologically inspired architectures like artificial neural networks, and more recently deep learning networks that have shown promising results in solving such tasks. However, the learning algorithms adopted by these architectures require multiple iterations and modifications of the connectivity weights across all layers of the network. The existence of similar multi-layered learning in the biological neuronal networks for efficient processing of input stimuli and classification of inputs has not been observed yet experimentally. An alternative learning architecture is to have a random neuronal ensemble with a mix of inhibitory and excitatory neurons that is then connected to another layer of perceptron type neurons, in a probabilistic manner, with learning restricted to the final perceptron layer. We describe this further in the schematic in Fig. 1, where a layered neuronal system with probabilistic connectivity at input and output of first layer, is connected to a second layer having neurons equipped with STDP, to solve the problem of input classification without any need for network modification/learning at the input layer. We experimentally validate this architecture by using neuronal ensembles cultured on a multi electrode array, to form the first layer of the Fig. 1. The multi-electrode array allows us to create complex spatio-temporal input stimulation patterns, that get encoded by the neuronal tissue which is then observed as responses at the electrodes for further analysis. We show through modeling and by fitting experimental data that probabilistic connections and a layered architecture as in Fig. 1, can provide a very robust platform to implement context dependent classification. Our data and results show the presence and usefulness of coarse-conjunctive tuning of neurons in the random ensemble, as observed in the brain for pattern recognition. These results also suggest the usefulness of Liquid State Machines as a framework for brain-computations1.
The architecture shown in Fig. 1 mimics the ability of the brain to combine (or bind) multiple features of a stimulus into single percepts (the ‘binding problem’). This is illustrated by single neurons coding for 3 different shapes (square, triangle and circle) and 3 different colors in the first layer (L1). If we now consider an object with two features having both color and shape information for e.g. a red square, the brain does not perceive these features separately but as a composite red square. The brain solves this problem by having neurons that encode shape and color information conjunctively (the conjunctive neurons). This type of encoding is called ‘conjunctive encoding’. If we consider all possible combinations between color and shape, the combinatorial problem is huge and would require far too many neurons to adequately represent all possible conjunctions. The problem can be partially solved by having a coding mechanism with distributed representations where each neuron encodes a mixture of input features that become active at different levels to different combinations of the two features viz. color and shape. Thus, a single neuron instead of encoding for a single conjunction, codes for multiple conjunctions, i.e., the neuron does coarse conjunctive encoding. This is illustrated by the conjunctive neurons in the second layer (L2) of Fig. 1.
Such a conjunction of activity patterns has been considered important for meaningful information processing, efficient storage, retrieval from partial cues, pattern completion and generalization behaviors in the brain2. Conjunctive neurons allow for hierarchical encoding of information by spatio-temporal convergence of inputs which show up as “Trial Type Cells”3. Conjunctive binding is a key operation in mechanistic models for hippocampus2, brain inspired computational models4 and VLSI architectures5,6. Rats develop a conjunctive code for context and order7 and for different objects within a visual scene8. Recordings of single hippocampal neurons show that encoding is coarse-conjunctive9. Multiunit activity could be grouped according to behavioral task in progress10. Cells in EC conjunctively encode position and head position information11. Different face features decoded from single neuron recordings in IT shows coarse tuning of neurons9. Firing of hippocampal cells which encode spatial map also correlate to task events10. It is also suitable for function approximation and generalization by an artificial neural network12.
Modeling studies suggest that combination of features in the stimulus input can be distinguished by a distribution of activation of many neurons. It is also conceivable that the output from many coarse conjunctive neurons converge to one or few ‘output’ neurons that in turn control behavior. In the mushroom body of the common fruitfly Drosophila melanogaster, structural layers of the kind illustrated in Fig. 1 exist. Output from ~2000 third order kenyon cells that encode odour stimuli, converge on ~21 structurally distinct olfactory bulb output neurons (OBONs)13 and a suppression of a single pair of OBON regulates aversive memory associations14. However, the way information is encoded and decoded across different layers before it converges on the output neuron is not known.
Neuronal cultures on multi-electrode arrays have been previously used to study neuronal networks. The ability to train neuronal cultures has been studied15. Different groups have used such cultures to demonstrate processing of spatio-temporal stimuli16–23. They have been used as a model to study the network basis of neurological disorders and recently to study the role of neurotransmitters in neuronal network dynamics24–26. Their activity has been modeled using connectivity maps and hidden markov models27. They have been used to construct simple computational systems. However, such systems have not been used to test different hypotheses about the network architectures for computing using neuronal circuits.
In this study, we have attempted to understand how coarse encoding arises and how features related to the input are encoded by a distributed system of neurons connected randomly using neurons cultured on multielectrode arrays. First, we show that responses from a neuronal ensemble grown on multi-electrode array show coarse-conjunctive encoding of multiple spatio-temporal inputs and then demonstrate their ability to do context dependent encoding, which can then be decoded/classified robustly using ‘perceptrons’ as proxy for the output neuron shown in layer 3 (L3) of Fig. 1. The inputs are paired electrical stimuli at different spatial locations, in different combinations separated by a time interval (spatio-temporal pattern), whose physical parameters were fixed much like the sensory stimuli in the cognition experiments where the perception of sensory stimuli with fixed physical features are studied for context dependency. The results show that neuronal ensembles with probabilistic ‘random’ connectivity can inherently do coarse-conjunctive encoding, without any specific learning or training. We discuss the relevance of such an architecture, where an interplay of random connectivity and layered architecture simplifies the pattern classification tasks.
Methods
Neuronal Culture
All animal experiments were performed in accordance with guidelines, rules and regulations of the Institutional Animal Ethics Committee (IAEC) for animal experiments of the Indian Institute of Science, Bangalore, India constituted as per article number 13 of the CPCSEA (Committee for the purpose of Control and Supervision of Experiments on Animals, http://cpcsea.nic.in) rules, laid down by Government of India.
Neuronal culture growth and maintenance was using standard procedures22,28. Briefly, dissociated neuronal cell cultures were prepared from hippocampus of 0–2 day old wistar rat pups on 120 MEA from MultiChannel Systems. Micro-dissected hippocampus was digested in papain solution and plated on electrode region of the MEA coated with laminin. The dishes were flooded with 1 ml of medium after the cells had adhered to the substrate, and stored with ethylene-propylene membrane lids in a 65% RH incubator at 37 °C, 5% CO2.
We used antibiotic/antimycotic drugs to control contamination. Feedings consisted of 50% medium replacement twice per week. The medium was used with glial conditioning (ara-C) after 7 days.
The culture dish was placed in a separate incubator which maintained an ambient of 5% CO2 at 37 °C while doing recordings and stimulations.
Recording and Stimulation
We used MEA-2100 System from MultiChannel Systems©, Germany for recording from and stimulating the cultures grown on the MEA. The hardware was used to record signals from 120 channels simultaneously at 50 kHz and to generate stimulus pulses at all electrodes under software control.
Analysis
The data was acquired from the device using MATLAB. Spike detection was done on the acquired data for further processing. This required filtering, artifact suppression and appropriate threshold crossing detection which was done on-line using MATLAB. Threshold for each electrode was estimated as 5x standard deviation (estimated using median values) and was applied on the absolute value of the signal.
For electrical stimulation we chose the parameters which have been shown to be effective in previous studies29. For each stimulus we used a bi-phasic voltage pulse of amplitude 500 mV and a pulse width of 500 µs in each phase.
Experimental Protocols
Input Patterns
A spatio-temporal input coding strategy was adopted22. 8 electrodes which evoked maximum response from the culture on stimulation were selected as input electrodes and were labeled A,B,C,D,E,F,G,H. These gave us spatially distinct sites for stimulation. Spatio-temporal patterns were created by pairing two of these electrodes at a time resulting in 56 different patterns (AB, BA, AC…). The time interval between pulses was set as 0.5 ms and 3 ms. The patterns were applied to the culture with a time interval of 250 ms.
Output decoding
We defined the output vector from the culture for each pattern as a 120 element binary vector indicating the occurrence of a spike in a 100 ms post stimulus window.
Xjk is the output pattern for the culture for the kth presentation of input pattern j. Here sMjk is the spike occurrence indicator for electrode M and is defined as sMjk = 1 if at least one spike occurs in the time window 5 ms to 100 ms after the jth input pattern is presented to the culture kth time.
A perceptron is a simple processing element which does a weighted sum of its inputs and generates a binary (1/0) output if the sum is greater than a threshold value. It can be described by the following expression.
Here Ojk is defined as the output of the perceptron j with a weight vector Wj for the kth presentation of input pattern j.
The weight vector describes a hyper plane which separates the set of outputs which the perceptron is trained to identify from the rest. These set of weights are learned using the perceptron training algorithm, the delta rule30.
The decoder is an array of such perceptrons which can be used to assign a class to an output vector.
Results
We stimulated the culture with 56 spatio-temporal input patterns and recorded the responses. These were generated using 8 electrodes (labeled A, B C D..H) Pairing two at a time with a time delay of 0.5 ms22. We defined the first electrode to be stimulated to set the ‘context’ in which subsequent stimuli are processed.
We then looked at responses at each electrode for these patterns and found them to be coarsely tuned with multiple electrodes responding probabilistically to the 56 input patterns (Fig. 2c). With an array of perceptrons, we were able to classify the output codes which showed them to be linearly separable22 (Fig. 4a). This method has been shown to be equivalent to other classification methods like logistic regression31. Figure 3b shows coarse tuning at two electrodes with responses to multiple input patterns. The output response from a single electrode (responses represented by blue dots in Fig. (3b)) cannot distinguish the different input patterns (DF, DG, DH, HA, HB, HC). The probability of the responses at a single electrode show conjunctive and disjunctive behavior based on timing and order of inputs as a result of excitatory and inhibitory connections from the inputs (Fig. 3a). Looking at all the electrodes, we found that a significant number of electrodes show this kind of response leading to distinct population codes (inferred based on them being linearly classifiable by perceptrons). The input patterns become distinguishable as small clusters upon increasing the number of output electrodes (as in example demonstrated using Fig. 3c). A minimum number of output electrodes are thus necessary to separate the input patterns. With 120 electrodes the input patterns were separable and classifiable. Thus coarse-conjunctive coding results in unique population codes.
The paired input patterns could be grouped into 8 sets ([AB, AC…], [BA, BC…]). 8 linear perceptrons were trained to classify them successfully showing the grouping of output responses based on first electrode stimulated in the sequence.
In order to study context dependent grouping of the inputs, we left out one of the patterns within each group where the first stimulus sets the context (e.g. AH in group [AB, AC, AD, AE, AF, AG, AH]) and trained the output perceptrons.
To be able to identify the group of input, we had 8 perceptrons. Each perceptron was trained to respond to a presentation of a pattern belonging to a particular input group. When the training was done, one set of patterns was randomly left out (say [AH, BH, CH, DH, EH, FH, GH]). After training, to check the ability of the perceptron to identify a novel input pattern from the network response, we presented the pattern (say AH) to all the perceptrons and evaluated their response. Then using a winner-take-all strategy, the pattern was assigned to the group corresponding to the perceptron that shows highest activation (In the example, this should be A*). This was repeated for other left out patterns (BH, CH etc, 45 samples each) and the classification accuracy for each of these were noted. If 80% of the samples of a pattern were correctly grouped, we say that the perceptron layer was able to identify the novel group correctly. The number of such groups was noted. This was repeated by leaving out other set of patterns (say [AG, BG, CG…], [AF, BF, CF…]) and similar analysis was performed. Figure 4b presents an average number of correct groups thus identified by each culture.
The fact that the output generated by a pattern AH was grouped into [A*] group instead of [H*] group indicates that the network response is strongly influenced by the first stimulus in the sequence rather than a co-occurrence of A & H. Together with the results that each pattern is distinct (56 patterns were linearly separable and patterns within each group were linearly separable) but also can be grouped (A*, B*..) while being able to classify a novel pattern, shows the ability of coarse-coded conjunctive scheme in neuronal cultures to create unique descriptions suitable for pattern classification and pattern generalization. The ability of the perceptron to do this, shows that the network dynamics and resulting response is such that a neuron in the next layer is able to group inputs correctly.The results on the ability to correctly classify untrained patterns emphasizes that the classification ability is not just due to a mapping to high dimension and demonstrates a ‘context’ dependent response to the second input and shows the inherent network property to generate such responses.
To check if the coding is suitable for probabilistic connections between layers as in the brain, we made the connections between the output electrodes and decoder perceptrons probabilistic and evaluated the classification performance. We mimicked the possible connectivity in a neuronal architecture by randomly connecting a perceptron in the output layer to a fraction of output electrodes (Fig. 1). The performance was robust and degraded gracefully as number of connections were reduced (Fig. 4c). The result indicates that the code generated by coarse-conjunctive neurons is distributed enough to allow a neuron randomly connected to a set of neurons in this layer to learn an arbitrary linearly dependent function.
We created a random network model for a mechanistic description of the stimulus responses for spatio-temporal input patterns from neuronal cultures. We viewed the network as a two layer network with an input layer consisting of stimulated neurons and output layer consisting of neurons directly connected to input neurons. This allows us to view cultured network as a layered architecture. Such network structures are used in studying computational capabilities of neuronal systems, brain-inspired computational frameworks and artificial neural networks. We studied how our experimental setup could mimic these computational models. The membrane and synaptic time constants were constrained biologically. Crucially, this points to the possibility of studying computational properties and learning capabilities of biological layered networks using cultures grown on multi-electrode arrays and investigate if computations done using artificial neuronal networks and brain-inspired frameworks can be done using biological systems.
The model as shown in Fig. 5, had 120 neurons, with 80% being excitatory and the rest inhibitory. Each neuron was supposed to mimic an electrode and we expected this two layered architecture to explain the observed spiking probability. The probability of firing of a neuron was calculated as a sigmoid function of weighted sum of inputs coming into the neurons.
The connectivity between the neurons in the model was set using two methods.
In the first method, we had a global parameter p which defines probability of connection between any two neurons in the network. We tuned this parameter so that when the input patterns are applied, the output generated by the network has similar overall properties like being linearly separable, sequence dependence and grouping when paired stimuli are applied. We then analyzed the network generated this way for a number of connections received by each neuron to allow it to mimic the observed behavior of the biological network. This provided further validation of schema for computation using layered architectures with random connectivity.
In the next method, we estimated the functional connectivity and the connection weights between the input and output electrodes in the neuronal culture by fitting the model outputs for different paired stimuli to match the probability of firing of output electrodes in experiments using a combination of genetic algorithms and gradient descent. The genetic algorithm tuned whether or not a connection exists between neurons while the gradient descent tuned the connection weights. Using this approach, we had a network which had firing probabilities at different electrodes close to the experimental data. The validity of the model-fit was established by using the model to generate output vectors and analyzing them in the same way as experimental data. We then compared the connectivity in this network with that of the network generated by the first method to see whether the number of connections are similar. In the first method, the connection probability is used to manipulate the connectivity, while in the second method the experimental data is used to do so. Since both methods can now recreate the overall experimental results in simulation, we were more confident of the model network explaining the observed behavior and use connection probability as a parameter to further study how connection probability might affect network performance.
Using the functional connectivity so obtained, we got further insights about the structure of the network. Figure 6 shows a histogram of number connections between input and output electrodes for a randomly generated connectivity between electrodes and those estimated using fitting the model to the data. They are in agreement to an extent that on an average, an output neuron has a functional interaction with 3 input electrodes for the network in the culture. Also, the higher number of functional connections estimated when the delay between pairing is 0.5 ms indicates that for these networks the dominant cause for generating conjunctive neurons would be through overlap of EPSP’s from multiple inputs.
We then varied the number of connections between the neurons to see how it affects the classification performance. This is shown in Fig. 7 where the connectivity parameter (p) is varied and the classification and grouping abilities of the model network is studied in the same way as with the biological network. As expected, we found that a minimum degree of random connectivity is required for generating sufficient number of coarse-conjunctive neurons. Interestingly, with the parameter value at 0.1 where the model shows a 100% classification ability for 56 classes, the number of distinct groups possible was around 6 which was similar to the observation across multiple neuronal cultures as presented in Fig. 4.
Discussion
We discuss the relevance of the above findings in the context of computing mechanisms in the brain. Currently, it is not clear how the functional connectivity in the brain changes and to what degree, in order to learn to perform some action. Also learning of a precise weight at different layers in a neuronal network would be a difficult challenge without accurate feedback signals and would require many repetitions of the training as experienced by researchers working with deep learning networks. It is also not clear how the equivalent error correcting mechanisms would work in a biological neuronal network.
In the following discussion, we show that with an interplay of random connectivity and a layered structure, neuronal circuits can solve such problems without needing to learn a large number of synaptic weights. We show how our experiments and modeling studies support this hypothesis for neurons cultured on multi-electrode arrays.
Linear separability as a key intermediate step for problem solving
Identification of the correct features from the data and transforming the inputs to a linearly separable space has been established as a key intermediate step in problem solving in machine learning. The ‘kernel’ in the support vector machine based classification, ‘hidden layer’ in artificial neural networks, the ‘random network’ in LSM’s, all use this same principle. Once the problem has been thus translated, the required arbitrary function to be learned is a linear combination of these outputs by a single neuron obtained by tuning its input synaptic weights, without needing large scale modification of the preceding input network (Fig. 1). Such learning of a linear combination of inputs, has been shown to be theoretically possible for a biological neuron equipped with STDP mechanism32. Specifically, classification and pattern recognition can be seen as a special case of thresholding of these linearly combined outputs. In our study, using 56 input patterns we have shown that the output of the neuronal culture shows such a transformation property (Fig. 4a). The output of the culture, which encodes the input stimulus into a higher dimensional representation, are linearly separable via perceptrons, and learn functions like classification, grouping, sequence detection and novel pattern recognition (Fig. 4b). Previously we have shown that such a biological neuronal network in culture on multi-electrode arrays can translate linearly un-separable inputs to a high dimensional linearly separable space22.
Conjunctive neurons create linearly separable population codes
The generation of linearly separable population codes can be explained using the schema for a hypothetical network shown in Fig. 1. Experimentally, we show that each neuron in the randomly interconnected network shows a conjunctive code (Fig. 3b). This most likely arises out of pairing of excitatory and inhibitory pre-synaptic inputs when two electrodes in the array are stimulated within a time window (Fig. 3a). Both excitatory and inhibitory connections are required for the neurons to show both an increase and decrease in firing probability as a result of pairing. Such connections also allow the neurons to detect the order of firing. Our results on the neuronal culture show that single neurons receiving random connections show ‘conjunctive encoding’ which are sensitive to electrodes being stimulated, their timing and the order of pairing, akin to the ‘conjunctive neurons’ demonstrated in vivo. The additional observation of ‘disjunctive encoding’ suggests the presence of both excitation and inhibition and their importance in the generation of a variety of conjunctive neurons with arbitrary inputs (Fig. 3a). A linearly separable population code can emerge from a sufficient collection of such randomly connected neurons. This finding emerges from our analysis of the output data, by using random subsets of output electrode data for classification (Fig. 4). These results emphasize the importance and sufficiency of randomly connected neurons to create such population codes without needing any specific learning/training of these networks (Fig. 7).
Neurons show coarse conjunctive coding
The results shown in Fig. 3b indicate that single neurons can show a coarse conjunctive response, i.e., each neuron is responsive to pairing of multiple spatio-temporal inputs. The presence of coarse-conjunctive neurons has been shown in the mammalian brain and its importance and advantages have been highlighted in theoretical studies2 . Coarse-conjunctive codes makes the encoding of the inputs robust as schematically illustrated for hypothetical network in Fig. 1. With such a code, a larger number of patterns can be represented by the network without needing a conjunctive neuron for every feature in the input space (Fig. 1). A decoding neuron in the final layer (L3), only partially connected to such a population of coarse encoding neurons from preceding layers, can still have sufficient information for decoding. Our analysis with random connectivity between the neuronal culture and output layer perceptron demonstrates this to be true for neuronal cultures on MEA (Fig. 4c). Such a scheme is suitable for structured yet probabilistic connections as found in biological neuronal systems.
Coarse conjunctive encoding emerges out of random connections without specific learning
Distinct coarse-conjunctive neurons can emerge out of random connectivity between two layers in a network. Our modeling study inspired by our experimental data shows this to be true (Figs 5 and 6). Our analysis with the model also shows that a minimal connectivity is required for generation of such a code (Fig. 7). Conjunctivity arises due to firing of inputs within a time window and depends on the electrodes from which it receives connections, which can be random. The ability to detect the timing and order of firing depends on the inherent time delays in the circuit and the presence of excitatory and inhibitory connections. Neurons show coarse-conjunctive encoding as it receives inputs from more than two electrodes. Each neuron has a distinct coarse-conjunctive tuning curve due to the random nature of connections. A sufficient number of such connections create a set of neurons which can project the inputs into a high-dimensional linearly separable space.
The robust nature of the encoding allows the subsequent layer of neurons with partial connectivity to learn an arbitrary function
An intermediate layer receiving random connectivity from a previous layer generates a robust encoding using coarse conjunctive neurons. As a result of such a code, a perceptron, probabilistically connected to this layer is able to identify the input pattern or a group of inputs. By extension, a neuron equipped with STDP should be able to achieve the same. Significantly, to learn a new class, instead of a large-scale change to all the synaptic weights in the network, only the weights of a single target output neuron connected randomly to the preceding coarse conjunctive neurons, needs to be modified.
In conclusion, we have shown that random neuronal networks in a culture, generate coarse-conjunctive outputs and unique population codes that are linearly separable for different input sequences without any specific training of the culture. The findings have physiological relevance in giving us some preliminary understanding of how neuronal networks in the brain might sift through information and implicitly classify them intrinsically, via linearly separable, highly redundant, coarse conjunctive encodings of the input stimulus, without needing explicit training/learning at all functional layers during information flow. Such an encoding ability might have a great utilitarian role in simplifying the learning process by needing modification of only a few final neuronal layers, as opposed to the entire network. However, this conjecture requires further experimental analysis of neuronal recordings from the brain in vivo.
Acknowledgements
This work was supported by Ministry of Electronics and Information Technology, Government of India. We would like to thank Center for Nanoscience and Engineering at Indian Institute of Science Bangalore for support for carrying out this work. We also thank Robert Bosch Centre for Cyber-Physical Systems , at Indian Institute of Science Bangalore for support.
Author Contributions
Jude Baby George performed the experiments, analyzed the data and wrote the manuscript. Grace Mathew Abraham and Zubin Rashid cultured and maintained neuronal cultures and helped do the experiments. BharadwajAmrutur and SujitSikdar analyzed the results and conclusions and wrote the manuscript.
Competing Interests
The authors declare that they have no competing interests.
Footnotes
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Maass W, Natschlager T, Markram H. Real-time computing without stable states: A new framework for neural computation based on perturbations. Neural Computation. 2002;14:2531–2560. doi: 10.1162/089976602760407955. [DOI] [PubMed] [Google Scholar]
- 2.O’Reilly, R. C. & Rudy, J. W. Conjunctive representations in learning and memory: principles of cortical and hippocampal function. Psychological Review108, 311–345 (2001). [DOI] [PubMed]
- 3.Deadwyler SA, Berger TW, Opris I, Song D, Hampson RE. Neurons and networks organizing and sequencing memories. Brain Research. 2015;1621:335–344. doi: 10.1016/j.brainres.2014.12.037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Eliasmith C, et al. A large-scale model of the functioning brain. Science. 2012;338:1202–1205. doi: 10.1126/science.1225266. [DOI] [PubMed] [Google Scholar]
- 5.Benjamin BV, et al. Neurogrid: A mixed-analog-digital multichip system for large-scale neural simulations. Proceedings of the IEEE. 2014;102(5):699–716. doi: 10.1109/JPROC.2014.2313565. [DOI] [Google Scholar]
- 6.Merolla PA, et al. A million Spiking-neuron integrated circuit with a scalable communication network and interface. Science. 2014;345:668–673. doi: 10.1126/science.1254642. [DOI] [PubMed] [Google Scholar]
- 7.Morris AM, Weeden CS, Churchwell JC, Kesner RP. The role of the dentate gyrus in the formation of contextual representations. Hippocampus. 2013;23:162–168. doi: 10.1002/hipo.22078. [DOI] [PubMed] [Google Scholar]
- 8.Save, E., Poucet, B., Foreman, N. & Buhot, M. C. Object exploration and reactions to spatial and non-spatial changes in hooded rats following damage to parietal cortex or hippocampal formation. Behavioral Neuroscience, 106, 447–456 (1992). [PubMed]
- 9.Young MP, Yamane Y. Sparse population coding of faces in the infero-temporal cortex. Science. 1992;256:1327–1331. doi: 10.1126/science.1598577. [DOI] [PubMed] [Google Scholar]
- 10.Deadwyler SA, Bunn T, Hampson RE. Hippocampal ensemble activity during spatial delayed-nonmatch-to-sample performance in rats. The Journal of Neuroscience. 1996;16:354–372. doi: 10.1523/JNEUROSCI.16-01-00354.1996. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Sargolini F, et al. Conjunctive representation of position, direction, and velocity in entorhinal cortex. Science. 2006;312:758–762. doi: 10.1126/science.1125572. [DOI] [PubMed] [Google Scholar]
- 12.Cong, B., Encoding neural networks to compute the atmospheric points spread function. In Proceedings International Conference on Information Technology: Coding and Computing (Cat. No.PR00540), 344–349 (2000).
- 13.Aso, Y et al. The neuronal architecture of the mushroom body provides a logic for associative learning. eLife, 10.7554/eLife.04577 (2014). [DOI] [PMC free article] [PubMed]
- 14.Ueoka Y, Hiroi M, Abe T, Tabata T. Suppression of a single pair of mushroom body output neurons in Drosophila triggers aversive associations. FEBS Open Bio. 2017;7:562–576. doi: 10.1002/2211-5463.12203. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Tateno T, Jimbo Y. Activity-dependent enhancement in the reliability of correlated spike timings in cultured cortical neurons. Biological Cybernetics 80. 1999;1:45–55. doi: 10.1007/s004220050503. [DOI] [PubMed] [Google Scholar]
- 16.DeMarse, T. B., & Dockendorf, K. P. Adaptive flight control with living neuronal networks on microelectrode arrays, IEEE Joint Conference on Neural Networks, 10.1109/IJCNN.2005.1556108 (2005).
- 17.Novellino, A. et al. Connecting neurons to a mobile robot: An in vitro bidirectional neural interface, Computational Intelligence and Neuroscience10.1155/2007/12725 (2007). [DOI] [PMC free article] [PubMed]
- 18.Bakkum, J. D., Chao, Z. C. & Potter, S. M. Spatio-temporal electrical stimuli shape behavior of an embodied cortical network in a goal-directed learning task. Journal of Neural Engineering5, 310–323 (2008). [DOI] [PMC free article] [PubMed]
- 19.Dockendorf KP, Park I, He P, Príncipe JC, DeMarse TB. Liquid state machines and cultured cortical networks: The separation property. Biosystems. 2009;95:90–97. doi: 10.1016/j.biosystems.2008.08.001. [DOI] [PubMed] [Google Scholar]
- 20.Ortman RL, Venayagamoorthy K, Potter SM. Input separability in living liquid state machines. Proceedings of the 10th international conference on Adaptive and natural computing algorithms. 2011;1:220–229. doi: 10.1007/978-3-642-20282-7_23. [DOI] [Google Scholar]
- 21.Isomura, T., Kotani, K. & Jimbo, Y. Cultured Cortical Neurons Can Perform Blind Source Separation According to the Free-Energy Principle. PLoSComputBiol, 1110.1155/2007/12725 (2015). [DOI] [PMC free article] [PubMed]
- 22.George JB, et al. Input coding for neuro-electronics hybrid systems. Biosystems. 2014;126:1–11. doi: 10.1016/j.biosystems.2014.08.002. [DOI] [PubMed] [Google Scholar]
- 23.George, J. B., Abraham, G. M., Amrutur, B. &Sikdar S. K., Robot navigation using neuro-electronic hybrid systems robot navigation using neuro-electronic hybrid systems. 28th International Conference on VLSI Design, 93–95 (2015).
- 24.Niedringhaus M, Chen X, Conant K, Dzakpasu R. Synaptic potentiation facilitates memory-like attractor dynamics in cultured in vitro hippocampal networks. PLoS One. 2013;8:e57144. doi: 10.1371/journal.pone.0057144. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Fong, M., Newman, J. P., Potter, S. M., and Wenner, P. M. Upward synaptic scaling is dependent on neurotransmission rather than spiking, Nature Communications 6 (2015). [DOI] [PMC free article] [PubMed]
- 26.Li Y, Chen X, Dzakpasu R, Conant K. Dopamine dependent effects on basal and glutamate stimulated network dynamics in cultured hippocampal neurons. J Neurochem. 2016;140:550–560. doi: 10.1111/jnc.13915. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Dimitris X, et al. Revealing ensemble state transition patterns in multi-electrode neuronal recordings using hidden markov models. IEEE Transactions on Neural Systems and Rehabilitation Engineering. 2011;19:345–355. doi: 10.1109/TNSRE.2011.2157360. [DOI] [PubMed] [Google Scholar]
- 28.Potter SM, Demarse TB. A new approach to neural cell culture for long-term studies. Journal of Neuroscience Methods. 2001;110:17–24. doi: 10.1016/S0165-0270(01)00412-5. [DOI] [PubMed] [Google Scholar]
- 29.Wagenaar DA, Pine J, Potter SM. Effective parameters for stimulation of dissociated cultures using multi-electrode arrays. Journal of Neuroscience Methods. 2004;138:27–37. doi: 10.1016/j.jneumeth.2004.03.005. [DOI] [PubMed] [Google Scholar]
- 30.Widrow, B. & Hoff, M. E. Adaptive switching circuits. IRE WESCON Convention Record, New York, 4, 96–104 (1960).
- 31.Tu JV. Advantages and disadvantages of using artificial neural networks versus logistic regression for predicting medical outcomes. Journal of Clinical Epidemiology. 1996;49:1225–1231. doi: 10.1016/S0895-4356(96)00002-9. [DOI] [PubMed] [Google Scholar]
- 32.Legenstein RA, Naeger C, Maass W. What Can a Neuron Learn with Spike-Timing-Dependent Plasticity? Neural Computation. 2005;17:2337–2382. doi: 10.1162/0899766054796888. [DOI] [PubMed] [Google Scholar]