Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2018 Jan 1.
Published in final edited form as: Lang Cogn Neurosci. 2016 Aug 9;32(3):274–285. doi: 10.1080/23273798.2016.1218033

A null model for cortical representations with grandmothers galore

Gabriel Kreiman 1
PMCID: PMC5710804  NIHMSID: NIHMS893053  PMID: 29204455

Abstract

There has been extensive discussion in the literature about the extent to which cortical representations can be described as localist or distributed. Here we discuss a simple null model that encompasses a family of related architectures describing the transformation of signals throughout the parts of the visual system involved in object recognition. This family of models constitutes a rigorous first approximation to explain the neurophysiological properties of ventral visual cortex. This null model contains both distributed and local representations throughout the entire hierarchy of computations and the responses of individual units are meaningful and interpretable when encoding is adequately defined for each computational stage.


How does the brain represent information? A cartoonish answer was entertainingly and endearingly depicted in a recent Pixar film, Inside Out, where we get to peek inside the mind of a young girl, and find five characters, each representing basic emotions: Joy, Sadness, Fear, Disgust and Anger. The notion of a little character inside the brain to represent each emotion could easily be extended by adding more and more characters to represent other sensations, thoughts, actions, decisions, even grandmothers. There could be one character that gets activated whenever we see a chair, another one for grandma, another one whenever we listen to Beethoven, another one whenever we are hungry and another one when we raise our right arm. Despite the seemingly naïve nature of this scheme, after replacing “characters” by “neurons”, this idea seems to be at the heart of how lay people and many Cognitive Scientists and Neuroscientists think about brain representations.

I argue here that this cartoon version of cortical representations is neither completely wrong nor completely right. We need to dig deeper into what we mean by representations and translate those definitions into the biophysical language of neurons. What emerges is an even simpler, and yet far more wonderful and elegant biologically plausible description of how brains store information, a null model for cortical representations. I will focus the discussion here on visual information, specifically on how we see and recognize shapes, objects and faces. I restrict the discussion to visual objects because we know more about the visual system and the transformation of inputs along the visual hierarchy than about any other modality or any other aspect of cognition. This should not be interpreted to imply that we fully understand visual object processing; there remain fundamental unanswered questions the field. Despite these lacunae, we are beginning to develop biologically inspired computational models that provide a reasonable approximation and first sketch of visual processing.

I will argue that the term “grandmother cell” and several derivatives and synonyms have not been well defined (Gross, 2002). In the Cognitive Science literature, the question about information representation has elicited significant discussion between so called “localist” representations (McClelland & Rumelhart, 1981) and “distributed” representations (McClelland, Rumelhart, & Group, 1986); see (Bowers, 2009) for a clear and detailed review and discussion of the literature. It seems clear to most investigators that there is no one-to-one map between a single neuron and perception (i.e. it cannot be the case that there is only one neuron in the entire brain that responds to a given photograph of grandma and it cannot be the case that a given neuron responds only to one particular photograph of grandma and to no other possible photograph of grandma or other stimuli). Once we abandon the idea of a one-to-one map, one formulation of the question about how information is represented in the brain asks whether single neurons store “meaningful things” or not. For example, it could be argued that the word grandmother constitutes a meaningful thing and we can ask whether there are neurons that respond selectively to grandmother and not to other stimuli. Toy computational models have been built where there are units that represent letters or words. Yet, when discussing real brains, this apparently simple question is not sufficiently well defined. Do we mean to imply that there is a single neuron in the entire brain that responds to grandmother? What do we mean by “neuronal response”? How selective does the neuron need to be in this definition? Do we mean to imply that the neuron(s) would respond to grandmother above baseline and to no other possible stimulus in the world? Would the neuron(s) respond to the word “grandma” and the misspelled word “grandmotter”? Would the neuron respond to a picture of grandma, to any picture of grandma, to any grandma, to grandma's clothes, to her gait, her voice, her accent, etc.? The universe of related questions is infinite and so is the list of discussions that can be generated around such poor definitions. Instead, I will provide a brief, and admittedly incomplete, description of the cascade of processes that lead to representing visual information, arguing that this involves both implicit and explicit representations and that ultimately defining these operations is the crux of the problem to understand cortical representations.

Basic tenets of representation

The basic requirements for representing information in the brain have been lucidly articulated by several authors before (e.g. (Barlow, 1972; Crick & Koch, 2003; Parker & Newsome, 1998) among others). To constrain the discussion here, we consider an experiment where a visual stimulus is presented (e.g. a picture of grandma), the subject reports his/her perception (e.g. indicating yes/no whether he/she recognized grandma), and we scrutinize the responses of multiple neurons. I will re-state the main postulates by copying, expanding, rephrasing and adding examples to the ideas in Parker and Newsome (Parker & Newsome, 1998), to define the basic tenets of representation of visual information in cortex:

  1. Reflection of perception. The neuronal responses elicited by presentation of the stimulus should be directly comparable to perception and behavior as reported by the subject in terms of magnitude, timing, duration and specificity. Even though we obviously cannot see grandma without a retina, this tenet implies that the retinal ganglion cells are not part of the explicit visual representation of grandma.

  2. Timing. The neuronal responses should occur before or at the moment of perception (and not afterwards). This is a basic and fundamental definition of causality and has implications for how neuronal responses are defined, analyzed and interpreted. Visual recognition of an object flashed on the screen takes place within approximately 150 ms of stimulus onset (Kirchner & Thorpe, 2006; Potter & Levy, 1969; Thorpe, Fize, & Marlot, 1996) and therefore the neuronal responses should occur within this time frame.

  3. Selectivity. The neurons in question should signal relevant and selective information when and only when the organism perceives the stimulus (in the scenario described here, a two-alternative forced choice discrimination of whether grandma is present). There are abundant studies documenting selective responses to stimuli throughout visual cortex, e.g. in primary visual cortex (Hubel & Wiesel, 1962), area MT (Baker, Petersen, Newsome, & Allman, 1981) and inferior temporal cortex (Desimone, Albright, Gross, & Bruce, 1984). Throughout the brain, neurons fire spontaneously even in the absence of their preferred stimuli. Here we focus on elevated responses beyond this baseline firing.

  4. Stability. The neuronal representation should be stable over prolonged periods of time. The representation of grandma should be the same today and a month from now. This is not to say that the brain cannot acquire or forget information. Plasticity is a fundamental property of cortical circuits but there should exist a stable representation of information in cortex. Recent studies have shown that the representation of visually selective information can be very stable, e.g. (Bansal et al., 2012; Bondar, Leopold, Richmond, Victor, & Logothesis, 2009; McMahon, Jones, Bondar, & Leopold, 2014; Tolias et al., 2007).

  5. Single trial interpretability. Differences in the firing patterns of the candidate neurons (or a subset thereof) to different external stimuli should be sufficiently reliable in a statistical sense to account for, and be reconciled with, the precision of the organism's responses in single trials. In other words, we should be able to reliably read out information about the stimulus from the neuronal responses, in single trials, and with a linear read-out mechanism. The imposition of linearity here is important; the retina contains all the information about the picture of grandma but it cannot be linearly read out. Machine learning techniques have been extensively used to extract selective visual information in single trials, e.g. (Hung, Kreiman, Poggio, & DiCarlo, 2005). The imposition of single trials interpretability is critical: the brain cannot average across trials to make a decision (the question of noise in neuronal responses is further elaborated upon in the Discussion).

  6. Predictive of perceptual fluctuations. Fluctuations in the firing of some set of the candidate neurons to the repeated presentation of identical external stimuli should be predictive of the observer's judgment on individual stimulus presentations including errors and other behavioral changes, e.g. (Newsome, Britten, & Movshon, 1989). For example, in the context of bistable perception, an observer's subjective report for the presence or absence of the stimulus should correlate with the neuronal responses, as observed in inferior temporal cortex (Sheinberg & Logothetis, 1997) but not in primary visual cortex (Leopold & Logothetis, 1996).

  7. Invariance. The representation should be robust to certain transformations of the input. For example, in the context of recognizing an object, recognition is essentially unaffected by a wide range of changes in the object's size, illumination, rotation and other transformations. Robustness is not complete: it is possible to disrupt recognition if the image transformation is sufficiently large. Yet, recognition is remarkably robust to image changes and the neuronal representation should reflect this degree of invariance (e.g. (Deco & Rolls, 2004; DiCarlo, Zoccolan, & Rust, 2012; Ito, Tamura, Fujita, & Tanaka, 1995; Riesenhuber & Poggio, 1999; Sary, Vogels, & Orban, 1993)).

  8. Task independence. The firing patterns of the neurons in question should not be affected by how behavior is reported. Grandma is grandma, regardless of whether we ask subjects to indicate yes/no by pressing buttons, to verbally report her name or to passively view the picture. There are of course plenty of neurons in cortex that are strongly modulated by the task demands, but there should exist a stable representation that is task-independent. For example, to a first approximation, neurons in inferior temporal cortex carry information about object shapes that is conveyed to pre-frontal cortex where investigators have described strong modulation depending on the task demands (Freedman, Riesenhuber, Poggio, & Miller, 2001; Meyers, Freedman, Kreiman, Miller, & Poggio, 2008). Task demands can modulate the representation (e.g. visual attention can significantly enhance neuronal activity throughout visual cortex (Reynolds & Chelazzi, 2004)) but it should still be possible to decode the information about grandma in a task-invariant manner.

  9. Susceptibility to stimulation. Direct interference with the firing patterns of some set of the candidate neurons (e.g. by electrical, chemical or optogenetic stimulation) should lead to measurable changes at the perceptual level. These perceptual changes should be contingent on the spatial and temporal specificity of the external manipulation. Current injection has been shown to bias behavioral responses in monkeys in a content-specific fashion in different parts of visual cortex, e.g. (Afraz, Kiani, & Esteky, 2006; Salzman, Britten, & Newsome, 1990).

  10. Susceptibility to lesions. Susceptible to Temporary or permanent removal of all or part of the candidate set of neurons should lead to a measurable perceptual deficit, e.g. (Afraz, Boyden, & DiCarlo, 2015; Dean, 1976; Holmes, 1918).

Starting from the very beginning

When photons impinge on the eyes, retinal ganglion cells (RGC) respond vigorously when there is a change in luminosity within their receptive fields. To get right to the point: is this a grandmother-like representation or a distributed one? The question is never even discussed among retinal physiologists and hardly makes sense. We have a rather elaborate understanding of how photons are transformed into electrical signals and we are beginning to elucidate the circuitry that leads to RGC firing at a very fine level of detail (e.g. (Field et al., 2010)). The picture of grandma is implicitly represented by the activity of a large number of RGCs; one could rephrase this statement to argue that the information about a complex visual stimulus such as grandma is distributed over multiple RGCs. At the same time, the representation of light changes in a specific location is explicitly represented by RGCs. The RGCs therefore behave like grandmother cells in terms of explicitly representing changes in luminosity within their receptive field. Coarsely speaking, we can represent the firing rate r of a given RGC as r=g(i=1Nwixi), where g is a non-linearity (e.g. a threshold), wi denote the synaptic weights of the N units that project onto the RGC, and xi denote the activities of the N inputs. Although admittedly oversimplified, this formulation forms the foundation of the basic intuition of how neurons integrate information and remains at the heart of models of network activity (Dayan & Abbott, 2001).

RGCs project to neurons in the lateral geniculate nucleus (LGN), and these cells in turn project to primary visual cortex (V1). Neurons in primary visual cortex respond vigorously and selectively to oriented bars presented within their receptive field. Neurons in primary visual cortex do not respond uniquely to a single orientation. A neuron that shows maximal response to a bar of, say, 45-degree orientation will also respond to bars of 44-degree and 46-degree orientation. There is a tuning curve with a certain width that describes how selective the neuron is for different orientations. Investigators have described simple and complex cells in V1, the latter pooling information from multiple simple cells to achieve robustness: whereas the responses of a simple cell are sensitive to the exact location of the oriented bar within its receptive field, a complex cell will yield approximately the same response when the bar is translated to different positions within the receptive field. Again, in the study of primary visual cortex the discussion about grandmother cell-like representations is notably absent. There are multiple computational models that provide a reasonable account of how orientation tuning emerges in V1, e.g. (Carandini et al., 2005; Hubel & Wiesel, 1962; Olshausen & Field, 1996; Priebe & Ferster, 2012). The fundamental questions to completely understand the responses of neurons in layer 4 of primary visual cortex involve fully elucidating the activity of the relevant inputs (xi in the formulation above), the strength of the relevant connections (wi in the formulation above) and the specific nature of the computations including thresholds, time constants, synaptic adaptation, synaptic depression and dendritic non-linearities. Information about grandma remains distributed over a large number of V1 cells. We cannot linearly decode the presence of grandma from an ensemble of V1 neurons. However, the presence of an oriented bar at a specific location can be linearly and robustly decoded in single trials from the activity of a small number of V1 neurons.

In sum, in the pathway from the retina to V1, information is represented both in a distributed way and in a grandmother-like way. If we define neuronal response as an enhanced firing rate beyond baseline firing and we define a “meaningful thing” as a stimulus with the right contrast, color, orientation and spatial frequency localized within the receptive field of a given V1 neuron, then this neuron behaves like a grandmother cell. The representation is redundant because there are multiple similar neurons and is achieved by virtue of integrating the activity over a distributed representation of the same stimulus at the level of the LGN.

The ventral visual cortex

Ascending through the visual hierarchy, we have a decent idea of the structures and anatomical pathways that transform visual information along the so-called what pathway involved in object recognition in macaque monkeys (Connor, Brincat, & Pasupathy, 2007; Felleman & Van Essen, 1991; Markov et al., 2012). Yet, we understand much less about the details of the computations that transform information from one visual area to the next.

The ventral visual stream can be described by an approximate hierarchy: V1 neurons project to area V2, V2 neurons in turn project to area V4, V4 neurons project to inferior temporal cortex (ITC). In addition to these projections there are abundant horizontal connections within each area and top-down connections (e.g. V2 projects back to V1). Moving from V1 to ITC, there is a progressive increase in the size of the receptive fields, in the degree of complexity of the type of features represented at each stage and also in the degree of tolerance to image transformations. For example, some neurons in V4 are particularly sensitive to the extent and type of curvature of the visual stimulus (Pasupathy & Connor, 2002). Although the specific circuits that give rise to selectivity along the ventral stream are not clearly understood, it is possible to approximate the type of selective responses to different shapes using the same type of ideas articulated above by non-linearly pooling inputs from simpler responses (Cadieu et al., 2007).

Inferior temporal cortex constitutes a vast expanse of the ventral stream. Lesions in inferior temporal cortex typically lead to specific deficits in visual object recognition (e.g. (Afraz, Boyden, & DiCarlo, 2015; Benton & Wav Allen, 1972; Dean, 1976)). Neurons in inferior temporal cortex respond to complex shapes including abstract ones like folded paperclips or fractal patterns and real world objects including faces (Desimone, Albright, Gross, & Bruce, 1984; Logothetis, Pauls, & Poggio, 1995; Logothetis & Sheinberg, 1996; Naya, Sakai, & Miyashita, 1996; Tanaka, 1996; Tsao, Freiwald, Tootell, & Livingstone, 2006). Their responses also show a significant degree of robustness to a large number of stimulus transformations including changes in preferred stimulus position within the receptive field, scale, illumination, rotation, and many others (Ito, Tamura, Fujita, & Tanaka, 1995; Logothetis, Pauls, & Poggio, 1995; Sary, Vogels, & Orban, 1993; Tovee, Rolls, & Azzopardi, 1994). It is possible to linearly decode in single trials the identity of objects from the activity of small ensembles of ITC neurons (Hung, Kreiman, Poggio, & DiCarlo, 2005).

Consider an ITC neuron that responds selectively to a complex shape, say a face. To be clear, what we mean in practice is that the investigator records the activity of this neuron and counts the number of spikes in a given time window in response to presentation of a battery of multiple stimuli, say on the order of several hundred different shapes. The stimuli that contain faces elicit not only a response above baseline but also a response that is statistically stronger than the responses to the other stimuli in the set. If we were able to trace the input connections of this ITC neuron, and one day in the not too distant future we may be able to do this, we would likely observe that the inputs are distributed over multiple V4 cells. As a null hypothesis, we imagine the transformation of V4-like responses into ITC-like responses to obey the same principles that govern the transformation of multiple LGN inputs converging to give rise to orientation selectivity in V1. In other words, at each stage of processing, there is a transformation from a distributed representation to a grandmother-like representation. Each stage of processing thus connotes a distributed representation and also a localist one, depending on what particular aspects of the stimulus we are considering. Many of the discussions about localist versus distributed representations in the brain have centered on high-level “meaningful things” such as grandmothers, words, or the concept of love or fear. Unfortunately, for most of these high-level things, we still do not really understand the degree of selectivity and invariance of the relevant neurons, let alone the inputs that can give rise to such representations. In the context of visual information, the simplest null model is that we can build a representation of progressively more complex shapes by adequate pooling and filtering signals from the previous processing stage. This leads to an elegant null model for cortical representations, with grandmothers galore as long as we define what is represented at each stage and also with clearly distributed representations.

Computational models of visual recognition

The brief qualitative description of the transformation of visual inputs in the previous section has been formalized in the form of a theoretical framework and quantitatively instantiated into algorithms for visual recognition (DiCarlo, Zoccolan, & Rust, 2012; Fukushima, 1980; Mel, 1997; Olshausen, Anderson, & Van Essen, 1993; Perret, Oram, Hietanen, & Benson, 1994; Riesenhuber & Poggio, 1999; Rolls, 1991; Serre et al., 2007). These biologically inspired algorithms are characterized by a bottom-up hierarchy of linear and non-linear computations that start at the pixel level and progressively build a complex representation of the visual inputs. Ascending through the visual hierarchy, units display larger receptive fields, selectivity to more complex visual features and increased tolerance to changes in those features. The specific implementation details vary across different models but the general principles are the ones articulated in the previous section: units pooling information from previous layers, weighted linear sums followed by thresholding, non-linearities that give rise to tolerance. As a reasonable first approximation, these models assume that “cortex is cortex” and hence that the same mathematical operations are performed at each level (except that each level operates on different inputs and therefore conveys different outputs).

There have been multiple implementations of this type of network architecture (e.g. (Miconi, Groomes, & Kreiman, 2015; Riesenhuber & Poggio, 1999; Serre et al., 2007; Wyatte, Curran, & O'Reilly, 2012; Yamins et al., 2014) among many others). These architectures also constitute the basic foundation of deep convolutional networks, which have been successful in many engineering applications, most notably in computer vision (e.g. (Krizhevsky, Sutskever, & Hinton, 2012)).

This family of models constitutes only a first order approximation to the complexities of the cortical circuitry; for example, most of these models involve exclusively feed-forward projections and it is well known that there are abundant horizontal and feedback connections throughout neocortex (Douglas & Martin, 2004). Despite this and many other simplifications, these models have been shown to capture basic properties of neuronal responses throughout the ventral visual stream (Riesenhuber & Poggio, 1999; Serre et al., 2007).

An attractive feature of computational models is that we can inspect the activity of every unit and we can explain with perfect accuracy all the inputs and outputs to any possible stimulus. The question of localist versus distributed representations has been discussed before in the context of certain computational models (e.g. (Bowers, 2009; McClelland & Rumelhart, 1981; McClelland, Rumelhart, & Group, 1986)) but those computational models have not been directly linked to neurophysiological measurements. Is information in deep hierarchical architectures for visual recognition represented in a localist manner or a distributed one? Each unit in these models is selective and represents a meaningful thing. In the typical computational implementation of these models, every unit shows selectivity. The biological implementation could differ substantially: there is a large number of different types of neurons in every patch of cortex and we still lack a fundamental understanding of the functional properties of these neuronal types and whether all of them show the same type and degree of selectivity. Each unit is part of a small group of units (compared to the total number of units in each area) that codes for that meaningful thing, we can precisely pinpoint what that meaningful thing is for each unit, injecting current into those units leads to an enhanced representation of the features they encode and removing those units from the model leads to a visual scotoma for those features. In other words, every unit in the model is a grandmother unit, as long as we define grandmother in terms of the unit's specific preferences. At the same time, each unit is part of a large ensemble that projects onto the next stage to build a more complex feature. There is no real dichotomy between localist and distributed representations in these models. All the computations are clearly defined in terms of a few well-defined rules, a simple, yet beautiful null model for cortical representations.

Discussion

It does not make sense to discuss distributed versus localist representations in the brain without defining clearly the identity of the neurons in question and the specific type of information that is represented. Accepting the family of models succinctly described above as a plausible first-order description for the functions of ventral visual cortex, one could argue that we have only described the representation of visual features that range from luminance changes to oriented bars to more complex shapes but we have not specifically described how to represent chairs, dogs, grandmothers and other objects. However, these models provide a very explicit description of how to represent those shapes. Indeed, those models are routinely tested in object recognition tasks that involve dozens to millions such objects (e.g. (Krizhevsky, Sutskever, & Hinton, 2012; Serre et al., 2007)). In a typical scenario, a set of training images is presented to the models to simulate the activity of every unit in response to every image. Next, a machine learning approach is used to train the weights of a classifier to map the activity of units in one or more layers of the model to the image labels (e.g. label “grandmother” for the corresponding set of pictures that include grandma). Finally, the model is tested with a different set of images to evaluate how well it can label those new images. In many cases, a linear kernel is used for the classifier. In this case, a given linear classifier unit takes the distributed input from several units in an earlier layer, multiplies the activity of those units by suitable weights and thresholds the results, finally deciding in a binary way whether the image should be labeled “grandmother” or not. These classifier units are localist par excellence; the inputs to those units contain a clearly distributed representation of visual information. This process is repeated once and again throughout the hierarchical model.

While in some instantiations of these models there may be a single classifier unit for each label, in the brain there is a significant degree of redundancy (Barlow, 1972). The notion of a localist representation does not depend on having a one-to-one map between units and meaningful things. This is a typical misconception of the definition of grandmother cells that has been cleared in multiple earlier discussions. In the original parable by Lettvin, there were 18,000 such cells that responded uniquely to “mother” (Gross, 2002). At no stage of processing do we have only one neuron representing information. There are multiple RGCs that represent luminance changes in a given point in space, multiple V1 neurons that respond to a certain orientation in a given location, and multiple ITC neurons that represent complex visual features at a given location.

There are many important nuances to the representation of information in the brain that deserve further discussion. Once we allow for the notion established in the previous paragraph that the map between neurons and meaningful things is not one-to-one, we may ask how many neurons are involved in representing each feature. Neurophysiological recordings typically show that representations tend to be sparse. Sparseness can be loosely defined as using only a small number of neurons out of the entire set of possible neurons to represent a particular feature (for a more formal definition, see equation 4 in (Olshausen & Field, 1996)). Sparseness can also be loosely defined by the related but distinct notion that a neuron responds to only a small fraction of all possible stimuli. In computational models, introducing sparseness as a constraint can alter the resulting representation and lead to the development of biologically plausible feature tuning properties. For example, there has been considerable discussion about the idea that information is sparsely represented in primary visual cortex (e.g. (Olshausen & Field, 1996)). Sparseness is an attractive property because it leads to a more efficient representation that is easier to decode in later stages by virtue of the independence of the inputs, and it may also have energetic efficiency implications (Laughlin, van Steveninck, & Anderson, 1998). Some investigators draw a sharp distinction between localist representations, sparse distributed representations and dense distributed representations (e.g. (Bowers, 2009)). In the null model outlined here, the distinction is merely a quantitative one. Each visual feature at each location is represented by a relatively small number of units (but more than one unit).

Two essential causal manipulations provide evidence that we are in the right track towards understanding the representation of visual information: disturbing the representation through small lesions or by injecting currents into local circuits. Lesions in V1 lead to localized scotomas in the visual field and lesions in higher visual areas lead to specific visual deficits such as the inability to detect color, motion or complex shapes (e.g. (Dean, 1976; Zeki, 1990, 1991)). Eliminating a single neuron is unlikely to lead to clear behavioral manifestations, presumably due to the robustness and redundancy in the representation. We currently do not have the technical ability to selectively lesion or inactivate all the neurons that respond to a specific feature. Yet, a recent study demonstrated that optogenetic silencing of a relatively small group of neurons in a patch of cortex containing neurons selective for face gender can lead to a small but significant impairment in gender discrimination (Afraz, Boyden, & DiCarlo, 2015). Advocates of distributed representations would be right to point out that a large number of neurons is removed in these lesion studies; future technological developments will allow a tighter titration of the behavioral consequences of eliminating specific types and different numbers of neurons from cortical circuits. The specificity of the ensuing effects in lesion studies supports the notion that there is a localized, interpretable representation of visual information in cortical circuits. This specificity in the consequences of local lesions is also consistent with the type of effects one would expect from silencing local groups of units in the computational models outlined in the previous section.

Another important causal manipulation that can shed light on the nature of the neural representations involves examining the effects of electrical stimulation. Injecting current through high-impedance microelectrodes placed extracellularly probably alters the activity of hundreds to several thousand neurons in the vicinity of the electrode, depending on the type of electrode, the current intensity, current injection pattern, brain area and tissue excitability (Logothetis et al., 2010; Tehovnik, 1996), in addition to indirect effects on other brain areas. Several studies have demonstrated that perceptual and behavioral effects can be elicited via current injection in this fashion. Stimulating V1 leads to perception of localized phosphenes (e.g. (Brindley & Donaldson, 1972; Dobelle & Mladejovsky, 1974; Tehovnik, Slocum, Smirnakis, & Tolias, 2009)). Similarly, stimulating other parts of visual cortex can lead to specific visually-triggered behaviors including motion detection enhancement upon stimulating area MT (Salzman, Britten, & Newsome, 1990) or face detection enhancement upon stimulating inferior temporal cortex (Afraz, Kiani, & Esteky, 2006). Recently, a series of elegant studies have advanced the possibility of stimulating individual neurons, albeit not in the visual system (e.g. (Brecht, Schneider, Sakmann, & Margrie, 2004), reviewed in (Doron & Brecht, 2015)). Rather surprisingly, these studies have demonstrated that stimulating single neurons can elicit both sensory percepts as well as motor outputs. For example, adding approximately 15 action potentials to baseline activity in the rat barrel cortex, which is involved in the representation of whisker tactile stimulation, led to significant and specific behavioral reports. The results of this type of experiment are typically weak and show a significant degree of variability. Furthermore, the mechanisms of cortical propagation in these so-called nanostimulation experiments are poorly understood; it has been suggested that the behavioral manifestations arise as a consequence of amplification via recurrent connections. Despite these considerations, it is difficult to argue that the activity of individual neurons is uninterpretable given the results of these micro and nanostimulation studies. Rather, the stimulation studies lend further support to the notion of local circuits with interpretable responses that can be activated via stimulation to elicit specific perceptual sensations or motor outputs. The results of these stimulation experiments are also consistent with the consequences of activating local groups of units in the type of computational models outlined in the previous section.

Information from the top layers of the ventral visual stream is conveyed to multiple brain areas, most notably the medial temporal lobe and frontal cortex. Neurophysiological recordings in the human medial temporal lobe have described a plethora of interesting selective responses to places (Ekstrom et al., 2003), object categories (Kreiman, Koch, & Fried, 2000b; Mormann et al., 2011) and specific individuals or landmarks (Quian Quiroga, Reddy, Kreiman, Koch, & Fried, 2005). These responses could be interpreted as additional examples of a localist representation for high-level information. However, as fascinating as those signals are, the neuronal responses in the medial temporal lobe structures do not fulfill all the requirements outlined above required for a representation of visual information, most notably the fact that removing those neurons does not lead to visual impairments (Squire, Stark, & Clark, 2004). Rather, the medial temporal lobe structures play a fundamental role in interpreting the degree of familiarity and novelty of current stimuli, associating multi-modal information in the context of emotional valence and prior knowledge, to form and retrieve memories (Brown & Aggleton, 2001; Cameron, Yashar, Wilson, & Fried, 2001; Eichenbaum, 2004; Gelbard-Sagiv, Mukamel, Harel, Malach, & Fried, 2008; Kreiman, 2007; Kreiman, Koch, & Fried, 2000a; Mormann et al., 2014; Rutishauser, Mamelak, & Schuman, 2006; Rutishauser, Schuman, & Mamelak, 2008; Squire, Stark, & Clark, 2004).

Information from the ventral visual cortex is not only conveyed to the medial temporal lobe but also to frontal cortex structures that play a critical role in interpreting the visual inputs in a task-dependent manner to make decisions leading to behavioral outputs (Miller & Cohen, 2001). Neurons in pre-frontal cortex show dynamic tuning to task-dependent variables. Recently, it has been argued that each neuron in pre-frontal cortex is tuned to multiple different aspects of an object sequence memory task rather than a single one (Rigotti et al., 2013). This so-called mixed-selectivity does not imply that the responses of individual units cannot be interpreted. This question goes back to the definition of “meaningful things”. If a neuron shows an increased firing rate during the presentation of a specific stimulus and also before a given motor response, it may be challenging to label the neuron as “purely visual” or “purely motor”. However, the inadequacy of these labels pre-defined by the investigators does not preclude from decoding the properties of neuronal responses. The neuron in question still shows a reproducible response that can be correlated to the sensory events, decisions and motor outputs during the task. The interpretational challenges arise only because of our attempt to force specific anthropomorphic descriptions based on language to the changes in firing rate. These language-based descriptions of neuronal preferences have been relatively successful in early sensory areas but it seems likely, and far more powerful, that we will require mathematically defined operations to explain the responses in higher parts of cognition. Even within the complex and fascinating realm of high-level cognitive information represented in frontal cortex, there is still hope for interpreting the activity of individual neurons.

The apparent “mixed” selectivity alluded to in the previous paragraph is not the only conceptual barrier to interpreting the activity of cortical neurons. Another important problem arises when we consider the response dynamics and the dependence on behavioral tasks. Neuronal responses throughout the visual system show complex temporal dynamics that evolve on scales of tens to hundreds of milliseconds (e.g. (Hung, Kreiman, Poggio, & DiCarlo, 2005; Meyers, Freedman, Kreiman, Miller, & Poggio, 2008; Richmond, Optican, & Spitzer, 1990; Ringach, Hawken, & Shapley, 1997; Smith, Majaj, & Movshon, 2005; Woloszyn & Sheinberg, 2009)). Additionally, the responses throughout visual cortex can be significantly modulated by task demands (e.g. (Bansal et al., 2014; Baylis & Rolls, 1987; Gilbert, Li, & Piech, 2009; Ramalingam, McManus, Li, & Gilbert, 2013; Rigotti et al., 2013; Vogels, Sary, & Orban, 1995; Woloszyn & Sheinberg, 2009)). The family of models described in the previous section does not really account for neural dynamics (other than the serial passing of signals from one layer to the next) nor does it clearly describe how the same neural representation can be used for multiple different tasks. Furthermore, if the interpretation of spike trains from a given neuron evolves over time due to internal dynamics and/or due to task demands, it is unclear how post-synaptic neurons can decode such time-varying inputs. Changes over long temporal scales (e.g. due to learning) may be easier to decipher since they may be accompanied by concomitant modifications in the synaptic weights, yet we need to be able to also understand changes in the meaning of spike trains that occur within tens to hundreds of milliseconds. In the extreme version, the same firing rate of a given neuron could convey distinct types of information at two different time points. Understanding these dynamics and task-dependent effects opens the doors to interesting questions both for localist and distributed representations and will require significant extensions to the simplified family of models described here.

A prominent example of dynamics in neural network models is illustrated in attractor networks. Attractor networks constitute particularly interesting computational models that are rather distinct form the family of bottom-up architectures described in the previous section (e.g. (Hopfield, 1982)). Briefly, units are interconnected in an all-to-all fashion and the state of the network is defined by the activity of all units. Each unit participates in representing every possible input and every input is represented by all units. In some sense, each unit in this type of attractor network is not representing any meaningful thing in and of itself. This seems to be a major departure from the hierarchical feed-forward models described here as a null model. Even though there has not been clear compelling biological evidence for or against the possibility of an attractor network type of neocortical circuits, given the dense interconnectivity of cortical circuits, it is not inconceivable that these ideas may help us better understand the dynamics of computations in cortex. Still, these networks typically do not explicitly specify how and when information is read out from these models. Brains need a way of converting the representation of visual information into decisions and actions (e.g. to indicate that a particular image does or does not contain a picture of grandma). We need an additional step to peek inside the attractor network, examine its state and determine whether that state corresponds to one of the stored patterns. The simplest way to think about how this read-out mechanism could be implemented with biological circuits is to wire all of the units in the network to another set of units that are capable of discriminating different states. These read-out units would behave as localist units deciphering the distributed the representation in the attractor network.

Another important issue that should be discussed is the apparent unreliability of neuronal responses. Upon repeated presentation of an identical stimulus, neurons seem to respond capriciously, and the variance in the spike counts can be as high as the mean spike count. The high degree of trial-to-trial variability in the spike trains elicited in response to a given stimulus has puzzled neuroscientists trying to understand how neurons encode information for decades (Perkel, Gerstein, & Moore, 1967). This degree of variability seems to pose a challenge to a localist representation where the activity of each neuron in each trial matters and should be interpretable. One possibility is that the activity of each neuron can be described as a localist representation only on average, but this seems like an artificial and convoluted argument; the brain must function in individual trials. A more attractive possibility is that the apparently high noise levels are a consequence of heterogeneity in the experimental conditions. Even when investigators think that the conditions are identical from one trial to the next, there exist multiple internal and small external variables that can be different. In support of this argument, cortical responses can show smaller degrees of trial-to-trial variability when there is better control of experimental conditions (e.g. (Bair & Koch, 1996; Churchland et al., 2010)). Furthermore, several studies examining responses in primary or early sensory neurons where there may be less fluctuation in internal conditions have demonstrated that variability can be very low, in some cases, as low as the theoretical limits imposed by the discrete nature of spikes (e.g. (Berry & Meister, 1998; Kreiman, Krahe, Metzner, Koch, & Gabbiani, 2000; van Steveninck, Lewen, Strong, Koberle, & Bialek, 1997)). Additionally, variability can also be significantly reduced in slice recordings (Holt, Softky, Koch, & Douglas, 1996; Mainen & Sejnowski, 1995; Stevens & Zador, 1998). In sum, under conditions where it is possible to exert better control over the experimental conditions, trial-to-trial variability is significantly reduced and the responses of individual neurons become more interpretable in single trials.

Even if we accept the rudimentary sketch proposed as a null model for the cortical representation of visual information, we still have very little to say about a myriad of high-level and abstract ideas such as how the brain encodes “love”, or “I would like to talk to grandma today”, or “I miss grandma's delicious food”. In a completely speculative and admittedly naïve tone, we can imagine that evolution works by duplicating, copying, improving and refining existing circuits. Hence, the basic computational principles uncovered when scrutinizing the representation of visual information could well extrapolate to other aspects of cognition. Needless to say, there is a vast and fascinating uncharted territory awaiting for neuroscientists to probe the representation of high-level cognitive information.

Acknowledgments

This work was supported by NSF (1358839, CCF-1231216) and NIH (R01EY026025)).

References

  1. Afraz A, Boyden ES, DiCarlo JJ. Optogenetic and pharmacological suppression of spatial clusters of face neurons reveal their causal role in face gender discrimination. Proc Natl Acad Sci U S A. 2015;112(21):6730–6735. doi: 10.1073/pnas.1423328112. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Afraz SR, Kiani R, Esteky H. Microstimulation of inferotemporal cortex influences face categorization. Nature. 2006;442(7103):692–695. doi: 10.1038/nature04982. [DOI] [PubMed] [Google Scholar]
  3. Bair W, Koch C. Temporal precision of spike trains in extrastriate cortex of the behaving macaque monkey. Neural Computation. 1996;8:1185–1202. doi: 10.1152/jn.90980.2008. [DOI] [PubMed] [Google Scholar]
  4. Baker JF, Petersen SE, Newsome WT, Allman JM. Visual response properties of neurons in four extrastriate visual areas of the owl monkey (Aotus trivirgatus): a quantitative comparison of medial, dorsomedial, dorsolateral, and middle temporal areas. J Neurophysiol. 1981;45(3):397–416. doi: 10.1152/jn.1981.45.3.397. [DOI] [PubMed] [Google Scholar]
  5. Bansal AK, Madhavan R, Agam Y, Golby A, Madsen JR, Kreiman G. Neural Dynamics Underlying Target Detection in the Human Brain. Journal of Neuroscience. 2014;34(8):3042–3055. doi: 10.1523/JNEUROSCI.3781-13.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Bansal AK, Singer JM, Anderson WS, Golby A, Madsen JR, Kreiman G. Temporal stability of visually selective responses in intracranial field potentials recorded from human occipital and temporal lobes. Journal of Neurophysiology. 2012;108:3073–3086. doi: 10.1152/jn.00458.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Barlow HB. Single units and sensation: a neuron doctrine for perception. Perception. 1972;1:371–394. doi: 10.1068/p010371. [DOI] [PubMed] [Google Scholar]
  8. Baylis GC, Rolls ET. Responses of neurons in the inferior temporal cortex in short-term and serial recognition memory tasks. Experimental brain research. 1987;65(3):614–622. doi: 10.1007/bf00235984. [DOI] [PubMed] [Google Scholar]
  9. Benton AL, Wav Allen MW. Prosopagnosia and Facial Discrimination. Journal of neurological Science. 1972;15(5) doi: 10.1016/0022-510x(72)90004-4. [DOI] [PubMed] [Google Scholar]
  10. Berry MJ, Meister M. Refractoriness and neural precision. Journal of Neuroscience. 1998;18(6):2200–2211. doi: 10.1523/JNEUROSCI.18-06-02200.1998. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bondar IV, Leopold DA, Richmond BJ, Victor JD, Logothesis NK. Long-term stability of visual pattern selective responses of monkey temporal lobe neurons. PLoS One. 2009;9(12) doi: 10.1371/journal.pone.0008222. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Bowers JS. On the biological plausibility of grandmother cells: Implications for neural network theories in psychology and neuroscience. Psychol Rev. 2009;116(1):220–251. doi: 10.1037/a0014462. doi:2009-00258-008 [pii] 10.1037/a0014462. [DOI] [PubMed] [Google Scholar]
  13. Brecht M, Schneider M, Sakmann B, Margrie TW. Whisker movements evoked by stimulation of single pyramidal cells in rat motor cortex. Nature. 2004;427:704–710. doi: 10.1038/nature02266. [DOI] [PubMed] [Google Scholar]
  14. Brindley GS, Donaldson PE, et al. The extent of the region of occiptal cortex that when stimulated gives phosphenes fixed in the visual field. Journal of Physiology. 1972;225(2):2. [PubMed] [Google Scholar]
  15. Brown MW, Aggleton JP. Recognition memory: What are the roles of the perirhinal cortex and hippocampus? Nature Reviews Neuroscience. 2001;2(1):51–61. doi: 10.1038/35049064. [DOI] [PubMed] [Google Scholar]
  16. Cadieu C, Kouh M, Pasupathy A, Connor C, Riesenhuber M, Poggio T. A model of V4 shape selectivity and invariance. [gk comment] Journal of Neurophysiology. 2007;98(3):1733–1750. doi: 10.1152/jn.01265.2006. [DOI] [PubMed] [Google Scholar]
  17. Cameron KA, Yashar S, Wilson CL, Fried I. Human hippocampal neurons predict how well word pairs will be remembered. Neuron. 2001;30:289–298. doi: 10.1016/S0896-6273(01)00280-X. [DOI] [PubMed] [Google Scholar]
  18. Carandini M, Demb JB, Mante V, Tolhurst DJ, Dan Y, Olshausen BA, Rust NC. Do we know what the early visual system does? J Neurosci. 2005;25(46):10577–10597. doi: 10.1523/JNEUROSCI.3726-05.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Churchland MM, Yu BM, Cunningham JP, Sugrue LP, Cohen MR, Corrado GS, Shenoy KV. Stimulus onset quenches neural variability: a widespread cortical phenomenon. Nat Neurosci. 2010;13(3):369–378. doi: 10.1038/nn.2501. doi:nn.2501 [pii] 10.1038/nn.2501. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Connor CE, Brincat SL, Pasupathy A. Transformation of shape information in the ventral pathway. Curr Opin Neurobiol. 2007;17(2):140–147. doi: 10.1016/j.conb.2007.03.002. [DOI] [PubMed] [Google Scholar]
  21. Crick F, Koch C. A framework for consciousness. Nat Neurosci. 2003;6(2):119–126. doi: 10.1038/nn0203-119. [DOI] [PubMed] [Google Scholar]
  22. Dayan P, Abbott L. Theoretical Neuroscience. Cambridge: MIT Press; 2001. [Google Scholar]
  23. Dean P. Effects of inferotemporal lesions on the behavior of monkeys. Psychological Bulletin. 1976;83(1):41–71. doi: 10.1037/0033-2909.83.1.41. [DOI] [PubMed] [Google Scholar]
  24. Deco G, Rolls ET. A neurodynamical cortical model of visual attention and invariant object recognition. Vision Res. 2004;44(6):621–642. doi: 10.1016/j.visres.2003.09.037. [DOI] [PubMed] [Google Scholar]
  25. Desimone R, Albright TD, Gross CG, Bruce C. Stimulus-selective properties of inferior temporal neurons in the macaque. Journal of Neuroscience. 1984;4(8):2051–2062. doi: 10.1523/JNEUROSCI.04-08-02051.1984. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. DiCarlo JJ, Zoccolan D, Rust NC. How does the brain solve visual object recognition? Neuron. 2012;73(3):415–434. doi: 10.1016/j.neuron.2012.01.010. doi:S0896-6273(12)00092-X [pii] 10.1016/j.neuron.2012.01.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Dobelle WH, Mladejovsky MG. Phosphenes produced by electrical stimulation of human occipital cortex, and their application to the developement of a prosthesis for the blind. Journal of Physiology. 1974;243(2):23. doi: 10.1113/jphysiol.1974.sp010766. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Doron G, Brecht M. What single-cell stimulation has told us about neural coding. Philos Trans R Soc Lond B Biol Sci. 2015;370(1677):20140204. doi: 10.1098/rstb.2014.0204. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Douglas RJ, Martin KA. Neuronal circuits of the neocortex. Annu Rev Neurosci. 2004;(27):419–451. doi: 10.1146/annurev.neuro.27.070203.144152. [DOI] [PubMed] [Google Scholar]
  30. Eichenbaum H. Hippocampus: cognitive processes and neural representations that underlie declarative memory. Neuron. 2004;44(1):109–120. doi: 10.1016/j.neuron.2004.08.028. [DOI] [PubMed] [Google Scholar]
  31. Ekstrom AD, Kahana MJ, Caplan JB, Fields TA, Isham EA, Newman EL, Fried I. Cellular networks underlying human spatial navigation. Nature. 2003;425(6954):184–187. doi: 10.1038/nature01964. [DOI] [PubMed] [Google Scholar]
  32. Felleman DJ, Van Essen DC. Distributed hierarchical processing in the primate cerebral cortex. Cereb Cortex. 1991;1:1–47. doi: 10.1093/cercor/1.1.1. [DOI] [PubMed] [Google Scholar]
  33. Field GD, Gauthier JL, Sher A, Greschner M, Machado TA, Jepson LH, Chichilnisky EJ. Functional connectivity in the retina at the resolution of photoreceptors. Nature. 2010;467(7316):673–677. doi: 10.1038/nature09424. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Freedman DJ, Riesenhuber M, Poggio T, Miller EK. Categorical representation of visual stimuli in the primate prefrontal cortex. Science. 2001;291:312–316. doi: 10.1126/science.291.5502.312. [DOI] [PubMed] [Google Scholar]
  35. Fukushima K. Neocognitron: a self organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biological Cybernetics. 1980;36(4):193–202. doi: 10.1007/978-3-642-46466-9_18. [DOI] [PubMed] [Google Scholar]
  36. Gelbard-Sagiv H, Mukamel R, Harel M, Malach R, Fried I. Internally Generated Reactivation of Single Neurons in Human Hippocampus During Free Recall. Science. 2008 doi: 10.1126/science.1164685. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Gilbert CD, Li W, Piech V. Perceptual learning and adult cortical plasticity. J Physiol. 2009;587(Pt 12):2743–2751. doi: 10.1113/jphysiol.2009.171488. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Gross CG. Genealogy of the Grandmother Cell. The Neuroscientist. 2002;8(5):512–518. doi: 10.1177/107385802237175. [DOI] [PubMed] [Google Scholar]
  39. Holmes G. Disturbances of Visual Orientation. Br J Ophthalmol. 1918;2(10):506–516. doi: 10.1136/bjo.2.10.506. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Holt GR, Softky WR, Koch C, Douglas RJ. Comparison of discharge variability in vitro and invivo in cat visual cortex neurons. Journal of Neurophysiology. 1996;75(5):1806–1814. doi: 10.1152/jn.1996.75.5.1806. [DOI] [PubMed] [Google Scholar]
  41. Hopfield JJ. Neural networks and physical systems with emergent collective computational abilities. PNAS. 1982;79:2554–2558. doi: 10.1073/pnas.79.8.2554. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Hubel DH, Wiesel TN. Receptive fields, binocular interaction and functional architecture in the cat's visual cortex. J Physiol. 1962;160:106–154. doi: 10.1113/jphysiol.1962.sp006837. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Hung CP, Kreiman G, Poggio T, DiCarlo JJ. Fast Read-out of Object Identity from Macaque Inferior Temporal Cortex. Science. 2005;310:863–866. doi: 10.1126/science.1117593. [DOI] [PubMed] [Google Scholar]
  44. Ito M, Tamura H, Fujita I, Tanaka K. Size and position invariance of neuronal responses in monkey inferotemporal cortex. J Neurophysiol. 1995;73(1):218–226. doi: 10.1152/jn.1995.73.1.218. [DOI] [PubMed] [Google Scholar]
  45. Kirchner H, Thorpe SJ. Ultra-rapid object detection with saccadic eye movements: visual processing speed revisited. Vision Res. 2006;46(11):1762–1776. doi: 10.1016/j.visres.2005.10.002. doi:S0042-6989-(05)00511-0[pii] 10.1016/j.visres.2005.10.002. [DOI] [PubMed] [Google Scholar]
  46. Kreiman G. Single neuron approaches to human vision and memories. Curr Opin Neurobiol. 2007;17(4):471–475. doi: 10.1016/j.conb.2007.07.005. [DOI] [PubMed] [Google Scholar]
  47. Kreiman G, Koch C, Fried I. Imagery neurons in the human brain. Nature. 2000a;408:357–361. doi: 10.1038/35042575. [DOI] [PubMed] [Google Scholar]
  48. Kreiman G, Koch C, Fried I. Category-specific visual responses of single neurons in the human medial temporal lobe. Nat Neurosci. 2000b;3(9):946–953. doi: 10.1038/78868. [DOI] [PubMed] [Google Scholar]
  49. Kreiman G, Krahe R, Metzner W, Koch C, Gabbiani F. Robustness and Variability of Neuronal Coding by Amplitude Sensitive Afferents in the Weakly Electric Fish Eigenmannia. Journal of Neurophysiology. 2000;84(1):189–204. doi: 10.1152/jn.2000.84.1.189. [DOI] [PubMed] [Google Scholar]
  50. Krizhevsky A, Sutskever I, Hinton G. ImageNet Classification with Deep Convolutional Neural Networks; Paper presented at the IPS, Montreal.2012. [Google Scholar]
  51. Laughlin SB, van Steveninck RRR, Anderson JC. The metabolic cost of neural information. Nat Neurosci. 1998;1(1):36–41. doi: 10.1038/236. [DOI] [PubMed] [Google Scholar]
  52. Leopold DA, Logothetis NK. Activity changes in early visual cortex reflect monkeys' percepts during binocular rivalry. Nature. 1996;379:549–553. doi: 10.1038/379549a0. [DOI] [PubMed] [Google Scholar]
  53. Logothetis NK, Augath M, Murayama Y, Rauch A, Sultan F, Goense J, Merkle H. The effects of electrical microstimulation on cortical signal propagation. Nat Neurosci. 2010;13(10):1283–1291. doi: 10.1038/nn.2631. [DOI] [PubMed] [Google Scholar]
  54. Logothetis NK, Pauls J, Poggio T. Shape representation in the inferior temporal cortex of monkeys. Current Biology. 1995;5(5)(95):552–563. 00108–4. doi: 10.1016/s0960-9822. [DOI] [PubMed] [Google Scholar]
  55. Logothetis NK, Sheinberg DL. Visual object recognition. Annual Review of Neuroscience. 1996;19:577–621. doi: 10.1146/annurev.ne.19.030196.003045. [DOI] [PubMed] [Google Scholar]
  56. Mainen ZF, Sejnowski TJ. Reliability of spike timing in neocortical neurons. Science. 1995;268:1503–1506. doi: 10.1126/science.7770778. [DOI] [PubMed] [Google Scholar]
  57. Markov NT, Ercsey-Ravasz MM, Ribeiro Gomes AR, Lamy C, Magrou L, Vezoli J, Kennedy H. A Weighted and Directed Interareal Connectivity Matrix for Macaque Cerebral Cortex. Cereb Cortex. 2012 doi: 10.1093/cercor/bhs270. doi:bhs270 [pii] 10.1093/cercor/bhs270. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. McClelland JL, Rumelhart DE. An Interactive Activation Model of context effects in letter perception: I. An account of basic findings. Psychological Review. 1981;88:375–407. doi: 10.1037/0033-295x.88.5.375. [DOI] [PubMed] [Google Scholar]
  59. McClelland JL, Rumelhart DE Group, PDP Research. Parallel distributed processing: Psychological and biological models. Vol. 2. Cambridge, MA: MIT Press; 1986. [Google Scholar]
  60. McMahon DB, Jones AP, Bondar IV, Leopold DA. Face-selective neurons maintain consistent visual responses across months. Proc Natl Acad Sci U S A. 2014;111(22):8251–8256. doi: 10.1073/pnas.1318331111. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Mel B. SEEMORE: Combining color, shape and texture histogramming in a neurally inspired approach to visual object recognition. Neural Computation. 1997;9:777. doi: 10.1162/neco.1997.9.4.777. [DOI] [PubMed] [Google Scholar]
  62. Meyers E, Freedman DJ, Kreiman G, Miller EK, Poggio T. Dynamic Population Coding of Category Information in ITC and PFC. Journal of Neurophysiology. 2008;100:1407–1419. doi: 10.1152/jn.90248.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Miconi T, Groomes L, Kreiman G. There's Waldo! A Normalization Model of Visual Search Predicts Single-Trial Human Fixations in an Object Search Task. Cereb Cortex. 2015 doi: 10.1093/cercor/bhv129. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Miller EK, Cohen JD. An integrative theory of prefrontal cortex function. Annu Rev Neurosci. 2001;24:167–202. doi: 10.1146/annurev.neuro.24.1.167. [DOI] [PubMed] [Google Scholar]
  65. Mormann F, Ison M, Quiroga RQ, Koch C, Fried I, Kreiman G. Visual cognitive adventures of single neurons in the human medial temporal lobe. In: Fried I, Rutishauser U, Cerf M, Kreiman G, editors. Single neuron studies of the human brain Probing cognition. Cambridge, MA: MIT Press; 2014. [Google Scholar]
  66. Mormann F, Dubois J, Kornblith S, Milosavljevic M, Cerf M, Ison M, Koch C. A category-specific response to animals in the right human amygdala. Nat Neurosci. 2011;14(10):1247–1249. doi: 10.1038/nn.2899. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Naya Y, Sakai K, Miyashita Y. Activity of primate inferotemporal neurons related to a sought target in a paired-association task. PNAS. 1996;93:2664–2669. doi: 10.1073/pnas.93.7.2664. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Newsome WT, Britten KH, Movshon JA. Neuronal correlates of a perceptual decision. Nature. 1989;341:52–54. doi: 10.1038/341052a0. [DOI] [PubMed] [Google Scholar]
  69. Olshausen BA, Anderson CH, Van Essen DC. A neurobiological model of visual attention and invariant pattern recognition based on dynamic routing of information. J Neurosci. 1993;13(11):4700–4719. doi: 10.1523/JNEUROSCI.13-11-04700.1993. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Olshausen BA, Field DJ. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature. 1996;381(6583):607–609. doi: 10.1038/381607a0. [DOI] [PubMed] [Google Scholar]
  71. Parker AJ, Newsome WT. Sense and the single neuron: Probing the physiology of perception. Annual Review of Neuroscience. 1998;21:227–277. doi: 10.1146/annurev.neuro.21.1.227. [DOI] [PubMed] [Google Scholar]
  72. Pasupathy A, Connor CE. Population coding of shape in area V4. Nat Neurosci. 2002;5(12):1332–1338. doi: 10.1038/nn972. [DOI] [PubMed] [Google Scholar]
  73. Perkel DH, Gerstein GL, Moore GP. Neuronal Spike Trains and Stochastic Point Processes. Biophysical Journal. 1967;7(67):391–418. 86596–2. doi: 10.1016/S0006-3495. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Perret DI, Oram MW, Hietanen JK, Benson PJ. Issues of representation in object vision. In: Farah Mj, Ratcliff G., editors. The neuropsychology of high-level vision. New Jersey: Lawrence Erlbaum Associates; 1994. pp. 33–61. [Google Scholar]
  75. Potter MC, Levy E. Recognition memory for a rapid sequence of pictures. J Exp Psychol. 1969;81(1):10–15. doi: 10.1037/h0027470. [DOI] [PubMed] [Google Scholar]
  76. Priebe NJ, Ferster D. Mechanisms of neuronal computation in mammalian visual cortex. Neuron. 2012;75(2):194–208. doi: 10.1016/j.neuron.2012.06.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Quian Quiroga R, Reddy L, Kreiman G, Koch C, Fried I. Invariant visual representation by single neurons in the human brain. Nature. 2005;435:1102–1107. doi: 10.1038/nature03687. [DOI] [PubMed] [Google Scholar]
  78. Ramalingam N, McManus JN, Li W, Gilbert CD. Top-down modulation of lateral interactions in visual cortex. J Neurosci. 2013;33(5):1773–1789. doi: 10.1523/JNEUROSCI.3825-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Reynolds JH, Chelazzi L. Attentional modulation of visual processing. Annu Rev Neurosci. 2004;27:611–647. doi: 10.1146/annurev.neuro.26.041002.131039. [DOI] [PubMed] [Google Scholar]
  80. Richmond BJ, Optican LM, Spitzer H. Temporal encoding of two-dimensional patterns by single units in primate primary visual cortex. I. Stimulus-response relations. Journal of Neurophysiology. 1990;64(2):351–369. doi: 10.1152/jn.1990.64.2.351. [DOI] [PubMed] [Google Scholar]
  81. Riesenhuber M, Poggio T. Hierarchical models of object recognition in cortex. Nat Neurosci. 1999;2(11):1019–1025. doi: 10.1038/14819. [DOI] [PubMed] [Google Scholar]
  82. Rigotti M, Barak O, Warden MR, Wang XJ, Daw ND, Miller EK, Fusi S. The importance of mixed selectivity in complex cognitive tasks. Nature. 2013;497(7451):585–590. doi: 10.1038/nature12160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Ringach DL, Hawken MJ, Shapley R. Dynamics of orientation tuning in macaque primary visual cortex. Nature. 1997;387(6630):281–284. doi: 10.1038/387281a0. [DOI] [PubMed] [Google Scholar]
  84. Rolls ET. Neural organization of higher visual functions. Curr Opin Neurobiol. 1991;1:274–278. doi: 10.1016/0959-4388(91)90090-t. [DOI] [PubMed] [Google Scholar]
  85. Rutishauser U, Mamelak AN, Schuman EM. Single-trial learning of novel stimuli by individual neurons of the human hippocampus-amygdala complex. Neuron. 2006;49(6):805–813. doi: 10.1016/j.neuron.2006.02.015. [DOI] [PubMed] [Google Scholar]
  86. Rutishauser U, Schuman EM, Mamelak AN. Activity of human hippocampal and amygdala neurons during retrieval of declarative memories. Proc Natl Acad Sci U S A. 2008;105(1):329–334. doi: 10.1073/pnas.0706015105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Salzman CD, Britten KH, Newsome WT. Cortical microstimulation influences perceptual judgments of motion direction. Nature. 1990;346:174–177. doi: 10.1038/346174a0. [DOI] [PubMed] [Google Scholar]
  88. Sary G, Vogels R, Orban GA. Cue-invariant shape selectivity of macaque inferior temporal neurons. Science. 1993;260:995–997. doi: 10.1126/science.8493538. [DOI] [PubMed] [Google Scholar]
  89. Serre T, Kreiman G, Kouh M, Cadieu C, Knoblich U, Poggio T. A quantitative theory of immediate visual recognition. Progress In Brain Research. 2007;165C(06):33–56. 65004–8. doi: 10.1016/S0079-6123. [DOI] [PubMed] [Google Scholar]
  90. Sheinberg DL, Logothetis NK. The role of temporal areas in perceptual organization. Proceedings of the National Academy of Sciences, USA. 1997;94:3408–3413. doi: 10.1073/pnas.94.7.3408. [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Smith MA, Majaj NJ, Movshon JA. Dynamics of motion signaling by neurons in macaque area MT. Nat Neurosci. 2005;8(2):220–228. doi: 10.1038/nn1382. [DOI] [PubMed] [Google Scholar]
  92. Squire LR, Stark CE, Clark RE. The Medial Temporal Lobe. Annu Rev Neurosci. 2004;27:279–306. doi: 10.1146/annurev.neuro.27.070203.144130. [DOI] [PubMed] [Google Scholar]
  93. Stevens CF, Zador AM. Input synchrony and the irregular firing of cortical neurons. Nat Neurosci. 1998;1(3):210–217. doi: 10.1038/659. [DOI] [PubMed] [Google Scholar]
  94. Tanaka K. Inferotemporal cortex and object vision. Annual Review of Neuroscience. 1996;19:109–139. doi: 10.1146/annurev.ne.19.030196.000545. [DOI] [PubMed] [Google Scholar]
  95. Tehovnik EJ. Electrical stimulation of neural tissue to evoke behavioral responses. J Neurosci Methods. 1996;65(1):1–17. doi: 10.1016/0165-0270(95)00131-x. [DOI] [PubMed] [Google Scholar]
  96. Tehovnik EJ, Slocum WM, Smirnakis SM, Tolias AS. Microstimulation of visual cortex to restore vision. Prog Brain Res. 2009;175(09):347–375. 17524–6. doi: 10.1016/S0079-6123. [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Thorpe S, Fize D, Marlot C. Speed of processing in the human visual system. Nature. 1996;381:520–522. doi: 10.1038/381520a0. [DOI] [PubMed] [Google Scholar]
  98. Tolias AD, Ecker AS, Siapas AG, Hoenselaar A, Keliris GA, Logothesis N. Recording Chronically From the Same Neurons in Awake, Behaving Primates. Journal of Neurophysiology. 2007 doi: 10.1152/jn.00260.2007. doi:10.1152. [DOI] [PubMed] [Google Scholar]
  99. Tovee MJ, Rolls ET, Azzopardi P. Translation invariance in the responses to faces of single neurons in the temporal visual cortical areas of the alert macaque. J Neurophysiol. 1994;72(3):1049–1060. doi: 10.1152/jn.1994.72.3.1049. [DOI] [PubMed] [Google Scholar]
  100. Tsao DY, Freiwald WA, Tootell RB, Livingstone MS. A cortical region consisting entirely of face-selective cells. Science. 2006;311(5761):670–674. doi: 10.1126/science.1119983. [DOI] [PMC free article] [PubMed] [Google Scholar]
  101. van Steveninck R, Lewen GD, Strong Sp, Koberle R, Bialek W. Reproducibility and variability in neural spike trains. Science. 1997;275:1805–1808. doi: 10.1126/science.275.5307.1805. [DOI] [PubMed] [Google Scholar]
  102. Vogels R, Sary G, Orban GA. How task-related are the responses of inferior temporal neurons? Vis Neurosci. 1995;12(2):207–214. doi: 10.1017/s0952523800007884. [DOI] [PubMed] [Google Scholar]
  103. Woloszyn L, Sheinberg DL. Neural dynamics in inferior temporal cortex during a visual working memory task. J Neurosci. 2009;29(17):5494–5507. doi: 10.1523/JNEUROSCI.5785-08.2009. doi:29/17/5494[pii] 10.1523/JNEUROSCI.5785-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Wyatte D, Curran T, O'Reilly R. The limits of feedforward vision: recurrent processing promotes robust object recognition when objects are degraded. J Cogn Neurosci. 2012;24(11):2248–2261. doi: 10.1162/jocn_a_00282. [DOI] [PubMed] [Google Scholar]
  105. Yamins DL, Hong H, Cadieu CF, Solomon EA, Seibert D, DiCarlo JJ. Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proc Natl Acad Sci U S A. 2014;111(23):8619–8624. doi: 10.1073/pnas.1403112111. [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Zeki S. A century of cerebral achromatopsia. Brain. 1990;113(Pt 6):1721–1777. doi: 10.1093/brain/113.6.1721. [DOI] [PubMed] [Google Scholar]
  107. Zeki S. Cerebral akinetopsia (visual motion blindness). A review. Brain. 1991;114(Pt 2):811–824. doi: 10.1093/brain/114.4.2021. [DOI] [PubMed] [Google Scholar]

RESOURCES