Abstract
Advances in neural signal and image acquisition as well as in multivariate signal processing and machine learning are enabling a richer and more rigorous understanding of the neural basis of human decision-making. Decision-making is essentially characterized behaviorally by the variability of the decision across individual trials—e.g., error and response time distributions. To infer the neural processes that govern decision-making requires identifying neural correlates of such trial-to-trial behavioral variability. In this paper, we review efforts that utilize signal processing and machine learning to enable single-trial analysis of neural signals acquired while subjects perform simple decision-making tasks. Our focus is on neuroimaging data collected noninvasively via electroencephalograpy (EEG) and functional magnetic resonance imaging (fMRI). We review the specific frame-work for extracting decision-relevant neural components from the neuroimaging data, the goal being to analyze the trial-to-trial variability of the neural signal along these component directions and to relate them to elements of the decision-making process. We review results for perceptual decision-making and discrimination tasks, including paradigms in which EEG variability is used to inform an fMRI analysis. We discuss how single-trial analysis reveals aspects of the underlying decision-making networks that are unobservable using traditional trial-averaging methods.
Keywords: Decision-making, electroencephalography, functional magnetic resonance imaging, machine learning, single-trial analysis
I. Introduction
OUR ability to make rapid decisions based on transient stimuli is a unique aspect of our brains’ capacity to process information. Recent advances in signal processing and neuroimaging have begun to provide tools for answering questions related to how we make rapid decisions by enabling the identification of cortical networks underlying rapid decision-making in the human brain. Typically, such studies utilize functional magnetic resonance imaging (fMRI) and have identified areas in frontal, parietal, and thalamic cortex in which metabolic activity correlates with decision-related variables. However, decision-making is a dynamic process, and the localized activations found with fMRI must be part of cortical networks defined by the relative timing of these activations and their causality. Electroencephalography (EEG) offers the ability to resolve neural processing at the millisecond level, however, with a relatively poor spatial resolution. Combinations of these modalities have been recently developed enabling more precise identification of cortical networks underlying rapid decision-making, including those processes involved in integration of contextual information and accumulation of multiple sources of evidence.
At the core of the scientific and engineering challenge is how to analyze neuroimaging data on a trial-by-trial basis, since perceptual decision-making is in fact characterized by variations in behavior across trials, even given nominally identical stimuli. In this paper, we review experimental results in animals and humans with respect to perceptual decision-making, focusing on the role neuroimaging has to play in shedding light on the underlying information processing. We also review how signal processing and machine learning play a role, specifically in our ability to identify meaningful neural signatures, on trial-by-trial basis, to characterize the underlying network processes. We also review our own specific results, which focus on using EEG, fMRI, and both simultaneously for inferring the networks underlying perceptual decision-making.
II. Decision-Making and Neuroimaging
Perceptual decision-making is the process by which incoming sensory information is combined and used to influence our behavior [1]. A number of ground-breaking electrophysiology studies have already laid the foundation upon which future research on perceptual decision-making will be conducted [2], [3].
A. Perceptual Decision-Making in Animals
Experiments in primates have already established a relationship between behavioral choices and neuronal activity in sensory regions. In a motion discrimination task during which monkeys had to discriminate the direction of motion from random dot kinetograms consisting of varying amounts of coherent motion, Newsome et al. showed that the activity of direction-selective neurons in middle temporal area (MT) can provide an adequate account of behavioral performance [4]-[6] and that trial-to-trial variability in these neuronal signals could predict the monkeys’ actual choices [7]. In addition, electrical microstimulation of MT neurons biased the monkeys choices towards the neurons’ preferred direction [8]-[10]. Interestingly, this pattern of neural responses appears to extend even to highly complex visual stimuli such as faces [11], [12].
Similarly, in the somatosensory domain, Romo et al. used a task in which monkeys had to discriminate the vibration frequency of two sequentially presented tactile stimuli and report which one was the highest. They showed that, on average, the responses in primary somatosensory cortex (S1) correlated with the behavioral performance of the monkeys and the trial-to-trial fluctuations in these responses predicted the animals’ choices [13]. Furthermore, microstimulation of S1 neurons, in the absence of physical stimulation, was sufficient to reproduce the behavioral patterns seen under normal conditions [14], [15]. These findings support the idea that neuronal signals in areas such as MT and S1 provide the sensory evidence upon which monkeys base their decisions regarding motion and vibrotactile discrimination, respectively.
How, then, can activity in these areas be used to influence decision-making? Theoretical models of decision-making have modeled binary perceptual choices as a race-to-barrier diffusion process [16]-[20]. The diffusion model for two-choice decisions in particular assumes that in the decision process, evidence is integrated over time to one of two decision thresholds corresponding to the two choices. The rate of accumulation is called drift rate, and it is determined by the quality of the sensory information. The better the information quality, the larger the drift rate toward the appropriate decision boundary and the faster and more accurate the response.
In the monkey brain, a region in the intraparietal sulcus, commonly known as the lateral intraparietal (LIP) area, is shown to exhibit some of the characteristics proposed by the aforementioned models. Specifically, electrophysiological recordings in LIP during the motion discrimination task [21]-[24] have shown that, for choices made towards the response field of the neuron under consideration, neural activity increased in a ramp-like fashion, consistent with a temporal integration process. The rate at which responses increased was proportional to the amount of coherent motion in the stimulus. This buildup of activity is believed to represent the accumulated difference in firing rates of two opposing pools of direction selective MT neurons. Just prior to the response, neural responses in LIP achieved a common firing rate, independent of motion strength, consistent with the idea of a common decision threshold. More recently, these results were extended from a two- to a four-choice direction discrimination task [25]. The same pattern of activity was reported in the frontal eye fields (FEFs) and the dorsolateral prefrontal cortex (DLPFC) [26] and is seen as evidence that all of these areas are involved in the conversion of an analog motion representation into a binary decision variable.
Likewise, during vibrotactile frequency discrimination, regions in the secondary somatosensory (S2), medial premotor cortex (MPC), ventral premotor cortex (VPC), and DLPFC were reported as being involved in decision formation [27]-[30]. As with the LIP, responses in the MPC and VPC were also shown to form a decision by computing the difference between the responses of S2 neurons that code for each of the two frequencies used for stimulation [29], [30]. Furthermore, it is believed that some S2 neurons have the capacity to combine past and present sensory information to form a decision [28] and that short-term memory representation in DLPFC might also play a role in decision-making [27], [31].
B. Perceptual Decision-Making in Humans
The principles that have emerged from the electrophysiology work in animals have inspired scientists to study the neural correlates of perceptual decision-making in humans using noninvasive neuroimaging techniques such as the fMRI and the EEG.
Heekeren et al. used fMRI to study perceptual decision-making during a face-versus-house discrimination task [32]. Consistent with previous neuroimaging studies [33]-[37], they identified face and house selective regions in the ventral temporal cortex that, much like MT in a motion discrimination task, provide the sensory evidence upon which a decision is later computed. Specifically, they showed that these regions were modulated by the amount of degradation in the face and house stimuli, respectively. Inspired by the work of Romo et al., recent fMRI studies have also used vibrotactile frequency tasks to identify regions that encode the sensory evidence needed for a somatosensory decision. Consistent with electrophysiological data in monkeys, they found that the somatosensory cortex and S1 in particular encode the incoming sensory information [38], [39].
As summarized above, animal work has provided evidence that the brain combines this sensory information using a difference-based comparator operation. Heekeren et al. directly tested whether a comparison operation is also at work in the human brain using the face-versus-house discrimination task outlined above [32]. They proposed that higher level decision areas should show the greatest activity on easy trials (clear images of faces and houses) rather than on hard ones (degraded images) and, in addition, they should correlate with the difference between the signals from the face and house selective regions that represent the sensory evidence. The only region that exhibited this pattern of activity was the posterior portion of the left DLPFC. Therefore, even for complex object categories, the brain appears to integrate the outputs of different pools of selectively tuned neurons in order to make perceptual decisions.
Another recent fMRI study also showed how the fMRI signal can be used to examine the process of accumulation of sensory evidence [40]. Pictures were revealed gradually over the course of several seconds, and participants indicated the time of recognition with a button press. There was a gradual buildup in fMRI signal peaking in correspondence with the time of recognition in inferior temporal, frontal, and parietal regions, suggesting that these regions might be involved in accumulating sensory evidence.
Contrary to electrophysiology studies in monkeys that have thus far claimed that “to see and decide is, in effect, to plan a motor-response” [41], Heekeren et al. also investigated whether decisions may be transformed into actions independent of the motor effectors used to indicate the choice [42]. They used the same motion discrimination task used in the monkey studies and asked participants to indicate their choice with either a button press or an eye movement. They found that the area in the left DLPFC they identified earlier [32] exhibited decision-related activity (as highlighted above) independent of the motor system used to indicate the choice. They concluded “that humans may have evolved a more abstract decision-making network, thereby allowing a more flexible link between decision and action” [32].
Despite fMRI’s excellent spatial resolution, the slow scanning rates and the low-pass nature of the hemodynamic signal place significant constraints on its capacity to provide high temporal resolution information. To overcome this limitation, advanced multivariate analysis techniques have been developed to also study the temporal characteristics of perceptual decision-making in humans, as we highlight in the next sections.
III. Multivariate Processing and Neuroimaging
A. Capability of Various Neuroimaging Modalities
All noninvasive neuroimaging modalities make only indirect measures of neuronal activity. fMRI records the local hemodynamic response (blood oxygenation level) within the brain following increased metabolic activity of neurons. This can be recorded with millimeter spatial resolution over the entire brain, which results in thousands of simultaneously recorded voxels. Hemodynamic response is sluggish (1–5 s), and so activity is typically sampled only every 1 or 2 s.1
EEG measures electrical potentials that build up at the surface of the scull resulting from large-scale ionic currents in the brain. These currents originate primarily at the dendrites of neurons reflecting neuronal input activity. Surface electrical potentials are easy to record and capture primarily cortical activity with centimeter spatial resolution. Currents from deeper structures decay significantly and are too broadly distributed on the scull to be localized with much accuracy. However, EEG signals do track fast changes of neural activity and are thus recorded with millisecond temporal resolution. Useful signal may be found at frequencies of up to 80 Hz, which is much higher than the less than 1 Hz temporal resolution of fMRI.
Magnetic encephalography (MEG) detects static magnetic fields generated by the same ionic currents that underlie the EEG signal. MEG therefore has a similar temporal resolution. However, because the brain is engulfed by the cerebrospinal fluid, direct currents measured in EEG are partially shunted, while magnetic field lines cross unaltered, leading to a somewhat better localization of cortical current sources when using MEG. As a corollary, fast synchronous neuronal activity that has a shorter length scale is easier to detect and localize, and so the useful frequencies in MEG may be as high as 100 Hz. The disadvantage of MEG is its high cost, which is comparable to MRI with its much higher spatial resolution.
B. Decoding Problem is Similar for Different Modalities
While the various functional neuroimaging modalities may have different spatial and temporal resolutions, when considering activity on a second by second basis, they all provide a comparable number of separate measurements. fMRI measures perhaps once a second a volume of approximately 50×50×20 voxels. EEG or MEG measure about 100 channels at perhaps 500 samples per second. In either case, one is left with approximately 50 000 measurements per second. We will refer to these in the following simply as “samples” reflecting a sampling of brain activity either in space (fMRI voxels) or space–time (EEG electrodes and samples). Interestingly, the amount of noise in the two modalities are also comparable. Typically, for either modality, trials have to be repeated several times before significant differences between two experimental conditions can be seen (maybe, N = 100). A simple power analysis indicates that the signal-to-noise ratio (SNR) in the individual dimensions is therefore about −13 dB (see Fig. 1 for other values of N).
Fig. 1.

Power analysis: number of independent measures N required at a given SNR to achieve statistical significant difference in mean activity between two experimental conditions. Curves for various significance levels (p-values) are computed assuming Normal distributed data. Both fMRI and EEG require about the same number of trials to show comparable significance levels, thus suggesting that their SNR is comparable. Multivariate analysis increases N, in effect, by using multiple samples, thus requiring fewer trials for averaging. Machine-learning techniques are used to determine over which samples to average and with which weighting.
Rarely do individual samples carry sufficient information to distinguish between different states of the brain. To recover significant activity, conventional analysis has relied on averaging over many repeated trials of the same stimulus and task conditions under the assumption that the relevant neuronal activity is reproducible from one trial to the next. While this approach has been highly successful in the past, we now know that it severely underestimates the amount of information that can be extracted from these simultaneous recordings. The reason is that activity may be distributed over very many samples and may not be strong enough in any one sample to allow a distinction between different neural processes. Instead, when carefully combining multiple samples with multivariate analysis techniques, one can identify neural activity that correlates well with observed behavioral responses despite rather low SNR.
Given the different nature of fMRI and EEG, one would think that the analysis techniques to accomplish this differ drastically between the two modalities. Surely the correlation structure of neighboring samples and the trial-to-trial variability differ significantly, and this should affect how one is to combine these high-dimensional data. However, despite the significant difference of fMRI and EEG, there has been a remarkable convergence in the techniques that have been used to extract behavioral correlates. The successful multivariate analysis methods are based on modern machine-learning techniques, which aim to uncover a pattern in high-dimensional spaces based on a small set of examples. The linear decoding techniques often amount to cleverly averaging over the large dimensional space, thus accomplishing what trial-averaging typically does, i.e., use a large number of independent measurements to increase SNR (see Fig. 1).
What is most remarkable is that the successful techniques for decoding neuroimaging signals are very similar to the techniques that have been proposed in machine learning for the goal of classifying complex dynamical patterns. The basic notion of what is known as reservoir computing [43], [44] is to represent a stimulus in a high dimensional space capturing many of its nonlinear features but then use a simple weighted sum of these features to classify the pattern into a few categories. It appears that the specific choice of nonlinear features is not very important as long as there is a large number of distinct features to assure that all relevant invariances are captured and that the representation is diverse enough to distinguish between different stimuli [45].
The neuroimaging signals we measure surely are a highly nonlinear representation of the perceptual stimuli or decision processes we aim to decode. The relevant parameters for decoding are then the dimensionality of the problem, the amount of noise, and the invariance and diversity in the feature space. In that regard, the challenge of extracting neural correlates is strikingly similar for the different neuroimaging modalities: The goal is to find in an approximately 104-dimensional space neural activity that correlates with a few possible decision outcomes (often just two) based on about 102 exemplars of signals with at best −10 dB in SNR. The only potential difference in the modalities is then the invariance and diversity of neural signals. Only here is where domain knowledge of the different modalities is required.
C. Single-Trial Analysis and the Decoding Problem
The original emphasis on single-trial analysis of neuroimaging data comes from the ambitious goal of “reading the brain,” that is, interpreting neural signals at the speed in which they develop so as to translate thought into action [46]. Aside from the potential practical benefits of such technology,2 there is an important benefit to the study of perceptual and cognitive processing. At a minimum, by combining multiple measures to an aggregate measure with higher SNR, one may be able to discover signals that have thus far eluded detection. Equivalently, it may be possible to obtain similar effect sizes with fewer trials, thus allowing the experimenter to segment the data into multiple response categories, e.g., distinguish fast responses from slow responses to identify the correlates of response-time variability [53]. Indeed, even when subjects are given an identical task with an identical stimulus, they never repeat the exact same behavior at the exact same speed and with the same neuronal signals. Uncovering the systematic covariation of neural signals with variations in behavioral response (speed, accuracy, strength, etc.) may shed light on internal states such as arousal, attention, motivation, etc., which are difficult to control experimentally. In addition, in some experiments, the number of stimulus dimensions may be too large to systematically explore. Quantifying the correlation of single-trial activity with a large set of stimulus features may uncover relationships that would otherwise be impossible to explore due to a limited sample size. Finally, single-trial activity could be used to correlate distinct neuroimaging modalities to each other [54] and even to identify common signals in the activity detected for different subjects [55].
The methods that have been developed for single-trial analysis can be grouped into two categories: 1) methods in which the presented stimulus is used to predict the resulting neural activity—we call this the forward model; and 2) methods that try to determine from the neural signal what stimulus was presented to the subject or how the subject responds. We call this direct decoding and note that forward models are often also inverted to similarly decode brain activity. These two approaches will be discussed next, followed by a discussion of the central problem of regularization.
D. Direct Decoding of Discrete States
The most straightforward method of “reading the brain” is to try to distinguish between two externally observable events based on some appropriate sum of the observed brain activity. For instance, say a subject on a given trial is presented with a target stimulus—call this the positive class. The negative class is the case in which the subject was presented a stimulus without a target. Assume that one records for the jth trial in sample i the activity xij. To decide on the class, a linear classifier takes a weighted sum of this activity xij
| (1) |
If this sum is larger than some threshold yj > θ, then this trial j is considered to belong to the positive class; otherwise it is considered to belong to the negative class. The magnitude and sign of the weights indicates how much each sample should contribute to the decision and in which direction. The goal is to find weights wi such that the classification into positive and negative class is correct for most trials.
Essentially, the goal is to determine from the magnitude of the measured brain activity whether the subject did or did not perceive the target. If we are able to do so, then we can argue that we have found the neural correlate of perception, or, equivalently, we can argue that we can “read” the brain signals to determine whether the subject did or did not perceive a target. The same could be done with signals associated with decision-making by trying to correctly classify what choice a subjects is about to make, or by predicting whether a response to a task will be correct or incorrect.
Given the D ≈ 104 dimensions of the samples, it is easy to find a suitable set of weights that give the correct classification on, say, N = 100 “training” exemplars. The challenge lies in finding weights that will perform well on “testing” examples that have not been used during “learning,” i.e., weights that will generalize well. There are a variety of learning algorithm for this binary classification problem, which vary in their ability to generalize despite the small number of training exemplars and in their ability to tolerate noisy exemplars. Support-vector machines (SVMs) perform well when the number of degrees of freedom is large compared to the number of training exemplars—as in the present case with D > N. These algorithms focus on exemplars that are closest to the dividing boundary between the two classes—these are called the support vectors. However, without modifications, SVMs do not perform well in the presence of significant noise, as the boundary elements will invariably be misleading (for most brain-reading applications SNR < −20 dB). Alternatively, one can use logistic regression (LR), which is naturally suited to noisy data, as the transition from one class to the other is assumed to be gradual following a logistic function. On the other hand, in the case of perfectly separable data, LR has to be modified to prevent a transition boundary that is infinitely sharp. Regularization can be used to trade off between perfect separability, for which SVM dominates, with low SNR for which LR dominates (see Section III-G). In our applications, we have sometimes also used Fisher linear discriminants (FLDs), which aim to increase SNR [56]. While FLD is simple to compute, it is not very robust since the required covariance estimates are dominated by outliers, which are common in neuroimaging data.
Direct decoding using linear classification as outlined above has been used with great success in fMRI, and several reviews are already available [46], [57]. The following are just a few examples of the type of activity that could be extracted: Brain activity of a listener was used to detect the identity of one of two speakers [58]. The geometry of a visual stimulus can be determined from visual cortex activity by predicting areas of high and low contrast on the image [59]. Decoders have also been used to distinguish spatial memories by distinguishing between two locations to which a subject navigated within a virtual environment despite identical visual appearance of the locations [60]. Concepts evoked by simple line drawings could be categorized without the use of visual cortex based on activity broadly distributed over the entire brain, and this categorization generalized well across subjects [61]. Free choice decisions could be detected several seconds ahead of an overt behavioral response [62], and even ahead of the conscious realization by the subject that he/she had reached a decision [63].
The voxels that contribute significant discriminant activity in these classifications indicate which brain areas are involved in a specific task. The results from these studies have changed many existing notions of localization of brain activity. In many instances, it has become clear that processing involved larger areas that are overlapping, suggesting a higher degree of multitasking than previously thought.
We have used linear classification of EEG signals on a single-trial basis to analyze rapid perceptual decisions. Linear classifiers using about 1 s of data can determine with high accuracy (>90%) when a subject has seen an object of interest in an image presented only 50–100 ms [64]. We have been able to distinguish between processing of images that contain a face from images with other objects [65]. We have also used linear classification to detect when a subject perceived to have made a mistake in a rapid perceptual decision task [66]. In Section IV, we will discuss in detail how linear classification in EEG has been used to identify the time course and spatial distribution of activity associated with perceptual decision-making.
E. Linear Versus Nonlinear Features in EEG/MEG
It is surprising that so much can be accomplished by linearly combining the raw magnitude of fMRI or EEG signals. Basic pre-possessing methods, of course, are beneficial.3 However, in EEG and MEG, the time axis does require special consideration, resulting in the use of nonlinear features. In these modalities, the temporal accuracy of the signal is better than the temporal variability of the underlying neural processing. Thus, aligning the time samples from one trial to the next is problematic. This can be addressed to some degree by locking activity to different external events such as the time of stimulus presentation or the time of a behavioral response (see Section IV for an example of this). Alternatively, one can aim to find a signal that is not locked in time. By this, we mean that the signal has no consistent sign but instead oscillates at some frequency with arbitrary phase. Therefore, instead of analyzing magnitude, a linear property of the signal, one will analyze power, a second-order feature of the signal.
Powers—or, more often, log-powers—are computed using Fourier or wavelet analysis. Essentially, the samples in time are converted into time-frequency samples, resulting typically in the same number of independent degrees of freedom. Traditional analysis has identified various relevant frequency bands such as the alpha band (8–12 Hz) and the gamma band (25–50 Hz), among others. In EEG, the power in the alpha band is very pronounced, and it has been found to decrease with the active engagement in a task and increase with active inhibition of unattended stimuli [67]. Single-trial classification of power in the alpha band has been used successfully to control the movement of a pointer on the screen in two dimensions (left versus right and up versus down) [48]. Gamma-band activity has often been associated with perceptual processing [68], but gamma oscillations are often transient and more localized in space, and thus harder to detect.4
Other nonlinear features include autoregressive (AR) model parameters that capture the shape of the power spectrum. AR coefficients are popular in brain–computer interface (BCI) research, as they give a low-dimensional parametrization of the entire spectrum without committing to any preconceived notion of a relevant frequency band. AR coefficients have often been used to classify EEG signals into more than two categories by comparing the coefficients to prototypes from various categories [49]. EEG signals have also been classified into more than two categories (in this case, spoken words) by reducing the time course of the evoked EEG activity into its strongest sinusoidal components [69]. Other popular nonlinear features are various coherence measures, power modulation, and more fanciful nonlinear features, such as various entropy and information measures, fractal dimension, Lyapunov exponents, etc. While some of these parametrizations are rather popular, in particular AR coefficients, they have not found the universal applicability of simple power measures of various frequency bands.
F. Decoding Using a Forward Model
Direct decoding as outlined above is often limited by the number of categories they can recover—in most cases, just two. To study cognition, it may be necessary to identify more than binary categories, and certainly in perception, continuous stimulus properties are of great importance. An alternative approach to direct decoding is to build a “forward model,” that is, a model that predicts brain activity that will be evoked by a given stimulus. One can then compare the predicted activity to that previously generated by various stimuli and identify the stimulus as the one that is closest to the prediction.
To identify a significant number of categories with this approach, one requires techniques that can predict brain signals with sufficient accuracy. In fMRI, this approach has allowed identification of thousands of categories [59]. However, this has required significant knowledge of the stimulus domain, which is perhaps not surprising, as one has to replicate the highly nonlinear mapping from the stimulus domain to the response of the brain. Typically, stimuli sj have to be decomposed into relevant features—say, φm(sj) represents the mth feature for the jth stimulus. In [70], for instance, images are decomposed into local signal power at multiple scales and multiple orientations. These features are then combined linearly to predict the activity of individual voxels in the brain
| (2) |
Aside from reconstructing the activity associated with images, this technique has also been used to predict brain activity associated with the meaning of nouns [71]. To do so, noun words were represented by 25 semantic features quantifying the relationship of the word to various senses such as “see,” “hear,” “taste” or actions such as “run,” “push,” “enter,” and so on). In this case, the feature value quantified the average distance of the stimulus word to the feature word within a large corpus of written text. With this technique, it was possible to identify 60 different new nouns such as “airplane” or “celery.” Evidently, the feature set used here is extremely complex and required a strong hypothesis as to the relevant feature dimensions. The value of the prediction approach outlined here lies precisely in its ability to validate scientific hypotheses as to the relevant dimensions the brain uses to encode information.
G. Regularization and Generalization
A central challenge of the decoding and prediction problems in (1) and (2) is to constrain the large number of free coefficients wi and aim. Regularization refers to the general problem of constraining or reducing the effective number of degrees of freedom in a parametrization or regression task.5 Without regularization, the degrees of freedom are severely underconstrained given the large size of the feature space, D = 104 or more, and the small number exemplars, typically N = 102 or less. Various techniques for regularization have been developed in the machine-learning community to address this problem, and they will only be broadly categorized here. All techniques essentially rely on various prior assumptions about the classification or prediction problem.
1) Finite Noise
Finding a good classifier or predictor often involves minimizing a cost function, which quantifies the errors in prediction or classification. Regularization often involves adding to this cost function an extra “penalty term,” which gives preference to some parameter values over others. For instance, a quadratic penalty term L2 = Σi|wi|2 will assign a cost to the nonzero coefficient, so that reducing the cost will reduce the coefficient to zero unless the training data provide sufficient evidence that a specific coefficient is really required. Penalized logistic regression [56] and ridge regression [72] use this technique. L2 quantifies the norm of the weight vector; hence it is also called L2 norm regularization. In LR, the L2 norm regularization is particularly useful: LR models the transition from one class to the other as a sigmoid logistic function. The norm of the weight vector corresponds to the sharpness of this transition boundary. Thus, penalizing the norm limits the sharpness of the transition boundary. This, in effect, implements the assumption that the data carry some noise even in the case that the training samples are perfectly separable (which is always the case because D > N).
2) Smoothness
If one can assume that the activity in neighboring samples is similar, then it is reasonable to assume that the corresponding weights should be similar as well. It is easy to incorporate such neighborhood relationships into a quadratic penalty term by defining the corresponding expected covariance matrix for the weights as Rww and then incorporating this covariance in the norm as (see, e.g., [73]).
3) Sparseness
The problem with a quadratic penalty term, however, is that large coefficients are penalized more than small coefficients, and so the solutions will favor a distributed representation with many small values. Alternatively, one can use L1 = Σi|wi| as penalty, which implicitly assumes that weights are sparse, i.e., most weights are zero or very small with a few very large values. Sparse priors and LASSO techniques [74] use this approach. This will encourage extraction of more localized activity and will ignore samples that do not contribute significant predictive/discriminant activity. Smoothness constraints are difficult to implement for sparse priors. Instead, in cases where different samples are expected to behave similarly, one can enforce constrain using sparse penalty terms that apply to groups of coefficients together (such as the group-LASSO technique [75] or elastic nets [76] with combine L1 and L2 norms). One should note that various feature selection algorithm exist that conceptually aim to do the same thing as sparse priors, i.e., select a few dimensions that will contribute significantly to the classification while giving all other features zero weight. We recommend to use sparse priors as a more systematic and modern approach to perform feature selection.
4) Factorized Parametrization
An alternative approach to regularization is to parametrize the weight vector with a smaller number of free parameters. For instance, one might assume that in EEG or MEG, the time and space axis are independent. If that is the case, then the weight vector can be factorized as a bilinear model with T + K parameters instead of TK (where T is the number of samples in time and K the number of electrodes) [73]. If there are more than two dimensions to organize the weight vector (for instance, there may be an additional frequency dimension), one can use more complex factorizations such as the PARAFAC model [77]. These models can lead to a drastic reduction in the number of degrees of freedom, thus reflecting rather strong assumptions on the parameter space that have to be well justified.
5) Spatial Components
An additional approach to reducing the dimensionality of the problem is to combine spatial dimensions into a subspace and then perform linear classification or prediction within that subspace. The most notable of these approaches are common spatial pattern (CSP), independent component analysis (ICA), and principal component analysis (PCA). CSP combines electrodes linearly such that the resulting aggregate signals have maximal or minimal power for two conditions, respectively [56]. These powers are then used as features for classification. CSP therefore uses information about the two different classes to combine electrodes into some optimal spatial components. ICA and PCA, in contrast, combine signals linearly without class information based on other statistical criteria [56]. These techniques may be useful in neuroimaging studies that have an insufficient number of labels or simply lack labels for categorization.
IV. Identifying the Neural Correlates of Decision-Making Using Multivariate Single-Trial Analysis of EEG and/or fMRI
A. Single-Trial EEG Reveals Temporal Evolution of Decision-Making
One of the first studies to use single-trial analysis of the EEG to explore the temporal characteristics of perceptual decision-making in humans was by Philiastides and Sajda [65]. Motivated from previous work in primates [4], [7], we quantified the relationship between neural activity and behavioral output during a simple “face”/“car” decision-making task. The difficulty of the task was manipulated by changing the spatial phase coherence of the stimuli in a range that spanned psychophysical threshold.
Using the LR approach (see Section III-C) for different time windows and phase coherence levels, we identified two EEG components whose single-trial amplitudes discriminated maximally between faces and cars [Fig. 2(a)]. The early component was consistent with the well-known face-selective N170 [78]-[82], and its temporal onset appeared to be unaffected by task difficulty. The late component (>300 ms after the stimulus) systematically shifted later in time and became more persistent (i.e., broader) as a function of task difficulty. Both of these components indexed decision accuracy in that a high magnitude discriminator output value indicated an easy trial, while values near zero indicated more difficult decisions.
Fig. 2.
Single-trial EEG components correlating with decision accuracy on our “face”/“car” categorization task. (a) Discriminant component maps for the early (≈170 ms) and late (≈300 ms) components. Trials are aligned to the onset of visual stimulation (time 0 ms) and sorted by response time (black sigmoidal curves). Only face trials are shown here. Each row in these maps represents the discriminator output for a single face trial with the mean of all car trials subtracted (i.e., ). (b) Comparison of one subject’s psychometric function (gray, solid line) with neurometric functions obtained from the early component (light gray dotted line), the late component (dark gray dotted line), and a combination of the two (black solid line).
To compare neuronal to psychophysical performance, we constructed neurometric functions by considering the data in the early and late time windows. Specifically, receiver operating characteristic (ROC) analysis was used to quantify the discriminator’s performance at each phase coherence level. The area under the ROC curves was plotted against the corresponding phase coherence levels to construct the neurometric functions. We showed that the neurometric curves were good predictors of behavioral performance as captured by the psychometric function [Fig. 2(b)]. Neurometric functions from the late component were a better match to the psychophysical data than those from the early one. Choice probability analysis [7] also revealed that the late component was a significant predictor of the content of the subject’s final decision.
Situated somewhere between the early and late components (≈220 ms poststimulus), there was a third component, the strength of which systematically increased with increasing task difficulty. This component was a good predictor of the onset time of the late component. We originally speculated that this component reflects a top-down influence of attention on decision-making rather than a mere bottom-up processing of the stimulus evidence. To substantiate this claim, we ran a variant of the original behavioral paradigm, where the same stimuli were colored red or green and the subjects were either cued to perform a “green”/“red” discrimination or the original “face”/“car” discrimination [83]. This manipulation allowed us to vary the difficulty of the task while leaving the stimulus evidence unchanged in order to treat confounding factors relating to early, bottom-up, processing of the stimulus. For example, for images with the same overall phase coherence level, we compared the single-trial amplitudes of the difficulty component during challenging face/car decisions with those of trivial color decisions. We found that the trial-to-trial amplitudes of the difficulty component were significantly reduced when the subjects were simply discriminating the color of the stimulus.
This variant of the experiment has also yielded additional evidence on the role of the early and late components. We showed that while the early component remained unaffected by task demands (in that the single-trial amplitudes of this component to face versus car stimuli remained unchanged during color discrimination), the late component was largely diminished when subjects were making a color decision. More recently, we demonstrated that the late component can also be used to index the quality of evidence used in the decision process as quantified by the drift rate parameter in a diffusion model simulation [84]. Specifically, the results showed that dividing the data from nominally identical stimuli on the basis of the late-component single-trial amplitudes (but not those of the early one) produced differences in the drift-rate estimates. Though trial-to-trial variability in drift rate is often assumed, prior to this work, there has been no independent way to measure its neurophysiological correlates.
B. EEG-Informed fMRI Reveals Spatiotemporal Characteristics of Decision-Making
Despite significant progress made in understanding perceptual decision-making in humans using EEG and fMRI in isolation, the spatial localization restrictions of EEG and the temporal resolution constraints of fMRI suggest that only a combination of these modalities can provide a full spatiotemporal characterization of this process.
One way this fusion could be achieved is by simultaneous EEG/fMRI measurements, where trial-to-trial variability in EEG components can be used to construct parametrically modulated fMRI regressors. Major technical challenges for simultaneous acquisition include 1) removal of large magnetic field gradients and radio-frequency (RF) pulses used to produce the MR images from the EEG [85], 2) special EEG amplifier design to remove the dc components without allowing the gradients to saturate the input stage [86], 3) novel EEG electrode design to minimize artifact formation [85], [87], 4) removal of cardiac-related artifacts (ballistocardiogram) [85], [88], and 5) removal of motion artifacts in the EEG, which are usually amplified when subjects are placed in an MR scanner [89]. Our group has overcome most of the technical difficulties outlined above and has been able to develop a truly simultaneous EEG and fMRI recording system [54], [90]-[94], which includes novel signal processing for artifact removal [95] and a discriminant-based multivariate analysis framework for integrating single-trial variability of EEG with fMRI [54].
Specifically, we perform single-trial discrimination to identify task-relevant components yτ, with τ indexing the time window of interest [Fig. 3(a)]. The discriminator output will have dimensions T×N, where T is the total number of training samples and N the total number of trials. Finally, to achieve more robust single-trial estimates for yτ, we typically average across all training samples to obtain
| (3) |
where i is used to index trials and j training samples. Hypothetical single-trial discriminator amplitudes can be seen in Fig. 3(b). can then be used to construct parametric fMRI regressors for each component of interest [Fig. 3(c)]. These in turn are convolved with a prototypical hemodynamic response function [Fig. 3(d)] in order to be used as fMRI data predictors in the context of a general linear model (GLM). Identifying the brain regions that correlate with each of these regressors can enable a more comprehensive characterization of the cortical network involved in different neurocognitive processes. We have already evaluated our system for yielding unique and meaningful single-trial activations in a simple auditory oddball task [54] and are currently in the process of using the system for perceptual decision-making paradigms.
Fig. 3.
EEG-derived fMRI regressors. (a), (b) The discriminator output (D.O.) is to used to acquire single-trial information for several temporally distinct EEG components. (c) The onset time and duration (δ) of each of the regressor events are determined by the onset time (τ) and duration (δ) of the EEG components. The amplitude of each regressor event will be based on the output of the linear discriminator yτ as defined in (3). (d) EEG-derived regressors are convolved with a canonical hemodynamic response function prior to the GLM analysis.
In the absence of simultaneous EEG/fMRI measurements, an EEG-informed fMRI approach can be used instead. In this case, the neural characterization proceeds in two steps. Initially, an EEG study is performed to identify components of interest. The discriminator output associated with each component and each experimental condition is then averaged across trials
| (4) |
where once again τ is used to index the component (time) of interest and C is used to index the different experimental conditions. In a second step, the same experiment is repeated while fMRI data are collected. The average discriminator output per component and experimental condition obtained from (4) will now be used to model the fMRI data. Importantly, is now a scalar—that is, in the absence of single-trial information during the fMRI session, all like trials will be modeled in the same way. Though intertrial variability is ultimately concealed in this formulation, important information regarding the localization of each of the EEG components, that would otherwise be unattainable using EEG or fMRI alone, can now be obtained.
To demonstrate the efficacy of this approach, we applied this methodology for the perceptual decision-making work presented in the previous section, where different EEG components were found to respond uniquely to different experimental manipulations/conditions. That is, all EEG-derived regressors from (4) were uncorrelated, a requirement that every sensible fMRI experimental design should satisfy.
As highlighted in the previous section, the strength of our early EEG component was proportional to the stimulus evidence (i.e., stronger for easy than hard trials), and it remained unchanged during the face/car and color discriminations. The late EEG component also responded proportionally to the stimulus evidence during the face/car discrimination, but it was stronger across all difficulty levels relative to the early one. Unlike the early component, however, it was virtually eliminated during the color discrimination. In contrast to both the early and late components, the strength of the difficulty component was inversely proportional to the amount of stimulus evidence (i.e., stronger for hard than easy trials).
As a result of these observations, we repeated this experiment in the scanner; and using (4), we constructed three parametric fMRI regressors, one for each of the early, difficulty, and late components. To modulate the heights of the corresponding regressor events, we estimated the relative strengths of our components with respect to the difficulty (i.e., low [L] versus high [H] coherence) and the type of task (i.e., face versus car [FC] or red versus green [RG]) (i.e., , , , , τ = {early,difficulty,late}).
Fig. 4 summarizes our results. For the early component, we identified significant correlations with activity in areas implicated in early visual processing of objects/faces such as the fusiform face area (FFA), the superior temporal sulcus (STS). Both the FFA and STS have previously been implicated in early visual processing of faces using neuroimaging [34], [96]-[98] and field potentials recorded directly from the cortical surface [99]-[101]. These results implicate the early component in early visual processing of the incoming sensory evidence.
Fig. 4.

Spatiotemporal processing timing diagram resulting from an EEG-informed fMRI analysis. Color bar represents z-scores for each signature (corrected for multiple comparisons). (Adapted from [102].)
The difficulty component, correlated with activity in brain regions that are typically associated with the human attentional network such as the supplementary and frontal eye fields, the anterior cingulate cortex, the DLPFC, and the anterior insula [1], [32]. These observations support the hypothesis that there exists an attentional control system that exerts top-down influence on decision-making.
Finally, we found activations correlating with the late component in the lateral occipital complex (LOC) and in the right ventrolateral prefrontal cortex (rVLPFC). Aside from its involvement in object categorization [103]-[107], the LOC has been implicated in “perceptual persistence” [108], [109], a process in which a percept assembled by lower visual areas is allowed to remain in the visual system, via feedback pathways, as a form of iconic memory [110]-[112].
The brief stimulus durations used in these experiments suggest that perceptual persistence is a likely mechanism by which rapid object decision-making takes place. That is, for brief presentations, the accumulation of evidence is not based on the decaying stimulus traces themselves but rather on a durable representation of the stimulus retained in short-term memory. This interpretation explains why the late component was a better predictor of overall behavioral performance than the early one, why it correlated strongly with drift rate in a diffusion model simulation, and why it disappeared when a demanding face versus car discrimination was no longer required (e.g., during color discrimination).
Taken together, these results reaffirm that the EEG-informed fMRI approach is a promising new tool in mapping out the spatiotemporal characteristics of different neurocognitive processes, such as perceptual decision-making, in humans.
V. Conclusion
Though perceptual decision-making is perhaps rather simplistic relative to the complex decisions we make everyday, understanding the neural processes governing even the most simple decisions will shed light on how we make decisions which involve context, reward and value. It is clear, however, that a marriage of tools and methods from both neuroscience and engineering is required in that this undertaking involves analysis of a tremendous amount of high-dimensional data, systems for acquiring the data in real-time, and sophisticated and insightful paradigms for manipulating our behavior and interpreting the results. In summary, we are just at the beginning of having the necessary tools and techniques to develop a deep understanding of how we make a decision—for example, on whether to read this paper or not.
Acknowledgments
This work was supported by the National Institutes of Health under Grant EB004730 and by the Defense Advanced Research Projects Agency under Contract NBCHC080029.
Biographies

Paul Sajda (S’91-M’93-SM’06) received the B.S. degree in electrical engineering from the Massachusetts Institute of Technology (MIT), Cambridge, in 1989 and the M.S. and Ph.D. degrees in bioengineering from the University of Pennsylvania (UPenn), Philadelphia, in 1992 and 1994, respectively.
In 1994, he joined the David Sarnoff Research Center, where he went on to become Head of the Adaptive Image and Signal Processing Group. He is currently an Associate Professor of biomedical engineering and radiology at Columbia University, New York, where he is Director of the Laboratory for Intelligent Imaging and Neural Computing. His research focuses on neural engineering, neuroimaging, computational neural modeling, and machine learning applied to image understanding.
Dr. Sajda is a Fellow of the American Institute of Medical and Biological Engineering. He has received several awards for his research, including a National Science Foundation CAREER Award (2002), the Sarnoff Technical Achievement Award (1996), The Pollack Award for outstanding dissertation research in bioengineering (UPenn, 1994), The Flexner Award for outstanding thesis research in the neurosciences (UPenn, 1993), and the Adler Award for outstanding undergraduate research in electrical engineering (MIT, 1989). He is an Associate Editor of the IEEE Transactions on Biomedical Engineering and a member of the IEEE Technical Committee on Neuroengineering.

Marios G. Philiastides received the B.S. degree in biomedical engineering from Rutgers—The State University, New Brunswick, NJ, in 2001, the M.S. degree in electrical engineering from Stanford University, Stanford, CA, in 2003, and the Ph.D. degree in biomedical engineering from Columbia University, New York, in 2007.
He is currently a Research Scientist with the Max Planck Institute for Human Development, Berlin, Germany. His research focuses on neuroimaging and computational modeling of visual and cognitive function with emphasis on perceptual and reward-based decision-making.

Lucas C. Parra (M’99-SM’07) received the Ph.D. degree in physics from the Ludwig-Maximilian University, Germany, in 1996.
He is a Professor of biomedical engineering at The City College, City University of New York. Previously, he was Head of the Adaptive Image and Signal Processing Group, Sarnoff Corporation (1997–2003), and a Member of the Machine learning and the Imaging departments at Siemens Corporate Research (1995–1997). His areas of expertise include machine learning, acoustic array processing, emission tomography, and electroencephalography. His current research in biomedical signal processing focuses on functional brain imaging and computational models of the central nervous system.
Footnotes
Because measuring magnetic resonance is very costly and requires significant infrastructure, some have proposed to image hemodynamic activity using near-infrared light, which does penetrate the scull. While this functional near-infrared technology is more accessible, its spatial resolution and penetration depth are limited by light diffusion, and its temporal resolution remains limited by the slow hemodynamic response.
Human studies have used EEG signal to extract neural activity in real-time with sufficient accuracy to control a cursor on the screen in one or two dimensions [47], [48] or to control the grip of a robotic hand [49]. Animal studies using electrodes implanted in various cortical areas have demonstrated two- and three-dimensional control of a robotic arm [50]. The obvious goal of this work is to develop a brain-controlled communication or prosthetic device to overcome the impairments resulting from spinal-cord injuries (for an overview, see [51]). Our team has used single-trial analysis to augment the performance of unimpaired subjects at a perceptual decision-making task, specifically, visual search [52].
In fMRI and MEG, usually activity is first corrected to account for head motion and in fMRI activity is sometimes filtered in time to account for the slow hemodynamic response. For EEG, the data are filtered to remove slow drifts (<1 Hz) and power-line inductive noise (60/50 Hz).
Fast oscillations are spatially more localized, presumably because fast firing can only be coherent for smaller neuronal populations. Thus the approach of spatial averaging as outlined here may not improve SNR significantly. Gamma activity is typically also more transient in time, and thus trial averaging will require many more samples unless the transient oscillations are induced reliably at a given point in time.
Note the opposing uses of this word. In statistical testing, “degrees of freedom” refers to the number of independent measurements minus the number of parameters that have been fit to the data. Instead, in the context of parameter fitting and machine-learning, “degrees of freedom” stands for the number of free parameters that are to be constrained by the measurements. We prefer the latter use of the word.
Contributor Information
Paul Sajda, Department of Biomedical Engineering, Columbia University, New York, NY 10027 USA, (psajda@columbia.edu).
Marios G. Philiastides, Max Plank Institute for Human Development, Berlin, Germany, (marios@mpib-berlin.mpg.de)
Lucas C. Parra, Department of Biomedical Engineering, The City College of New York, New York, NY 10031 USA, (parra@ccny.cuny.edu).
REFERENCES
- [1].Heekeren HR, Marrett S, Ungerleider L. The neural systems that mediate human perceptual decision making. Nat. Rev. Neurosci. 2008 Jun;9(6):467–479. doi: 10.1038/nrn2374. [DOI] [PubMed] [Google Scholar]
- [2].Gold JI, Shadlen M. The neural basis of decision making. Annu. Rev. Neurosci. 2007;30:535–574. doi: 10.1146/annurev.neuro.29.051605.113038. [DOI] [PubMed] [Google Scholar]
- [3].Romo R, Salinas E. Flutter discrimination: Neural codes, perception, memory and decision making. Nat. Rev. Neurosci. 2003;4(3):203–218. doi: 10.1038/nrn1058. [DOI] [PubMed] [Google Scholar]
- [4].Britten KH, Shadlen MN, Newsome WT, Movshon JA. The analysis of visual motion: A comparison of neuronal and psychophysical performance. J. Neurosci. 1992;12(12):4745–4765. doi: 10.1523/JNEUROSCI.12-12-04745.1992. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [5].Britten KH, Shadlen MN, Newsome WT, Movshon JA. Responses of neurons in macaque MT to stochastic motion signals. Vis. Neurosci. 1993;10(12):4745–4765. doi: 10.1017/s0952523800010269. [DOI] [PubMed] [Google Scholar]
- [6].Shadlen MN, Newsome WT. Motion perception: Seeing and deciding. Proc Nat. Acad. Sci. 1996;93:628–633. doi: 10.1073/pnas.93.2.628. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [7].Britten KH, Newsome WT, Shadlen MN, Celebrini S, Movshon JA. A relationship between behavioral choice and visual responses of neurons in macaque MT. Vis. Neurosci. 14:87–100. doi: 10.1017/s095252380000715x. [DOI] [PubMed] [Google Scholar]
- [8].Salzman CD, Britten KH, Newsome WT. Cortical microstimulation influences perceptual judgements of motion direction. Nature. 1990 Jul;346(6280):174–177. doi: 10.1038/346174a0. [DOI] [PubMed] [Google Scholar]
- [9].Salzman CD, Murasugi CM, Britten KH, Newsome WT. Microstimulation in visual area mt: Effects on direction discrimination performance. J. Neurosci. 1992 Jun;12(6):2331–2355. doi: 10.1523/JNEUROSCI.12-06-02331.1992. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [10].Ditterich J, Mazurek ME, Shadlen M. Microstimulation of visual cortex affects the speed of perceptual decisions. Nat. Neurosci. 2003 Aug;6(8):891–898. doi: 10.1038/nn1094. [DOI] [PubMed] [Google Scholar]
- [11].Keysers C, Xiao D-K, Foldiak P, Perrett DI. The speed of sight. J. Cogn. Neurosci. 2001 Jan;13(1):90–101. doi: 10.1162/089892901564199. [DOI] [PubMed] [Google Scholar]
- [12].Afraz SR, Kiani R, Esteky H. Microstimulation of inferotemporal cortex influences face categorization. Nature. 2006 Aug;442(7103):692–695. doi: 10.1038/nature04982. [DOI] [PubMed] [Google Scholar]
- [13].Salinas E, Hernandez A, Zainos A, Romo R. Periodicity and firing rate as candidate neural codes for the frequency of vibrotactile stimuli. J. Neurosci. 2000 Jul;20(14):5503–5515. doi: 10.1523/JNEUROSCI.20-14-05503.2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [14].Romo R, Hernandez A, Zainos A, Salinas E. Somatosensory discrimination based on cortical microstimulation. Nature. 1998 Mar;392(6674):387–390. doi: 10.1038/32891. [DOI] [PubMed] [Google Scholar]
- [15].Romo R, Hernandez A, Zainos A, Brody CD, Lemus L. Sensing without touching: Psychophysical perfrormance based on cortical microstimulation. Neuron. 2000 Apr;26(1):273–278. doi: 10.1016/s0896-6273(00)81156-3. [DOI] [PubMed] [Google Scholar]
- [16].Ratcliff R. A theory of memory retrieval. Psychol. Rev. 1978;85(2):59–108. [Google Scholar]
- [17].Ratcliff R, Tuerlinckx F. Estimating parameters of the diffusion model: Approaches to dealing with contaminant reaction times and parameters variability. Psychon. Bull. Rev. 2002;9(3):438–481. doi: 10.3758/bf03196302. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [18].Mazurek ME, Roitman JD, Ditterich J, Shadlen MN. A role for neural integrators in perceptual decision making. Cereb. Cortex. 2003 Nov;13(11):1257–1269. doi: 10.1093/cercor/bhg097. [DOI] [PubMed] [Google Scholar]
- [19].Bogacz R. Optimal decision making theories: Linking neurobiology with behaviour. Trends. Cogn. Sci. 2007 Mar;11(3):118–125. doi: 10.1016/j.tics.2006.12.006. [DOI] [PubMed] [Google Scholar]
- [20].Grossberg S, Pilly PK. Temporal dynamics of decision-making during motion perception in the visual cortex. Vision Res. 2008 Jun;48(12):1345–1373. doi: 10.1016/j.visres.2008.02.019. [DOI] [PubMed] [Google Scholar]
- [21].Shadlen MN, Newsome WT. Neural basis of perceptual decision making in the parietal cortex (area LIP) of the rhesus monkey. J. Neurophysiol. 2001 Oct;86(4):1916–1936. doi: 10.1152/jn.2001.86.4.1916. [DOI] [PubMed] [Google Scholar]
- [22].Roitman JD, Shadlen MN. Response of neurons in the lateral intraparietal area during a combined visual discrimination reaction time task. J. Neurosci. 2002 Nov;22(21):9475–9489. doi: 10.1523/JNEUROSCI.22-21-09475.2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [23].Huk AC, Shadlen MN. Neural activity in the macaque parietal cortex reflects temporal integration of visual motion signals during perceptual decision making. J. Neurosci. 2005 Nov;25(45):10420–10436. doi: 10.1523/JNEUROSCI.4684-04.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [24].Hanks TD, Ditterich J, Shadlen M. Microstimulation of macaque area lip affects decision-making in a motion discrimination task. Nat. Neurosci. 2006 May;9(5):682–689. doi: 10.1038/nn1683. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [25].Churchland AK, Kiani R, Shadlen MN. Decision-making with multiple alternatives. Nat. Neurosci. 2008 Jun;11(6):693–702. doi: 10.1038/nn.2123. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [26].Kim JN, Shadlen MN. Neural correlates of decision making in the dorsolateral prefrontal cortex of the macaque. Nat. Neurosci. 1999 Feb;2(2):176–185. doi: 10.1038/5739. [DOI] [PubMed] [Google Scholar]
- [27].Romo R, Brody C, Hernandez A, Lemus L. Neuronal correlates of parametric working memory in prefrontal cortex. Nature. 1999 Jun;399(6735):470–473. doi: 10.1038/20939. [DOI] [PubMed] [Google Scholar]
- [28].Romo R, Hernandez A, Zainos A, Lemus L, Brody CD. Neuronal correlates of decision-making in secondary somatosensory cortex. Nat. Neurosci. 2002 Nov;5(11):1217–1224. doi: 10.1038/nn950. [DOI] [PubMed] [Google Scholar]
- [29].Romo R, Hernandez A, Zainos A. Neuronal correlates of decision-making in ventral premotor cortex. Neuron. 2004 Jan;41(1):165–173. doi: 10.1016/s0896-6273(03)00817-1. [DOI] [PubMed] [Google Scholar]
- [30].Hernandez A, Zainos A, Romo R. Temporal evolution of a decision-making process in medial premotor cortex. Neuron. 2002 Mar;33(6):959–972. doi: 10.1016/s0896-6273(02)00613-x. [DOI] [PubMed] [Google Scholar]
- [31].Brody CD, Hernandez A, Zainos A, Romo R. Timing and neural encoding of somatosensory parametric working memory in macaque prefrontal cortex. Cereb. Cortex. 2003 Nov;13(11):1196–1207. doi: 10.1093/cercor/bhg100. [DOI] [PubMed] [Google Scholar]
- [32].Heekeren HR, Marrett S, Bandettini PA, Ungerleider LG. A general mechanism for perceptual decision-making in the human brain. Nature. 2004;431:859–862. doi: 10.1038/nature02966. [DOI] [PubMed] [Google Scholar]
- [33].Haxby JV. The functional organization of human extrastriate cortex: A PET-rCBF study of selective attention to faces and locations. J. Neurosci. 1994;14:6336–6353. doi: 10.1523/JNEUROSCI.14-11-06336.1994. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [34].Kanwisher N, McDermott J, Chun MM. The fusiform face area: A module in human extrastriate cortex specialized for face perception. J. Neurosci. 1997;17:4302–4311. doi: 10.1523/JNEUROSCI.17-11-04302.1997. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [35].McCarthy G, Puce A, Gore J, Allison T. Face-specific processing in the human fusiform gyrus. J. Cogn. Neurosci. 1997b;9:605–611. doi: 10.1162/jocn.1997.9.5.605. [DOI] [PubMed] [Google Scholar]
- [36].Epstein R, Kanwisher N. A cortical representation of the local visual environment. Nature. 1998;392:598–601. doi: 10.1038/33402. [DOI] [PubMed] [Google Scholar]
- [37].Ishai A, Ungerleider LG, Martin A, Schouten JL, Haxby JV. Distributed representation of objects in the human ventral visual pathway. Proc. Nat. Acad. Sci. 1999;96:9379–9384. doi: 10.1073/pnas.96.16.9379. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [38].Pleger B, Ruff CC, Blankenburg F, Bestmann S, Wiech K, Stephan KE, Capilla A, Friston KJ, Dolan RJ. Neural coding of tactile decisions in the human prefrontal cortex. J. Neurosci. 2006 Nov;26(48):12596–12601. doi: 10.1523/JNEUROSCI.4275-06.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [39].Preuschhof C, Heekeren HR, Taskin B, Schubert T, Villringer A. Neural correlates of vibrotactile working memory in the human brain. J. Neurosci. 2006;26:13231–13239. doi: 10.1523/JNEUROSCI.2767-06.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [40].Ploran EJ, Nelson SM, Velanova K, Donaldson DI, Petersen SE, Wheeler ME. Evidence accumulation and the moment of recognition: Dissociating perceptual recognition processes using fmri. J. Neurosci. 2007 Oct;27(44):11912–11924. doi: 10.1523/JNEUROSCI.3522-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [41].Rorie AE, Newsome WT. A general mechanism for perceptual decision-making in the human brain? Trends Cogn. Sci. 2005;9:41–43. doi: 10.1016/j.tics.2004.12.007. [DOI] [PubMed] [Google Scholar]
- [42].Heekeren HR, Marrett S, Ruff DA, Bandettini PA, Ungerleider LG. Involvement of human left dorsolateral prefrontal cortex in perceptual decision making is independent of response modality. Proc. Nat. Acad. Sci. 2006;103(26):10023–10028. doi: 10.1073/pnas.0603949103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [43].Maass W, Natschlager T, Markram H. Real-time computing without stable states: A new framework for neural computation based on perturbations. Neural Comput. 2002 Nov;14(11):2531–2560. doi: 10.1162/089976602760407955. [DOI] [PubMed] [Google Scholar]
- [44].Jaeger H, Maass W, Principe J. Special issue on echo state networks and liquid state machines. Neural Netw. 2007 [Google Scholar]
- [45].Buonomano D, Maass W. State-dependent computations: Spatiotemporal processing in cortical networks. Nature Rev. Neurosci. 2009;10(2):113–125. doi: 10.1038/nrn2558. [DOI] [PubMed] [Google Scholar]
- [46].Haynes J-D, Rees G. Decoding mental states from brain activity in humans. Nat. Rev. Neurosci. 2006 Jul;7(7):523–534. doi: 10.1038/nrn1931. [DOI] [PubMed] [Google Scholar]
- [47].Wolpaw JR, Birbaumer N, McFarland DJ, Pfurtscheller G, Vaughan TM. Brain-computer interfaces for communication and control. Clin. Neurophysiol. 2002 Jun;113(6):767–791. doi: 10.1016/s1388-2457(02)00057-3. [DOI] [PubMed] [Google Scholar]
- [48].Wolpaw JR, McFarland DJ. Control of a twodimensional movement signal by a noninvasive brain-computer interface in humans. PNAS. 2004 Dec;101(51):17849–17854. doi: 10.1073/pnas.0403504101. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [49].Pfurtscheller G, Neuper C. Motor imagery and direct brain-computer communication. Proc. IEEE. 2001 Jul;89:1123–1134. [Google Scholar]
- [50].Andersen RA, Burdick JW, Musallam S, Pesaran B, Cham JG. Cognitive neural prosthetics. Trends Cogn. Sci. 2004 Nov;8(11):486–493. doi: 10.1016/j.tics.2004.09.009. [DOI] [PubMed] [Google Scholar]
- [51].Birbaumer N. Brain-computer-interface research: Coming of age. Clin. Neurophysiol. 2006 doi: 10.1016/j.clinph.2005.11.002. [DOI] [PubMed] [Google Scholar]
- [52].Parra LC, Christoforou C, Gerson AD, Dyrholm M, Luo A, Wagner M, Philiastides MG, Sajda P. Spatiotemporal linear decoding of brain state: Application to performance augmentation in high-throughput tasks. IEEE, Signal Process. Mag. 2008 Jan;25:95–115. [Google Scholar]
- [53].Gerson AD, Parra LC, Sajda P. Cortical origins of response time variability during rapid discrimination of visual objects. Neuroimage. 2005;28(2):342–353. doi: 10.1016/j.neuroimage.2005.06.026. [DOI] [PubMed] [Google Scholar]
- [54].Goldman RI, Wei C-Y, Philiastides MG, Gerson AD, Friedman D, Brown TR, Sajda P. Single-trial discrimination for integrating simultaneous EEG and fMRI: Identifying cortical areas contributing to trial-to-trial variability in the auditory oddball task. Neuroimage. 2009 Aug;47(1):136–147. doi: 10.1016/j.neuroimage.2009.03.062. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [55].Hasson U, Nir Y, Levy I, Fuhrmann G, Malach R. Intersubject synchronization of cortical activity during natural vision. Science. 2004 Mar;303(5664):1634–1640. doi: 10.1126/science.1089506. [DOI] [PubMed] [Google Scholar]
- [56].Parra LC, Spence CD, Gerson AD, Sajda P. Recipes for the linear analysis of EEG. Neuroimage. 2005;28(2):326–341. doi: 10.1016/j.neuroimage.2005.05.032. [DOI] [PubMed] [Google Scholar]
- [57].Norman KA, Polyn SM, Detre GJ, Haxby JV. Beyond mind-reading: Multi-voxel pattern analysis of fmri data. Trends Cogn. Sci. 2006 Sep;10(9):424–430. doi: 10.1016/j.tics.2006.07.005. [DOI] [PubMed] [Google Scholar]
- [58].Formisano E, De Martino F, Valente G. Multivariate analysis of fmri time series: Classification and regression of brain responses using machine learning. Magn. Res. Imag. 2008 Sep;26(7):921–934. doi: 10.1016/j.mri.2008.01.052. [DOI] [PubMed] [Google Scholar]
- [59].Miyawaki Y, Uchida H, Yamashita O, Sato MA, Morito Y, Tanabe HC, Sadato N, Kamitani Y. Visual image reconstruction from human brain activity using a combination of multiscale local image decoders. Neuron. 2008 Dec;60(5):915–929. doi: 10.1016/j.neuron.2008.11.004. [DOI] [PubMed] [Google Scholar]
- [60].Hassabis D, Chu C, Rees G, Weiskopf N, Molyneux PD, Maguire EA. Decoding neuronal ensembles in the human hippocampus. Curr. Biol. 2009 Apr;19(7):546–554. doi: 10.1016/j.cub.2009.02.033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [61].Shinkareva SV, Mason RA, Malave VL, Wang W, Mitchell TM, Just MA. Using fMRI brain activation to identify cognitive states associated with perception of tools and dwellings. PLoS One. 2008;3(1):e1394. doi: 10.1371/journal.pone.0001394. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [62].Haynes J-D, Sakai K, Rees G, Gilbert S, Frith C, Passingham RE. Reading hidden intentions in the human brain. Curr. Biol. 2007 Feb;17(4):323–328. doi: 10.1016/j.cub.2006.11.072. [DOI] [PubMed] [Google Scholar]
- [63].Soon CS, Brass M, Heinze H-J, Haynes J-D. Unconscious determinants of free decisions in the human brain. Nat. Neurosci. 2008 May;11(5):543–545. doi: 10.1038/nn.2112. [DOI] [PubMed] [Google Scholar]
- [64].Gerson AD, Parra LC, Sajda P. Cortically-coupled computer vision for rapid image search. IEEE Trans. Neural Syst. Rehab. Eng. 2006 Jun;14:174–179. doi: 10.1109/TNSRE.2006.875550. [DOI] [PubMed] [Google Scholar]
- [65].Philiastides MG, Sajda P. Temporal characterization of the neural correlates of perceptual decision making in the human brain. Cereb. Cortex. 2006;16(4):509–518. doi: 10.1093/cercor/bhi130. [DOI] [PubMed] [Google Scholar]
- [66].Parra L, Alvino C, Tang A, Pearlmutter B, Young N, Osman A, Sajda P. Linear spatial integration for single-trial detection in encephalography. Neuroimage. 2002;17:223–230. doi: 10.1006/nimg.2002.1212. [DOI] [PubMed] [Google Scholar]
- [67].Klimesch W, Sauseng P, Hanslmayr S. Eeg alpha oscillations: The inhibition-timing hypothesis. Brain Res. Rev. 2007 Jan;53(1):63–88. doi: 10.1016/j.brainresrev.2006.06.003. [DOI] [PubMed] [Google Scholar]
- [68].Rodriguez E, George N, Lachaux JP, Martinerie J, Renault B, Varela FJ. Perception’s shadow: Long-distance synchronization of human brain activity. Nature. 1999 Feb;397(6718):430–433. doi: 10.1038/17120. [DOI] [PubMed] [Google Scholar]
- [69].Suppes P, Lu ZL, Han B. Brain wave recognition of words. Proc. Nat. Acad. Sci. USA. 1997 Dec;94(26):14965–14969. doi: 10.1073/pnas.94.26.14965. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [70].Kay KN, Naselaris T, Prenger RJ, Gallant JL. Identifying natural images from human brain activity. Nature. 2008;452:352–356. doi: 10.1038/nature06713. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [71].Mitchell TM, Shinkareva SV, Carlson A, Chang K-M, Malave VL, Mason RA, Just MA. Predicting human brain activity associated with the meanings of nouns. Science. 2008 May;320(5880):1191–1195. doi: 10.1126/science.1152876. [DOI] [PubMed] [Google Scholar]
- [72].Hoerl AE, Kennard RW. Ridge regression: Applications to nonorthogonal problems. Technometrics. 1970 Feb;12(1):69–82. [Google Scholar]
- [73].Dyrholm M, Christoforos C, Parra LC. Bilinear discriminant component analysis. J. Machine Learn. Res. 2007;8:1097–1111. [Google Scholar]
- [74].Tibshirani R. Regression shrinkage and selection via the lasso. J. Royal Statist. Soc. B. 1996;58(1):2670288. [Google Scholar]
- [75].Meier L, van de Geer S, Buhlmann P. The group lasso for logistic regression. J. Royal Statist. Soc. B. 2008;70(1):53–71. [Google Scholar]
- [76].Zou H, Hastie T. Regularization and variable selection via the elastic net. J. Royal Statist. Soc. B. 2005;67(2):301–320. [Google Scholar]
- [77].Morup M, Hansen LK, Hermann CS, Parnas J, Arnfred SM. Parallel factor analysis as an exploratory tool for wavelet transformed event-related eeg. NeuroImage. 2006;29(3):938–947. doi: 10.1016/j.neuroimage.2005.08.005. [DOI] [PubMed] [Google Scholar]
- [78].Jeffreys DA. Evoked studies of face and object processing. Vis. Cogn. 1996;3:1–38. [Google Scholar]
- [79].Bentin S, Allison T, Puce A, Perez A, McCarthy G. Electrophysiological studies of face perception in humans. J. Cogn. Neurosci. 1996;8:551–565. doi: 10.1162/jocn.1996.8.6.551. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [80].Halgren E, Raij T, Marinkovic K, Jousmaki V, Hari R. Cognitive response profile of the human fusiform face area as determined by MEG. Cereb. Cortex. 2000;10:69–81. doi: 10.1093/cercor/10.1.69. [DOI] [PubMed] [Google Scholar]
- [81].Liu J, Higuchi M, Marantz A, Kanwisher N. The selectivity of the occipitotemporal M170 for faces. Neuroreport. 2000;11:337–341. doi: 10.1097/00001756-200002070-00023. [DOI] [PubMed] [Google Scholar]
- [82].Rossion B, Joyce C, Cottrell GW, Tarr MJ. Early laterization and orientation tuning for face, word, object processing in the visual cortex. Neuroimage. 2003;20:1609–1624. doi: 10.1016/j.neuroimage.2003.07.010. [DOI] [PubMed] [Google Scholar]
- [83].Philiastides MG, Ratcliff R, Sajda P. Neural representation of task difficulty and decision-making during perceptual categorization: A timing diagram. J. Neurosci. 2006 Aug;26(35):8965–8975. doi: 10.1523/JNEUROSCI.1655-06.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [84].Ratcliff R, Philiastides MG, Sajda P. Quality of evidence for perceptual decision making is indexed by trial-to-trial variability of the EEG. Proc. Nat. Acad. Sci. 2009;106:6539–6544. doi: 10.1073/pnas.0812589106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [85].Ives JR, Warach S, et al. Monitoring the patient’s EEG during echo planar MRI. Electroencephalogr. Clin. Neurophysiol. 1993;87(6):417–420. doi: 10.1016/0013-4694(93)90156-p. [DOI] [PubMed] [Google Scholar]
- [86].Huang-Hellinger FR, Breiter HC, et al. Simultaneous functional magnetic resonance imaging and electrophysiological recording. Human Brain Map. 1995;3(1):13–25. [Google Scholar]
- [87].Logothetis NK, Pauls J, et al. Neurophysiological investigation of the basis of the fMRI signal. Nature. 2001;412(6843):150–157. doi: 10.1038/35084005. [DOI] [PubMed] [Google Scholar]
- [88].Muri RM, Felblinger J, et al. Recording of electrical brain activity in a magnetic resonance environment: Distorting effects of the static magnetic field. Magn. Res. Med. 1998;39(1):18–22. doi: 10.1002/mrm.1910390105. [DOI] [PubMed] [Google Scholar]
- [89].Hill RA, Chiappa KH, et al. EEG during MR imaging: Differentiation of movement artifact from paroxysmal cortical activity. Neurology. 1995;45(10):1942–1943. doi: 10.1212/wnl.45.10.1942-a. [DOI] [PubMed] [Google Scholar]
- [90].Goldman RI, Cohen MS, et al. Proc. Int. Soc. Magn. Res. Med. Denver, CO: 2000. Combining EEG and functional MRI: Cleaning up the electrical signals. [Google Scholar]
- [91].Goldman RI, Stern JM, et al. Acquiring simultaneous EEG and functional MRI. Clin. Neurophysiol. 2000;111(11):1974–1980. doi: 10.1016/s1388-2457(00)00456-9. [DOI] [PubMed] [Google Scholar]
- [92].Cohen MS, Goldman RI, et al. Proc. Org. Human Brain Map. Brighton, U.K.: 2001. Simultaneous EEG and fMRI made easy. [Google Scholar]
- [93].Goldman RI, Cohen MS. Simultaneous EEG and fMRI of alpha rhythm. Neuroreport. 2002;13(18):2487–2492. doi: 10.1097/01.wnr.0000047685.08940.d0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [94].Sajda P, Goldman RI, Philiastides MG, Gerson AD, Brown TR. A system for single-trial analysis of simultaneously acquired EEG and fMRI; Proc. 3rd Int. IEEE/EMBS Conf. Neural Eng.; May 2-5, 2007.pp. 287–290. [Google Scholar]
- [95].Dyrholm M, Goldman R, Sajda P, Brown TR. Removal of BCG artifacts using a non-Kirchhoffian overcomplete representation. IEEE Trans. Biomed. Eng. 2009;56(2):200–204. doi: 10.1109/TBME.2008.2005952. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [96].Puce A, Allison T, Bentin S, Gore JC, McCarthy G. Temporal cortex activation in humans viewing eye and mouth movements. J. Neurosci. 1998;18(6):2188–2199. doi: 10.1523/JNEUROSCI.18-06-02188.1998. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [97].Haxby JV, Hoffman EA, Gobbini MI. The distributed human neural system for face perception. Trends Cogn. Sci. 2000;4(6):223–233. doi: 10.1016/s1364-6613(00)01482-0. [DOI] [PubMed] [Google Scholar]
- [98].Hoffman E, Haxby JV. Distinct representation of eye gaze and identity in the distributed human neural system for face perception. Nat. Neurosci. 2000;3:80–84. doi: 10.1038/71152. [DOI] [PubMed] [Google Scholar]
- [99].Allison T, Puce A, Spencer DD, McCarthy G. Electrophysiological studies of human face perception: Potentials generated in occipitotemporal cortex by face and nonface stimuli. Cereb. Cortex. 1999;9:415–430. doi: 10.1093/cercor/9.5.415. [DOI] [PubMed] [Google Scholar]
- [100].Puce A, Allison T, McCarthy G. Electrophysiological studies of human face perception—III. Effects of top-down processing on face-specific potentials. Cereb. Cortex. 1999;9:445–458. doi: 10.1093/cercor/9.5.445. [DOI] [PubMed] [Google Scholar]
- [101].Puce A, Allison T, McCarthy G. Electrophysiological studies of human face perception—II. Response properties of face specific potentials generated in occipitotemporal cortex. Cereb. Cortex. 1999;9:431–444. doi: 10.1093/cercor/9.5.431. [DOI] [PubMed] [Google Scholar]
- [102].Philiastides MG, Sajda P. EEG-informed fMRI reveals spatiotemporal characteristics of perceptual decision making. J. Neurosci. 2007 Nov;27(48):13082–13091. doi: 10.1523/JNEUROSCI.3540-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [103].James TW, Humphrey GK, Gati JS, Menon RS, Goodale MA. The effects of visual object priming on brain activation before and after recognition. Curr. Biol. 2000;10:1017–1024. doi: 10.1016/s0960-9822(00)00655-2. [DOI] [PubMed] [Google Scholar]
- [104].James TW, Humphrey GK, Gati JS, Menon RS, Goodale MA. Differential effects of viewpoint on object-driven activation in dorsal and ventral streams. Neuron. 2002;35:793–801. doi: 10.1016/s0896-6273(02)00803-6. [DOI] [PubMed] [Google Scholar]
- [105].Grill-Spector K, Kushnir T, Hendler T, Malach R. The dynamics of object-selective activation correlate with recognition performance in humans. Nat. Neurosci. 2000;3:837–843. doi: 10.1038/77754. [DOI] [PubMed] [Google Scholar]
- [106].Grill-Spector K, Kourtzi Z, Kanwisher N. The lateral occipital complex and its role in object recognition. Vis. Res. 2001;41:1409–1422. doi: 10.1016/s0042-6989(01)00073-6. [DOI] [PubMed] [Google Scholar]
- [107].Grill-Spector K, Knouf N, Kanwisher N. The fusiform face area subserves face perception, not generic within-category identification. Nat. Neurosci. 2004;7(5):555–562. doi: 10.1038/nn1224. [DOI] [PubMed] [Google Scholar]
- [108].Ferber S, Humphrey GK, Vilis T. The lateral occipital complex subserves the perceptual persistence of motion-defined groupings. Cereb. Cortex. 2002;35:793–801. doi: 10.1093/cercor/13.7.716. [DOI] [PubMed] [Google Scholar]
- [109].Large ME, Aldcroft A, Vilis T. Perceptual continuity and the emergence of perceptual persistence in the ventral visual pathway. J. Neurophysiol. 2005;93:3453–3462. doi: 10.1152/jn.00934.2004. [DOI] [PubMed] [Google Scholar]
- [110].Di Lollo V. Temporal characteristics of iconic memory. Nature. 1977;267:241–243. doi: 10.1038/267241a0. [DOI] [PubMed] [Google Scholar]
- [111].Coltheart M. The persistences of vision. Philos. Trans. Royal Soc. B. 1980;290:57–69. doi: 10.1098/rstb.1980.0082. [DOI] [PubMed] [Google Scholar]
- [112].VanRullen R, Koch C. Visual selective behavior can be triggered by a feed-forward process. J. Cogn. Neurosci. 2003;15(2):209–217. doi: 10.1162/089892903321208141. [DOI] [PubMed] [Google Scholar]


