Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2018 Feb 1.
Published in final edited form as: J Math Psychol. 2016 Jul 25;76(B):80–93. doi: 10.1016/j.jmp.2016.06.008

Integrating Theoretical Models with Functional Neuroimaging

Michael S Pratte 1,2, Frank Tong 2
PMCID: PMC5342841  NIHMSID: NIHMS801640  PMID: 28286346

Abstract

The development of mathematical models to characterize perceptual and cognitive processes dates back almost to the inception of the field of psychology. Since the 1990s, human functional neuroimaging has provided for rapid empirical and theoretical advances across a variety of domains in cognitive neuroscience. In more recent work, formal modeling and neuroimaging approaches are being successfully combined, often producing models with a level of specificity and rigor that would not have been possible by studying behavior alone. In this review, we highlight examples of recent studies that utilize this combined approach to provide novel insights into the mechanisms underlying human cognition. The studies described here span domains of perception, attention, memory, categorization, and cognitive control, employing a variety of analytic and model-inspired approaches. Across these diverse studies, a common theme is that individually tailored, creative solutions are often needed to establish compelling links between multi-parameter models and complex sets of neural data. We conclude that future developments in model-based cognitive neuroscience will have great potential to advance our theoretical understanding and ability to model both low-level and high-level cognitive processes.

Introduction

The general goal of cognitive psychology has been to understand psychological processes at what Marr (1982) would call the algorithmic or representational level (see Love, 2015). In order to explore the algorithms and representational structures that might underlie processes such as attention or memory, cognitive psychologists often propose formal theoretical models of these processes, and test them by assessing predicted patterns in behavior. For decades, doing so has provided remarkable insights into how the mind works. During the same time, the field of neuroscience has made strides in understanding what Marr calls the implementation level, or, how these processes are implemented in the biological machinery that makes up the brain. More recently, the merger of these fields into a unified cognitive neuroscience has resulted in part from the development of new neuroimaging techniques, such as functional magnetic resonance imaging (fMRI), which have made investigating the biological substrates of human cognition possible. The more targeted approach of combining theoretical modeling and neuroscience has been termed computational neuroscience, a field that is often credited as originating from Marr’s work. In this review, we consider a newly emerging endeavor that has arisen from the merger of cognitive psychology, theoretical modeling and neuroscience: Using theoretical models in conjunction with human neuroimaging to study psychological processes.

Advances in mathematical and computational approaches have played a key role in fMRI since its invention. Some of these developments include analytical approaches for extracting relevant information from the BOLD signal across the temporal domain, such as the use of temporal phase-encoded designs (Engel, 2012; Engel et al., 1994; Sereno et al., 1995) and the development of de-convolution approaches for fast-event related designs (Boynton, Engel, Glover, & Heeger, 1996; Buckner et al., 1996; Glover, 1999). Developments in inferential statistical techniques have produced a number of tools that have helped make fMRI mapping studies so successful, especially regarding techniques to account for what is arguably the most serious multiple comparisons problem in psychology (see Nichols & Hayasaka, 2003). Given the multivariate nature of fMRI data, correlation-based approaches (Haxby et al., 2001), machine learning techniques (Kamitani & Tong, 2005; Norman, Polyn, Detre, & Haxby, 2006; Tong & Pratte, 2012) and voxel-based modeling approaches (Brouwer & Heeger, 2009; Kay, Naselaris, Prenger, & Gallant, 2008; Serences & Saproo, 2009) have been used to capture the complexities of these high-dimensional data sets, providing powerful new ways of identifying perceptual information contained in brain activity patterns (Harrison & Tong, 2009; Haynes & Rees, 2005; Kamitani & Tong, 2005; Serences & Boynton, 2007), as well as evidence of semantic tuning properties (Huth, Nishimoto, Vu, & Gallant, 2012; Mitchell et al., 2008). The correlational structure of activity patterns has also been used to characterize object representations in the ventral temporal lobe (Haxby et al., 2001; Kriegeskorte et al., 2008). Correlation-based approaches have also proven useful for delineating the functional connectivity of the human brain (e.g., Honey et al., 2009), including resting state networks (Fox et al., 2005), and more recent studies of brain connectivity have benefitted from the application of graph theoretical models (Bullmore & Sporns, 2009) and other model-based approaches (Tavor et al., 2016). Such advances in analytic methods continue to expand the ways in which fMRI can be used to study the brain.

More recently, research on new mathematical approaches for fMRI has evolved beyond the goal of simply developing more powerful analytic methods, to that of integrating and testing cognitive models. This model-based approach to cognitive neuroscience represents an exciting development that goes beyond the simpler goals of “brain mapping”, identifying correlations between individual differences and brain activity, or information-based approaches to characterize cortical function. Instead, the goal of model-based cognitive neuroscience lies in describing the perceptual or cognitive processes that underlie behavior in a mathematically precise manner, and determining the neural processes that underlie these computations.

Cognitive process models have a long history in the study of human performance. For example, models based on signal detection theory have served as the foundation for studying perception (Green & Swets, 1966), attention (Lu & Dosher, 1998) and memory (Kintsch, 1967). Early cognitive research also demonstrated that stochastic accumulator models can accurately predict patterns of choice reaction times across numerous behavioral paradigms (Ratcliff & Rouder, 1998; Stone, 1960). While some cognitive process models have focused on identifying and quantifying a few key parameters to capture patterns of cognitive performance, other models rely on general learning principles to train complex networks with numerous parameters to perform a cognitive task. For example, neural network models (e.g., McClelland & Rogers, 2003) have been developed to characterize high-level processes including speech perception (McClelland & Elman, 1986), categorization (Ashby & Maddox, 1993; Nosofsky & Palmeri, 1997), cognitive control (Botvinick, Braver, Barch, Carter, & Cohen, 2001), and human memory (Polyn, Norman, & Kahana, 2009).

One might expect that the application of theoretical models to neuroimaging data would be a naturally obvious and fruitful endeavor. However, most cognitive neuroscientists have not rushed to meet this challenge until recently. Why has this been the case? A central challenge lies in establishing strong links between the parameters of a cognitive model and particular brain responses embedded within a cognitive experiment. Cognitive models typically rely on latent constructs of presumed psychological processes that must somehow be translated into a predicted pattern of brain responses. If the model leads to clear predictions regarding how the univariate BOLD responses should change over time, such models may be more straightforward to test using standard fMRI analysis procedures. Earlier applications of model-based fMRI have relied on such approaches to identify the neural correlates of reward prediction error (e.g., J. P. O’Doherty, Dayan, Friston, Critchley, & Dolan, 2003; J.P. O’Doherty et al., 2004; J. P. O’Doherty, Hampton, & Kim, 2007) and response conflict (e.g., Botvinick, Cohen, & Carter, 2004). However, fMRI data is very high-dimensional, such that establishing links between cognitive models and the information contained in multivariate brain activity patterns is considerably more challenging. Even if a model can be positively related to an information-based metric of brain processing, the next step of determining whether a particular model provides a compelling fit of the high-dimensional brain data can be difficult to demonstrate.

In this review, we highlight several recent studies that have successfully combined theoretical models with fMRI data, addressing diverse questions spanning lower-level perceptual processes to higher-level cognitive processes. A central theme across these studies is the goal of identifying compelling relationships between brain, model and behavior (see Figure 1A), often with the cognitive model serving as the intermediary for mapping between brain and behavior. However, as we will see, there are many possible options and approaches for establishing these links, as a model fitted to behavioral data might be used to predict brain responses, or brain data might be incorporated into a model to predict behavior. Moreover, intermediate processing steps may take place before links are established, such as methods to reduce the high dimensionality of brain data to lower-dimensional measures that can be more directly related to model predictions.

Figure 1.

Figure 1

Approaches To Linking Models With Brain and Behavior. A) The goal of a model-based cognitive neuroscience is to develop theory by forming links between brain activity, behavior, and a theoretical model that describes these data. B) In the approach taken by Brouwer and Heeger (2011) and Pratte et al. (2013), a multivariate pattern analysis is first used to transform the brain activity data into a lower-dimensional space, and the model is then applied to these more interpretable data. C) In the approach taken by Kragel et al. (2015) the brain activity signal is directly incorporated within the model, and statistical tests are used to determine whether this inclusion produces a more accurate account of the behavioral data than a model without the neural data. D) Mack et al. (2013) first fitted a model to behavioral data, and used the fitted model to decode patterns of brain activity. They then compared this decoding performance across different models to determine which model provided the best correspondence with the neural data. E) Ide et al. (2013) first fited a model to behavioral data, and then used a regression analysis to identify fMRI voxels that co-varied with various model constructs over the course of the experiment.

We begin this review by discussing an application of the normalization model to the visual perception of orientation (Brouwer & Heeger, 2011). In this work, fMRI data from early cortical visual areas was first transformed into interpretable constructs using a multivariate modeling approach, and the normalization model was then fitted to the resulting measurements. We then considered an application of models of visual attention to fMRI data (Pratte, Ling, Swisher, & Tong, 2013). Here, a multivoxel pattern classification approach was used to transform multivariate fMRI data, obtained from multiple levels of the visual hierarchy, into an interpretable measure of information representation, and the theoretical model was fitted to the result. Whereas in both of the earlier studies, a formal model was fitted to information decoded from the mutlivariate fMRI signal (Figure 1B), we next consider a study in which the time course of the fMRI signal on individual trials was incorporated within a theoretical model of behavioral memory performance (Figure 1C), to determine whether this neural signal can lead to more accurate predictions of free recall performance (Kragel, Morton, & Polyn, 2015). An fMRI study of categorization highlights yet another approach (Figure 1D), by assessing the degree to which competing models of behavioral categorization performance can account for observed patterns of neural data (Mack, Preston, & Love, 2013). Finally, we review a study of cognitive control that demonstrates how the application of a theoretical model to fMRI data can reveal new insights about neural processing that would have been impossible without the model (Ide, Shenoy, Yu, & Li, 2013). Here, a model of behavioral performance in the stop-signal task was incorporated within the fMRI analysis (Figure 1E), and the results suggest that the function of the anterior cingulate is more specific than has been suggested by previous studies.

This collection of studies demonstrates both the feasibility and potential of model-based cognitive neuroscience. The approaches are remarkably diverse, both in how the data are used to inform the model, and in the technical solutions employed to establish compelling relationships between brain, model and cognitive performance (see Figure 1). Of particular interest is that fact that none of the reviewed works share exactly the same strategy to link a theoretical model with functional imaging data. Rather, these examples underscore how individual studies have relied on clever innovations that are custom-built for a particular model or experimental paradigm. As such, we neither foresee nor prescribe a one-size-fits-all approach to model-based cognitive neuroscience. Rather, we believe that the diversity in attempts to integrate modeling and functional brain imaging will forward the advance of theoretical models with a momentum that would not happen if these fields remained isolated from one another.

Perception: The Normalization Model

A major goal in the study of visual perception is to understand how neural responses to a stimulus are affected by the presence of other stimuli in the visual field. One of the most widely studied models of such stimulus interactions, the normalization model (see Carandini & Heeger, 2012 for a review) has been rigorously tested using behavioral, neurophysiological, and more recently, human neuroimaging approaches. The normalization model posits that the response of a neuron to a stimulus will depend on its sensitivity to the stimulus, and further that this response is divisively reduced by the concurrent activity of other neighboring neurons. The activity of these other neurons is driven largely by the presence of other stimuli, such that the model accounts for a myriad of behavioral and neural effects resulting from stimulus interactions. The purported benefits of such a normalization process to the perceptual system include reducing the overall neural and metabolic demands within a local region, de-correlating the responses of neurons with different stimulus preferences, and providing for more efficient neural coding.

Before considering interactions between multiple stimuli, a simpler case of normalization can be understood by considering how the response of a single neuron changes as a stimulus changes size. Neural responses in many brain areas, such as the lateral geniculate nucleus (LGN) and striate cortex (V1), increase monotonically as a function of stimulus contrast. These responses typically follow a sigmoidally shaped response function, with a signature compressive effect at high stimulus contrasts (Figure 2A). For example, consider a neuron in the LGN that responds to stimulation within its receptive field, in a manner that depends on the stimulus contrast (C). According to the normalization model, the activity within a pool of nearby cells that also respond to this stimulus (Cm) will divisively dampen the response of the target neuron R(C),

R(C)=Cσ+Cm

where parameter σ determines the shape of the contrast-response function1. The population of neurons in the denominator (Cm) is referred to as the normalization pool, and the summed responses of these neurons act to normalize the response of the target neuron. When the response of a neuron is plotted as a function of contrast, increased activity in the neuron’s normalization pool will act to suppress responses of the target neuron by an amount that depends on the stimulus contrast. For example, Figure 2A shows that the response of an LGN neuron is suppressed as the stimulus is enlarged from 2° to 20°, growing beyond that neuron’s receptive field (Bonin, Mante, & Carandini, 2005). According to the normalization model, a larger stimulus will activate more nearby LGN neurons that make up the target neuron’s normalization pool, thereby suppressing the response of the target neuron. When the stimulus is made to be smaller than 2°, the response again decreases as the stimulus no longer occupies the entire receptive field. Overall, the relationship between stimulus size and the response of the neuron is non-monotonic at high stimulus contrasts, and the normalization model provides an accurate account of the shape of the contrast-response function, and the complex ways in which its shape changes across many stimulus manipulations.

Figure 2.

Figure 2

Normalization Model of Perception. A) Responses of an LGN neuron to a stimulus within its receptive field, plotted as a function of stimulus contrast (Data from Bonin et al., 2005, Figure adapted from Carandini and Heeger, 2012). Colors denote responses for various stimulus sizes; lines show fits of the normalization model. B) Example of cross-orientation suppression stimuli used by Brouwer and Heeger (2011), where targets of various contrast are presented alone (top) or with a superimposed mask (bottom). C) Orientation tuning curves that serve as basis functions for the forward encoding model. The orientation-tuned response of each voxel is modeled as a weighted sum of these response curves, with weighs estimated from the data. D) Target-tuned responses in V1 as a function of target contrast presented alone (open points) or with a superimposed mask (filled points) from Brouwer and Heeger (2011). E) Mask-tuned responses in V1 as a function of target contrast when the mask was not presented (open points) and when it was superimposed on the target (filled points). Panels D and E adapted from Brouwer and Heeger (2011).

The normalization model not only accounts for the neural interactions that may be evoked by a single stimulus, it provides a parsimonious account of the non-linear interactions that can take place when multiple stimuli are presented, such as superimposed gratings of differing orientation (see Figure 2B). Consider a target population of neurons with responses that depend on the contrast (C) and the orientation (θ) of a stimulus within the population’s receptive field, where responses to various orientations are described by some function f(θ). The responses of nearby neurons will also depend on the contrast and orientation of the stimulus according to their own tuning curves fm(θ). These tuning curves are incorporated into the normalization model by assuming that the response of a neuron to a particular stimulus is weighted by the neuron’s tuning function and divisively normalized by the response of neighboring neurons to that same stimulus,

R(C,θ)=f(θ)Cσ+fm(θ)Cm

A major goal of the normalization model is to account for neural responses to simultaneously presented stimuli. A prominent example is cross-orientation suppression, in which responses to a preferred target orientation are suppressed by the presence of a masking grating with an orthogonal orientation. For example, consider the case when a vertical target stimulus of various contrasts is presented alone, or superimposed with a horizontal mask of 50% contrast (Figure 2B). The normalization model is easily amended to account for this situation. Let C1 and C2 denote the contrast of target and mask, respectively, and θ1 and θ2 denote their orientations. The total response of the target population to both stimuli becomes:

R(C1,C2,θ1,θ2)=f(θ1)C1+f(θ2)C2σ+fm(θ1)C1,m+fm(θ2)C2,m

The model simply states that the total response reflects the summed response to both target and mask, but that this response is dampened by the summed responses of neurons in the normalization pool to both stimuli.

Brouwer and Heeger (2011) tested the predictions of the normalization model for cross-orientation suppression in humans using fMRI, following the steps shown in Figure 1B. Each measured fMRI voxel reflects the pooled response from many neurons, and in an early visual area such as the primary visual cortex (V1), most neurons exhibit strong selectivity for local orientation. The orientation-selective response of a voxel that encompasses many such neurons can be described by a combination of the tuning curves of the underlying neurons. Brouwer and Heeger (2011) developed a forward modeling approach (Brouwer & Heeger, 2009; Kay et al., 2008) in which the responses of a voxel to different orientations are modeled as a linear combination of orientation-tuned basis functions. For example, Brouwer and Heeger modeled orientation responses using 6 evenly spaced raised cosine profiles to cover the space of all orientations (Figure 2C), and the amplitude of response for each of these 6 tuned populations could then be estimated for each voxel by using linear regression applied to a subset of the fMRI data. The fitted model is then inverted and applied to voxel responses in another data set, providing an overall measure of how strongly particular orientations are activated across the entire population of voxels in V1. This analysis allowed the researchers to separately measure the strength of V1 responses to various orientations. Consequently, in a cross-orientation scenario in which two stimuli are presented simultaneously, the combined response of voxels in V1 could be decomposed to provide estimated responses to each stimulus separately.

In their experiment, Brouwer and Heeger (2011) showed vertically oriented target stimuli of various contrasts either alone, or superimposed with a horizontally oriented mask stimulus of 50% contrast (Figure 2B). Orientation preferences were first determined for voxels in V1, and then this estimated forward model was used to provide separate measures of V1 responses to target and mask stimuli. Figure 2D shows the resulting contrast response functions for V1 responses to the target orientation, and the fitted normalization model. In the absence of a mask, the target-tuned response rises with target contrast in a way that is accounted for well by the model. The addition of the superimposed horizontal mask reduces target-tuned responses, and this gross effect is the signature of cross-orientation suppression. Specifically, the effect of the mask is to reduce target-tuned responses for low-contrast targets, but it has a weaker suppressive influence when the target grating’s contrast exceeds that of the masking grating, as predicted by the normalization model.

Figure 2E shows V1 responses to the mask-tuned orientation, as a function of target contrast. In the absence of a mask (Target Only condition), these responses are invariant to target contrast as expected. However, the presence of the horizontal mask leads to large horizontally tuned responses for low target contrasts, and these responses systematically decrease as the target contrast is increased. This negative-going curve is also predicted by the normalization model: as the target-tuned cells become more strongly activated by increasing target contrast, these target-tuned responses will act to normalize responses to the mask, leading to greater inhibition of the mask-tuned cells at high target contrasts.

The normalization model provides a simple and accurate account of how fMRI responses behave in this rather complex stimulus paradigm. This study relied on a sophisticated approach to decompose multivariate V1 activity into target-selective and mask-selective responses. This approach therefore provides a starting point for further investigations of the normalization model, such as its application to other feature domains (e.g., Moradi & Heeger, 2009) and to higher-level processes such as attention (e.g., Herrmann, Heeger, & Carrasco, 2012).

Attention: The Perceptual Template Model

The effect of attending to a particular stimulus is often characterized as enhancing the perceived quality of the attended stimulus. A prominent example of this effect is when the to-be-attended stimulus is embedded in external noise, such as trying to focus on a person’s voice at a crowded party or trying to see through a rain-spattered windshield. In an effort to describe the mechanisms that underlie our ability to enhance the representation of an attended stimulus in the presence of perceptual noise, Lu and Dosher (1998) constructed a model based on core concepts of signal detection theory and signal processing, which they termed the Perceptual Template Model (PTM). The PTM model assumes that attending to a noisy stimulus has the effect of increasing the signal-to-noise ratio of the representation of that stimulus. As such, attention can do one of two things: Increase the strength of the incoming input (amplification), or decrease the irrelevant noise (noise reduction). Whereas the amplification mechanism simply increases responses to all of the incoming information (i.e., both signal and noise), the noise-reduction mechanism selectively decreases responses to the external noise. To do so, the model assumes that there exists a perceptual template of the relevant signal, which allows the brain to discriminate between signal and noise. Since its inception, this model has been applied and tested behaviorally in many contexts (see Carrasco, 2011 for a review). More recently, the PTM model has been used with fMRI to determine the neural correlates of these purported attentional processes.

In the vast majority of fMRI studies, the effect of attention is to increase the mean BOLD signal, for example, within retinotopic visual areas corresponding to the attended part of space (e.g., Buracas & Boynton, 2007; Kastner, Pinsk, De Weerd, Desimone, & Ungerleider, 1999). The PTM model, however, makes the opposite prediction for a specific region of the stimulus space: If attention acts as a noise-reduction mechanism in a brain region, then attending to a low-contrast stimulus embedded in high levels of external noise should reduce the total response to signal plus noise, as responses to the noise are reduced but responses to the signal are not affected. If the mean BOLD signal in a region reflects the total activity of cells responding to both signal and noise, then attention should reduce mean BOLD responses for low-contrast signals presented in noise. Lu, Li, Tjan, Dosher, and Chu (2011) found support for this prediction, showing that attending to low-contrast gratings in high levels of external noise led to reduced mean BOLD responses in V1, whereas attention increased responses to the low-contrast grating when it was presented alone.

These results illustrate that it is possible to test theoretical models with fMRI by making predictions about mean BOLD, especially by exploring predicted interactions with parametric manipulations of the stimulus. However, the formal PTM model specifies how amplification and noise reduction affect the signal-to-noise ratio of stimulus information, rather than the total neural response as is measured with mean BOLD. In particular, consider a stimulus with contrast C that is embedded in visual noise (termed external noise) with contrast NE. The goal is to predict some measure of the sensitivity to a signal relative to the noise (d′) for a neural population tuned to some property of the stimulus (e.g. orientation). Sensitivity varies as a function of the contrast of the stimulus and the contrast of the noise. In the absence of attention, sensitivity is proportional to the stimulus strength, and lowered by external noise:

d(C,NE)=CNE2+NI2

where NI is the level of internal noise inherent in the perceptual system, and determines the shape of the response which varies as a function of stimulus contrast (C) or external noise contrast (NE). Within the denominator, external and internal noise sources are combined, and the combined effects act to degrade the representation of the stimulus.

According to the PTM model, there are two ways that attention can act on the incoming signal and noise. An amplification mechanism (a) will multiplicatively enhance the entire input signal, thus increasing responses to both the signal and the external noise. As a consequence, amplification can improve the representation of the stimulus if external noise ranges from negligible levels to levels not much greater than that of internal noise. In contrast, a selective noise-reduction mechanism (r) will act solely to decrease the level of external noise without affecting the stimulus:

d(C,NE)=aC(aNEr)2+NI2

This model makes strong predictions about how attention should affect sensitivity to a signal in a population of cells. A noise-reduction mechanism would significantly improve the quality under conditions of high external noise by reducing responses to the noise, but in the absence of noise, a noise-reduction mechanism would have nothing to do and would lead to no net benefit. This pattern may be contrasted with the effects of a non-selective amplification mechanism, which will improve visual processing in low levels of external noise and comparatively high levels of internal noise: This is achieved by boosting the representation of both signal and external noise, such that internal noise has less of an impact. However, amplification will be ineffective when external noise levels are higher than internal noise, as both the signal and noise will be increased, leading to no net benefit.

In order to assess the correspondence between this model’s predictions and fMRI responses, Pratte et al. (2013) adopted a multivariate pattern classification approach to measure the amount of stimulus information contained at multiple levels of the cortical visual hierarchy about the viewed stimulus category. The accuracy of fMRI classification served as a measure of discrimination sensitivity to link to the predictions of the model (see Figure 1B).

The stimuli were comprised of line drawings of faces, houses, shoes and chairs embedded in varying levels of external noise (Figure 3A). Participants either attended to the stimuli by performing a one-back matching task on the images, or they performed a demanding task on letters presented at fixation, effectively withdrawing attention away from the object images. To measure neural sensitivity to the stimuli in a group of voxels, a multivariate pattern classifier was trained to identify which stimulus type (face, house, shoe or chair) was being viewed in a given block, based on the measured fMRI activity patterns in each cortical visual area. The ability to predict the viewed stimulus category based on activity patterns served as a measure of stimulus-specific sensitivity. This measure of classification accuracy was assessed separately for each external noise level and each attention condition.

Figure 3.

Figure 3

Perceptual Template Model of Attention. A) Example of stimuli used in Pratte et al. (2013), where here a house image is embedded in various levels of external noise. B) Accuracy of a pattern classifier trained on V1 responses to identify the category of a viewed object (face, house, chair or shoe) plotted as a function of external noise level, separately for when the objects were attended (filled points) or un-attended (open points). Lines denote the fitted perceptual template model, and stars indicate significant attention effects at each noise level. C) Classification accuracy for activity patterns in the FFA. D) Estimates of noise reduction effects of attention in various visual areas obtained by fitting the PTM model to classification data from various regions of interest, arranged loosely from early visual areas to higher-level visual areas. E) Estimates of amplification effects of attention. Figures adapted from Pratte et al. (2013)

Figure 3B shows classification performance for V1 as a function of external noise level, separately for when the stimulus was attended and when attention was withdrawn from the object images. As expected, adding external noise to the stimulus systematically lowered classification performance, regardless of whether the stimulus was attended to or not. The influence of attention, however, depended on the external noise level: Attention enhanced classification performance when stimuli were embedded in high levels of noise, but had no effect for stimuli presented in low or moderate noise levels. This pattern of results implies that a pure noise-reduction mechanism operates in V1: Attention is only effective in the presence of high levels of external noise. This can be contrasted with the results found in the fusiform face area (FFA, Figure 3C), an area in the ventral temporal cortex that shows strong selectivity for faces (Kanwisher, McDermott, & Chun, 1997). Unlike V1, attention increased the sensitivity of the FFA to stimuli presented in both low- and high-noise regimes. The attentional enhancement observed across all noise levels suggests that both noise-reduction and amplification mechanisms impact the contents of visual representations in these high-level brain regions.

The lines in Figures 3B and 3C show predictions of the fitted PTM model, which captures the effects of both external noise and attention on fMRI classification performance quite well. The model was fitted to data from several brain regions in the visual cortex, providing estimates of noise reduction and amplification in each. Figures 3D and 3E show the resulting parameter estimates of noise reduction and amplification, respectively. Whereas noise reduction appears to occur throughout the early visual cortex, amplification is absent in early areas V1 and V2, but becomes more prominent in higher-level areas. These results have important implications for how attention may operate to improve visual processing, first by relying on attentional templates tuned to the target to filter out external noise, and subsequently by amplifying these filtered responses to enhance overall responses, effectively overcoming the effects of internal noise.

Despite tackling a very different problem, the methodological approach of Pratte et al. shares similarities with the previous study by Brouwer & Heeger’s (2011). Both relied on an analytic approach to isolate stimulus-specific components of the fMRI responses, and then compared these observed responses with those predicted by a fitted cognitive model. In Pratte et al.’s study, a multivariate classification approach was used to estimate the amount of stimulus information in a visual area, which could then be mapped onto the predictions and specific parameters of the PTM model. We believe that using multivariate pattern analyses as a transformational step between the dense spatiotemporal information in the original fMRI data and the predictions of a formal model provides a powerful framework for future studies. However, as we will discuss below, there are other promising ways to link fMRI responses with theoretical models.

Memory: The Context Maintenance and Retrieval Model

Theoretical modeling has played a central role in the study of memory (see Raaijmakers & Shiffrin, 2002 for a review). At the same time, human neuroimaging has been used successfully to examine various mnemonic processes including long-term memory (see Henson, 2005; Wagner et al., 1998), free recall (Polyn, Natu, Cohen, & Norman, 2005) and visual working memory (Ester, Serences, & Awh, 2009; Harrison & Tong, 2009; Marois, 2016; Sprague, Ester, & Serences, 2014; Xu & Chun, 2006). However, bridging these approaches by applying formal memory models to neural data has been largely absent. One reason may be that, unlike the relatively simple normalization and perceptual template models considered above, memory models are often extremely complex. For example, neural network models of memory contain multiple layers, many units and numerous weighted connections, such that its performance cannot be expressed in a closed-form equation. As a consequence, their model predictions can only be approximated via simulation, and their parameters must be fitted with advanced minimization routines such as differential evolution, or after extensive training of the network. With such complexity, applying these models even to simple behavioral data can be challenging, such that the task of linking model predictions to neural data is not straightforward.

Recently however, Kragel et al. (2015) successfully established links between a well-studied model of free recall and fMRI measures of brain activity, providing insights into neural mechanisms underlying memory processing that would not have been possible otherwise. A challenge in understanding human memory is the organization and structure that people naturally impose upon any set of learned items. This structure is particularly evident in studies that allow for free recall, in which participants are presented with a list of items to study and are then asked to recall as many words as they can in whatever order they choose. Several characteristics of free recall data have been pivotal for theory development, such as the prominence of primacy and recency effects. One particularly interesting feature of free recall is the strong influence of temporal context: Upon recalling a word at test, people are more likely to recall words that were nearby in the study list (with the subsequent word more likely to be reported than the preceding word). This pattern of behavioral data has been interpreted to suggest that during the study of a word, the characteristics of the word are stored along with the temporal context in which the word was presented, which includes previous study words. At retrieval, the activation of a memory trace for a particular word will also activate the context with which that item was stored, and this heightened activity helps bring to mind other studied items that occurred nearby in time.

This notion of temporal reinstatement has been formalized in the Context Maintenance and Retrieval Model (CMR, Polyn et al., 2009), a member of a large class of temporal context models (Howard & Kahana, 2002). According to the CMR model, a studied item can be represented by a feature vector, as is standard in many models of memory. Additionally, a separate vector represents the context at the time of study for a given item. The context surrounding a word at study includes the previously studied words, and will be more greatly influenced by more recent words. Consequently, this context vector changes gradually over the course of the study epoch, and serves to link the memory representations of temporally proximal words. At test, recalling an item will re-activate the context state associated with that item, thus increasing the probability of recalling words that were studied nearby.

Kragel et al. (2015) posited that changes in brain activity during the recall period may reflect these dynamics, driven by two different processes: 1) More or less brain activity in an area might correspond to how well a word is being recalled, termed retrieval success, and 2) Activity may be related not to the item strength, but rather to the degree with which the context surrounding the item is being activated, termed temporal reinstatement.

There is nothing inherent in the experimental design of a free recall task that can be used to decompose the neural signal into these two separate processes of temporal reinstatement and general memory performance. Instead, Kragel et al. devised an approach of first fitting a standard CMR model to the behavioral data (i.e. the recall events) collected during fMRI scanning, in which parameters were found by maximizing the likelihood of the behavioral data given the model. They then fitted enhanced versions of the model in which the fMRI signal from a given voxel was allowed to modulate a specific parameter of the CMR model over the time course of the recall period (see Figure 1C). A voxel’s average response to each recall event could either modulate the overall success rate of memory performance during the course of recall, or it could modify the strength of temporal reinstatement during recall. If the resulting model, informed by this modulatory source of fMRI signal, was more likely to have produced a participant’s particular behavioral data than the model without fMRI modulation included, then the voxel was deemed to be related to the particular process (accounting for model complexity using AIC). By allowing the response of each voxel to affect one, both or neither model parameter, the researchers could identify voxels that were involved in memory performance, temporal reinstatement, or both.

The results of this analysis provided novel evidence to implicate distinct regions of the medial temporal lobe in temporal reinstatement (Figure 4A) and retrieval success (Figure 4B). A spatial gradient can be observed along the posterior-anterior axis, whereby more anterior voxels are involved in retrieval success and more posterior voxels co-vary with the level of temporal reinstatement. The authors used a statistical model comparison approach to assess the merit of various models, to determine which regions of interest were related to retrieval success (RS), temporal reinstatement (TR), or both processes. The model comparison results further supported the notion of posterior-anterior gradient both in the hippocampus (Figure 4C) and the medial temporal cortex (Figure 4D). Taken together, these results suggest that when posterior regions showed increased neural activity, there was an increased likelihood that the next recalled item would be one that was studied nearby in time. In contrast, activity in anterior regions did not relate to the temporal organization of recalled items; instead, lower activity indicated a failure in retrieval success.

Figure 4.

Figure 4

Context Maintenance and Retrieval Model of Memory. A) Voxels that lead to a better-fitting behavioral model when their signals modulate the temporal reinstatement parameter of the model. B) Voxels that produce a better behavioral model when they modulate the retrieval success parameter. C) Results of an ROI analysis in which the ROIs are ordered from their posterior-to-anterior position within the Hippocampus. Bars denote the relative evidence (weighted AIC) for each of the three fitted models: 1) a model in which the ROI signal modulates the temporal reinstatement parameters (TR, red), 2) the ROI signal modulates the retrieval success parameters (RS, blue), or 3) the ROI signal modulates both parameters (Joint, orange). D) Same ROI analysis as in (C) but for ROIs within the medial temporal lobe. Figures adapted from Kragel et al. (2015).

The ability to tease apart these cognitive processes relied on incorporating fMRI activity within a theoretical model of contextual reinstatement. The result of doing so provides new evidence for the role of the medial temporal lobe in episodic memory and free recall, by linking BOLD activity in these regions to formalized computational mechanisms of temporal context. These results also provide support for the CMR model: If the model were poorly aligned with cognitive/neural processes, then it is unlikely that its parameters would map onto brain responses in such a structured and coherent manner. Other temporal context models may fair better or worse at linking the BOLD signal to behavior, and the modeling approach developed by Kragel et al. provides a powerful framework for addressing such model-based questions.

The approach taken by Kragel et al. to establish links between a theoretical model and neural data is quite different than those discussed previously, where a model was fitted to amplitudes of fMRI population responses or to the accuracy of fMRI decoding. Instead, Kragel et al. incorporated fMRI signals to modulate latent processes within a model of behavioral data, and identified voxels for which doing so led to better behavioral predictions. This approach serves as a promising framework for future studies, and provides a compelling demonstration that there are many ways to accomplish a model-based approach to cognitive neuroscience.

Categorization: Exemplar vs. Prototype Models

In the studies considered so far, a specific model was applied to some combination of behavioral and neural data, and the resulting fits and parameter estimates were used to make inferences about the underlying mechanisms. However, formal theoretical modeling provides for a more powerful approach of fitting multiple models to the data, such that the merits of different cognitive theories can be compared quantitatively. Admittedly, the task of developing multiple models to link to neural data presents a greater challenge, especially to ensure judicious and equitable implementations of each model. However, a recent study by Mack et al. (2013) demonstrates how doing so can lead to powerful ways to test cognitive theories.

People are highly adept at learning categorical boundaries in multidimensional spaces, and can accurately categorize new instances according to these learned categories. There are two prominent classes of models that describe how people might do this, namely prototype and exemplar-based models (see Ashby & Maddox, 2005 for a review). Prototype theories posit that as we encounter multiple instances from different categories, we construct a prototype for each category. For example, the prototype may be constructed as the average of all stimuli experienced in that category. Upon seeing a new exemplar of a category, the values of its features are then compared with the prototypes for each category. Greater similarity (or proximity) to one of the prototypes then provides the basis for categorization of the novel exemplar. In contrast, exemplar-based theories posit that people store particular instances from each category as we experience them. Upon seeing a new instance that is to be categorized, the values of its features are compared with all previous exemplars stored in memory for each category, and the current instance is categorized based on its similarity with all exemplars stored in memory.

Mack et al. created stimuli that took on one of two values of color, shape, size, and spatial location. The set of possible stimuli were divided into two categories, such that the average values of each stimulus dimension varied across categories, and provided reasonable prototypes of each category. Participants were trained to categorize a subset of the stimuli, and were then tested with new instances from each category while undergoing fMRI scanning. According to prototype theories, the degree to which a new instance belongs to each category is based on how similar the new instance is to the average, or prototype, stored in memory for each category. If xkm denotes these averages for the kth category and mth feature dimension, and the feature values of a new item at test are ym for the mth feature dimension, then the similarity between the item and the kth category prototype is

sk=exp(cm=14wm|ymxkm|)

where parameter c controls how quickly similarity decreases as the distance between a prototype and a new item increases, and parameters wm allow for particular dimensions to be differentially weighted. For example, if more attention were paid to one feature dimension over the others, leading to greater weighting of that feature, then that feature would have a greater influence on the similarity evaluation (sk) between an item (ym) and the prototypes (xkm) for each category. The exponential function leads to a similarity value of 1.0 when the instance matches the prototype exactly, and the exponential decay of this similarity metric provides for a generally good characterization of human similarity judgments (Shepard, 1987).

The exemplar-based model follows a similar structure, however similarity is based on the summed distance between a new instance and all previously stored items in a category, rather than distance to the average item. If xjm are the mth feature values for the jth previously studied item, then the total similarity between a new item ym and the Kth category can be specified by

sk=exp(cjKm=14wm|ymxjm|)

where parameters and wm govern overall similarity and feature weighting, respectively.

For each item to be categorized at test, these equations describe the model predictions regarding the similarity (or representational match) between the representation of that item and that of each category. These similarity values can be used to compute a predicted probability that an item is placed within each category, and these predictions can be compared with observed categorization performance to evaluate the predictive power of the models. In their experiment, Mack et al. (2013) found that both models fitted the participants’ behavioral performance equally well, as is often the case (Rouder & Ratcliff, 2004). Although the prototype and exemplar models predict different representational match values for each item (Figure 5A), the coarseness of the behavioral measure (category A or category B) appeared to lack the sensitivity to discriminate between the more subtle differences between the models in their representational match predictions.

Figure 5.

Figure 5

Exemplar and Prototype Models of Categorization. A) Categorization models predict a certain amount of representational match between an internal representation and a particular item from a learned category. Here the match for the exemplar (green) and prototype (blue) models are shown for various items used by Mack et al. (2013). B) The level of agreement (measured as mutual information) between the representational match as predicted by each model, and the level of representational match as measure in the fMRI signal. Values of zero would indicate no correspondence between a model’s predictions and the data; larger values indicate a higher correspondence. C) Searchlight analysis to identify areas where the differences in brain activity to the different items at test were well predicted by the exemplar model. Figures adapted from Mack et al. (2013).

Mack et al. (2013) evaluated whether neural activity as measured with fMRI multivariate pattern analysis might provide a more sensitive measure of variations in representational match in this categorization task. To do so, they first fit the prototype model to a participant’s behavioral data, and used the model to label each test trial by its predicted representational match value (Figure 5A). In a cross-validation procedure, they applied a regression analysis to a subset of the data, mapping the weighted sum of voxel activity onto the model-defined representational match value for each trial (using 1000 voxels chosen from the entire brain using a localizer analysis). This fitted linear model was then used to predict the representational match values of trials not included in the training data set, based on the corresponding voxel activity patterns. This procedure was repeated such that representational match values were predicted for every trial, based on the neural signal. The accuracy of these predictions served as an index of how closely the neural activity patterns match the prototype model’s definition of representational match. Prediction accuracy above chance would indicate that the neural signal carries category information as defined by the prototype model. Critically, the same procedure was conducted for the exemplar model, providing predication accuracy for representational match values as defined by distance from the exemplars. Comparing how well the neural data could be used to predict match values as defined by the two models provides a way to determine which model more closely matches the neural representation (see Figure 1D).

The correspondence between model-based representational match values and those predicted from brain activity was captured by a measure of mutual information between predicted and observed values. This mutual information metric was significantly greater than zero for both models (Figure 5B), suggesting that brain activity (from voxels chosen across the entire brain) conveys some information about the similarity between a presented item and the previously learned category representations as defined by either model. Critically however, this measure was higher for the exemplar model than for the prototype model, implying that the exemplar model provides a better description of category representations in the human brain. A follow up searchlight analysis was conducted to identify brain areas that showed activation differences to different items that were in line with item differences predicted by the exemplar model (Figure 5C). The results indicated that regions of the lateral occipital complex, inferior parietal lobe, as well as the right inferior frontal lobe, exhibited responses consistent with the predictions of the exemplar-based model.

These results underscore the potential power of combining theoretical models with neuroimaging techniques. Although these models make unique predictions about behavioral data, the resolution of such data is not always sufficient to provide convincing evidence in favor of one model or another. By using multivariate pattern analysis to construct an appropriate link between the neural data and model parameters, the cortical activity patterns observed in this study provided a measurement that could be used to adjudicate between the categorization models. The ability to make a coherent link between theories and brain imaging data underlies the success of this and other model-based neuroscience approaches. Mack et al. (2013) demonstrate that doing so can provide insights into processing that would not have otherwise been possible.

Error Monitoring in The Stop Signal Task

A variety of cognitive functions have been ascribed to the dorsal anterior cingulate cortex (dACC), based on studies conducted using fMRI, EEG, and other methods. Prominent theories have proposed roles for the dACC in error detection, conflict monitoring, error anticipation, volatility monitoring, action-outcome learning, reward, and the coding of errors in predicting future outcomes, to name a few (see e.g. Botvinick et al., 2004; Hayden, Heilbronner, Pearson, & Platt, 2011; Rushworth, Walton, Kennerley, & Bannerman, 2004). Teasing apart which of these roles the dACC is primarily responsible for during a task is extremely challenging, as most tasks that are designed to tap into one of these processes may also evoke some combination of the others. For example, tasks designed to elicit errors often do so by incorporating rare events within the experimental design, such that events that produce behavioral errors are also rare. Consequently, brain responses to these rare events might reflect the fact that they are unexpected, or alternatively, because they tend to elicit error responses. Fortunately, formal cognitive models have been devised that can provide separate measures these various processes, based on the structure of the task and the resulting behavioral data. The key to determining how these processes map on to particular brain regions, such as the dACC, is to somehow leverage the models in order to isolate the neural activity associated with specific components of these overlapping mental processes.

Ide et al. (2013) set out to isolate these often co-occurring mental processes, such as error detection and conflict monitoring, in order to determine what brain areas are involved in which particular functions. They utilized a classic cognitive control task, known as the stop-signal task (Logan, Cowan, & Davis, 1984; see Verbruggen & Logan, 2008), in which participants make a button press in response to a “go” signal, such as a green dot. On a minority of trials, however, a “stop” signal (e.g. a red dot) appears some time following the go signal, instructing participants to refrain from making a response to the preceding go stimulus. Replicating previous results, Ide et al. (2013) found that the dACC as well as many surrounding regions showed a stronger BOLD response on stop trials than on go trials (yellow/orange colors in Figure 6). However, this heightened activity might reflect any combination of many underlying cognitive processes. For example, higher activity on stop trials might reflect the processing of conflict between the prepotent “go” response and the required stopping behavior, or it might reflect the experience of surprise associated with the rare event of a stop trial.

Figure 6.

Figure 6

Bayesian Model of Stop Signal Task. Contrasting the BOLD signal to identify regions of greater activity for stop than for go trials (red/orange) reveals extensive activation in both cortical and sub-cortical regions. The more nuanced model-based analysis, however, reveals that only a small region (magenta color) within the dorsal anterior cingulate cortex (indexed by cross hairs, labeled dACC) is related to the process of comparing expected outcomes with actual outcomes, termed surprise. Figure adapted from Ide et al. (2013).

The authors applied a Bayesian rational decision-making model to account for behavioral performance in the stop-signal task, with dynamic updating based on trial-by-trial learning (Pradeep Shenoy, Angela, & Rao, 2010; P. Shenoy & Yu, 2011). The degree to which a participant expects a stop signal to occur on a given trial can be specified by a latent variable (Pstop) that varies from trial to trial depending on the stimulus history and the participant’s past behavior. For example, following a long sequence of go trials, people tend to make progressively faster responses. This behavior is thought to reflect a lowered expectation of an impending stop trial after each successive go trial, such as by the buildup of an automatic, prepotent response. In the model this behavior is instantiated as a relaxing of the decision criterion to make a “go” response, and this dynamic latent process is measured by parameter Pstop. Another parameter, Perror, reflects the likelihood that a participant will make an error on the upcoming trial, dependent upon the recent stimulus history and the participant’s past behavior. By specifying how these and several other latent processes are linked to the behavioral data, the model is able to predict the engagement of these processes for each trial across an experiment.

Ide et al. (2013) fit this model to behavioral data collected while participants underwent fMRI scanning. The resulting fitted model provided trial-by-trial predictions regarding the levels of multiple latent constructs, such as Pstop and Perror. These values were then used as independent variables in a multiple regression model, in which the BOLD signal from individual voxels served as the dependent variable. The results of the regression analysis could then be used to identify voxels that co-varied with each particular latent process, such as the participant’s expectations of a stop trial (see Figure 1E). One critical analysis involved comparing the participant’s presumed expectations about a trial, as specified by the model, with the actual outcome of the trial. In particular, the difference between the latent parameter Pstop and the actual stimulus condition (stop vs. go) on each trial provided a measure of the degree of mismatch between participant’s expectations and the actual stimulus outcome, termed surprise. The magenta color in Figure 6 shows voxels that were modulated specifically by this measure of surprise, and the results of this whole-brain analysis revealed that voxels sensitive to surprise are located squarely within the dACC (marked by cross hairs). This analysis suggested that the dACC’s sensitivity to surprise on stop trials is independent of other cognitive processes associated with stop trials, such as cognitive control and error monitoring. In fact, the researchers performed several additional analyses to determine what other factors may co-vary with dACC activity, but found none.

The results of this fMRI modeling work suggest that the dACC is involved in comparing our expectations about the environment with actual outcomes, and that accounting for this cognitive process may preclude the need to invoke additional cognitive processes such as error monitoring or response conflict. Because these processes are so comingled in typical tasks, contrasts between conditions in previous studies likely reflect a conglomeration of cognitive processes. However, Ide et al.’s use of a theoretical model in conjunction with fMRI provided a more nuanced and powerful approach for ascribing particular functions to the anterior cingulate.

Discussion

Human cognitive neuroscience and theoretical modeling of cognitive processes have led to remarkable advances in our understanding of mental function, but largely as isolated endeavors. The integration of these approaches is a natural step forward, and the recent proliferation of review articles (Forstmann, Wagenmakers, Eichele, Brown, & Serences, 2011), books (Forstmann & Wagenmakers, 2015), and special journal issues speaks to the growing interest in model-based cognitive neuroscience. Establishing a link between a model’s constructs and non-invasive measures of human brain activity presents many challenges. The studies we have reviewed here highlight the diversity of recent approaches to meet these challenges, and we take their success as evidence that there are many promising paradigms for building a model-based cognitive neuroscience.

The studies reviewed here differ not only in the particular methods used, but differ in more fundamental ways with respect to how the models and neural data were combined (see Figure 1). For example, some studies fitted a theoretical model directly to the neural data, and used the resulting parameter estimates and model fits to learn about the underlying psychological and neural processes (Brouwer & Heeger, 2011; Pratte et al., 2013). Other studies first fitted a model to the behavioral data and then used the resulting fitted model to make inferences about the neural signals (Ide et al., 2013). The opposite approach has also proven effective, in which fMRI signals from a voxel or brain region of interest are incorporated within a model in order to make better predictions of behavioral performance (Kragel et al., 2015). Finally, Mack et al. (2013) fitted multiple models to the behavioral data, and determined which model produced the highest correspondence with patterns of brain activity. The approaches taken in these studies is not an exhaustive list, but the diversity clearly shows that there are many possible paths for integrating theoretical modeling and cognitive neuroscience. Going forward, the development of new ways to link models and neural data has great potential for advancing our theoretical and empirical understanding of brain function. Although we are hesitant to make strong prescriptions for how one ought to approach model-based neuroscience, it may be helpful to consider the similarities of an approach with previously successful efforts. Although no two studies reviewed here completely converged in their methods, there are some commonalities among them that we highlight below as promising avenues for future work.

One approach for establishing links between high-dimensional fMRI data and parameters of interest in a cognitive model is to construct a lower-dimensional measure of the information contained in fMRI activity patterns. This approach was taken in the studies of the normalization model of perception (Brouwer & Heeger, 2011) and the perceptual template model of attention (Pratte et al., 2013). Both studies utilized multivariate pattern analyses as an intermediary step to link the multivariate BOLD responses with the theoretical model. For example, Pratte et al. (2013) applied pattern classification to determine how well the activity patterns in individual visual areas could predict which of four possible object categories the participant was viewing on individual stimulus blocks. The resulting prediction accuracy provided a measure of how robustly an object was represented within a particular visual area, which is precisely the targeted construct for the theoretical model. Adopting a multivariate approach in this way necessitates several considerations. First, the stimulus, experimental paradigm, and task may require some modifications to be well-suited for multivariate pattern classification (see Tong & Pratte, 2012 for details). In addition, special care must be taken to ensure that the results of such a complex intermediary analysis serve to measure the intended construct. In the studies considered here that leveraged multivariate pattern analysis (Pratte et al., 2013), multivariate regression (Mack et al., 2013) or a multivariate forward-modeling approach (Brouwer & Heeger, 2011), such considerations were built into the design and analysis of the experiment. However, demonstrating that these complex analysis methods are measuring the neural information that they are intended to extract will be critical for future studies that utilize such techniques. Nonetheless, we believe that multivariate approaches such as pattern classification analyses (Kamitani & Tong, 2005; Norman et al., 2006; Tong & Pratte, 2012) and forward modeling approaches (Brouwer & Heeger, 2009; Kay et al., 2008) will provide for powerful ways to link high-dimensional neuroimaging data with model parameters that isolate important cognitive constructs.

Another commonality among recent model-based neuroscience work is the use of temporally dynamic models. For example, in the work of Kragel et al. (2015), the Context Maintenance and Retrieval model specifies latent constructs such as the strength of temporal reinstatement over the course of free recall. By incorporating the amplitude of the fMRI signal at each trial into specific parameters of the model, they could determine whether the neurally informed model provided a better prediction of free recall performance. Models such as this, in which processes are proposed to vary over the course of an experiment, provide excellent candidates for a model-based neuroscience. For such models, the behavioral data alone is often not sufficient to identify how processes might vary over time. However, the corresponding neural measurements provide potentially informative information about the temporal dynamics of such latent processes. Although some dynamic cognitive phenomena occur on slow time scales that can be measured with fMRI (e.g., Davis, Love, & Preston, 2012; J. P. O’Doherty et al., 2003; Serences & Saproo, 2009), such as variations in performance across trials (e.g., Ide et al., 2013), the sluggish nature of the hemodynamic BOLD response may limit the ability to study fast cognitive processes that typically occur over a timeframe of a few hundred milliseconds. However, other methods that measure neural signals in real time, such as EEG and MEG, provide a promising avenue to investigate faster dynamic processes in the same manner.

Formal models have provided many insights into cognitive processes, but progress can sometimes stall following the development of more advanced and nuanced models, when the behavioral data fails to clearly distinguish between competing models. The comparison of prototype and exemplar models of categorization is a striking example, as these models aim to explain the same phenomena, and differ only subtly in the nature of their underlying latent representations. This feature has made the models particularly difficult to differentiate using behavioral measures alone, as the latent constructs produce similar behavior. However, by using the fMRI signal to tap into the underlying latent representation, Mack et al. (2013) were able to compare the models with a granularity not possible by studying behavior alone. We doubt that this is a special case, but rather, we suspect that neural measurements might be informative to the study of a variety of proposed latent constructs that are difficult to measure accurately with behavior alone. The challenge will be to design paradigms and analytic approaches that link neural signals to the proposed latent process in a testable manner. The studies reviewed herein provide evidence that doing so is highly feasible.

It is often said that no model is correct, but that some models are better than others, and among the better ones, some may be considered adequate with respect to their predictive power and the plausibility of their assumptions. Each of the studies reviewed here included various checks to ensure that the utilized model was at least adequate, such that combining the model with neural data allowed for clearly interpretable inferences to be drawn. These inferences would not have been possible with a poorly specified model, and the often-demonstrated match between predicted and observed neural data speaks to the potential power of these models. Looking forward, we believe that one of the most promising avenues for a model-based neuroscience will lie in using cognitive neuroscience data in order to test, constrain, and compare different models, with the eventual goal of developing the next generation of better theoretical models.

Our review was selective, aiming to demonstrate the breadth of topics for which a model-based cognitive neuroscience has proven successful, and the diversity of approaches taken. Other notable achievements in combining theoretical models with functional imaging have been made in recent years, with respect to the study of information accumulation (Turner, van Maanen, & Forstmann, 2015), reinforcement learning (Badre & Frank, 2012; J. P. O’Doherty et al., 2007), and with complex cognitive architectures such as ACT-R (Borst & Anderson, 2014). In this review, we chose to focus on fMRI approaches to neuroscience, and on theoretical models of perception and cognition, but it is worth noting that the more general ideas underlying model-based neuroscience have the potential to impact all areas of neuroscience and psychology.

We anticipate that future advances in theoretical understanding will be bolstered by advances in technology. Developments in the methods used to estimate model parameters, such as Bayesian estimation (e.g., Turner et al., 2013), will allow researchers to test more sophisticated models than are currently possible. Likewise, advances in neuroimaging technologies, such as high-field fMRI (Duyn, 2012), will allow for more fine-grained measures of human brain activity. Spurred by progress on multiple fronts, the future of model-based neuroscience holds great promise for linking theoretical processes with their underlying neural instantiations.

Acknowledgments

This research was supported by National Science Foundation grant BCS-1228526 to FT and National Institutes of Health Fellowship F32-EY022569 to MP, and was facilitated by National Institutes of Health center grant P30-EY-008126 to the Vanderbilt Vision Research Center.

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

1

These models often include parameters such as an intercept, scale and non-linearity to account for the various measurements that are used. These parameters are omitted here for simplicity.

References

  1. Ashby FG, Maddox WT. Relations between prototype, exemplar, and decision bound models of categorization. Journal of Mathematical Psychology. 1993;37(3):372–400. [Google Scholar]
  2. Ashby FG, Maddox WT. Human category learning. Annual Review of Psychology. 2005;56:149–178. doi: 10.1146/annurev.psych.56.091103.070217. [DOI] [PubMed] [Google Scholar]
  3. Badre D, Frank MJ. Mechanisms of hierarchical reinforcement learning in cortico-striatal circuits 2: evidence from fMRI. Cerebral Cortex. 2012;22(3):527–536. doi: 10.1093/cercor/bhr117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bonin V, Mante V, Carandini M. The suppressive field of neurons in lateral geniculate nucleus. Journal of Neuroscience. 2005;25(47):10844–10856. doi: 10.1523/JNEUROSCI.3562-05.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Borst JP, Anderson JR. Using the ACT-R Cognitive Architecture in combination with fMRI data. In: Forstmann BU, Wagenmakers EJ, editors. An Introduction to Model-Based Cognitive Neuroscience. New York: Springer; 2014. [Google Scholar]
  6. Botvinick MM, Braver TS, Barch DM, Carter CS, Cohen JD. Conflict monitoring and cognitive control. Psychological Review. 2001;108(3):624–652. doi: 10.1037/0033-295x.108.3.624. [DOI] [PubMed] [Google Scholar]
  7. Botvinick MM, Cohen JD, Carter CS. Conflict monitoring and anterior cingulate cortex: an update. Trends in Cognitive Sciences. 2004;8(12):539–546. doi: 10.1016/j.tics.2004.10.003. [DOI] [PubMed] [Google Scholar]
  8. Boynton GM, Engel SA, Glover GH, Heeger DJ. Linear systems analysis of functional magnetic resonance imaging in human V1. Journal of Neuroscience. 1996;16(13):4207–4221. doi: 10.1523/JNEUROSCI.16-13-04207.1996. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Brouwer GJ, Heeger DJ. Decoding and reconstructing color from responses in human visual cortex. Journal of Neuroscience. 2009;29(44):13992–14003. doi: 10.1523/JNEUROSCI.3577-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Brouwer GJ, Heeger DJ. Cross-orientation suppression in human visual cortex. Journal of Neurophysiology. 2011;106(5):2108–2119. doi: 10.1152/jn.00540.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Buckner RL, Bandettini PA, O’Craven KM, Savoy RL, Petersen SE, Raichle ME, et al. Detection of cortical activation during averaged single trials of a cognitive task using functional magnetic resonance imaging. Proceedings of the National Academy of Sciences of the United States of America. 1996;93(25):14878–14883. doi: 10.1073/pnas.93.25.14878. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Bullmore E, Sporns O. Complex brain networks: graph theoretical analysis of structural and functional systems. Nature Reviews Neuroscience. 2009;10(3):186–198. doi: 10.1038/nrn2575. [DOI] [PubMed] [Google Scholar]
  13. Buracas GT, Boynton GM. The effect of spatial attention on contrast response functions in human visual cortex. Journal of Neuroscience. 2007;27(1):93–97. doi: 10.1523/JNEUROSCI.3162-06.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Carandini M, Heeger DJ. Normalization as a canonical neural computation. Nature Reviews Neuroscience. 2012;13(1):51–62. doi: 10.1038/nrn3136. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Carrasco M. Visual attention: the past 25 years. Vision Research. 2011;51(13):1484–1525. doi: 10.1016/j.visres.2011.04.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Davis T, Love BC, Preston AR. Learning the exception to the rule: model-based FMRI reveals specialized representations for surprising category members. Cerebral Cortex. 2012;22(2):260–273. doi: 10.1093/cercor/bhr036. [DOI] [PubMed] [Google Scholar]
  17. Duyn JH. The future of ultra-high field MRI and fMRI for study of the human brain. Neuroimage. 2012;62(2):1241–1248. doi: 10.1016/j.neuroimage.2011.10.065. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Engel SA. The development and use of phase-encoded functional MRI designs. Neuroimage. 2012;62(2):1195–1200. doi: 10.1016/j.neuroimage.2011.09.059. [DOI] [PubMed] [Google Scholar]
  19. Engel SA, Rumelhart DE, Wandell BA, Lee AT, Glover GH, Chichilnisky EJ, et al. fMRI of human visual cortex. Nature. 1994;369(6481):525. doi: 10.1038/369525a0. [DOI] [PubMed] [Google Scholar]
  20. Ester EF, Serences JT, Awh E. Spatially global representations in human primary visual cortex during working memory maintenance. Journal of Neuroscience. 2009;29(48):15258–15265. doi: 10.1523/JNEUROSCI.4388-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Forstmann BU, Wagenmakers EJ. An Introduction to Model-Based Cognitive Neuroscience. New York: Springer; 2015. [Google Scholar]
  22. Forstmann BU, Wagenmakers EJ, Eichele T, Brown S, Serences JT. Reciprocal relations between cognitive neuroscience and formal cognitive models: opposites attract? Trends in Cognitive Sciences. 2011;15(6):272–279. doi: 10.1016/j.tics.2011.04.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Fox MD, Snyder AZ, Vincent JL, Corbetta M, Van Essen DC, Raichle ME. The human brain is intrinsically organized into dynamic, anticorrelated functional networks. Proceedings of the National Academy of Sciences of the United States of America. 2005;102(27):9673–9678. doi: 10.1073/pnas.0504136102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Glover GH. Deconvolution of impulse response in event-related BOLD fMRI. Neuroimage. 1999;9(4):416–429. doi: 10.1006/nimg.1998.0419. [DOI] [PubMed] [Google Scholar]
  25. Green DM, Swets JA. Signal Detection Theory and Psychophysics. New York: Wiley; 1966. [Google Scholar]
  26. Harrison SA, Tong F. Decoding reveals the contents of visual working memory in early visual areas. Nature. 2009;458(7238):632–635. doi: 10.1038/nature07832. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Haxby JV, Gobbini MI, Furey ML, Ishai A, Schouten JL, Pietrini P. Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science. 2001;293(5539):2425–2430. doi: 10.1126/science.1063736. [DOI] [PubMed] [Google Scholar]
  28. Hayden BY, Heilbronner SR, Pearson JM, Platt ML. Surprise signals in anterior cingulate cortex: neuronal encoding of unsigned reward prediction errors driving adjustment in behavior. Journal of Neuroscience. 2011;31(11):4178–4187. doi: 10.1523/JNEUROSCI.4652-10.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Haynes JD, Rees G. Predicting the stream of consciousness from activity in human visual cortex. Current Biology. 2005;15(14):1301–1307. doi: 10.1016/j.cub.2005.06.026. [DOI] [PubMed] [Google Scholar]
  30. Henson R. A mini-review of fMRI studies of human medial temporal lobe activity associated with recognition memory. Quarterly Journal of Experimental Psychology. B, Comparative and Physiological Psychology. 2005;58(3–4):340–360. doi: 10.1080/02724990444000113. [DOI] [PubMed] [Google Scholar]
  31. Herrmann K, Heeger DJ, Carrasco M. Feature-based attention enhances performance by increasing response gain. Vision Research. 2012;74:10–20. doi: 10.1016/j.visres.2012.04.016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Honey CJ, Sporns O, Cammoun L, Gigandet X, Thiran JP, Meuli R, et al. Predicting human resting-state functional connectivity from structural connectivity. Proceedings of the National Academy of Sciences of the United States of America. 2009;106(6):2035–2040. doi: 10.1073/pnas.0811168106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Howard MW, Kahana MJ. A distributed representation of temporal context. Journal of Mathematical Psychology. 2002;46:269–299. [Google Scholar]
  34. Huth AG, Nishimoto S, Vu AT, Gallant JL. A continuous semantic space describes the representation of thousands of object and action categories across the human brain. Neuron. 2012;76(6):1210–1224. doi: 10.1016/j.neuron.2012.10.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Ide JS, Shenoy P, Yu AJ, Li CS. Bayesian prediction and evaluation in the anterior cingulate cortex. Journal of Neuroscience. 2013;33(5):2039–2047. doi: 10.1523/JNEUROSCI.2201-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Kamitani Y, Tong F. Decoding the visual and subjective contents of the human brain. Nature Neuroscience. 2005;8(5):679–685. doi: 10.1038/nn1444. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Kanwisher N, McDermott J, Chun MM. The fusiform face area: a module in human extrastriate cortex specialized for face perception. Journal of Neuroscience. 1997;17(11):4302–4311. doi: 10.1523/JNEUROSCI.17-11-04302.1997. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Kastner S, Pinsk MA, De Weerd P, Desimone R, Ungerleider LG. Increased activity in human visual cortex during directed attention in the absence of visual stimulation. Neuron. 1999;22(4):751–761. doi: 10.1016/s0896-6273(00)80734-5. [DOI] [PubMed] [Google Scholar]
  39. Kay KN, Naselaris T, Prenger RJ, Gallant JL. Identifying natural images from human brain activity. Nature. 2008;452(7185):352–355. doi: 10.1038/nature06713. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Kintsch W. Memory and decision aspects of recognition learning. Psychological Review. 1967;74(6):496–504. doi: 10.1037/h0025127. [DOI] [PubMed] [Google Scholar]
  41. Kragel JE, Morton NW, Polyn SM. Neural activity in the medial temporal lobe reveals the fidelity of mental time travel. Journal of Neuroscience. 2015;35(7):2914–2926. doi: 10.1523/JNEUROSCI.3378-14.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Kriegeskorte N, Mur M, Ruff DA, Kiani R, Bodurka J, Esteky H, et al. Matching categorical object representations in inferior temporal cortex of man and monkey. Neuron. 2008;60(6):1126–1141. doi: 10.1016/j.neuron.2008.10.043. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Logan GD, Cowan WB, Davis KA. On the ability to inhibit simple and choice reaction time responses: a model and a method. Journal of Experimental Psychology: Human Perception and Performance. 1984;10(2):276–291. doi: 10.1037//0096-1523.10.2.276. [DOI] [PubMed] [Google Scholar]
  44. Love BC. The algorithmic level is the bridge between computation and brain. Topics in Cognitive Science. 2015;7(2):230–242. doi: 10.1111/tops.12131. [DOI] [PubMed] [Google Scholar]
  45. Lu ZL, Dosher BA. External noise distinguishes attention mechanisms. Vision Research. 1998;38(9):1183–1198. doi: 10.1016/s0042-6989(97)00273-3. [DOI] [PubMed] [Google Scholar]
  46. Lu ZL, Li X, Tjan BS, Dosher BA, Chu W. Attention extracts signal in external noise: a BOLD fMRI study. Journal of Cognitive Neuroscience. 2011;23(5):1148–1159. doi: 10.1162/jocn.2010.21511. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Mack ML, Preston AR, Love BC. Decoding the brain’s algorithm for categorization from its neural implementation. Current Biology. 2013;23(20):2023–2027. doi: 10.1016/j.cub.2013.08.035. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Marois R. The Brain Mechanisms of Working Memory: An Evolving Story. In: Jolicoeur P, Lefebvre C, Martinez-Trujillo J, editors. Mechanisms of Sensory Working Memory: Attention and Perfomance XXV. Elsevier; 2016. [Google Scholar]
  49. Marr D. Vision. A Computational Investigation into the Human Representation and Processing of Visual Information. New York: Freeman; 1982. [Google Scholar]
  50. McClelland JL, Elman JL. The TRACE model of speech perception. Cognitive Psychology. 1986;18(1):1–86. doi: 10.1016/0010-0285(86)90015-0. [DOI] [PubMed] [Google Scholar]
  51. McClelland JL, Rogers TT. The parallel distributed processing approach to semantic cognition. Nature Reviews Neuroscience. 2003;4(4):310–322. doi: 10.1038/nrn1076. [DOI] [PubMed] [Google Scholar]
  52. Mitchell TM, Shinkareva SV, Carlson A, Chang KM, Malave VL, Mason RA, et al. Predicting human brain activity associated with the meanings of nouns. Science. 2008;320(5880):1191–1195. doi: 10.1126/science.1152876. [DOI] [PubMed] [Google Scholar]
  53. Moradi F, Heeger DJ. Inter-ocular contrast normalization in human visual cortex. Journal of Vision. 2009;9(3):13, 11–22. doi: 10.1167/9.3.13. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Nichols T, Hayasaka S. Controlling the familywise error rate in functional neuroimaging: a comparative review. Statistical Methods in Medical Research. 2003;12(5):419–446. doi: 10.1191/0962280203sm341ra. [DOI] [PubMed] [Google Scholar]
  55. Norman KA, Polyn SM, Detre GJ, Haxby JV. Beyond mind-reading: multi-voxel pattern analysis of fMRI data. Trends in Cognitive Sciences. 2006;10(9):424–430. doi: 10.1016/j.tics.2006.07.005. [DOI] [PubMed] [Google Scholar]
  56. Nosofsky RM, Palmeri TJ. An exemplar-based random walk model of speeded classification. Psychological Review. 1997;104(2):266–300. doi: 10.1037/0033-295x.104.2.266. [DOI] [PubMed] [Google Scholar]
  57. O’Doherty JP, Dayan P, Friston K, Critchley H, Dolan RJ. Temporal difference models and reward-related learning in the human brain. Neuron. 2003;38(2):329–337. doi: 10.1016/s0896-6273(03)00169-7. [DOI] [PubMed] [Google Scholar]
  58. O’Doherty JP, Dayan P, Schultz J, Deichmann R, Friston K, Dolan RJ. Dissociable roles of ventral and dorsal striatum in instrumental conditioning. Science. 2004;304(5669):452–454. doi: 10.1126/science.1094285. [DOI] [PubMed] [Google Scholar]
  59. O’Doherty JP, Hampton A, Kim H. Model-based fMRI and its application to reward learning and decision making. Annals of the New York Academy of Sciences. 2007;1104:35–53. doi: 10.1196/annals.1390.022. [DOI] [PubMed] [Google Scholar]
  60. Polyn SM, Natu VS, Cohen JD, Norman KA. Category-specific cortical activity precedes retrieval during memory search. Science. 2005;310(5756):1963–1966. doi: 10.1126/science.1117645. [DOI] [PubMed] [Google Scholar]
  61. Polyn SM, Norman KA, Kahana MJ. A context maintenance and retrieval model of organizational processes in free recall. Psychological Review. 2009;116(1):129–156. doi: 10.1037/a0014420. [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Pratte MS, Ling S, Swisher JD, Tong F. How attention extracts objects from noise. Journal of Neurophysiology. 2013;110(6):1346–1356. doi: 10.1152/jn.00127.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Raaijmakers JGW, Shiffrin RM. Models Of Memory. In: Pashler H, Medin D, editors. Stevens’ Handbook of Experimental Psychology, Third Edition, Volume 2: Memory and Cognitive Processes. New York: John Wiley & Sons, Inc; 2002. [Google Scholar]
  64. Ratcliff R, Rouder JN. Modeling Response Times For Two-Choice Decisions. Psychological Science. 1998;9(5):347–356. [Google Scholar]
  65. Rouder JN, Ratcliff R. Comparing categorization models. Journal of Experimental Psychology: General. 2004;133(1):63–82. doi: 10.1037/0096-3445.133.1.63. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Rushworth MF, Walton ME, Kennerley SW, Bannerman DM. Action sets and decisions in the medial frontal cortex. Trends in Cognitive Sciences. 2004;8(9):410–417. doi: 10.1016/j.tics.2004.07.009. [DOI] [PubMed] [Google Scholar]
  67. Serences JT, Boynton GM. Feature-based attentional modulations in the absence of direct visual stimulation. Neuron. 2007;55(2):301–312. doi: 10.1016/j.neuron.2007.06.015. [DOI] [PubMed] [Google Scholar]
  68. Serences JT, Saproo S. Population response profiles in early visual cortex are biased in favor of more valuable stimuli. Journal of Neurophysiology. 2009;104(1):76–87. doi: 10.1152/jn.01090.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Sereno MI, Dale AM, Reppas JB, Kwong KK, Belliveau JW, Brady TJ, et al. Borders of multiple visual areas in humans revealed by functional magnetic resonance imaging. Science. 1995;268(5212):889–893. doi: 10.1126/science.7754376. [DOI] [PubMed] [Google Scholar]
  70. Shenoy P, Angela JY, Rao RP. A rational decision making framework for inhibitory control; Paper presented at the Advances in neural information processing systems.2010. [Google Scholar]
  71. Shenoy P, Yu AJ. Rational decision-making in inhibitory control. Frontiers in Human Neuroscience. 2011;5:48. doi: 10.3389/fnhum.2011.00048. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Shepard RN. Toward a universal law of generalization for psychological science. Science. 1987;237(4820):1317–1323. doi: 10.1126/science.3629243. [DOI] [PubMed] [Google Scholar]
  73. Sprague TC, Ester EF, Serences JT. Reconstructions of information in visual spatial working memory degrade with memory load. Current Biology. 2014;24(18):2174–2180. doi: 10.1016/j.cub.2014.07.066. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Stone M. Models for choice-reaction time. Psychometrika. 1960;25(3):251–260. [Google Scholar]
  75. Tavor I, Parker Jones O, Mars RB, Smith SM, Behrens TE, Jbabdi S. Task-free MRI predicts individual differences in brain activity during task performance. Science. 2016;352(6282):216–220. doi: 10.1126/science.aad8127. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Tong F, Pratte MS. Decoding patterns of human brain activity. Annual Review of Psychology. 2012;63:483–509. doi: 10.1146/annurev-psych-120710-100412. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Turner BM, Forstmann BU, Wagenmakers EJ, Brown SD, Sederberg PB, Steyvers M. A Bayesian framework for simultaneously modeling neural and behavioral data. Neuroimage. 2013;72:193–206. doi: 10.1016/j.neuroimage.2013.01.048. [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Turner BM, van Maanen L, Forstmann BU. Informing cognitive abstractions through neuroimaging: the neural drift diffusion model. Psychological Review. 2015;122(2):312–336. doi: 10.1037/a0038894. [DOI] [PubMed] [Google Scholar]
  79. Verbruggen F, Logan GD. Response inhibition in the stop-signal paradigm. Trends in Cognitive Sciences. 2008;12(11):418–424. doi: 10.1016/j.tics.2008.07.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Wagner AD, Schacter DL, Rotte M, Koutstaal W, Maril A, Dale AM, et al. Building memories: remembering and forgetting of verbal experiences as predicted by brain activity. Science. 1998;281(5380):1188–1191. doi: 10.1126/science.281.5380.1188. [DOI] [PubMed] [Google Scholar]
  81. Xu Y, Chun MM. Dissociable neural mechanisms supporting visual short-term memory for objects. Nature. 2006;440(7080):91–95. doi: 10.1038/nature04262. [DOI] [PubMed] [Google Scholar]

RESOURCES