Abstract
When a behaviorally relevant stimulus has been previously associated with reward, behavioral responses are faster and more accurate compared to equally relevant but less valuable stimuli. Conversely, task-irrelevant stimuli that were previously associated with a high reward can capture attention and distract processing away from relevant stimuli (e.g., seeing a chocolate bar in the pantry when you are looking for a nice, healthy apple). Although increasing the value of task-relevant stimuli systematically up-regulates neural responses in early visual cortex to facilitate information processing, it is not clear whether the value of task-irrelevant distractors influences behavior via competition in early visual cortex or via competition at later stages of decision-making and response selection. Here, we measured functional magnetic resonance imaging (fMRI) in human visual cortex while subjects performed a value-based learning task, and we applied a multivariate inverted encoding model (IEM) to assess the fidelity of distractor representations in early visual cortex. We found that the fidelity of neural representations related to task-irrelevant distractors increased when the distractors were previously associated with a high reward. This finding suggests that value-driven attentional capture begins with sensory modulations of distractor representations in early areas of visual cortex.
Task-irrelevant distractors can automatically capture attention when they were previously associated with a high reward. This study shows that these high-valued distractors increased the fidelity of neural representations in retinotopically organized regions of visual cortex, showing that value-driven attentional capture begins with early sensory modulations of distractor representations in human visual system.
Introduction
In most real-world situations, stimuli that are visually salient—such as a camera flash in a theater or a green object in a sea of red—automatically capture attention [1–4]. Likewise, distractors that are distinguished only by their value, not their visual salience, also capture visual attention—even on occasions when high-valued distractors are unactionable and irrelevant to current behavioral goals (e.g., seeing a piece of cake on another table, but your waiter tells you it is no longer available) [5–10]. In the laboratory, the value associated with an irrelevant distractor interferes with the processing of task-relevant visual information, resulting in increased response times (RTs) and sometimes reduced accuracy in a variety of tasks ranging from simple visual discrimination to more complex scenarios in which the value of multiple competing items must be compared [5–8,10–17]. Importantly, these behavioral effects of value-based attentional capture are underexpressed and overexpressed in patients with attention-deficit hyperactivity disorder and addiction, respectively [18,19]. Although previous work has shown that the value of task-relevant visual information increases neural activity in areas of early visual cortex [20–26], it is unclear how the learned value of irrelevant distractors modulates cortical responses in these cortical regions. Thus, here we aimed to examine the involvement of early visual cortex in supporting attentional control guided by the learned value associated with task-irrelevant distractors.
Several theories of reinforcement learning have emphasized the essential role of reward in prioritizing sensory information [27–34]. For example, stimuli associated with high reward are thought to better compete for sensory representations such that those stimuli become visually salient and automatically capture attention, even though these highly rewarded stimuli might be physically nonsalient [7]. Based on this previous work, some have proposed that dopaminergic neurons in the midbrain and in the ventral striatum relay value-related signals to cortical regions within the early visual cortex, resulting in the potentiation of visual representations related to high-reward visual stimuli [35–37]. According to this proposal, we hypothesized that the learned value associated with irrelevant distractors should enhance distractor representations in early visual cortex and that this enhancement should be spatially restricted to the distractor locations, much like value-based modulations of task-relevant visual stimuli [23,25]. Alternatively, we might expect no value-based modulations or even a reduction in the response to distractors, as recent studies have shown that higher reward can lead to the suppression of distractor-related neural representations (see more details in the Discussion section; also [38–41]).
To test these alternative accounts, we recruited human participants to perform a value-based decision-making task and measured their brain activity in visual cortex using functional magnetic resonance imaging (fMRI). Subjects were required to select 1 of 2 task-relevant options while ignoring a third irrelevant and unactionable distractor that was rendered in a color that had been previously associated with a variable level of reward. We hypothesized that the previously assigned value of the distractor color would modulate evoked responses in early visual cortex and that this reward-based modulation would be specific to the spatial location of the distractor stimulus. To evaluate spatially selective modulations, we used an inverted encoding model (IEM) to reconstruct a representation of each stimulus using activation patterns of hemodynamic responses from retinotopically organized visual areas V1, V2, and V3. We chose this multivariate analysis over the univariate analysis of fMRI data because we were interested in examining how modulations of large-scale activation patterns across entire visual areas supported value-driven attention. The IEM method is useful here because it exploits modulations of all voxels in areas of early visual cortex, including attention-related increases and decreases in the hemodynamic response, as well as shifts in the position of voxel receptive fields [20,42–55]. We found that distractors previously associated with a high value slowed choice RTs. These high-value distractors were also represented with higher fidelity in extrastriate visual areas V2 and V3. Importantly, these value-based modulations of behavior and of neural representations depended on target selection history—these modulations were only observed when participants had previously selected and learned the value of irrelevant distractors. Together, these results suggest that the influence of high-value distractors on attentional capture begins with an early modulation of sensory responses and that this value-driven attentional capture occurs when participants have learned the value associated with the visual feature of the distractor.
Results
High-valued distractors automatically capture attention
In the present study, we used fMRI to measure activity in retinotopically organized visual areas V1, V2, and V3 while human participants (N = 15) performed a 2-alternative value-based decision-making task with changing reward associations [6] (Fig 1). On each trial, 3 stimuli were presented, each rendered in a different color. Two of the stimuli were presented in fixed target locations, and subjects had to choose between them. The third stimulus, termed a “distractor,” was presented in another fixed location that subjects could never select. Participants learned that different rewards (1 or 9 cents) were associated with the colors of visual stimuli presented at the two target locations. Importantly, the distractor was not actionable and was thus completely irrelevant with respect to evaluating the relative value of the two possible targets. Across trials, the colors of the targets and the distractor changed randomly so that the distractor color on a given trial could match the color of a previously selected target that yielded either a low or a high monetary reward. Additionally, the pairings between color and reward changed across miniblocks of 8 trials so that values assigned to different colors could be counterbalanced. Thus, for behavioral and fMRI analyses, we sorted trials based on incentive values assigned to the colors of distractors (i.e., low- or high-valued distractor). The incentive value was always defined. However, a given color may not have been selected on previous trials. Therefore, the current value of the distractor was not always known to the participant. We thus used the selection of previous choices to examine the influence of learning reward contingencies on value-driven attention. To do so, we coded whether the distractor in a given trial was selected as a target in the previous 3 trials (i.e., selected or unselected; see Materials and methods). Note that we used the previous 3 trials in the main analysis because this data-sorting criterion yielded the most balanced number of trials across different experimental conditions (and we examine different sorting schemes in S2 Fig, detailed below).
Overall, subjects selected higher-valued targets more often than lower-valued targets (Fig 2A, p ≤ 1 × 10−6, 2-tailed, resampling test). This indicates that subjects were able to learn the values assigned to the different colors. Next, we fit the choice preference data as a function of differential target value with a cumulative Gaussian function (Fig 2B; see a similar fitting procedure in [6]). We found no effect of distractor value (high − low distractor value) on the standard deviation (or sigma) and the mean (or mu) of the cumulative Gaussian function on trials in which the current distractors were previously selected (p = 0.9420 and 0.0784 for sigma and mu, respectively, 2-tailed) or unselected (Fig 2B; p = 0.5637and 0.8206 for sigma and mu, respectively, 2-tailed). Since we used a 2-parameter cumulative Gaussian model to fit 3 data points for each experimental condition, we conducted an additional analysis in which we used the reduced model that only varied sigma to ensure the reliability of the result. First, a nested F test showed that this reduced model performed as well as the full model in which both sigma and mu were optimized, suggesting that this reduced model is more parsimonious than the full model (F[4, 3] = 0.9643, p = 0.4206). Second, we found that fitting the data with this reduced model yielded consistent results in which there was no significant distractor-value modulation on the sigma value for either selected (p = 0.9428) or unselected condition (and p = 0.5365). The null distractor-value effect in the choice preference data is inconsistent with our previous study using a variant of this value-based learning task, in which we found robust distractor-value modulations on the sigma value of the choice preference function [6]. We think this is due to the fact that in the current design, the value assignment for each color changed much more frequently than the task design used in our previous study (every 8 versus 36 trials) [6]. That said, the null distractor effect on the choice preference data is also consistent with a large body of literature demonstrating smaller and more variable distractor-value effects on task accuracy [11,56,57].
Although there was no distractor-value modulation on the choice preference data, RTs differed significantly across different distractor types (Table 1). We observed a significant effect of distractor value (high − low distractor value) on RTs on trials in which the current distractor was previously selected (Fig 2D; p ≤ 1 × 10−6, 2-tailed). However, there was no distractor-value modulation on trials in which the current distractors were previously unselected (p = 0.2756, 2-tailed). Moreover, the magnitude of the distractor-value modulation was significantly higher for the current distractor that was previously selected versus unselected (p = 0.0102, 1-tailed). These RT results show that the distractor value captures attention, leading to a relative decrease in the speed with which subjects processed task-relevant targets [5–8,13–17].
Table 1. Cumulative Gaussian best-fit variables describing choice preference data and RTs for different distractor types shown in Fig 2.
Behavioral Measurements | Distractor Types: Distractor Value and Selection History (Mean ± SEM) | |||
---|---|---|---|---|
Low and Unselected | High and Unselected | Low and Selected | High and Selected | |
Sigma | 23.99 ± 8.01 | 29.87 ± 9.48 | 32.66 ± 7.09 | 33.56 ± 7.81 |
Mu | 0.85 ± 2.41 | 0.17 ± 2.80 | 1.31 ± 1.58 | 1.46 ± 1.13 |
RTs (ms) | 600 ± 20 | 612 ± 15 | 592 ± 18 | 643 ± 19 |
Abbreviation: RT, response time
The reward history of distractors modulates neural representations in early visual cortex
To examine the influence of the distractor value on spatially specific distractor- and target-related neural representations in early visual cortex, we employed a multivariate analysis of fMRI data—an IEM (Materials and methods; Fig 3) [20,50,52,58]. The IEM exploits the spatial tuning of neuronal populations in visual cortex to reconstruct model-based representations of target and distractor stimuli based on population-level activity measured via fMRI. As expected, we found that these reconstructions peaked at the center of each of the 3 locations (Fig 4A; sorted as unselected target, selected target, and distractor). Qualitatively, the reconstructed activation at the distractor location was highest when the distractor colors matched the target colors that had been selected and rewarded with a higher value in the previous trials (i.e., the high-valued and previously selected distractor, the top right of the Fig 4A), compared to all the other distractor types.
To quantify this effect, we computed the mean activation level in the reconstructed stimulus representations over the space occupied by the distractors (Fig 4A, see Materials and methods; also see [20]). Then, we used a nonparametric resampling method (i.e., resampling subjects with replacement) to evaluate the impact of distractor value (high versus low distractor values) on the mean activation of the distractor representation. We did this separately for trials in which the current distractor had been previously selected or unselected in preceding trials to determine whether distractor-value modulations depended on the selection history associated with the color of the distractor.
First, we analyzed the data averaged across V1–V3 (Fig 4B). We found a significant distractor-value modulation (high > low value) for the distractor that was previously selected (p = 1 × 10−3, 2-tailed) but a null result for the distractor that was previously unselected (p = 0.4956, 2-tailed, resampling test). We directly evaluated this effect and found that selection history significantly increased distractor-value modulation (p = 0.0243, 1-tailed, resampling test). We then repeated these tests separately for individual visual areas. We found significant distractor-value modulations for the previously selected distractor in extrastriate visual areas V2 and V3 (p = 0.0011 and p = 0.0052, passing the Holm-Bonferroni-corrected thresholds of 0.0167 and 0.025, respectively, 2-tailed) but not in the primary visual cortex V1 (p = 0.3318, 2-tailed). In V2 and V3, we confirmed that selection history had a significant effect on distractor-value modulation (p = 0.0086 and p = 0.0374, respectively, 1-tailed). Similar to the data averaged across V1–V3, there was no significant distractor-value modulation for the previously unselected distractors in any visual area (p = 0.2031, p = 0.6263, and p = 0.9230, for V1, V2, and V3, respectively, 2-tailed). In sum, we used an IEM to evaluate spatially specific representations of behaviorally irrelevant stimuli with an associated reward history. We found that the value associated with irrelevant visual features is encoded in spatially specific activation in early visual areas V2 and V3.
To address whether the distractor-value modulations were driven by knowledge about the value associated with a given color, we split trials into early and late phases based on trial position relative to the start of each miniblock (in which the value–color assignments changed). If value learning matters, we should see robust distractor-value modulations only in the late phase but not in the early phase of each miniblock. Consistent with this prediction, we found a significant main effect of learning on distractor-value modulations in early visual cortex (late > early phases; p = 0.0167, 2-tailed) (see S1 Fig). This learning effect was driven by a significant distractor-value modulation in the late phase and only on trials in which the colors of the distractors matched the colors of the previously selected targets (i.e., selected: p = 0.0038, passing a Holm-Bonferroni threshold of 0.0125, 2-tailed). There was no significant distractor-value modulation in the late phase of the miniblocks on trials in which the color of the distractors did not match the color of the previously selected targets (i.e., unselected: p = 0.3844, 2-tailed). Importantly, there was also no significant distractor-value modulation in the early phase for either of the two distractor types (p = 0.1976 and 0.1800 for selected and unselected, respectively, 2-tailed). The null result for the early phase also speaks against the contribution of reward-based priming; as priming effects should rely only on the most recent choices, we should have seen distractor-value modulations regardless of the amount of learning that subjects acquired in the miniblock.
To further test the possibility of priming effects influencing our results, we conducted an auxiliary analysis in which we sorted trials by coding whether the current distractor was selected immediately in the previous trial or was selected at least once in the 2 or 3 previous trials. To ensure that trials included in these different data-sorting approaches came from a similar range of trial positions relative to the start of each miniblock, we only included trials between the fourth and the eighth (i.e., the last) trials of each miniblock. We found that distractors that were selected on the immediately preceding trial did not induce a significant distractor-value modulation in the data (collapsed across V1–V3, p = 0.1367, 2-tailed). This suggests that distractor-value modulations were unlikely to be driven entirely by reward-mediated priming effects (S2 Fig, left). However, when we sorted trials based on past selections over the previous 2 and 3 trials, we observed a significant distractor-value modulation (p = 0.013 and p < 0.001 for the data sorted based on the 2 and 3 previous trials passing Holm-Bonferroni thresholds of 0.025 and 0.0167, respectively). Since these two data sorting approaches included trials in which the same color targets could be selected more than once, these results emphasize the importance of learning the reward-color contingencies in producing value-driven modulations as opposed to reward-mediated priming effects. Note that qualitatively similar, albeit weaker, results were observed in individual visual areas V1–V3 (see S2 Fig).
Target selection and target value are encoded in early visual cortex
As shown in Fig 3A, stimulus representations were generally higher for selected targets compared to unselected targets. To quantify this effect, we computed the mean activation level in the reconstructed stimulus representations over the space occupied by the selected and unselected targets (Fig 5A). For the data collapsed across V1–V3, we observed a significant target selection modulation (selected > unselected targets: p = 0.0011 for data collapsed across distractor types; p = 0.0642, 0.0003, 0.0228, and 0.0022 for low-valued and unselected, high-valued and unselected, low-valued and selected, and high-valued and selected distractors, with the Holm-Bonferroni-corrected thresholds of 0.05, 0.0125, 0.025, and 0.0167, respectively, 2-tailed). These target-selection modulations were significant in all visual areas (p = 0.0189, 4.600 × 10−4, and 5.600 × 10−4, V1, V2, and V3, respectively; Holm-Bonferroni-corrected, 2-tailed).
Next, we evaluated the impact of distractor value on the differential activity between selected and unselected targets. We found no influence of distractor value on target representations (high- versus low-valued distractors) on trials in which the current distractor was previously selected (p = 0.2303, 2-tailed) or on trials in which the current distractor was unselected (p = 0.4463, 2-tailed). Similar null results were also observed when the data were analyzed separately in V1, V2, and V3 (p = 0.1639–0.8710 and 0.0744–0.9419 for the selected and unselected conditions, 2-tailed). These are consistent with the null distractor-value effects on the choice preference data (Fig 2A and 2B).
Previous studies have reported that the relative value of targets is encoded in early visual cortex [23–25]. To test this, we analyzed the target-selection modulation data both when the selected and unselected targets had the same value (i.e., selected = unselected targets) and when the selected target had a higher value compared to the unselected target (i.e., selected > unselected targets). As shown in Fig 5B, we found significant target-selection modulations only when the selected targets had a higher value compared to the unselected targets in all visual areas (p = 0.0055, 4 × 10−6, and 1 × 10−6, passing the Holm-Bonferroni-corrected thresholds of 0.0125, 0.0100, and 0.0083 for V1, V2, and V3, respectively, 2-tailed), but no significant target modulations when selected and unselected targets had the same value (p = 0.0437–0.0756, which did not pass the Holm-Bonferroni-corrected threshold of 0.0167, 2-tailed). In addition, on trials in which participants selected the higher-valued target, the target-selection effect was significantly stronger in V3 than in V1 (p = 0.0021, passing the Holm-Bonferroni-corrected threshold of 0.0167, 2-tailed). However, there was not a significant difference between V3 and V2 (p = 0.1165, 2-tailed) or between V2 and V1 (p = 0.1274, 2-tailed). Taken together with the previous section, our results show that the encoding of target value and distractor value can occur in parallel in early areas of visual cortex.
Discussion
Visual stimuli that are not physically salient but are paired with high-reward values are known to automatically capture attention, even when those stimuli are behaviorally irrelevant and unactionable [5–9]. Although a recent study reported that neural responses associated with distractors scale with the learned value [35], it is unclear whether these modulations were tied specifically to the location of the distractor and whether distractor response modulations led to attenuated target responses. Using a multivariate spatial reconstruction analysis of fMRI data, we show here that retinotopically organized regions in extrastriate visual areas V2 and V3 are modulated by the reward history of irrelevant visual stimuli. Importantly, the spatial reconstructions of these stimuli indicate that reward-based modulations occur precisely at the location of the distractor and that there is little associated impact on responses to simultaneously presented targets. Taken together, our results suggest that value-driven attentional capture may begin with early value-based modulation of sensory responses evoked by the distractor.
Importantly, these attentional capture effects on task-irrelevant distractors suggest that value-driven attention is not simply due to an increase in the non-stimulus-specific arousal levels induced by the reinforcement process, consistent with the finding that perceptual learning can occur even when task-irrelevant subthreshold stimuli are paired with rewards [36,59]. The fact that we observed value-based modulations that were spatially specific to the distractor location strongly supports this argument.
At first glance, our results seem to contradict several recent studies that observed a reward-based suppression of neural representations associated with distractors in sensory cortices [38–40,60]. However, in many of these studies, the reward manipulation was not specifically tied to the distractor, and distractor suppression was inferred based on modulations of neural responses related to the task-relevant targets [39,40,60]. Thus, these recent results are actually in line with the current data, in which the reconstruction activation of selected targets was higher than unselected targets and low-valued distractors. That said, another recent study reported that a high-valued distractor induced weaker neural representations in early visual cortex compared to the low-valued distractor [38]. However, they found that this was true only when the distractor was physically more salient than the target in a perceptually demanding task [38]. They reasoned that the high sensory competition between low-salience targets and high-salience distractors required top-down attentional suppression of the high-valued distractors [38]. However, this was not the case in the current experiment, in which all stimuli were suprathreshold and matched for luminance. Thus, in the context of our experimental design, we did not find evidence for distractor suppression at either the behavioral or neural level.
In the present study, we showed that an association between reward and color can induce neural modulations in early visual areas V1–V3. This is somewhat surprising, given evidence that neurons in higher visual areas (such as V4, V8, VO1, and inferior temporal cortex) are selectively tuned to chromatic information and responsible for processing color-based top-down modulations [48,58,61–65]. We suggest that value-based modulations in early visual areas may reflect top-down feedback signals from these higher visual areas, where the association between color and reward might be computed. Related to this idea, we found significant distractor-value modulations only in extrastriate visual cortex but not in V1, which may reflect a reentrant signal backpropagated to earlier visual areas. The more robust effects in higher visual areas were also observed for the task-relevant target reconstructions, consistent with previous reports [20,43,50,52,66]. Overall, this pattern of data supports theoretical frameworks suggesting that visual cortex operates as a priority map that indexes the rank-ordered importance of different sensory inputs [20,23–25,39,40,50,52,67,68]. That said, the assumption that the color–reward association can only be computed in higher visual areas has to be considered with caution, because previous studies have found that reward learning can shape neural activity in early visual cortex [24,25,69], and others have also found that primary and extrastriate visual areas contain neuronal populations with an inhomogeneous spatial distribution of color selectivity [42,70].
It is possible that the voluntary allocation of feature-based attention could be one selection mechanism that is guided by learned value. For instance, as subjects learned the value associated with each color, they might have selectively attended to higher-valued colors as each miniblock progressed. Since feature-based attention has been shown to operate globally across the visual scene, the prioritization of higher-valued colors could in turn enhance the neural representations of visual stimuli rendered in those colors even when they appeared outside the focus of attention [70–78]. This global effect of feature-based attention could induce attentional capture to the distractor location when the feature of the distractor matches the prioritized feature in the top-down set—akin to classic contingent capture effects [79–81]. Indeed, there has been an on-going debate about whether value-driven attention should be considered as another form of top-down attention [9,82–86]. However, value-driven attention capture has many characteristics that are distinct from top-down attention [83,85–87]. For example, value-driven capture can happen involuntarily and can be detrimental to behavioral goals, two effects that were demonstrated via the effect of distractor value on RTs in the present study. Therefore, even though value-based modulations of distractor-related neural representations might involve some aspects of top-down feature-based attention, we believe that these modulations reflect involuntary value-driven attentional capture. Also note that although some subjects might choose to voluntarily search for a color previously associated with a high reward, this seems unlikely because this strategy would hurt overall performance and slow the process of evaluating the value of the two relevant targets, as the targets will not always be rendered in the color that the subject is searching for.
Recently, it has been suggested that value-driven attentional capture is valence-free because both monetary gain and loss could produce similar attentional capture effects on RTs [88]. Based on this result, one might speculate that value-based modulations on the distractor-related neural representations shown here could potentially be independent of valence. However, a recent study has suggested that the value-based modulations of object-based stimulus representations in object-selective cortex depend on valence [60]. That said, as discussed earlier, in this study monetary gain and loss were not tied to the distractor stimulus itself [60]; thus, it is hard to conclude from this study alone whether value-based modulations on the distractor-related neural representations are valence-free. Future studies could employ our experimental approach in combination with the manipulation of monetary gain and loss to address this question further.
In summary, we demonstrate that the learned value of irrelevant distractors automatically captures attention and that this interferes with the processing of relevant visual information. This value-driven attentional capture results in increased RTs and heightened distractor representations in retinotopically organized areas of extrastriate visual cortex. Together, our findings suggest that value-driven attentional capture begins with early sensory modulations of distractor representations in visual cortex. Moreover, the modulations of both relevant targets and irrelevant distractors support a recent reframing of the classic dichotomy between bottom-up and top-down biasing factors in favor of a trichotomy that emphasizes a crucial role of value learning on the processing of relevant and irrelevant visual information [9].
Materials and methods
Ethics statement
Participants were recruited from the University of California, San Diego (UCSD), community, and all participants provided written informed consent as required by the local institutional review board at UCSD (IRB# 081318).
Participants
Sixteen neurologically healthy human observers with normal color vision and normal or corrected-to-normal acuity participated in the present study. They then completed 1 scanning session of the main experiment and 1 or 2 sessions of retinotopic mapping scans. Participants were compensated $20 per hour in the scanner, with additional monetary rewards that scaled with their behavioral performance in the value-based learning task (mean $13.13, SD 0.74). Data from 1 subject were excluded because of excessive movement artifacts during the retinotopy scans (>3 mm movement in more than half of the scans), leaving a total of 15 participants in the final analysis (age range 20–34 years old, mean age = 24.6 years ± 4.29 SD).
Stimuli and tasks
Visual stimuli were rear-projected onto on a 115-cm-wide flat screen placed approximately 440 cm from the participant’s eyes at the foot of the scanner bore using an LCD projector (1,024 × 768, 60 Hz, with a gray background, luminance = 8.68 cd/m2). The behavioral paradigms were programmed and presented via a laptop running Windows XP using MATLAB (Mathworks, Natick, MA, United States) and the Psychophysics Toolbox [89,90].
Value-based decision-making task
We adopted a value-based decision-making task that we recently used to show a robust effect of distractor reward history on behavior [6]. Each block started with an instruction period telling participants the locations of the two targets and the location of the irrelevant distractor. The position of each stimulus was indicated by different letter strings located inside three 3 circular placeholders equally spaced from one another (120° polar angle apart with an eccentricity of 3.02° visual angle; Fig 1). The placeholders remained visible for the entire run so that participants knew the precise target and distractor locations. The instruction period was followed by experimental trials in which 3 physically isoluminant checkerboard stimuli of different colors were presented (black paired with red, green, and blue; radius of 1.01° visual angle; spatial frequency of 1.98 cycles per degree visual angle). The stimuli were flickered on–off at 7.5 Hz for 1 s.
Participants were instructed to choose 1 of the 2 targets to maximize their reward and were told that the reward value associated with each color changed across the course of the scan. The reward values associated with each stimulus color were changed every 8 trials (a miniblock). Subjects were not explicitly informed about the length of this miniblock, but they were told that reward–color associations would change dynamically across a small chunk of trials. All 8 possible combinations of the 3 colors and 2 reward values (1 and 9 cents) were presented in each miniblock. The color assignments to each target and distractor stimulus were also counterbalanced within each miniblock. Trial order was pseudorandomized so that the colors of the visual stimuli at 3 stimulus locations swapped in an unpredictable fashion. The assignment of different values to each color was also randomized so that changes in color–reward associations were unpredictable.
Participants were instructed to choose 1 of the 2 targets using 2 fingers on the right hand, as indicated in a diagram displayed before the run started (Fig 1). Importantly, the distractor could never be chosen and was thus choice-irrelevant. After a 1.25-s delay following the offset of the stimulus array, participants received visual feedback indicating the value associated with the chosen target color (“1” or “9”; feedback duration = 0.25 s). If a response was not given before the stimulus offset, they would receive a letter “M” (“miss”) to indicate that no reward was earned on that trial. In a random 20% of trials, rewards were withheld to encourage participants to explore and learn the value of each color (done independently for each of the two targets). In these trials, “0” cents were given, indicating that participants received no reward. The feedback period was followed by a blank intertrial interval with a central fixation for 1.5 s.
Participants completed 6 total blocks with the distractor location remaining stable for 2 consecutive blocks to ensure that participants knew the exact position of the distractor stimulus. Across all blocks the distractor location was counterbalanced across the 3 possible stimulus positions. Each block lasted 4 min 57 s and contained 48 experimental trials and 20 pseudorandomly interleaved null trials. There was a blank period of 9 s at the end of each block. We counterbalanced stimulus configurations across participants to ensure our results were not influenced by any spatial bias. To sample data from the entire circular space across subjects, the stimulus arrays were rotated by 30° polar angle to form 4 configurations (15°-135°-255°, 45°-165°-285°, 75°-195°-315°, and 105°-225°-345°), and these 4 configurations were counterbalanced across subjects. Each subject viewed 1 of these 4 configurations for their entire scanning session.
Visuospatial mapping task
Participants also completed 4–7 blocks of a visuospatial mapping task (1 completed 4 blocks, 1 completed 7 blocks, and the rest completed 6 blocks). The data from this task were then used as an independent data set to train an IEM that was used to reconstruct spatial representations of the targets and distractors in the value-based learning task (see the fMRI analysis section below for more details). Participants were instructed to fixate centrally and to covertly attend to a checkerboard stimulus rendered at 100% Michelson contrast that pseudorandomly appeared at different locations on the screen (3-s duration; the same size, spatial frequency, and flicker frequency as the stimulus in the value-based learning task). The participant’s task was to detect a rare and brief dimming in contrast (19.57% target trials; 0.5-s duration; occurring between 0.5 and 2 s after stimulus onset). On each trial, the checkerboard stimulus was presented at 1 of 37 locations on a triangular grid (1.50° visual angle between vertices), covering a visual space that overlapped with the stimulus locations in the value-based learning task (the first panel in Fig 3A). To smoothly cover the entire circular space, we randomly rotated the entire triangular grid around its center by 0°, 20°, or 40° polar angle across different runs (blue, yellow, and red dots in the first panel in Fig 3A), so there were 111 different stimulus locations in total (see similar methods in [20]). On each run, there were a total of 37 nontargets (1 repeat per location) and 9 targets. Target locations were pseudorandomly drawn from the 37 locations (never repeated within each block). The magnitude of the contrast change was adjusted across trials so that accuracy was at approximately 76% (mean hit = 77.95%, SD = 12.23%). Each stimulus presentation was followed by an intertrial interval of 2–5 s (uniformly distributed). We pseudorandomly interleaved 10 null trials and included a blank period of 8.2 s at the end of the block. Each block lasted 6.28 min.
Behavioral analysis
We first sorted trials from the main value-based decision-making task based on target selection (i.e., target type: selected and unselected), target value (low and high value), distractor value based on previous target rewards associated with the color of the distractor (low and high value), and selection history (i.e., whether the distractor was previously unselected or selected at least once in 3 preceding trials). We chose the 3 previous trials as the analysis window because it yielded the most balanced number of trials between individual conditions. Note that because of the boundary between miniblocks (every 8 trials in which value–color assignments were the same), we could only go back 1 and 2 trials for the second and third trials, respectively. We excluded data from the first trial of every 8 trials in each miniblock to reduce the spillover effect from different sets of value–color assignments.
Next, we examined subjects’ choice preference. To do so, we labeled targets located clockwise (CW) and counterclockwise (CCW) to the distractor CW and CCW targets and computed the probability that participants chose CW over CCW targets and plotted as a function of CW target value and CCW target value (Fig 2A). Next, we plotted the choices as a function of differential target value (CW − CCW) separately for different distractor values and fit individual subjects’ data with the cumulative Gaussian function (Fig 2B). Specifically, we estimated the mean (or mu) and the standard deviation (or sigma) of the cumulative Gaussian function that best fit the choice preference data derived from different distractor values (see Table 1 for mean and SEM; [6]). To test distractor-value modulations on these values, we computed the bootstrap distribution of the difference in these values between the high- and low-distractor-value conditions (i.e., resampling subjects with replacement for 100,000 iterations) and calculated the percentage of values in this distribution that were larger or smaller than 0 to yield a 2-tailed p-value. We performed this statistical analysis separately for previously selected and unselected distractors (see above). Note that we use the 2-parameter cumulative Gaussian model that varied sigma and mu to be consistent with a previous study by our group, even though there were only 3 data points for each experimental condition [6]. To ensure that the fitting procedure provided reliable results, we also used a variant of the model in which we only optimized sigma for each experimental condition while fixing mu at 0. We then did the same resampling analysis on sigma obtained from this reduced model to test the effect of distractor value. We compared the performance of this reduced model with the full model (above) using the nested F test (see [91–93]). To do so, we compared the R2 values between these two models using the following equation:
(1) |
where Rfull2 and Rred2 were obtained from the best fits of the full and reduced models, respectively. Df1 is the number of values in the full model (8 free parameters: 4 sigma values and 4 mu values for high/low-valued previously selected/unselected distractors) minus the number of the values in the reduced model (4 sigma values for different distractor types: high/low-valued previously selected/unselected distractors). Df2 is the number of observations (3 differential target values × 4 distractor types) minus the number of the free parameters in the full model minus 1. The F distribution was then used to estimate the probability that the full model differed significantly from the reduced model.
Finally, we examined the effect of distractor value on RTs. First, we computed the mean RTs across different distractor values for individual subjects. Then, we computed the bootstrap distribution of the RT difference between the high and low distractor-value conditions (i.e., resampling subjects with replacement for 100,000 iterations) and calculated the percentage of values in this distribution that were larger or smaller than 0 (a 2-tailed p-value). We performed this statistical analysis separately for previously selected and unselected distractors. We then compared whether the effect of distractor value was significantly larger in the selected condition than in the unselected condition by a similar procedure that compared the two bootstrap distributions. Since we only observed significantly larger RT differences for previously selected targets, we knew the expected direction of the effect and therefore computed a 1-tailed p-value.
fMRI analysis
fMRI acquisition
All MRI data were acquired on a GE 3T MR750 scanner at the Keck Center for Functional Magnetic Resonance Imaging (CFMRI) at UCSD. Unless otherwise specified, all data were collected using a 32-channel head coil (Nova Medical). We acquired functional data using a multiband echo-planar imaging (EPI) protocol (Stanford Simultaneous Multi-Slice sequence). We acquired 9 axial slices per band at a multiband factor of 8, for 72 total slices (2 × 2 × 2 mm3 voxel size; 800-ms TR; 35-ms TE; 35° flip angle; 104 × 104 cm matrix size). Prior to each functional scan, 16 TRs were acquired as reference images for image reconstruction. Raw k-space data were reconstructed into NIFTI format image files on internal servers using scripts provided by CFMRI. In each session, we also acquired forward and reverse phase-encoding blips to estimate the susceptibility off-resonance field [94]. This was used to correct EPI signal distortion using FSL topup [95,96], the results of which were submitted to further preprocessing stages described below. In each session, we also acquired an accelerated anatomical using parallel imaging (GE ASSET on a FSPGR T1-weighted sequence; 1 × 1 × 1 mm3 voxel size; 8,136-ms TR; 3,172-ms TE; 8° flip angle; 172 slices; 1-mm slice gap; 256 × 192 cm matrix size). This same-session anatomical was coregistered to the functional data. It was also coregistered to a high-resolution anatomical from the retinotopic mapping session(s).
Retinotopic mapping
To identify regions of interest (ROIs) in early visual cortex, we used a combination of retinotopic mapping methods. Individual participants completed meridian mapping (1–2 blocks of approximately 5 min), during which they saw flickering checkerboard “bowties” along the horizontal and vertical meridians while fixating centrally. They also completed several scans of a polar angle mapping task (4–6 blocks of about 6 min), during which participants covertly attended to a rotating checkerboard wedge and detected brief contrast changes (see details in [20,52]). We identified retinotopically organized regions of visual areas V1, V2, and V3 using a combination of retinotopic maps of visual field meridians and polar angle preferences for each voxel in these visual areas and concatenated left and right hemispheres as well as dorsal and ventral aspects of individual areas [97,98]. Visual area borders were drawn on an inflated cortical surface created from a high-resolution anatomical scan (FSPGR T1-weighted sequence; 1 × 1 × 1 mm3; 8,136-ms TR; 3,172-ms TE; 8° flip angle; 172 slices; 1-mm slice gap; 256 × 192 cm matrix size) collected with an 8-channel head coil.
fMRI data preprocessing
Analysis was performed in BrainVoyager 20.2 (Brain Innovation, Maastricht, the Netherlands) supplemented with custom analysis scripts written in MATLAB R2016a (Mathworks, Natick, MA, USA). Using the distortion-corrected images, we first performed slice-time correction, affine motion correction, and temporal high-pass filtering. Then, the functional data were coregistered to the same-session anatomical and transformed to Talairach space. Each voxel’s time course was z-scored within each run. We then built a design matrix with individual trial predictors convolved with a double-gamma HRF (peak = 5 s, undershoot peak = 15 s; response undershoot ratio = 6; response dispersion = 1; undershoot dispersion = 1). We also included a baseline predictor. This allowed us to calculate single-trial beta weights using a general linear model (GLM). These beta weights served as input to the IEM described in the next section.
IEM
In order to create model-based reconstructions of target and distractor stimuli in the value-based learning task from individual ROIs, we employed an IEM for retinotopic space (see Fig 3; also see [42,48–50,52]). First, we computed a spatial sensitivity profile (i.e., an encoding model) for each voxel, parameterized as a weighted sum of experimenter-defined information channels (i.e., spatial filters in the second panel of Fig 3A) using an independent training data set acquired from the visuospatial mapping task (using only the nontarget trials). Then, we inverted the encoding models across all voxels to compute weights on the spatial information channels and used these weights to transform the fMRI data from the value-based learning task into an activation score. This inversion step provides one means of assessing how much information is encoded about a target or distractor stimulus by aggregating modulations across the entire population of voxels in a given visual area.
More specifically, the activation of each voxel is modeled as the weighted sum of 64 bivariate Gaussian-like spatial information channels arrayed in an 8 × 8 rectangular grid (see the second panel of Fig 3). The filter centers were equally spaced by 1.43° visual angle with a full-width half-maximum of 2° visual angle). The Gaussian-like function of each filter is described by:
(2) |
where r is the distance from the filter center and s is a size value indicating the distance between filter centers at which the filter returns to 0. We set values greater than s to 0 (s = 5.0332), resulting in a smooth filter at each position along the grid [50].
We then defined the idealized response of the information channels for each given training trial. To do this, we multiplied a discretized version of the stimulus (n trials × p pixels) by the 64 channels defined by Eq 1 (p pixels × k channels). We then normalized this result so that the maximum channel response is 1. This is C1 in the following equation:
(3) |
where B1 (n trials × m voxels) is the measured fMRI activity of each voxel during the visuospatial mapping task (i.e., beta weights, see fMRI preprocessing section), C1 (n trials × k channels) is the predicted response of each spatial filter (i.e., information channel normalized from 0 to 1), and W is a weight matrix (k channels × m voxels) that quantifies the contribution of each information channel to each voxel. Next, we used ordinary least-squares linear regression to solve for W with the following equation:
(4) |
Here, represents all estimated voxel sensitivity profiles, which we computed separately for each ROI. Next, we used and the measured fMRI activity of each voxel (i.e., beta weights) during each trial of the value-based learning task to estimate the activation of each information channel using the following equation (see Fig 3B):
(5) |
Here, represents the estimated activation of each information channel (n2 trials × k channels), which gives rise to the observed activation pattern across all voxels within that ROI (B2, n2 trials × m voxels). To visualize and coregister trials across 3 stimulus locations, we computed spatial reconstructions by multiplying the spatial profile of each filter by the estimated activation level of the corresponding channel (i.e., computing a weighted sum; the last panel of Fig 3B). We rotated the center position of the spatial filters on each trial of individual participants such that the resulting 2D reconstructions of the target and distractor stimuli share common positions across trials and participants (CCW target, CW target, and distractor locations centered at 30°, 150°, and 270° polar angle, respectively; 3.02° visual angle from the center of the 2D reconstruction). Next, we sorted trials based on selected and unselected target values (low and high), the reward history of the distractor (low and high), and whether the current distractor had been selected or unselected in the 3 previous trials in the same way as we did for the behavioral analysis. Then, we flipped all spatial reconstructions left to right on trials in which the selected target location was on the left (150°) so that the unselected and selected targets always shared common locations on the left and right of the reconstruction, respectively (150° and 30°). This step did not change the position of the distractor, so it stayed at a 270° polar angle. Finally, we averaged the 2D reconstructions across trials with the same trial types for individual participants and then averaged those reconstructions across participants, resulting in the grand-average spatial reconstructions shown in Fig 4A.
fMRI statistical analysis
Following a previous approach [20,53], we extracted the reconstruction activation for each trial type in individual participants by averaging the data within the circular space spanning the entire area of individual stimuli. This was used as our “reconstruction activation” measure. Like the behavioral analyses, all statistical analyses were conducted by resampling relevant values from each subject with replacement for 100,000 iterations and comparing these values across resampling iterations
In the main analysis, we examined the distractor-value modulation on the distractor reconstruction activation for data averaged across V1–V3. To do so, we computed the bootstrap distribution of the difference of the distractor reconstruction activation between the high and low distractor-value conditions and calculated the percentage of values in this distribution that were larger or smaller than 0 (2-tailed). We performed this statistical analysis separately for trials in which the current distractor was previously selected and unselected in preceding trials to examine whether the distractor-value modulation depended on the selection of previous choices. We then compared whether the effect of distractor value was significantly larger in the selected condition than in the unselected condition by a similar procedure that compared the two bootstrap distributions (1-tailed to the known direction of the difference). We repeated the same statistical procedures for individual visual areas and corrected for multiple comparisons using the Holm-Bonferroni method [99].
In addition, we ran 2 auxiliary analyses to examine the influence of the knowledge about the value of a given color and the possible effect of value-mediated priming on the distractor-related representations, respectively. In the first analysis, we divided the distractor-related representation data described above into trials that occurred early or late in each miniblock (second through fifth and sixth through eighth trials, respectively). We then examined the main effect of learning on the distractor-value modulation. To do so, we computed the bootstrap distribution of the difference of the distractor reconstruction activation between the high and low distractor-value conditions for the early and late phase. Then, we calculated the percentage of values between the two phases that were larger or smaller than 0 (2-tailed). We also computed the bootstrap distribution of the difference of the distractor reconstruction activation between the high and low distractor-value conditions and calculated the percentage of values in this distribution that were larger or smaller than 0 (2-tailed). This was done separately for the early and late phases as well as for trials in which the distractors were selected and unselected in previous trials. The Holm-Bonferroni method was used to correct for multiple comparisons.
The second auxiliary analysis was similar to the main analysis except that the selection of previous choices was based on whether the current distractor was selected at least once in 1, 2, or 3 preceding trials. To ensure that trials included in these different data-sorting approaches came from a similar range of trial positions relative to the start of each miniblock, we only included trials from the fourth through eighth trials of each miniblock. We examined the distractor-value modulation on the distractor-related representation in each of these data-sorting approaches by computing the bootstrap distribution of the difference of the distractor reconstruction activation between the high and low distractor-value conditions and calculated the percentage of values in this distribution that were larger or smaller than 0 (2-tailed). The Holm-Bonferroni method was used to correct for multiple comparisons. The procedure was first performed in the data collapsed across V1–V3, and the same analysis was then performed in individual visual areas.
Next, we tested the target selection modulation on the target reconstruction activation for data averaged across V1–V3. To do so, we computed the bootstrap distribution of the difference between the selected and unselected target reconstruction activation and calculated the percentage of values in this distribution that were larger or smaller than 0 (2-tailed). We first performed this on the data collapsed across all distractor types. Then, we assessed the target selection modulations separately for individual distractor values and corrected for multiple comparisons using the Holm-Bonferroni method. Then, we tested for the distractor-value modulation on the target-selection modulation by computing the bootstrap distribution of the difference of the target-selection modulations between the high and low distractor-value conditions and computing the percentage of values in this distribution that were larger or smaller than 0 (2-tailed). This was done separately for trials in which the current distractor was previously unselected and selected in preceding trials. We repeated the same statistical procedures for individual visual areas and corrected for multiple comparisons using the Holm-Bonferroni method.
Finally, we tested whether target-selection modulations depended on the relative value difference between selected and unselected targets, as suggested by previous studies [23–25]. For each target-value condition (same versus different target values) and each visual area, we computed the bootstrap distribution of the difference between the selected and unselected target reconstruction activation and calculated the percentage of values in this distribution that were larger or smaller than 0 (2-tailed). Here, we also corrected for multiple comparisons across different target value conditions and different visual areas using the Holm-Bonferroni method (6 comparisons). Since we found more robust target-selection modulations in higher visual areas in trials in which the selected and unselected targets had different values, we further tested whether the target-selection modulation in V3 was higher than that in V1, whether the target modulation in V2 was higher than that in V1, and whether the target modulation in V2 was higher than that in V1. To do so, we compared the target-selection modulation distributions across these visual areas (1-tailed, due to the known direction of the difference) and corrected for multiple comparisons using the Holm-Bonferroni method.
Supporting information
Acknowledgments
We thank Margaret Henderson for help with data processing, Chaipat Chunharas for assistance with data collection, and Edward Vul for useful discussions.
Abbreviations
- fMRI
functional magnetic resonance imaging
- IEM
inverted encoding model
- RT
response time
- n.s.
no significant difference
Data Availability
All data files and analysis scripts are available from the OSF database: https://osf.io/ky4jh/
Funding Statement
This study was supported by NEI R01-EY025872 to JTS (https://nei.nih.gov/), a James S. McDonnell Foundation Scholar Award to JTS (https://www.jsmf.org/), the Howard Hughes Medical Institute International student fellowship to SI (https://www.hhmi.org/developing-scientists/international-student-research-fellowships), a Royal Thai Scholarship from the Ministry of Science and Technology Thailand to SI (https://most.go.th/main/en), NSF GRFP to VAV (https://www.nsf.gov/), and NEI F32-EY028438 to TCS (https://nei.nih.gov/). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Yantis S, Jonides J. Abrupt visual onsets and selective attention: evidence from visual search. J Exp Psychol Hum Percept Perform. 1984;10(5): 601–621. 10.1037/0096-1523.10.5.601 [DOI] [PubMed] [Google Scholar]
- 2.Theeuwes J, Krueger L, Chun M, Pashler H. Perceptual selectivity for color and form. Percept Psychophys. 1992;51(6): 599–606. [DOI] [PubMed] [Google Scholar]
- 3.Egeth HE, Yantis S. Visual attention: control, representation, and time course. Annu Rev Psychol. 1997;48: 269–297. 10.1146/annurev.psych.48.1.269 [DOI] [PubMed] [Google Scholar]
- 4.Wolfe JM. Visual Search In: Pashler H., editor. Attention. London: University College London Press; 1998. [Google Scholar]
- 5.Anderson BA, Laurent PA, Yantis S. Value-driven attentional capture. Proc Natl Acad Sci U S A. 2011;108(25): 10367–71. 10.1073/pnas.1104047108 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Itthipuripat S, Cha K, Rangsipat N, Serences JT. Value-based attentional capture influences context-dependent decision-making. J Neurophysiol. 2015;114(1): 560–569. 10.1152/jn.00343.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Hickey C, Chelazzi L, Theeuwes J. Reward changes salience in human vision via the anterior cingulate. J Neurosci. Society for Neuroscience; 2010;30(33): 11096–103. 10.1523/JNEUROSCI.1026-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Anderson BA. Going for it: The economics of automaticity in perception and action. Curr Dir Psychol Sci. 2017;26(2): 140–145. 10.1177/0963721416686181 [DOI] [Google Scholar]
- 9.Awh E, Belopolsky A V., Theeuwes J. Top-down versus bottom-up attentional control: A failed theoretical dichotomy. Trends Cogn Sci. 2012;16(8): 437–443. 10.1016/j.tics.2012.06.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Libera C Della, Chelazzi L. Learning to attend and to ignore is a matter of gains and losses. Psychol Sci. 2009(6);20: 778–784. 10.1111/j.1467-9280.2009.02360.x [DOI] [PubMed] [Google Scholar]
- 11.Gluth S, Spektor MS, Rieskamp J. Value-based attentional capture affects multi-alternative decision making. Elife. 2018;7: 1–36. 10.7554/eLife.39659 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Moher J, Anderson BA, Song JH. Dissociable Effects of Salience on Attention and Goal-Directed Action. Curr Biol. Elsevier Ltd; 2015;25(15): 2040–2046. 10.1016/j.cub.2015.06.029 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Hickey C, van Zoest W. Reward-associated stimuli capture the eyes in spite of strategic attentional set. Vision Res. 2013;92: 67–74. 10.1016/j.visres.2013.09.008 [DOI] [PubMed] [Google Scholar]
- 14.Maclean MH, Diaz GK, Giesbrecht B. Irrelevant learned reward associations disrupt voluntary spatial attention. Atten Percept Psychophys. 2016;78(7): 2241–2252. 10.3758/s13414-016-1103-x [DOI] [PubMed] [Google Scholar]
- 15.Maclean MH, Giesbrecht B. Neural evidence reveals the rapid effects of reward history on selective attention. Brain Res. 2015;1606: 86–94. 10.1016/j.brainres.2015.02.016 [DOI] [PubMed] [Google Scholar]
- 16.MacLean MH, Giesbrecht B. Irrelevant reward and selection histories have different influences on task-relevant attentional selection. Attention, Perception, Psychophys. 2015;77(5): 1515–1528. 10.3758/s13414-015-0851-3 [DOI] [PubMed] [Google Scholar]
- 17.Krebs RM, Boehler CN, Egner T, Woldorff MG. The neural underpinnings of how reward associations can both guide and misguide attention. J Neurosci. 2011(26);31: 9752–9759. 10.1523/JNEUROSCI.0732-11.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Sali AW, Anderson BA, Yantis S, Mostofsky SH, Rosch KS. Reduced value-driven attentional capture among children with ADHD compared to typically developing controls. J Abnorm Child Psychol. Journal of Abnormal Child Psychology; 2018;46(6): 1187–1200. 10.1007/s10802-017-0345-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Anderson BA, Faulkner ML, Rilee JJ, Yantis S, Marvel CL, et al. Attentional bias for non-drug reward is magnified in addiction. 2014(6);21: 499–506. 10.1037/a0034575 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Sprague TC, Itthipuripat S, Vo VA, Serences JT. Dissociable signatures of visual salience and behavioral relevance across attentional priority maps in human cortex. J Neurophysiol. 2018;119(6): 2153–2165. 10.1152/jn.00059.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Zhang X, Zhaoping L, Zhou T, Fang F. Neural activities in V1 create a bottom-up saliency map. Neuron. 2012;73(1): 183–92. 10.1016/j.neuron.2011.10.035 [DOI] [PubMed] [Google Scholar]
- 22.Chen C, Zhang X, Wang Y, Zhou T, Fang F. Neural activities in V1 create the bottom-up saliency map of natural scenes. Exp Brain Res. 2016;234(6): 1769–1780. 10.1007/s00221-016-4583-y [DOI] [PubMed] [Google Scholar]
- 23.Serences JT. Value-based modulations in human visual cortex. Neuron. 2008;60(6): 1169–1181. 10.1016/j.neuron.2008.10.051 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Serences JT, Saproo S. Population response profiles in early visual cortex are biased in favor of more valuable stimuli. J Neurophysiol. 2010;104(1): 76–87. 10.1152/jn.01090.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Stănişor L, van der Togt C, Pennartz CMA, Roelfsema PR. A unified selection signal for attention and reward in primary visual cortex. Proc Natl Acad Sci U S A. National Academy of Sciences; 2013;110(22): 9136–41. 10.1073/pnas.1300117110 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Baruni JK, Lau B, Salzman CD. Reward expectation differentially modulates attentional behavior and activity in visual area V4. Nat Neurosci. 2015;18(11): 1656–1663. 10.1038/nn.4141 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Berridge KC, Robinson TE. What is the role of dopamine in reward: Hedonic impact, reward learning, or incentive salience? Brain Res Rev. 1998;28(3): 309–369. 10.1016/S0165-0173(98)00019-8 [DOI] [PubMed] [Google Scholar]
- 28.Ikeda T, Hikosaka O. Reward-dependent gain and bias of visual responses in primate superior colliculus. Neuron. 2003;39(4): 693–700. 10.1016/S0896-6273(03)00464-1 [DOI] [PubMed] [Google Scholar]
- 29.Ikemoto S, Panksepp J. The role of nucleus accumbens dopamine in motivated behavior: A unifying interpretation with special reference to reward-seeking. Brain Res Rev. 1999;31(1): 6–41. 10.1016/S0165-0173(99)00023-5 [DOI] [PubMed] [Google Scholar]
- 30.Redgrave P, Prescott TJ, Gurney K, Redgrave P, Prescott TJ. Is the short-latency dopamine response too short to signal reward error? Trends Neurosci. 1999; 22(4):146–51. [DOI] [PubMed] [Google Scholar]
- 31.Wise RA. Dopamine, learning and motivation. Nat Rev Neurosci. 2004;5(6): 483–494. 10.1038/nrn1406 [DOI] [PubMed] [Google Scholar]
- 32.Alcaro A, Huber R, Panksepp J. Behavioral functions of the mesolimbic dopaminergic system: An affective neuroethological perspective. Brain Res Rev. 2007;56(2): 283–321. 10.1016/j.brainresrev.2007.07.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Schultz W. Getting formal with dopamine and reward. Neuron. 2002;36(2): 241–263. Available from: https://www.cell.com/action/showPdf?pii=S0896-6273%2802%2900967-4 [DOI] [PubMed] [Google Scholar]
- 34.Schultz W, Dayan P, Montague PR. A neural substrate of prediction and reward. Science. 1997;275(5306): 1593–1600. 10.1126/science.275.5306.1593 [DOI] [PubMed] [Google Scholar]
- 35.Anderson BA, Laurent PA, Yantis S. Value-driven attentional priority signals in human basal ganglia and visual cortex. Brain Res. Elsevier; 2014;1587: 88–96. 10.1016/J.BRAINRES.2014.08.062 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Anderson BA. The attention habit: How reward learning shapes attentional selection. Ann N Y Acad Sci. 2016;1369(1): 24–39. 10.1111/nyas.12957 [DOI] [PubMed] [Google Scholar]
- 37.Anderson BA. Neurobiology of value-driven attention. Curr Opin Psychol. 2019;29: 27–33. 10.1016/j.copsyc.2018.11.004 [DOI] [PubMed] [Google Scholar]
- 38.Gong M, Jia K, Li S. Perceptual competition promotes suppression of reward salience in behavioral selection and neural representation. J Neurosci. 2017;37(26): 6242–6252. 10.1523/JNEUROSCI.0217-17.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Hickey C, Peelen M V. Reward selectively modulates the lingering neural representation of recently attended objects in natural scenes. J Neurosci. Society for Neuroscience; 2017;37(31): 7297–7304. 10.1523/JNEUROSCI.0684-17.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Hickey C, Peelen M V. Neural mechanisms of incentive salience in naturalistic human vision. Neuron. Cell Press; 2015;85(3): 512–518. 10.1016/j.neuron.2014.12.049 [DOI] [PubMed] [Google Scholar]
- 41.Gaspelin N, Luck SJ. The role of inhibition in avoiding distraction by salient stimuli. Trends Cogn Sci. 2018;22(1): 79–92. 10.1016/j.tics.2017.11.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Brouwer G, Heeger D. Decoding and reconstructing color from responses in human visual cortex. J Neurosci. 2009;29(44): 13992–14003. 10.1523/JNEUROSCI.3577-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Bressler DW, Fortenbaugh FC, Robertson LC, Silver MA. Visual spatial attention enhances the amplitude of positive and negative fMRI responses to visual stimulation in an eccentricity-dependent manner. Vision Res. 2013;85: 104–112. 10.1016/j.visres.2013.03.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Puckett AM, DeYoe EA. The attentional field revealed by single-voxel modeling of fMRI time courses. J Neurosci. 2015;35(12): 5030–5042. 10.1523/JNEUROSCI.3754-14.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Gouws AD, Alvarez I, Watson DM, Uesaki M, Rogers J, Morland AB. On the role of suppression in spatial attention: evidence from negative BOLD in human subcortical and cortical structures. J Neurosci. 2014;34(31): 10347–10360. 10.1523/JNEUROSCI.0164-14.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Muller NG, Kleinschmidt A. Dynamic interaction of object- and space-based attention in retinotopic visual areas. J Neurosci. 2003;23(30): 9812–9816. 10.1109/OCEANSE.2005.1511760 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Fischer J, Whitney D. Attention narrows position tuning of population responses in V1. Curr Biol. 2009;19(16): 1356–1361. 10.1016/j.cub.2009.06.059 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Brouwer GJ, Heeger DJ. Categorical clustering of the neural representation of color. J Neurosci. 2013;33(39): 15454–15465. 10.1523/JNEUROSCI.2472-13.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Itthipuripat S, Sprague T, Serences J. Functional MRI and EEG index complementary attentional modulations. J Neurosci. 2019;39(31): 6162–6179 10.1523/JNEUROSCI.2519-18.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Sprague TC, Serences JT. Attention modulates spatial priority maps in the human occipital, parietal and frontal cortices. Nat Neurosci. 2013;16(12): 1879–87. 10.1038/nn.3574 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Sprague TC, Ester EF, Serences JT. Reconstructions of information in visual spatial working memory degrade with memory load. Curr Biol. 2014;24(18): 1–7. 10.1016/j.cub.2014.07.066 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Vo VA, Sprague TC, Serences JT. Spatial tuning shifts increase the discriminability and fidelity of population codes in visual cortex. J Neurosci. 2017;37(12): 3386–3401. 10.1523/JNEUROSCI.3484-16.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Sprague TC, Ester EF, Serences JT. Restoring latent visual working memory representations in human cortex Article. Neuron. 2016;91(3): 694–707. 10.1016/j.neuron.2016.07.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Sprague TC, Adam KCS, Foster JJ, Rahmati M, Sutterer DW, Vo VA. Inverted encoding models assay population-level stimulus representations, not single-unit neural tuning. Eneuro. 2018;5(3): ENEURO.0098-18.2018. 10.1523/ENEURO.0098-18.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Tootell RB, Hadjikhani NK, Vanduffel W, Liu AK, Mendola JD, Sereno MI, et al. Functional analysis of primary visual cortex (V1) in humans. Proc Natl Acad Sci U S A. 1998;95(3): 811–7. Available from: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=PMC33802 10.1073/pnas.95.3.811 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Louie K, Khaw MW, Glimcher PW. Normalization is a general neural mechanism for context-dependent decision making. Proc Natl Acad Sci. 2013;110(15): 6139–6144. 10.1073/pnas.1217854110 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Chau BKH, Kolling N, Hunt LT, Walton ME, Rushworth MFS. A neural mechanism underlying failure of optimal choice with multiple alternatives. Nat Neurosci. 2014;17(3): 463–470. 10.1038/nn.3649 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Brouwer GJ, Heeger DJ. Decoding and reconstructing color from responses in human visual cortex. J Neurosci. 2009;29(44): 13992–14003. 10.1523/JNEUROSCI.3577-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Seitz AR, Kim D, Watanabe T. Rewards evoke learning of unconsciously processed visual stimuli in adult humans. Neuron. 2009;61(5): 700–707. 10.1016/j.neuron.2009.01.016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Barbaro L, Peelen M V., Hickey C. Valence, not utility, underlies reward-driven prioritization in human vision. J Neurosci. 2017;37(43): 1128–17. 10.1523/JNEUROSCI.1128-17.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Zeki SM. Functional organization of a visual area in the posterior bank of the superior temporal sulcus of the rhesus monkey. J Physiol. 1974;236(3): 549–573. 10.1113/jphysiol.1974.sp010452 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Conway BR, Moeller S, Tsao DY. Specialized color modules in macaque extrastriate cortex. Neuron. 2007;56(3): 560–573. 10.1016/j.neuron.2007.10.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Zeki S, Bartels A. The architecture of the colour centre in the human visual brain: new results and a review. Eur J Neurosci. 2000;12(1): 172–193. 10.1046/j.1460-9568.2000.00905.x [DOI] [PubMed] [Google Scholar]
- 64.Brewer AA, Liu J, Wade AR, Wandell BA. Visual field maps and stimulus selectivity in human ventral occipital cortex. Nat Neurosci. 2005;8(8): 1102–1109. 10.1038/nn1507 [DOI] [PubMed] [Google Scholar]
- 65.Hadjikhani N, Liu AK, Dale AM, Cavanagh P, Tootell RBH. Retinotopy and color sensitivity in human visual cortical area V8. Nat Neurosci. 1998;1(3): 235–241. 10.1038/681 [DOI] [PubMed] [Google Scholar]
- 66.Bressler DW, Silver MA. Spatial attention improves reliability of fMRI retinotopic mapping signals in occipital and parietal cortex. Neuroimage. 2010;53(2): 526–33. 10.1016/j.neuroimage.2010.06.063 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Serences JT, Yantis S. Selective visual attention and perceptual coherence. Trends Cogn Sci. 2006;10(1); 38–45. 10.1016/j.tics.2005.11.008 [DOI] [PubMed] [Google Scholar]
- 68.Sprague TC, Saproo S, Serences JT. Visual attention mitigates information loss in small- and large-scale neural codes. Trends Cogn Sci. 2015; 19(4): 215–226. 10.1016/j.tics.2015.02.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Shuler MG, Bear MF. Reward timing in the primary visual cortex. Science. 2006;311(5767): 1606–1609. 10.1126/science.1123513 [DOI] [PubMed] [Google Scholar]
- 70.Treue S, Martinez-Trujillo JC. Feature-based attention influences motion processing gain in macaque visual cortex. Nature. 1999;399(6736): 575–579. 10.1038/21176 [DOI] [PubMed] [Google Scholar]
- 71.Bichot NP, Bichot NP, Rossi AF, Desimone R. Parallel and serial neural mechanisms for visual search in macaque area V4. 2010;529(5721): 529–535. 10.1126/science.1109676 [DOI] [PubMed] [Google Scholar]
- 72.Maunsell JHR, Treue S. Feature-based attention in visual cortex. Trends Cogn Sci. 2006; 29(6): 317–322 10.1016/j.tins.2006.04.001 [DOI] [PubMed] [Google Scholar]
- 73.Saenz M, Buracas GT, Boynton GM. Global effects of feature-based attention in human visual cortex. Nat Neurosci. 2002;5(7): 631–632. 10.1038/nn876 [DOI] [PubMed] [Google Scholar]
- 74.Andersen SK, Hillyard SA, Müller MM. Attention facilitates multiple stimulus features in parallel in human visual cortex. Curr Biol. 2008;18(13): 1006–1009. 10.1016/j.cub.2008.06.030 [DOI] [PubMed] [Google Scholar]
- 75.Serences JT, Boynton GM. Feature-based attentional modulations in the absence of direct visual stimulation. Neuron. 2007;55(2): 301–312. 10.1016/j.neuron.2007.06.015 [DOI] [PubMed] [Google Scholar]
- 76.Ester EF, Serences JT, Awh E. Spatially global representations in human primary visual cortex during working memory maintenance. J Neurosci. 2009;29(48): 15258–15265. 10.1523/JNEUROSCI.4388-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Ester EF, Sutterer DW, Serences JT, Awh E. Feature-selective attentional modulations in human frontoparietal cortex. J Neurosci. 2016;36(31): 8188–8199. 10.1523/JNEUROSCI.3935-15.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78.Ester EF, Sprague TC, Serences JT. Parietal and frontal cortex encode stimulus-specific mnemonic representations during visual working memory. Neuron. 2015;87(4): 893–905. 10.1016/j.neuron.2015.07.013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Folk CL, Remington RW, Wright JH. The structure of attentional control: Contingent attentional capture by apparent motion, abrupt onset, and color. Journal of Experimental Psychology: Human Perception and Performance. 1994:20(2): 317–329. 10.1037/0096-1523.20.2.317 [DOI] [PubMed] [Google Scholar]
- 80.Folk CL, Remington RW, Johnston JC. Contingent attentional capture: A reply to Yantis (1993). Journal of Experimental Psychology: Human Perception and Performance. 1993;19(3): 682–685. 10.1037/0096-1523.19.3.682 [DOI] [Google Scholar]
- 81.Folk CL, Remington RW, Johnston JC. Involuntary covert orienting is contingent on attentional control settings. Journal of Experimental Psychology: Human Perception and Performance. 1992:18(4): 1030–1044. 10.1037/0096-1523.18.4.1030 [DOI] [PubMed] [Google Scholar]
- 82.Wolfe JM, Butcher SJ, Lee C, Hyle M. Changing your mind: On the contributions of top-down and bottom-up guidance in visual search for feature singletons. J Exp Psychol Hum Percept Perform. 2003;29(2): 483–502. 10.1037/0096-1523.29.2.483 [DOI] [PubMed] [Google Scholar]
- 83.Theeuwes J. Visual selection: Usually fast and automatic; Seldom slow and volitional. J Cogn. 2018;1(1): 1–15. 10.5334/joc.13 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 84.Theeuwes J. Visual selection: Usually fast and automatic; Seldom slow and volitional; A reply to commentaries. J Cogn. 2018;1(1): 1–4. 10.5334/joc.32 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Egeth H. Comment on Theeuwes’s characterization of visual selection. J Cogn. 2018;1(1): 1–3. 10.5334/joc.29 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Failing M, Theeuwes J. Selection history: How reward modulates selectivity of visual attention. Psychon Bull Rev. 2018;25(2): 514–538. 10.3758/s13423-017-1380-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Kim H, Anderson BA. Dissociable components of experience-driven attention. Curr Biol. 2019;29(5): 841–845.e2. 10.1016/j.cub.2019.01.030 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88.Wang L, Yu H, Zhou X. Interaction between value and perceptual salience in value-driven attentional capture. J Vis. 2013;13(3): 1–13. 10.1167/13.3.1 [DOI] [PubMed] [Google Scholar]
- 89.Brainard DH. The Psychophysics Toolbox. Spat Vis. 1997;10(4): 433–436. [PubMed] [Google Scholar]
- 90.Watson AB, Pelli DG. QUEST: a Bayesian adaptive psychometric method. Percept Psychophys. 1983;33(2): 113–120. 10.3758/BF03202828 [DOI] [PubMed] [Google Scholar]
- 91.Itthipuripat S, Ester EF, Deering S, Serences JT. Sensory gain outperforms efficient readout mechanisms in predicting attention-related improvements in behavior. J Neurosci. 2014;34(40): 13384–98. 10.1523/JNEUROSCI.2277-14.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 92.Itthipuripat S, Cha K, Byers A, Serences JT. Two different mechanisms support selective attention at different phases of training. PLoS Biol. 2017;15(6): 1–38. 10.1371/journal.pbio.2001724 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 93.Pestilli F, Carrasco M, Heeger DJ, Gardner JL. Attentional enhancement via selection and pooling of early sensory responses in human visual cortex. Neuron. 2011;72(5): 832–846. 10.1016/j.neuron.2011.09.025 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 94.Andersson JLR, Skare S, Ashburner J. How to correct susceptibility distortions in spin-echo echo-planar images: application to diffusion tensor imaging. Neuroimage. 2003;20(2): 870–888. 10.1016/S1053-8119(03)00336-7 [DOI] [PubMed] [Google Scholar]
- 95.Smith SM, Jenkinson M, Woolrich MW, Beckmann CF, Behrens TEJ, Johansen-Berg H, et al. Advances in functional and structural MR image analysis and implementation as FSL. Neuroimage. 2004;23(Suppl 1): S208–S219. 10.1016/J.NEUROIMAGE.2004.07.051 [DOI] [PubMed] [Google Scholar]
- 96.Jenkinson M, Beckmann CF, Behrens TEJ, Woolrich MW, Smith SM. FSL. Neuroimage. 2012;62(2): 782–790. 10.1016/j.neuroimage.2011.09.015 [DOI] [PubMed] [Google Scholar]
- 97.Engel SA, Rumelheart DE, Wandell BA. fMRI of human visual cortex. Nature. 1994;369(6481): 525 10.1038/369525a0 [DOI] [PubMed] [Google Scholar]
- 98.Swisher JD, Halko MA, Merabet LB, Mcmains SA, Somers DC. Visual topography of human intraparietal sulcus. J Neurosci. 2007;27(20): 5326–5337. 10.1523/JNEUROSCI.0991-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Dunn OJ. Multiple comparisons among means. J Am Stat Assoc. 1961;56: 52–64. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
All data files and analysis scripts are available from the OSF database: https://osf.io/ky4jh/