Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2019 Jun 14.
Published in final edited form as: J Cogn Neurosci. 2018 May 23;30(9):1281–1297. doi: 10.1162/jocn_a_01284

Attending to What and Where: Background Connectivity Integrates Categorical and Spatial Attention

Alexa Tompary 1, Naseem Al-Aidroos 2, Nicholas B Turk-Browne 3,4
PMCID: PMC6570400  NIHMSID: NIHMS1025595  PMID: 29791296

Abstract

Top–down attention prioritizes the processing of goal-relevant information throughout visual cortex based on where that information is found in space and what it looks like. Whereas attentional goals often have both spatial and featural components, most research on the neural basis of attention has examined these components separately. Here we investigated how these attentional components are integrated by examining the attentional modulation of functional connectivity between visual areas with different selectivity. Specifically, we used fMRI to measure temporal correlations between spatially selective regions of early visual cortex and category-selective regions in ventral temporal cortex while participants performed a task that benefitted from both spatial and categorical attention. We found that categorical attention modulated the connectivity of category-selective areas, but only with retinotopic areas that coded for the spatially attended location. Similarly, spatial attention modulated the connectivity of retinotopic areas only with the areas coding for the attended category. This pattern of results suggests that attentional modulation of connectivity is driven both by spatial selection and featural biases. Combined with exploratory analyses of frontoparietal areas that track these changes in connectivity among visual areas, this study begins to shed light on how different components of attention are integrated in support of more complex behavioral goals.

INTRODUCTION

Top–down attention prioritizes the processing of goal-relevant information throughout the visual system. This enables the extraction of the most useful information from complex and noisy input. One neural mechanism that supports this prioritization is the amplified response of visual areas that code for goal-relevant information. Such amplified responses have been observed when attention is directed to particular spatial locations (e.g., Connor, Gallant, Preddie, & Van Essen, 1996; Motter, 1993; Moran & Desimone, 1985) and when it is directed to specific features, objects, and categories (e.g., Furey et al., 2006; Serences, Schwarzbach, Courtney, Golay, & Yantis, 2004; O’Craven, Downing, & Kanwisher, 1999; Corbetta, Miezin, Dobmeyer, Shulman, & Petersen, 1990). Another neural mechanism that supports prioritization is increased coupling between areas. Such increased coupling occurs among goal-relevant visual regions and with frontoparietal areas that control attention, again for both spatial (Griffis, Elkhetali, Burge, Chen, & Visscher, 2015; Bosman et al., 2012; Gregoriou, Gotts, Zhou, & Desimone, 2009; Saalmann, Pigarev, & Vidyasagar, 2007) and nonspatial attention (Baldauf & Desimone, 2014; Al-Aidroos, Said, & Turk-Browne, 2012; Zhou & Desimone, 2011). Most of the studies above treat spatial and nonspatial attention as orthogonal to isolate their neural mechanisms. However, in real life, these two forms of attention are closely intertwined, such as when scanning the tables at a restaurant for a friend or when looking for a highway exit with a nearby gas station. This integration has been studied in the behavioral literature (e.g., Soto & Blanco, 2004; Goldsmith & Yeari, 2003), through response amplification in visual cortex (e.g., Müller & Kleinschmidt, 2003; Treue & Trujillo, 1999), and in terms of overlap in frontoparietal control areas (Shomstein & Yantis, 2004, 2006; Giesbrecht, Woldorff, Song, & Mangun, 2003). However, it is unclear how the integration of spatial and nonspatial attention is supported by coupling between areas.

In the current study, we examined changes in BOLD correlations between visual areas while attention is simultaneously allocated to both a location and a category. We presented a stream of faces and a stream of scenes, with one stream appearing to the left of fixation and the other to the right. Participants were instructed to detect repetitions in one stream, a task that could be performed using either spatial attention or categorical attention but would benefit from both. We used a background connectivity approach to quantify attentional modulation of coupling. This approach is conceptually similar to resting-state connectivity (Fox & Raichle, 2007), except if measured during tasks (Tompary, Duncan, & Davachi, 2015; Duncan, Tompary, & Davachi, 2014; Al-Aidroos et al., 2012; Norman-Haignere, McCarthy, Chun, & Turk-Browne, 2012; Summerfield et al., 2006). We examined the background connectivity of V4—a retinotopically organized occipital region that is modulated by attention (Cohen & Tong, 2015; Saenz, Buracas, & Boynton, 2002; McAdams & Maunsell, 1999)—with the fusiform face area (FFA) and the parahippocampal place area (PPA), ventral temporal regions sensitive to faces and scenes that are also modulated by attention (Serences et al., 2004; O’Craven et al., 1999).

This design differs from our previous study of categorical attention (Al-Aidroos et al., 2012), in which both face and scene streams were overlaid at one location. That study found stronger V4 correlations with FFA when faces were attended and with PPA when scenes were attended. By presenting the streams in different hemifields, we can exploit the contralateral organization of V4 to evaluate the role of spatial attention in this effect. On one hand, categorical attention might modulate coupling only between areas that code for spatially attended locations, similar to findings that spatial attention gates feature-based modulation of baseline activity (e.g., McMains, Fehd, Emmanouil, & Kastner, 2007). For example, when attending to faces on the right, FFA may show increased coupling only with the left V4. On the other hand, coupling might be modulated by categorical attention even for areas that code for spatially unattended locations, similar to findings that feature-based modulation of responses occurs globally (e.g., Andersen, Hillyard, & Müller, 2013; Saenz et al., 2002; Treue & Trujillo, 1999). In the example above, FFA would also show increased coupling with the right V4. Adjudicating between these hypotheses will help determine whether attentional modulation of coupling supports the integration of spatial and nonspatial attention.

METHODS

Participants

Fifteen naïve right-handed members of the Princeton community (8 men, ages 18–35 years, mean = 23.6 years) participated for monetary compensation ($20/hr). All participants had normal or corrected visual acuity and provided informed consent. The study protocol was approved by the institutional review board for human participants at Princeton University.

Overview of Procedure

Each participant completed two experimental sessions while undergoing fMRI scanning and eye tracking. Both sessions were completed within 10 days of each other. In the first session, participants completed a repetition detection task in which they selectively attended to particular spatial locations and categories. We used this main session to investigate how different attentional states altered background connectivity in visual cortex. In the second session, participants completed a retinotopic mapping task followed by a face/scene localizer task. We used these tasks to identify ROIs in occipital cortex (left and right V1–V4) and ventral temporal cortex (FFA and PPA), respectively.

First Session

The first session contained six runs: two “rest” runs followed by four “attention” runs. During each rest run, participants were shown a white circular fixation point (radius = 0.2°) centered on a gray background for 441 sec (Figure 1A). Participants were instructed to passively view the fixation point without performing any overt task. We used these rest runs to quantify baseline levels of connectivity.

Figure 1.

Figure 1.

Experimental design. (A) In Session 1, participants first completed two rest runs. They were instructed to fixate the dot in the center of the screen. (B) During attention runs, participants simultaneously viewed two streams of images and performed a repetition detection task on one stream. Faces were presented in one stream and scenes in the other. The location and category of the attended stream was manipulated such that participants attended a unique combination of spatial locations and categories in each run: faces on the left, faces on the right, scenes on the left, and scenes on the right. (C) In Session 2, participants completed six retinotopy scans: four polar scans with rotating wedges and two eccentricity scans with expanding/ contracting rings. These scans were used to identify V4 and ventral/dorsal V1–V3. (D) Participants then completed two functional localizers for FFA and PPA. These runs were identical to the attention runs, except only one stream was presented at a time. The blocks within each scan alternated between face and scene stimuli.

The attention runs were used to assess background connectivity in different attentional states (Figure 1B). Attention runs started with 9 sec of fixation followed by an on–off block design that alternated between twelve 18-sec stimulation blocks and twelve 18-sec fixation blocks (441 sec total). Each stimulation block consisted of a sequence of twelve 1-sec trials, separated by a 500-msec ISI. Each trial contained a face image and a scene image. Faces were drawn from a set of 24 photographs from the NimStim database (www.macbrain.org/resources.htm, neutral expressions). Scenes were drawn from a set of 24 photographs of single houses collected from the Internet and stock photograph CDs (Norman-Haignere et al., 2012). Images were converted to gray-scale, cropped using a circular mask, and resized to a 6° radius. The two possible locations for the images were the left and right quadrants of the upper visual field (centered 8° above the horizontal meridian and 8° to the left and right of the vertical meridian, respectively). These locations projected to the ventral stream in the visual processing hierarchy and to the contralateral right and left hemispheres, respectively.

Within each attention run, all images from a given category appeared in the same location (i.e., scenes on the left and faces on the right, or vice versa). We manipulated attentional state by instructing participants to perform a repetition detection task for one location and category, reporting whether the current task-relevant image was an immediate repeat of the last task-relevant image within a 1-sec response window. Participants reported repetitions by pressing a button on a button box and otherwise withheld responses. To ensure that they attended to the entire stream of 12 images per block, task-relevant repetition targets occurred either once or twice unpredictably (split 50/50). In blocks with one repetition, the repetition randomly occurred at any point in the block. When repetitions occurred twice, they were separated by at least one non-repetition trial, and the second repetition happened within the final quarter of the block. To increase the need for selection, repetitions also occurred in the task-irrelevant location and category, but they were to be ignored. Otherwise, the images within a block were selected randomly from the respective category set without replacement.

Counterbalancing the location of face and scene images and the relevant category for the repetition detection task across attention runs resulted in a 2 (attention to face vs. scene category) × 2 (attention to the left vs. right location) factorial design with four conditions: attention to faces in the left visual field, attention to faces in the right visual field, attention to scenes in the left visual field, and attention to scenes in the right visual field. Every participant completed one run of each type in an order that was pseudocounterbalanced across participants. Note that the stimulus configuration differed by attention condition: Faces appeared on the right and scenes appeared on the left for two of the conditions (attention to faces on the right, attention to scenes on the left), whereas faces appeared on the left and scenes appeared on the right for the other two conditions (attention to faces on the left and attention to scenes on the right).

Second Session

The second session contained eight runs: six “retinotopy” runs followed by two “localizer” runs. The retinotopy runs were used to identify visual areas—left and right V1–V4—containing topographic maps of the visual field. The stimuli and procedure were identical to the retinotopy task we have used previously (Al-Aidroos et al., 2012; see also, Arcaro, McMains, Singer, & Kastner, 2009). A dynamic and colorful circular checkerboard (4-Hz flicker, 14° radius) was shown at fixation to stimulate different visual receptive fields (Figure 1C). During “polar angle” runs, the checkerboard was masked so that only a 45° wedge was visible, and the mask rotated clockwise or counterclockwise at a rate of 9°/sec (40-sec period) creating the perception of a rotating wedge. These runs allowed us to identify the preferred polar angle of voxels in visual cortex. During “eccentricity” runs, the checkerboard was masked to create the perception of an expanding or contracting annulus. The expansion/contraction period was 40 sec, and the thickness and rate of expansion of the annulus increased logarithmically with eccentricity to approximate the cortical magnification factor of early visual cortex (from 0.3 to 1.5°). These runs allowed us to identify the preferred eccentricity of stimulation for voxels in visual cortex. All participants completed one clockwise and one counterclockwise polar angle run (order counterbalanced), followed by one expanding and one contracting eccentricity run (order counterbalanced), and followed by a second clockwise and a second counter-clockwise polar angle run (same order as initial polar angle runs). To ensure that participants were actively processing the visual display, they were instructed to detect when a central fixation point (radius = 0.25°) dimmed from white to gray (every 2–5 sec) using a button box.

The localizer runs were used to identify the FFA and PPA (Figure 1D). These runs were similar to the attention runs described above but differed in two ways. First, we presented one image at a time, always in the left visual field for one run and in the right visual field for the other run (order counterbalanced). Second, the category of these images varied across blocks, alternating between faces and scenes within run (starting category randomized). In all other respects, the localizer runs were identical to the attention runs, including stimulus sizes, possible locations, photographs, trial timings, block timings, and task (i.e., repetition detection). For the first two participants, the localizer runs were completed at the end of the first session.

Eye Tracking

A critical component of the attention runs was that participants fixated a centrally presented point while covertly attending to peripheral spatial locations for the repetition detection task. To be confident that stimuli fell in the anticipated retinotopic areas, we assessed fixation using a 60-Hz camera-based SMI iViewX MRI-LR eye tracker, mounted at the foot of the scanner bed. Calibration was performed at the beginning of each session and between runs when necessary. Drift correction was occasionally applied during fixation phases of the block (i.e., when only a fixation point was present). Data were recorded from whichever eye was tracked most accurately. The resulting gaze direction time courses were segmented into saccades, fixations, or noise (i.e., blinks or lost signal) using an algorithm similar to Nyström and Holmqvist (2010). Gaze direction could not be accurately recorded from 2/15 participants due to inconsistent pupil reflection or partially occluded pupils. We discarded eye-tracking data from seven runs total across all of the remaining participants because of poor or lost calibration. The main connectivity results were unaffected when participants and runs without eye-tracking data were excluded. Within successfully recorded runs, 14.5% of data were discarded because of blinks and artifacts. We assessed participants’ ability to maintain fixation by calculating the average horizontal distance of gaze position from fixation and the percentage of samples more than 2° from fixation.

Image Acquisition

fMRI data were acquired with a 3T scanner (Siemens Skyra) using a 16-channel head coil. Functional images during rest, attention, and localizer runs were acquired with a gradient-echo EPI sequence (repetition time [TR] = 1.5 sec, echo time = 28 msec, flip angle = 64°, matrix = 64 × 64, resolution = 3 × 3 × 3.5 mm), with 27 interleaved axial slices aligned to the anterior/posterior commissure line. TRs were time-locked to the presentation of images (or the beginning of the run, in the case of rest runs). Functional images during retinotopy runs were acquired with a similar sequence but at higher spatial resolution (TR = 2.0 sec, echo time = 40 msec, flip angle = 71°, matrix = 128 × 128, resolution = 2 × 2 × 2.5 mm), with 25 interleaved slices aligned parallel to the calcarine sulcus.

For each session and functional sequence, we collected T1 fast low-angle shot (FLASH) scans, aligned co-planar to the functional scan. Phase and magnitude field maps were acquired coplanar with the functional scans and with the same resolution to correct B0 field inhomogeneities. Finally, a high-resolution magnetization-prepared rapid acquisition gradient-echo (MPRAGE) anatomical scan was acquired for surface reconstruction and registration.

Image Preprocessing

fMRI data were analyzed using FSL (Smith et al., 2004), FreeSurfer (Dale, Fischl, & Sereno, 1999; Fischl, Sereno, & Dale, 1999), and MATLAB (MathWorks). Preprocessing began by removing the skull from images to improve registration. Data from the first six volumes of functional runs were discarded for T1 equilibration. Remaining functional images were motion-corrected using MCFLIRT, corrected for slice acquisition time, high-pass filtered with a 100-sec period cutoff, debiased using the field maps in FUGUE, and spatially smoothed (retinotopy runs: 3-mm FWHM; all other runs: 5-mm FWHM). The processed images were registered to the high-resolution MPRAGE and Montreal Neurological Institute standard brain (MNI152). Retinotopy scans were first aligned to the FLASH to improve registration to the MPRAGE.

ROIs

We identified FFA and PPA ROIs using the localizer runs. As in previous studies (Al-Aidroos et al., 2012; Norman-Haignere et al., 2012), we limited analyses to the right FFA because the right hemisphere is dominant for face processing (Verosky & Turk-Browne, 2012; Yovel, Tambini, & Brandman, 2008). We limited analyses to the right PPA to facilitate comparison with the right FFA, because its selectivity to spatial location is roughly equivalent (Schwarzlose, Swisher, Dang, & Kanwisher, 2008), although it likely exhibits less of a contralateral bias than FFA (Hemond, Kanwisher, & Op de Beeck, 2007). For each run, we fit a general linear model (GLM) with regressors for the face and scene blocks, modeled with 18-sec boxcar regressors that were convolved using FSL’s double-gamma hemodynamic response function. The temporal derivatives of these regressors were also included, as well as six regressors for different directions of head motion. We first contrasted parameter estimates for face versus scene blocks within the left and right visual field runs and then collapsed across visual field in a second-level GLM. We defined the FFA by choosing the voxel in the right lateral fusiform cortex most selective for face stimuli (i.e., face > scene blocks) and defined the PPA by choosing the voxel in the right collateral sulcus/parahippocampal cortex most selective for scene stimuli (i.e., scene > face blocks). In all analyses, BOLD signal from each ROI was extracted as a weighted average of the surrounding voxels, with weights determined by a Gaussian kernel with 5-mm FWHM centered on the peak voxel.

Based on the retinotopy runs, we drew ROIs in both hemispheres using FreeSurfer, including V4 and ventral and dorsal V1–V3. Retinotopy runs were prewhitened, corrected for hemodynamic lag (3 sec), and then phase-decoded to determine the preferred polar angle and eccentricity of stimulated voxels in visual cortex. We further accounted for hemodynamic lag by averaging phase estimates from clockwise and counterclockwise runs and from expansion and contraction runs. To facilitate ROI drawing, we plotted phase maps on 2-D surfaces segmented and flattened from the MPRAGE scans at the white matter/gray matter boundary. We identified boundaries between ROIs based on the relative locations of polar angle reversals and foveally stimulated regions (Arcaro et al., 2009; Wandell, Dumoulin, & Brewer, 2007).

Evoked Responses

We first assessed how attention modulated the amplitude of the BOLD responses in our ROIs. The preprocessed runs were fit with GLMs that captured the average evoked response in the attention blocks. We used a finite impulse response (FIR) model rather than a canonical hemodynamic response function because it avoided assumptions about the shape and timing of the response across voxels. Each model included 23 FIR regressors for the first 23 volumes of each block (including stimulation and fixation), with a delta function at the corresponding time point of all blocks and zeros elsewhere. The 24th volume of each block (just before the next block) was left out to serve as a baseline. We also included six regressors for the motion correction parameters from preprocessing. The resulting parameter estimates were extracted from each ROI, transformed to percent signal change, and submitted to statistical tests.

Background Connectivity

To assess how attention modulated coupling, we examined correlations in the BOLD signal across ROIs for different attention runs. We used a background connectivity approach, which removed evoked responses and other noise sources that could induce spurious correlations. This approach has previously been used successfully to measure how shifts in categorical attention regulate the strength of interactions between regions of visual cortex (Al-Aidroos et al., 2012; Norman-Haignere et al., 2012). First, the preprocessed data were submitted to a “nuisance” model. Specifically, we fit a GLM to each attention and rest run that included regressors for the global mean BOLD signal, the six motion parameters from preprocessing, and the BOLD signal from four seeds in the ventricles and four seeds in white matter (left/right, anterior/posterior). The residuals of this model were submitted to a second “evoked” GLM designed to remove evoked responses in a similar manner to above, except with an additional FIR regressor for the 24th volume in each block (because we were interested in removing all evoked activity, rather than measuring responses relative to baseline). In prior work, we found that removing the average response across blocks gave rise to similar findings compared with a model that removed separate responses to each block, thus capturing the mean and variance of evoked responses in each voxel (Al-Aidroos et al., 2012, Figure S5). We measured background connectivity by extracting the residuals of this model from each ROI and relating them to each other with Pearson correlation. We applied Fisher’s r-to-z transformation to all coefficients before statistical tests. All reported correlations, including in the figures, are presented as the original r values to facilitate interpretation and visualization.

Temporal Multivariate–Univariate Dependence Analysis

Finally, we conducted an exploratory analysis to identify potential control regions that could be responsible for modulating background connectivity in visual cortex. The measure of background connectivity described above provides a point estimate of the overall temporal relationship between brain areas. However, because none of the areas had perfect correlations, there was by definition variance in the extent to which individual time points supported this relationship. We reasoned that the activity of an additional area involved in controlling attention would fluctuate synchronously with this variance, with more activity (and greater modulation) during time points that strengthened the relationship among the other areas and less activity during time points that weakened it. This analysis involved two steps: (1) quantifying how much each time point contributed to background connectivity between visual ROIs and (2) relating this variance to the activity time course of other voxels in the brain.

To quantify the contribution of each time point, we adapted the multivariate–univariate dependence (MUD) analysis technique, which was developed to measure how much each voxel contributed to a spatial correlation computed across multiple voxels (Aly & Turk-Browne, 2016a, 2016b). Switching from spatial to temporal correlation, we measured how much each time point contributed to the background connectivity previously computed across time points for each attention run. For example, time points when two visual areas were both active (or both inactive) would support a positive temporal correlation, whereas time points when one was active and the other was inactive (or vice versa) would support a negative correlation. This can be quantified for all time points by first normalizing the time course of each region, subtracting the mean and dividing by the root sum-of-squares of the mean-centered data, and then taking the pointwise product of the two normalized time courses (for more details, see Aly & Turk-Browne, 2016a). These products correspond directly to the contribution of each time point to the overall correlation—their sum is the Pearson correlation coefficient (Wang, Cohen, Li, & Turk-Browne, 2015; Turk-Browne, 2013; Worsley, Chen, Lerch, & Evans, 2005).

The result of this first step is a normalized product time course for each pair of ROIs in all attention runs (Figure 7A). To identify potential control regions, we then used this time course as a regressor in a GLM of voxelwise activity in the residuals of the “evoked” model. We restricted analysis to frontal and parietal lobes (defined with MNI Structural Atlas) given the importance of frontoparietal cortex for top–down control (Noudoost, Chang, Steinmetz, & Moore, 2010; Serences & Yantis, 2007; Corbetta & Shulman, 2002) and to avoid circularity with the visual regions used to compute the product time courses. The resulting parameter estimates indicate the extent to which each voxel’s activity was correlated with the products—that is, whether the region was more (or less) active during time points that increased or decreased background connectivity. A separate GLM was fit for each attention run (face–right, face–left, scene–right, scene–left) and pair of regions (left V4–FFA, right V4–FFA, left V4–PPA, and right V4-PPA). Of the 16 GLMs per subject, we sorted the resulting statistical maps by spatial and categorical relevance, mirroring the conditions in the main background connectivity analyses: space+/category+, space+/category–, space–/category+, space–/category–. These maps were combined at the group level within condition, treating subject as a random effect, with the reliability of each voxel’s correlation with the product time course compared against zero using the randomise function in FSL. We corrected for multiple comparisons using cluster mass thresholding (cluster-forming threshold z = 3.0).

RESULTS

Behavior

Behavior in the repetition detection task during the attention runs was generally fast and accurate (Figure 2A, B). RTs for correctly detected repetitions were analyzed using a 2 (Space: left vs. right) × 2 (Category: faces vs. scenes) repeated-measures ANOVA. This analysis revealed a significant main effect of Space, F(1, 14) = 5.31, p = .04, reflecting a left visual field advantage (Kimura, 1966). The main effect of Category and the interaction were not significant (ps > .33).

Figure 2.

Figure 2.

Behavioral results. (A) Participants were faster to respond to repetitions in the left visual field. (B) Participants were more accurate with face versus scene repetitions in the LVF. *p < .05. Error bars reflect ±1 SEM.

The accuracy of repetition detection was measured with (Grier, 1971). The same 2 (space) × 2 (category) ANOVA revealed a significant interaction, F(1, 14) = 18.89, p < .001, but no reliable main effects of Category, F(1, 14) = 1.86, p = .20, or Space, F(1, 14) = 1.32, p = .27. The interaction reflects the fact that face repetitions were detected more accurately than scene repetitions in the left, t(l4) = 4.00, p = .001, but not the right visual field, t(l4) = −2.04, p = .06, consistent with the particularly strong left visual field bias in face processing (Verosky & Turk-Browne, 2012; Yovel et al., 2008; De Renzi, Perani, Carlesimo, Silveri, & Fazio, 1994). Reponses to repetitions in the unattended stream were rare (mean rate = 2.11%, SEM = 1.15%), confirming that participants were able to selectively attend to repetitions with the attended category and location.

Because stimuli were presented 8° in the periphery, we were attuned to the possibility of systematic biases in eye position across conditions. We considered two measures—average horizontal displacement of gaze location from fixation and the percentage of samples more than 2° from fixation (i.e., the inner boundary of our stimuli)—and subjected each measure to a 2 (Space) × 2 (Category) ANOVA. Participants tended to look slightly to the right of fixation on average (M = +1.01°, SD = 0.78), but this bias did not differ across attention runs. Indeed, there were no reliable main effects or interactions for either measure (ps > .32).

Evoked Responses

Although our primary interest was attentional modulation of connectivity, we first examined the more conventional index of attention—the amplitude of the BOLD response in the right and left V4, the right FFA, and the right PPA (Figure 3A). To quantify the amplitude of response (Figure 3B), we collapsed across ROIs and conditions and selected time points across all blocks in which the BOLD percent signal change reliably differed from rest. This resulted in a set of “stimulated” volumes (3–17; 1–2 and 18–24 were “nonstimulated”), over which we averaged the BOLD percent signal change within each ROI and condition.

Figure 3.

Figure 3.

Evoked responses. (A) Time course of FIR parameter estimates for all conditions in every region, modeling the average evoked BOLD response. (B) Average FIR parameter estimates over “stimulated” time points, as an index of response amplitude. Shaded regions and error bars reflect ±1 within-subject SEM.

First, we analyzed how V4 was modulated by attention. We computed a 2 (ROI: left/right V4) × 2 (Spatial attention: left vs. right) × 2 (Categorical attention: face vs. scene) ANOVA. We hypothesized that there would be a contralateral effect of Spatial attention, which was supported by an interaction between ROI and Spatial attention, F(1, 14) = 29.22, p < .001. This interaction was driven by greater BOLD response in the right V4 for the left versus right attention runs, t(14) = 3.57, p = .003, and in the left V4 for the right versus left attention runs, t(14) = 4.30, p < .001. This is consistent with the fact that V4 activity is modulated by spatial attention (Moran & Desimone, 1985). This ANOVA also revealed an interaction between ROI and Categorical attention, F(1, 14) = 7.05, p = .02. This was driven by a marginally increased BOLD response in the right V4 when attending to scenes versus faces, t(14) = 2.12, p = .05, but no corresponding difference in the left V4, t(14) = 0.33, p = .74. A complete report of the effects in this ANOVA can be found in Table 1A.

Table 1.

Evoked Responses

Effect DFn DFd F P

A. ANOVA Results: Left and Right V4
ROI 1 14   0.07   .70
Space 1 14   0.14   .72
Category 1 14   1.97   .18
ROI × Space 1 14 29.22 <.001*
ROI × Category 1 14   7.05   .02*
Space × Category 1 14   0.14   .72
ROI × Space × Category 1 14   1.78   .20
B. ANOVA Results: FFA and PPA
ROI 1 14   0.04   .85
Space 1 14   0.91   .36
Category 1 14   2.45   .14
ROI × Space 1 14   2.63   .13
ROI × Category 1 14 25.69 <.001*
Space × Category 1 14   0.71   .41
ROI × Space × Category 1 14   0.87   .37

(A) Modulation of the right and left V4 by attention. (B) Modulation of FFA and PPA by attention; DFd = degrees of freedom denominator of the F ratio; DFn = degrees of freedom for the numerator of the F ratio.

*

p < .05.

Next, we analyzed how ventral temporal regions were modulated by attention. We computed a 2 (ROI: FFA vs. PPA) × 2 (Spatial attention: left vs. right) × 2 (Categorical attention: face vs. scene) ANOVA. We hypothesized that the effect of categorical attention would depend on the selectivity of the regions, which was supported by an interaction between ROI and Categorical attention, F(1, 14) = 25.69, p < .001. This interaction was driven by greater BOLD response in FFA for face versus scene attention, t(l4) = 2.28, p = .04, and in PPA for scene versus face attention, t(l4) = 4.95, p < .001. This is consistent with past findings that categorical attention modulates FFA and PPA (O’Craven et al., 1999). A complete report of the effects in this ANOVA can be found in Table 1B.

Background Connectivity

We previously demonstrated that coupling between V4 and FFA/PPA is modulated by categorical attention (Al-Aidroos et al., 2012), thus we first examined how categorical and spatial attention interacted to modulate coupling in the same regions. We investigated background connectivity between four pairs of ROIs: left V4 and FFA, right V4 and FFA, left V4 and PPA, and right V4 and PPA. We labeled these connections differently for each attention run based on whether the V4 region coded for the attended location and whether the ventral temporal region coded for the attended category (Figure 4). This led to four connection types: both the V4 region and ventral temporal region coded for attended information (space+/category+), only the V4 region coded for attended information (space+/category−), only the ventral temporal region coded for attended information (space−/category+), and neither region coded for attended information (space−/category−). For example, when attending to faces on the right: left V4/FFA was labeled space +/category+, left V4/PPA was labeled space +/category−, right V4/FFA was labeled space−/category+, and right V4/PPA was labeled space−/category−.

Figure 4.

Figure 4.

Connectivity types. Connectivity was measured between four pairs of ROIs: left V4 and FFA, right V4 and FFA, left V4 and PPA, and right V4 and PPA. These pairs were labeled based on the relevance of the constituent regions to the attended stimuli. Left and right V4 were relevant to right and left attention, respectively. FFA and PPA were relevant to face and scene attention, respectively.

This labeling resulted in a factorial design that allowed us to evaluate how categorical attention modulated the coupling of regions that code for spatially attended and unattended locations. Using a 2 (Spatial relevance of connection: + vs. −) × 2 (Categorical relevance of connection: + vs. −) ANOVA, we found a main effect of Categorical relevance, F(1, 14) = 9.66, p = .008 (Figure 5A), replicating prior observations of increased V4 connectivity with FFA when attending to faces and with PPA when attending to scenes (Al-Aidroos et al., 2012). In contrast, there was no effect of spatial relevance, F(1, 14) = 2.51, p = .14.

Figure 5.

Figure 5.

Background connectivity. (A) Spatial and categorical attention enhanced background connectivity between the ventral temporal region that coded for the attended category and the V4 that coded for the attended location. (B) Background connectivity in each ROI pair by spatial and categorical selectivity. *p < .05. Error bars reflect within-subject SEM.

However, the main effect of Categorical relevance was qualified by an interaction between Spatial and Categorical relevance, F(1, 14) = 9.66, p = .008. This interaction was driven by two factors. First, there was greater connectivity with the V4 that coded for the attended location relative to the unattended location, but only when paired with the ventral temporal region that coded for the attended category (space+/category+ > space−/category+: t(14) = 2.69, p = .02) and not the ventral temporal region that coded for the unattended category (space+/category− > space−/category—: t(14) = –0.18, p = .86). Second, there was greater background connectivity with the ventral temporal region that coded for the attended category relative to the unattended category, but only when paired with the V4 region that coded for the attended location (space+/category+ > space+/category−: t(14) = 3.38, p = .004) and not the V4 region that coded for the unattended location (space−/category+ > space−/category−: t(14) = 0.85, p = .41). Together these results suggest that the effects of attention on background connectivity were spatially and categorically specific, enhancing connectivity only with the V4 that coded for the attended location and the ventral temporal region that coded for the attended category.

Accounting for Differences in Perceptual Input

One potential concern is that, in some conditions, attentional state is confounded with the stimulus presentation. For example, when comparing “space+/category+” connectivity and “space+/category−” connectivity between the left V4 and FFA, in the first condition, faces are presented on the right, houses are presented on the left, and the participant attends to the right, and in the second condition, houses are presented on the right, faces are presented on the left, and the participant attends to the right. These two conditions differ in the type of stimulus that is presented at the relevant location in addition to the type of attention employed. Thus, our findings could be explained by selectivity in how features of the stimulus are communicated between regions. For instance, increased background connectivity could reflect greater transmission of face-specific features from left V4 to FFA when faces are presented on the right, rather than modulations by attention.

We addressed this concern in two ways. First, we focused on comparisons where stimulus locations were held constant, but the spatial and categorical relevance of the ROIs differed by attention. We began by comparing coupling between space+/category+ versus space−/category− conditions. Critically, the ROI pairs included in the space−/category− condition comprise two regions that code for the unattended stimulus. For example, when attending to faces on the right, scenes are presented on the left. Therefore, if background connectivity captures bottom–up transfer of information about the stimulus, rather than attentional signals, we would expect equivalent connectivity between FFA/left V4 (space+/category+) and PPA/right V4 (space−/category−). In contrast, background connectivity between the spatially relevant V4 and categorically relevant temporal lobe was reliably greater than background connectivity between the irrelevant V4 and temporal lobe regions, t(14) = 2.77, p = .02. The other comparison that equates stimulus presentation is space+/category− versus space−/category+, where only one region in a given pair codes for the attended stimulus. Here, there is no reliable difference in background connectivity between these conditions, t(14) = −0.59, p = .56. This suggests that attention modulates coupling specifically in pairs of ROIs that code both for the category and location of the attended stimulus.

Second, we recomputed the same 2 (Spatial relevance of connection: + vs. −) × 2 (Categorical relevance of connection: + vs. −) ANOVA performed in the main findings, but restricted the analysis to attention runs that have the same configuration of stimuli. In runs with faces on the left and scenes on the right, we found a main effect of Categorical relevance, F(1, 14) = 6.19, p = .03, no effect of Spatial relevance, F(1, 14) = 1.646, p = .22, and a reliable interaction, F(1, 14) = 16.73, p = .001, consistent with the main findings. In runs with faces on the right and scenes on the left, we found a marginal effect of Categorical relevance, F(1, 14) = 3.81, p = .07, no effect of Spatial relevance, F(1, 14) = 2.04, p = .18, and no interaction, F(1, 14) = 0.29, p = .60. We directly compared these two subsets of the data in a 2 (Spatial relevance of connection: + vs. −) × 2 (Categorical relevance of connection: + vs. −) × 2 (Stimulus location: face–left/scene–right, face–right/scene–left) ANOVA and found a marginal three-way interaction, F(1, 14) = 3.24, p = .09, suggesting that the interaction between spatial and categorical relevance on background connectivity is at least partly driven by the attention runs where faces appeared on the left and scenes appeared on the right. However, the fact that this interaction was still reliable in half of our data, holding the stimulus configuration constant, suggests that it reflects differences in top–down attention rather than differences in perceptual input.

Hemisphere Effects in Ventral Temporal Regions

In all prior and subsequent analyses, FFA and PPA are constrained to the right hemisphere because of hemifield biases in FFA and for compatibility with prior work (see Methods). However, we expected that the left FFA and the left PPA would exhibit the same modulation of background connectivity with the right and left V4. When examining background connectivity between the left FFA/PPA and V4 in a 2 (Spatial relevance of connection: + vs. −) × 2 (Categorical relevance of connection: + vs. −) ANOVA, we find a main effect of Categorical relevance, F(1, 14) = 5.57, p = .03, and an interaction between Spatial relevance and Categorical relevance, F(1,14) = 13.97, p = .002, consistent with findings reported with the right FFA and the right PPA. To directly compare effects of the right and left hemisphere regions in the ventral temporal lobe, we submitted background connectivity values to a 2 (Hemisphere: right FFA/PPA, left FFA/PPA) × 2 (Spatial relevance of connection: + vs. −) × 2 (Categorical relevance of connection: + vs. −) ANOVA. No effect of Hemisphere or interaction was found (all Fs < 0.97, all ps > .34). This suggests that the attentional modulation of coupling with V4 is consistent across ventral temporal regions in the right and left hemispheres.

Intrinsic versus Evoked Contributions to Background Connectivity

To what extent do these modulations reflect spontaneous internal interactions between cortical regions versus correlations in BOLD activity evoked by external stimuli? The use of background connectivity above was an attempt to remove time-locked stimulus responses before examining correlations between regions. However, it remains possible that our model did not perfectly capture all evoked responses. To assess whether these responses contributed to our results, we performed two additional analyses.

We first evaluated whether background connectivity was modulated in the absence of evoked responses during the “nonstimulated” volumes (i.e., time points with responses that did not differ from baseline). The same 2 (Spatial relevance of connection: +vs. −) × 2 (Categorical relevance of connection: + vs. −) ANOVA restricted to these volumes revealed an interaction, F(1,14) = 9.91, p = .007, no effect of Spatial relevance, F(1, 14) = 1.37, p = .26, and no effect of Categorical relevance, F(1,14) = 0.01, p = .91. The interaction was driven by greater connectivity with the V4 that coded for attended location relative to the V4 that coded for the unattended location, but only when paired with the ventral temporal region that coded for the attended category (space+/category+ > space−/category+: t(14) = 2.30, p = .04) and not with the ventral temporal region that coded for the unattended category (space+/category− > space−/category−: t(14) = −0.27,p = .79), consistent with the findings observed when using all volumes. From these results, we can conclude that the modulation of background connectivity by spatial and categorical attention reflects intrinsic activity.

We next turned to another approach for assessing the contribution of evoked responses. The same task structure was used for every participant, and thus, evoked responses should be synchronized across participants (Hasson, Nir, Levy, Fuhrmann, & Malach, 2004). In contrast, intrinsic activity is idiosyncratic, and we would not expect it to be synchronized. Thus, examining correlations across rather than within participants can help diagnose evoked versus intrinsic activity. Specifically, if the observed modulation of background connectivity between V4 and FFA/PPA within participants reflects evoked activity, the same results should be obtained by correlating one participant’s V4 with other participants’ FFA/PPA. To perform this analysis, the right and left V4 background time courses from each participant were correlated with the average FFA and PPA background time courses over all other participants, separately for every ROI pair and attention run. The 2 (Spatial relevance of connection: + vs. −) × 2 (Categorical relevance of connection: + vs. −) ANOVA applied to these across-participant correlations revealed no main effect of Spatial relevance, F(1, 14) = 0.51, p = .49, no main effect of Categorical relevance, F(1, 14) = 0.38, p = .55, and no interaction, F(1, 14) = 0.05, p = .82. To verify that there were no alignment or other problems and that this analysis would have been sensitive to shared responses, the average correlation across conditions before removing the evoked activity in the background connectivity procedure was robust (mean r = .43; t(14) = 10.02, p < .001), and reliably greater than in the resulting background data (mean r = .019; t(14) = 1.59, p = .13; comparison: t(14) = 10.44, p < .001). Thus, these time courses contained neither across-participant correlations nor modulation of these correlations by attention. Critically, these same time courses produced the main results when the correlations were calculated within participant (Figure 5), emphasizing the role of intrinsic activity.

Relationship with Behavior

Al-Aidroos et al. (2012) found that individual differences in attentional modulation of V4–FFA/PPA background connectivity correlated with behavioral performance during the attention runs, as measured by greater d′ in the repetition detection task. We investigated this by defining a neural attentional modulation score for each participant ([space+/category+] − [space+/categ− + space−/ categ+ + space−/categ−]/3]. This score was not correlated with participants’ average behavioral A′ across attention runs, r(13) = −.23, p = .40. This null finding may reflect the fact that, in the current study, hit rates were higher on average (necessitating the use of A′ to characterize accuracy) and less variable than in the prior study—a ceiling effect that may have made it difficult to assess individual differences.

Areas V1–V3

We next conducted exploratory analyses to assess whether modulation of connectivity by spatial attention is specific to V4 or if similar effects could be seen earlier in the visual stream. We repeated the main background connectivity analysis in retinotopic areas V1–V3. These areas can be found in both the ventral and dorsal streams, with selectivity for the upper and lower visual field, respectively. Because spatial attention was directed to the upper visual field (where images appeared), we first focused on ventral areas (Figure 6A). We computed a 2 (Spatial relevance of connection: + vs. −) × 2 (Categorical relevance of connection: + vs. −) ANOVA for each ROI. We found no interactions in ventral V1 or V2 (Fs < 1.87, ps > .19). In ventral V3, there was a main effect of Spatial relevance, F(1, 14) = 6.07, p = .03, a marginal effect of Categorical relevance, F(1, 14) = 3.91, p = .07, and a marginal interaction, F(1, 14) = 3.31, p = .09. Because the lower visual field was always spatially irrelevant, we reasoned that there might be different, possibly opposite effects in dorsal areas (Figure 6B). Indeed, there was a marginal main effect of Categorical relevance in dorsal V1, F(1, 14) = 3.99, p = .07, and V2, F(1, 14) = 3.57, p = .08, but not dorsal V3, F(1, 14) = 0.86, p = .37, and no effects of spatial relevance (Fs < 1.13, ps > .31). There was a reliable interaction in dorsal V3, F(1, 14) = 11.56, p = .004, but not dorsal V1 or V2 (Fs < 0.69, ps > .42). Interestingly, and different from V4, the effects in dorsal regions were driven by lower connectivity for relevant connections.

Figure 6.

Figure 6.

Background connectivity in V1, V2, and V3. (A) The interaction between spatial and categorical relevance in V4 was marginally significant in ventral V3 but did not extend into V1 and V2. (B) In the dorsal stream, background connectivity with V1 and V2 was modulated by categorical relevance, but in the opposite direction of V4, with decreased coupling to the category-selective ventral temporal region. Overall, attention to the upper visual field (all conditions) had diverging effects relative to rest, with relatively enhanced connectivity for ventral areas and decreased connectivity for dorsal areas. Error bars reflect within-subject SEM.

To further explore how spatial attention to the upper and lower visual fields modulated connectivity, we compared overall background connectivity collapsed across the four conditions against the average of the two rest runs using a 3 (ROI: V1, V2, V3) × 2 (Stream: ventral, dorsal) × 2 (State: attention vs. rest). There was a main effect of ROI, F(2, 28) = 9.43, p < .001, with increasing background connectivity from V1 to V3, and a main effect of Stream, F(1, 14) = 45.88, p < .001, with stronger connectivity for ventral than dorsal streams; the main effect of State was not reliable, F(1, 14) = 0.24, p = .64. Critically, there was a highly robust interaction between Stream and State, F(1, 14) = 17.70, p < .001, with background connectivity during attention higher than during rest in the ventral stream and lower in the dorsal stream. There was also a marginal interaction between ROI, Stream, and State, F(2, 28) = 3.31, p = .05; no other interactions reached significance (Fs > 1.90, ps > .16).

Temporal MUD Analysis

Our primary hypotheses concerned modulation of interactions within the visual system. However, this modulation likely results from the deployment of top–down control regulated elsewhere in the brain. Previous studies have demonstrated that frontoparietal cortex supports such control (Noudoost et al., 2010; Serences & Yantis, 2007; Corbetta & Shulman, 2002), including by communicating with visual areas (Chadick & Gazzaley, 2011; Gregoriou et al., 2009; Saalmann et al., 2007). To explore the role of frontoparietal cortex in the modulation of background connectivity within visual cortex, we searched for voxels in the frontal and parietal lobes whose activity at a given time point predicted how much that time point contributed to the background connectivity between visual areas (Figure 7AB). By performing this analysis separately for each attention condition, we could evaluate which control structures support spatial attention, categorical attention, and their integration (Figure 7C; Table 2).

Figure 7.

Figure 7.

Temporal MUD analysis. (A) For each attention run, the normalized time courses from the right or left V4 were multiplied with the normalized time courses from FFA or PPA. The resulting time course represented the contribution of each time point to the background connectivity between the two regions (the sum of the products is the Pearson correlation coefficient). (A) The graphical intuition for this analysis is that if both ROI time courses have positive or negative normalized values at a given time point, then the time point falls in the first or third quadrant of a scatterplot relating the normalized activity of one region to another. Because of the mean-centering of both regions, the best-fit line passes through the origin and thus points in these quadrants support a positive slope. If one time course has a positive value and the other has a negative value, then the time point falls in the second or fourth quadrant, which supports a negative slope. The relative balance of points in quadrants 1/3 versus 2/4 thus determines the sign of the slope, and because of the variance normalization, the value of the slope is the Pearson correlation coefficient. (C) When examining voxels whose BOLD time courses correlated with the product time course for each condition, we observed significant clusters in the core attention network: bilateral IFG, bilateral IPS/SPL, and right TPJ. Contrasts corrected for multiple comparisons using cluster mass thresholding (p < .05; cluster-forming threshold z = 3).

Table 2.

Temporal MUD Results

Extent x y z p

Space+/Category+
R SPL/IPS 1234    34.9 −54.4    34.0 <.001
R IFG    913    44.7    13.6    15.7 <.001
R superior frontal gyrus    328    12.6    88.8    60.5    .009
R frontal pole    264    30.9    43.6    19.0    .016
L IFG    259 −39.4    23.7    11.9    .014
Posterior cingulate gyrus    184      0.1 −25.5    26.4    .031
Space+/Category–
R SPL/IPS    922    28.4 − 67.9    31.5    <.001
L SPL/IPS    281 −20.2 − 68.5    39.3    .006
R TPJ    131    61.2 − 38.3    16.9    .032
R insula/frontal operculum    102    34.7    24.5      3.3    .044
Space –/Category+
L postcentral gyrus    374 −64.2    −4.2    26.3    .001
R orbitofrontal cortex    155    32.2    30.3    −8.3    .010
Space –/Category–
R TPJ    268    66.2 − 34.1    20.0    .004
R precentral gyrus    268    54.2     4.1    34.1    .005
R lateral occipital cortex    204    43.5 −69.2    15.0    .007
L SPL/IPS    150 −29.8 −73.2    41.4    .023
R paracingulate gyrus    113      7.0    16.7    43.8    .032
R postcentral gyrus    101    62.0 −14.8    27.7    .045

Frontoparietal regions whose activity correlates with background connectivity within visual cortex. R and L indicate right and left hemisphere. Extent is size of the clusters in voxels. x, y, and z coordinates indicate the center of gravity in MNI space (mm). p corresponds to corrected significance of cluster.

In the space+/category+ condition—that is, the spatially relevant V4 area paired with the categorically relevant ventral temporal area—there were six frontoparietal clusters whose moment-by-moment activity tracked variation in background connectivity (corrected p < .05), including the right anterior intraparietal sulcus (IPS)/superior parietal lobule (SPL) and bilateral inferior frontal gyrus (IFG). These are core areas of the dorsal and ventral attention networks (Corbetta & Shulman, 2002). In the space+/category− condition, there were three clusters from this network: left and right IPS/SPL and right TPJ. In the space−/category+ condition, two clusters emerged: left postcentral gyrus and right OFC. In the space−/category− condition, six clusters emerged, including but not limited to the right TPJ, the left SPL/IPS, and the right lateral occipital cortex.

DISCUSSION

Based on prior work about feature-based attention, our key finding that categorical attention selectively modulated connectivity with spatially relevant regions may seem surprising. A number of studies have found that attending to a feature at one location enhances the response to that feature throughout the visual field (Andersen et al., 2013; Bondarenko et al., 2012; Cohen & Maunsell, 2011; Andersen, Fuchs, & Müller, 2009; Hayden & Gallant, 2009; McAdams & Maunsell, 1999; Treue & Trujillo, 1999; Chelazzi, Duncan, Miller, & Desimone, 1998; Chelazzi, Miller, Duncan, & Desimone, 1993). Furthermore, attention to motion or color in one location enhances the BOLD response to the same motion or color in a distant, unattended spatial location (Saenz et al., 2002). This spread of feature-based attention even extends to spatial locations with no stimulus present (Serences & Boynton, 2007). Behaviorally, top–down attention enhances sensitivity to features outside the attended location as well (Liu & Hou, 2011; Liu & Mance, 2011; White & Carrasco, 2011). We might then have expected that attention to faces or scenes in the right (or left) hemifield would enhance FFA and PPA connectivity with both the left and right V4. Such an effect would be comparable to work by Serences and Boynton (2007), who found that information about an attended feature (motion) is decodable from regions whose receptive fields were not stimulated by the task. However, this is inconsistent with the selectively contralateral effect we observed. One possible explanation for our results is that evoked and intrinsic signals fundamentally differ in their spatial specificity, with the former being spatially global and the latter being spatially local. Indeed, a prior fMRI study of feature-based attention found that intrinsic baseline shifts in anticipation of a stimulus were restricted to the expected location of the stimulus, whereas V4 responses evoked by the stimulus were modulated at both expected and unexpected locations (McMains et al., 2007). In our case, rather than focusing on a prestimulus period, we isolated intrinsic signals by regressing out stimulus-evoked responses before calculating background connectivity. Another possibility is our use of categorical attention, which is characterized by object membership (Vecera & Farah, 1994; Duncan, 1984), rather than attention toward low-level features like color and orientation (Duncan & Nimmo-Smith, 1996; Baylis & Driver, 1992). The neural signals that determine the spatial specificity of category-based attention may differ from those of feature-based attention, although both category-and feature-based attention interact with spatial attention behavior (Kravitz & Behrmann, 2011) and can be spatially global (Peelen, Fei-Fei, & Kastner, 2009).

Competition is another factor that may explain this finding (Moran & Desimone, 1985). Several studies have found enhanced responses to attended features in unattended locations only when there are competing stimuli within or outside the attended spatial location (Painter, Dux, Travis, & Mattingley, 2014; Zhang & Luck, 2009; Saenz, Buraĉas, & Boynton, 2003), although others have found such effects regardless of competition (Bartsch et al., 2015). In our experiment, the stimuli in the unattended location were a different category than those in the attended location (e.g., faces on the left while attending to scenes on the right) and likely did not compete with the attended category/location. The lack of competing category may minimize any global enhancements in BOLD signal or correlated activity that would be expected otherwise. Furthermore, behaviorally, spatial attention and feature-based attention seem to operate independently except when unattended distractors are competitive with attended targets (Leonard, Balestreri, & Luck, 2015; White, Rolfs, & Carrasco, 2015). This suggests that, although feature-based attention may globally enhance neural responses across the visual field, there may be later cognitive and neural processes that integrate over spatial and feature-based attention when needed. Switches in coupling between different levels of the visual system may be one such process. Finally, it is important to note that, because our results are based an interaction between the spatial and categorical relevance of regions, they cannot distinguish the possibility that spatial attention gates global responses driven by categorical attention from the reverse possibility that categorical attention narrows or filters the focus of spatial attention.

This finding is also consistent with our prior study that found effects of categorical attention on background connectivity but did not consider spatial attention (Al-Aidroos et al., 2012). In that study, face and scene stimuli were overlaid in the same spatial location and fixated centrally. Bilateral V4 showed stronger background connectivity with FFA when faces were attended and with PPA when scenes were attended. Because both the right and left V4 code for foveal stimuli, both regions were spatially relevant for the task. Thus, V4–FFA when faces were attended and V4–PPA when scenes were attended correspond to space+/category+ in the current study, and V4–PPA when faces were attended and V4–FFA when scenes were attended correspond to space+/category– in the current study. The fact that we found stronger connectivity for space+/category+ versus space+/category− in the current study thus replicates these prior findings. The current study also extends these findings by testing for effects of categorical attention on background connectivity with regions that do not code for the attended location (space−/category+ and space−/category−). That such effects were eliminated in unattended locations provides new insight into how attention modulates connectivity in the visual system and into how different varieties of attention are integrated in the brain.

In this experiment, we operationalized coupling between regions as correlations in time courses between two regions. However, correlated activity does not necessarily imply direct coupling or communication and instead could reflect other extrinsic or stimulus-driven sources of correlations. We attempted to minimize some of these other sources by regressing out signals related to stimulus timing, head motion, and nuisance regions before calculating the correlations. Past work has found that this procedure gives rise to correlated activity—or “background connectivity”—that is not driven by variance in BOLD signal across blocks or runs, another source of stimulus-driven correlations (Al-Aidroos et al., 2012). In that work, as well as in the present experiment, differences in intrinsic, subject-specific signals, rather than stimulus timing, underpin the observed modulation in background connectivity by attention. Such intrinsic signals still may not signify direct coupling between two regions but instead reflect coordinated activity moderated by a third region. However, this does not necessarily undercut our findings—rather, it raises interesting questions about the broader network of control regions modulating correlations within visual cortex. Indeed, models of cognitive control emphasize the role of control regions in establishing sensorimotor pathways (Miller & Cohen, 2001). We explore these questions with an exploratory analysis that identified regions whose BOLD signal tracked correlations between two visual regions (see below for further discussion).

One potential drawback to our design is that, although participants were instructed to attend a specific category and spatial location, participants could potentially complete the task by using only category or spatial attention. An ideal design would use composite images of faces and houses in both locations to necessitate both forms of attention. However, through behavioral piloting, we found that such a task was too difficult. Furthermore, any idiosyncratic strategies that participants might use in our current design would only work against finding interactions between categorical attention and spatial attention. In other words, we found that background connectivity was modulated by both spatial and categorical attention despite the fact that only one form of attention as necessary to perform the task.

Interestingly, we also found a broad decrease in background connectivity for dorsal visual areas during attention runs. There are a number of differences between the rest runs and attention runs that could explain this effect, including differences in stimulus configuration (central fixation point vs. fixation point and images in the upper visual field), differences in task (fixating the point vs. fixating the point and detecting stimulus repetitions in the upper visual field), and differences in attention (no attention vs. attention to the upper visual field). However, one intriguing interpretation of this finding is that dorsal connectivity was suppressed because the lower visual field, for which these areas code, was task irrelevant during the attention runs. In support of this account, suppression of dorsal connectivity with FFA/PPA was not observed in our prior study in which the stimuli were foveated (and thus attended in both upper and lower visual fields). In fact, and opposite from the current study, dorsal connectivity was enhanced relative to rest (Al-Aidroos et al., 2012; Figure 2C). Because this enhancement was observed in the same comparison (attention vs. rest) as the suppression in the current study, this difference of dorsal connectivity might be most parsimoniously attributed to the fact that dorsal areas coded for attended locations in the prior study but not the current study. Future studies that directly manipulate attention to upper versus lower visual fields in the same participants would be needed to definitively support this conclusion.

Although we primarily focused on how coupling within visual cortex supports combined spatial and nonspatial attention, we also conducted an exploratory analysis of control mechanisms that may drive such coupling. Prior studies have observed that frontoparietal areas couple with visual areas when they are relevant for current attentional goals (Griffis et al., 2015; Baldauf & Desimone, 2014; Chadick & Gazzaley, 2011). For example, we previously found frontoparietal regions that correlated with FFA when faces were attended and with PPA when scenes were attended (Al-Aidroos et al., 2012). Here we adopted a different approach, seeking to identify frontoparietal areas that track variance in the relationship (i.e., connectivity) between visual areas rather than in the activity of individual areas. This may provide additional leverage for understanding how top–down control signals modulate network activity within visual cortex and especially how different components of attention are integrated. For example, the right IPS/SPL and bilateral IFG were active during moments of high connectivity, but only between areas that were spatially and categorically relevant (e.g., left V4 and FFA, respectively, when attending to faces on the right). In contrast, bilateral IPS/SPL and the right TPJ also tracked connectivity in the visual system but only required spatial attention (e.g., left V4 and PPA in the example above). This dovetails with the identification of feature-specific attention signals in frontoparietal regions typically associated with spatial attention (Liu & Hou, 2013; Liu, Hospadaruk, Zhu, & Gardner, 2011), suggesting that modulation of signal in higher-level regions is related to different attentional goals. These findings lead to the prediction for future studies that the frontoparietal regions in particular may play an important role in integrating over spatial and feature-based attention.

Beyond attentional modulation of background connectivity between V4 and FFA/PPA, we also observed stimulus-evoked responses in these regions. Specifically, spatial attention but not categorical attention modulated responses in the right and left V4, whereas both spatial and categorical attention modulated responses in FFA and PPA. In examining connectivity, we took care to control for these evoked responses by removing them from the data. The fact that attentional modulation of background connectivity persisted into fixation periods without stimuli and did not synchronize across participants (despite identical stimulus timing) further suggests that our connectivity results do not depend on evoked responses. Nevertheless, these two mechanisms must interact to support top–down attention. One possibility is that good communication between regions may strengthen evoked responses downstream (Fries, 2005). Relatedly, connectivity may filter which information gets transmitted and lead to more selective processing in broadly tuned regions (Córdova, Tompary, & Turk-Browne, 2016). Here, connectivity with spatially selective V4 may aid FFA and PPA in processing category information from specific locations. Regardless, by focusing on covariance in neural activity typically ignored in analyses of evoked responses, our findings illustrate the distributed and integrative nature of attention.

Acknowledgments

This work was supported by NSF GRFP (A. T.), NSERC PDF and DG (N. A.), and NIH R01 EY021755 (N. T. B.).

REFERENCES

  1. Al-Aidroos N, Said CP, & Turk-Browne NB (2012). Top–down attention switches coupling between low-level and high-level areas of human visual cortex. Proceedings of the National Academy of Sciences, U.S.A, 109, 14675–14680. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Aly M, & Turk-Browne NB (2016a). Attention promotes episodic encoding by stabilizing hippocampal representations. Proceedings of the National Academy of Sciences, 113, E420–E429. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Aly M, & Turk-Browne NB (2016b). Attention stabilizes representations in the human hippocampus. Cerebral Cortex, 26, 783–796. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Andersen SK, Fuchs S, & Müller MM (2009). Effects of feature-selective and spatial attention at different stages of visual processing. Journal of Cognitive Neuroscience 23 238–246. [DOI] [PubMed] [Google Scholar]
  5. Andersen SK, Hillyard SA, & Müller MM (2013). Global facilitation of attended features is obligatory and restricts divided attention. Journal of Neuroscience 33 18200–18207. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Arcaro MJ, McMains SA, Singer BD, & Kastner S (2009). Retinotopic organization of human ventral visual cortex. Journal of Neuroscience, 29, 10638–10652. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Baldauf D, & Desimone R (2014). Neural mechanisms of object-based attention. Science, 344, 424–427. [DOI] [PubMed] [Google Scholar]
  8. Bartsch MV, Boehler CN, Stoppel CM, Merkel C, Heinze H-J, Schoenfeld MA, et al. (2015). Determinants of global color-based selection in human visual cortex. Cerebral Cortex, 25, 2828–2841. [DOI] [PubMed] [Google Scholar]
  9. Baylis GC, & Driver J (1992). Visual parsing and response competition: The effect of grouping factors. Perception & Psychophysics, 51, 145–162. [DOI] [PubMed] [Google Scholar]
  10. Bondarenko R, Boehler CN, Stoppel CM, Heinze H-J, Schoenfeld MA, & Hopf J-M (2012). Separable mechanisms underlying global feature-based attention. Journal of Neuroscience 32 15284–15295. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bosman CA, Schoffelen J-M, Brunet N, Oostenveld R, Bastos AM, Womelsdorf T, et al. (2012). Attentional stimulus selection through selective synchronization between monkey visual areas. Neuron, 75, 875–888. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Chadick JZ, & Gazzaley A (2011). Differential coupling of visual cortex with default or frontal-parietal network based on goals. Nature Neuroscience, 14, 830–832. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Chelazzi L, Duncan J, Miller EK, & Desimone R (1998). Responses of neurons in inferior temporal cortex during memory-guided visual search. Journal of Neurophysiology, 80, 2918–2940. [DOI] [PubMed] [Google Scholar]
  14. Chelazzi L, Miller EK, Duncan J, & Desimone R (1993). A neural basis for visual search in inferior temporal cortex. Nature, 363, 345–347. [DOI] [PubMed] [Google Scholar]
  15. Cohen EH, & Tong F (2015). Neural mechanisms of object-based attention. Cerebral Cortex, 25, 1080–1092. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Cohen MR, & Maunsell JHR (2011). Using neuronal populations to study the mechanisms underlying spatial and feature attention. Neuron 70 1192–1204. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Connor CE, Gallant JL, Preddie DC, & Van Essen DC (1996). Responses in area V4 depend on the spatial relationship between stimulus and attention. Journal of Neurophysiology, 75, 1306–1308. [DOI] [PubMed] [Google Scholar]
  18. Corbetta M, Miezin FM, Dobmeyer S, Shulman GL, & Petersen SE (1990). Attentional modulation of neural processing of shape, color, and velocity in humans. Science 248 1556. [DOI] [PubMed] [Google Scholar]
  19. Corbetta M, & Shulman GL (2002). Control of goal-directed and stimulus-driven attention in the brain. Nature Reviews Neuroscience, 3, 201–215. [DOI] [PubMed] [Google Scholar]
  20. Córdova NI, Tompary A, & Turk-Browne NB (2016). Attentional modulation of background connectivity between ventral visual cortex and the medial temporal lobe. Neurobiology of Learning and Memory, 134, 115–122. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Dale AM, Fischl B, & Sereno MI (1999). Cortical surface- based analysis: I. Segmentation and surface reconstruction. Neuroimage 9 179–194. [DOI] [PubMed] [Google Scholar]
  22. De Renzi E, Perani D, Carlesimo GA, Silveri MC, & Fazio F (1994). Prosopagnosia can be associated with damage confined to the right hemisphere—An MRI and PET study and a review of the literature. Neuropsychologia, 32, 893–902. [DOI] [PubMed] [Google Scholar]
  23. Duncan J (1984). Selective attention and the organization of visual information. Journal of Experimental Psychology: General, 113, 501–517. [DOI] [PubMed] [Google Scholar]
  24. Duncan J, & Nimmo-Smith I (1996). Objects and attributes in divided attention: Surface and boundary systems. Perception & Psychophysics, 58, 1076–1084. [DOI] [PubMed] [Google Scholar]
  25. Duncan K, Tompary A, & Davachi L (2014). Associative encoding and retrieval are predicted by functional connectivity in distinct hippocampal area CA1 pathways. Journal of Neuroscience, 34, 11188–11198. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Fischl B, Sereno MI, & Dale AM (1999). Cortical surface- based analysis: II. Inflation, flattening, and a surface-based coordinate system. Neuroimage 9 195–207. [DOI] [PubMed] [Google Scholar]
  27. Fox MD, & Raichle ME (2007). Spontaneous fluctuations in brain activity observed with functional magnetic resonance imaging. Nature Reviews Neuroscience, 8, 700–711. [DOI] [PubMed] [Google Scholar]
  28. Fries P (2005). A mechanism for cognitive dynamics: Neuronal communication through neuronal coherence. Trends in Cognitive Sciences 9 474–480. [DOI] [PubMed] [Google Scholar]
  29. Furey ML, Tanskanen T, Beauchamp MS, Avikainen S, Uutela K, Hari R, et al. (2006). Dissociation of face-selective cortical responses by attention. Proceedings of the National Academy of Sciences, U.S.A, 103, 1065–1070. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Giesbrecht B, Woldorff MG, Song AW, & Mangun GR (2003). Neural mechanisms of top-down control during spatial and feature attention. Neuroimage, 19, 496–512. [DOI] [PubMed] [Google Scholar]
  31. Goldsmith M, & Yeari M (2003). Modulation of object-based attention by spatial focus under endogenous and exogenous orienting. Journal of Experimental Psychology: Human Perception and Performance, 29, 897–918. [DOI] [PubMed] [Google Scholar]
  32. Gregoriou GG, Gotts SJ, Zhou H, & Desimone R (2009). High-frequency, long-range coupling between prefrontal and visual cortex during attention. Science, 324, 1207–1210. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Grier JB (1971). Nonparametric indexes for sensitivity and bias: Computing formulas. Psychological Bulletin, 75, 424. [DOI] [PubMed] [Google Scholar]
  34. Griffis JC, Elkhetali AS, Burge WK, Chen RH, & Visscher KM (2015). Retinotopic patterns of background connectivity between V1 and fronto-parietal cortex are modulated by task demands. Frontiers in Human Neuroscience 9 338. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Hasson U, Nir Y, Levy I, Fuhrmann G, & Malach R (2004). Intersubject synchronization of cortical activity during natural vision. Science, 303, 1634–1640. [DOI] [PubMed] [Google Scholar]
  36. Hayden BY, & Gallant JL (2009). Combined effects of spatial and feature-based attention on responses of V4 neurons. Vision Research, 49, 1182–1187. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Hemond CC, Kanwisher NG, & Op de Beeck HP (2007). A preference for contralateral stimuli in human object- and face-selective cortex. PLoS ONE, 2, e574. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Kimura D (1966). Dual functional asymmetry of the brain in visual perception. Neuropsychologia, 4, 275–285. [Google Scholar]
  39. Kravitz DJ, & Behrmann M (2011). Space-, object-, and feature-based attention interact to organize visual scenes. Attention, Perception & Psychophysics, 73, 2434–2447. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Leonard CJ, Balestreri A, & Luck SJ (2015). Interactions between space-based and feature-based attention. Journal of Experimental Psychology: Human Perception and Performance, 41, 11–16. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Liu T, Hospadaruk L, Zhu DC, & Gardner JL (2011). Feature-specific attentional priority signals in human cortex. Journal of Neuroscience, 31, 4484–4495. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Liu T, & Hou Y (2011). Global feature-based attention to orientation. Journal of Vision 11 8. [DOI] [PubMed] [Google Scholar]
  43. Liu T, & Hou Y (2013). A hierarchy of attentional priority signals in human frontoparietal cortex. Journal of Neuroscience, 33, 16606–16616. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Liu T, & Mance I (2011). Constant spread of feature-based attention across the visual field. Vision Research 51 26–33. [DOI] [PubMed] [Google Scholar]
  45. McAdams CJ, & Maunsell JHR (1999). Effects of attention on orientation-tuning functions of single neurons in macaque cortical area V4. Journal of Neuroscience, 19, 431–441. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. McMains SA, Fehd HM, Emmanouil T-A, & Kastner S (2007). Mechanisms of feature- and space-based attention: Response modulation and baseline increases. Journal of Neurophysiology, 98, 2110–2121. [DOI] [PubMed] [Google Scholar]
  47. Miller EK, & Cohen JD (2001). An integrative theory of prefrontal cortex function. Annual Review of Neuroscience 24, 167–202. [DOI] [PubMed] [Google Scholar]
  48. Moran J, & Desimone R (1985). Selective attention gates visual processing in the extrastriate cortex. Science, 229, 782–784. [DOI] [PubMed] [Google Scholar]
  49. Motter BC (1993). Focal attention produces spatially selective processing in visual cortical areas V1, V2, and V4 in the presence of competing stimuli. Journal of Neurophysiology, 70, 909–919. [DOI] [PubMed] [Google Scholar]
  50. Müller NG, & Kleinschmidt A (2003). Dynamic interaction of object- and space-based attention in retinotopic visual areas. Journal of Neuroscience, 23, 9812–9816. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Norman-Haignere SV, McCarthy G, Chun MM, & Turk-Browne NB (2012). Category-selective background connectivity in ventral visual cortex. Cerebral Cortex 22 391–402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Noudoost B, Chang MH, Steinmetz NA, & Moore T (2010). Top-down control of visual attention. Current Opinion in Neurobiology, 20, 183–190. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Nyström M, & Holmqvist K (2010). An adaptive algorithm for fixation, saccade, and glissade detection in eyetracking data. Behavior Research Methods, 42, 188–204. [DOI] [PubMed] [Google Scholar]
  54. O’Craven KM, Downing PE, & Kanwisher N (1999). fMRI evidence for objects as the units of attentional selection. Nature, 401, 584–587. [DOI] [PubMed] [Google Scholar]
  55. Painter DR, Dux PE, Travis SL, & Mattingley JB (2014). Neural responses to target features outside a search array are enhanced during conjunction but not unique-feature search. Journal of Neuroscience 34 3390–3401. [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Peelen MV, Fei-Fei L, & Kastner S (2009). Neural mechanisms of rapid natural scene categorization in human visual cortex. Nature, 460, 94–97. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Saalmann YB, Pigarev IN, & Vidyasagar TR (2007). Neural mechanisms of visual attention: How top-down feedback highlights relevant locations. Science, 316, 1612–1615. [DOI] [PubMed] [Google Scholar]
  58. Saenz M, Buracas GT, & Boynton GM (2002). Global effects of feature-based attention in human visual cortex. Nature Neuroscience, 5, 631–632. [DOI] [PubMed] [Google Scholar]
  59. Saenz M, Buracas GT, & Boynton GM (2003). Global feature-based attention for motion and color. Vision Research 43 629–637. [DOI] [PubMed] [Google Scholar]
  60. Schwarzlose RF, Swisher JD, Dang S, & Kanwisher N (2008). The distribution of category and location information across object-selective regions in human visual cortex. Proceedings of the National Academy of Sciences U.S.A. 105, 4447–4452. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Serences JT, & Boynton GM (2007). Feature-based attentional modulations in the absence of direct visual stimulation. Neuron, 55, 301–312. [DOI] [PubMed] [Google Scholar]
  62. Serences JT, Schwarzbach J, Courtney SM, Golay X, & Yantis S (2004). Control of object-based attention in human cortex. Cerebral Cortex 14 1346–1357. [DOI] [PubMed] [Google Scholar]
  63. Serences JT, & Yantis S (2007). Spatially selective representations of voluntary and stimulus-driven attentional priority in human occipital, parietal, and frontal cortex. Cerebral Cortex, 17, 284–293. [DOI] [PubMed] [Google Scholar]
  64. Shomstein S, & Yantis S (2004). Control of attention shifts between vision and audition in human cortex. Journal of Neuroscience, 24, 10702–10706. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Shomstein S, & Yantis S (2006). Parietal cortex mediates voluntary control of spatial and nonspatial auditory attention. Journal of Neuroscience, 26, 435–439. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Smith SM, Jenkinson M, Woolrich MW, Beckmann CF, Behrens TE, Johansen-Berg H, et al. (2004). Advances in functional and structural MR image analysis and implementation as FSL. Neuroimage, 23, S208–S219. [DOI] [PubMed] [Google Scholar]
  67. Soto D, & Blanco MJ (2004). Spatial attention and object- based attention: A comparison within a single task. Vision Research, 44, 69–81. [DOI] [PubMed] [Google Scholar]
  68. Summerfield C, Greene M, Wager T, Egner T, Hirsch J, & Mangels J (2006). Neocortical connectivity during episodic memory formation. PLoS Biology 4 e128. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Tompary A, Duncan K, & Davachi L (2015). Consolidation of associative and item memory is related to post-encoding functional connectivity between the ventral tegmental area and different medial temporal lobe subregions during an unrelated task. Journal of Neuroscience, 35, 7326–7331. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Treue S, & Trujillo JCM (1999). Feature-based attention influences motion processing gain in macaque visual cortex. Nature, 399, 575–579. [DOI] [PubMed] [Google Scholar]
  71. Turk-Browne NB (2013). Functional interactions as big data in the human brain. Science 342 580–584. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Vecera SP, & Farah MJ (1994). Does visual attention select objects or locations? Journal of Experimental PSychology: General, 123, 146–160. [DOI] [PubMed] [Google Scholar]
  73. Verosky SC, & Turk-Browne NB (2012). Representations of facial identity in the left hemisphere require right hemisphere processing. Journal of Cognitive Neuroscience, 24, 1006–1017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Wandell BA, Dumoulin SO, & Brewer AA (2007). Visual field maps in human cortex. Neuron 56 366–383. [DOI] [PubMed] [Google Scholar]
  75. Wang Y, Cohen JD, Li K, & Turk-Browne NB (2015). Full correlation matrix analysis (FCMA): An unbiased method for task-related functional connectivity. Journal of Neuroscience Methods 251 108–119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. White AL, & Carrasco M (2011). Feature-based attention involuntarily and simultaneously improves visual performance across locations. Journal of Vision, 11, 15. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. White AL, Rolfs M, & Carrasco M (2015). Stimulus competition mediates the joint effects of spatial and feature- based attention. Journal of Vision 15, 7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Worsley KJ, Chen J-I, Lerch J, & Evans AC (2005). Comparing functional connectivity via thresholding correlations and singular value decomposition. Philosophical Transactions of the Royal Society of London Series B Biological Sciences, 360, 913–920. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Yovel G, Tambini A, & Brandman T (2008). The asymmetry of the fusiform face area is a stable individual characteristic that underlies the left-visual-field superiority for faces. Neuropsychologia, 46, 3061–3068. [DOI] [PubMed] [Google Scholar]
  80. Zhang W, & Luck SJ (2009). Feature-based attention modulates feedforward visual processing. Nature Neuroscience, 12, 24–25. [DOI] [PubMed] [Google Scholar]
  81. Zhou H, & Desimone R (2011). Feature-based attention in the frontal eye field and area V4 during visual search. Neuron, 70, 1205–1217. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES