Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2019 Oct 1;199:245–260. doi: 10.1016/j.neuroimage.2019.05.068

Population receptive field estimates for motion-defined stimuli

Anna E Hughes 1,∗,1, John A Greenwood 1, Nonie J Finlayson 1,2, D Samuel Schwarzkopf 1,3
PMCID: PMC6693563  PMID: 31158480

Abstract

The processing of motion changes throughout the visual hierarchy, from spatially restricted ‘local motion’ in early visual cortex to more complex large-field ‘global motion’ at later stages. Here we used functional magnetic resonance imaging (fMRI) to examine spatially selective responses in these areas related to the processing of random-dot stimuli defined by differences in motion. We used population receptive field (pRF) analyses to map retinotopic cortex using bar stimuli comprising coherently moving dots. In the first experiment, we used three separate background conditions: no background dots (dot-defined bar-only), dots moving coherently in the opposite direction to the bar (kinetic boundary) and dots moving incoherently in random directions (global motion). Clear retinotopic maps were obtained for the bar-only and kinetic-boundary conditions across visual areas V1–V3 and in higher dorsal areas. For the global-motion condition, retinotopic maps were much weaker in early areas and became clear only in higher areas, consistent with the emergence of global-motion processing throughout the visual hierarchy. However, in a second experiment we demonstrate that this pattern is not specific to motion-defined stimuli, with very similar results for a transparent-motion stimulus and a bar defined by a static low-level property (dot size) that should have driven responses particularly in V1. We further exclude explanations based on stimulus visibility by demonstrating that the observed differences in pRF properties do not follow the ability of observers to localise or attend to these bar elements. Rather, our findings indicate that dorsal extrastriate retinotopic maps may primarily be determined by the visibility of the neural responses to the bar relative to the background response (i.e. neural signal-to-noise ratios) and suggests that claims about stimulus selectivity from pRF experiments must be interpreted with caution.

Keywords: Vision, Motion, Population receptive field analysis

1. Introduction

Motion perception is one of the fundamental dimensions of vision (Nakayama, 1985; Nishida, 2011), and it is now known that many areas of the brain are involved in motion processing (Dupont et al., 1994; Pitzalis et al., 2010; Sunaert et al., 1999; Tootell et al., 1997). Converging psychophysical, electrophysiological, and imaging evidence suggests that motion is processed in an hierarchical manner, with signals first being processed locally (within restricted spatial windows) in areas such as V1 and then combined at higher levels in the visual cortical hierarchy to generate global motion percepts over larger regions of the visual field (Adelson and Movshon, 1982; Braddick et al., 2001; Van Essen and Gallant, 1994; Williams and Sekuler, 1984).

Many psychophysical studies have used tasks involving the detection or discrimination of coherent motion to study the distinction between local- and global-motion processing (Britten et al., 1993, 1992; Newsome and Paré, 1988; Scase et al., 1996; Watamaniuk, 1993). Due to the aperture problem, direction-selective neurons in lower visual areas such as V1 are thought to process only local motion – the 1D motion orthogonal to the orientation of the edge that is passing through the receptive field (Adelson and Movshon, 1982; Marr and Ullman, 1981; Movshon, 1986; Wallach, 1935). To be able to process global motion, these 1D signals must be integrated over a relatively wide visual field area/region, a process thought to occur in higher visual areas, such as V5/MT ​+ ​to generate a global motion direction (Heeger et al., 1996; Simoncelli and Heeger, 1998). fMRI evidence has provided support for a distinction in the neural locations of local and global processing, showing that V1 was activated more by incoherent local noise than coherent global motion, perhaps because the noise stimulus led to the activation of neurons with a wider range of motion selectivities. The reverse pattern was seen in V5 and V3A, which both responded more to coherent motion compared to the noise stimulus (Braddick et al., 2001).

Other brain regions may be specialized for the detection of more complex motion patterns. One example is that of kinetic boundaries, where an edge is defined by differences in coherent motion direction on either side of the edge. A specific brain region known as the Kinetic Occipital area (KO), which is thought to include areas such as V3, V3A and V3B (Larsson and Heeger, 2006), may be specialized for detecting these boundaries (Van Oostende et al., 1997). However, other research suggests that KO may not be completely specialized for motion-boundary processing, with KO thought to also respond preferentially to stimuli such as form cues (Zeki et al., 2003). It has further been shown that other visual areas such as LO1, LO2 and V7 respond preferentially to motion boundaries, suggesting that motion-boundary processing may be more widely distributed across the visual cortex (Larsson et al., 2010).

Despite this debate over the specificity of the area responding to kinetic contours, it is clear that kinetic boundaries are relatively complex stimuli that are not processed at lower levels in the visual hierarchy: they produce little fMRI response in lower visual areas such as V1 and V2 (Van Oostende et al., 1997; Zeki et al., 2003), consistent with electrophysiological evidence that the majority of neurons in these areas do not respond to kinetic contours (Leventhal et al., 1998; Marcar et al., 2000).

The visual hierarchy also has multiple representations of the retina, laid out in topological maps that are commonly called retinotopic maps (Felleman and Van Essen, 1991; Sereno et al., 1995). While it was once thought that retinotopy was a property of lower level visual areas, it is now known that areas such as MT+ and MST also contain retinotopic maps (known as TO1 and TO2 respectively; (Amano et al., 2009), as do the frontal eye fields (Kastner et al., 2007) and even the default mode network (Knapen et al., 2018). Indeed, despite the large receptive field sizes within primate MT/V5, it is possible to track object position at the population level (Chen et al., 2015). This suggests that retinotopy is a general organizing principle within the cortex. It is therefore of interest to know whether retinotopic map properties vary according to the visual area under question, and particularly whether these properties are affected by the different functional selectivities of different regions of the brain.

One technique that has been used to analyse retinotopic maps via fMRI is population receptive field (pRF) analysis, providing an estimation of both the visual field position preferred by each voxel and the range of visual field locations where a stimulus evokes a response (Dumoulin and Wandell, 2008). pRFs can therefore be thought of as a statistical summary of the neuronal properties within a sampled region. Recent work has shown systematic differences in pRF sizes across different brain regions and eccentricities, with size increasing along the visual processing hierarchy and with increasing eccentricity (Alvarez et al., 2015; Amano et al., 2009; Dumoulin and Wandell, 2008; Haas et al., 2014; Harvey and Dumoulin, 2011; Schwarzkopf et al., 2014).

Across these different brain areas, several factors can affect pRF size and position. Retinotopy in early visual areas is primarily thought to be stimulus-driven, but there is evidence that higher level maps can also be attentionally-driven (Tootell et al., 1998). One study (Saygin and Sereno, 2008) used point light biological walkers moving in a ‘wedge’ stimulus to dissociate attention and stimulus effects in retinotopic mapping, and showed that V1 did not respond clearly when the distinction between stimulus and background was driven only attentionally, but it did respond when there was a visual difference between stimulus and background without attention directed to the stimulus. The opposite pattern was seen in frontal and parietal areas. The spatial tuning of pRFs is also affected by attentional load at fixation, with pRF size increasing and pRF location becoming more eccentric under high perceptual load (Haas et al., 2014). Similar results have also been reported in face-selective brain regions (Kay et al., 2015). This position modulation seems to occur across the entire visual field, not just at the attended location (Klein et al., 2014), and recent work suggests that these position shifts are the key mechanism by which attention enhances discriminability and representational quality of stimuli (Vo et al., 2017). It therefore seems that attention can affect cortical spatial tuning properties, in turn altering the visibility of stimulus differences.

Stimulus properties may also alter the measured properties of pRFs. In particular, the fMRI response of a voxel may be driven by different groups of cells depending upon the properties of the stimulus. While most studies have used simple luminance-defined stimuli (such as checkerboards) to generate pRF maps, more complex stimuli may be especially suited for generating maps in higher visual areas (Yildirim et al., 2018). A recent study that used a pRF mapping stimulus designed specifically to isolate orientation contrast showed reductions in measured pRF size in higher visual areas such as LO compared to the measurements made with standard luminance-based stimuli (Yildirim et al., 2018). pRF sizes have also been shown to vary based on the alignment and curvature of contours within mapping stimuli (Dumoulin et al., 2014). In the motion domain, there is evidence that MT ​+ ​may be more susceptible to stimulus configuration than earlier visual areas (Alvarez et al., 2015) but to date there has been no systematic investigation of the effects of different motion stimuli on pRF measurement.

Given the above variations due to stimulus properties and attentional state, in this study we asked whether stimuli thought to preferentially drive distinct stages of the motion-processing hierarchy can similarly alter the estimation of pRF parameters. In our initial experiment, we tested this with a moving bar stimulus similar to those commonly used in pRF mapping studies, defined by dots moving coherently. We then used different backgrounds in an attempt to differentially drive responses in different brain regions. In the bar-only condition, the bar was presented alone against a grey background. We predicted that this stimulus would generate a strong visual signal and enable the generation of pRF maps at all levels of the motion-processing hierarchy, much like a typical pRF mapping stimulus. In the ‘kinetic’ condition the bar was defined by kinetic boundaries. We predicted that if these stimuli are preferentially processed in KO/V3B, we might expect smaller pRF sizes and a higher proportion of voxels responding in this area. Finally, in the ‘global’ condition, the background consisted of incoherently moving dots. We therefore predicted that for the ‘global’ condition, higher visual areas that process stimuli in a more global manner should be able to distinguish between the bar and background and thus generate good pRF maps. In contrast, for these latter two conditions, we predicted that V1 would not be able to distinguish the bar from the background, leading to a reduced response.

In a second experiment, we asked whether any differential responses seen in the first experiment were a consequence of differences in the selectivity of these motion-selective regions, or whether they could be explained by other factors, such as the visibility of the stimulus. We compared a bar-only stimulus to two conditions with lower visibility bars: one motion-defined stimulus, where the bar was defined by transparent motion (against a non-transparent background), and a non-motion defined stimulus, where the bar was instead defined by differences in dot size. If pRF properties varied due to differential motion processing, we predicted that a different pattern of responses should arise for these two stimuli: for example, the ‘transparent’ stimulus should show lower responsivity than the ‘size defined’ stimulus in V1, but higher responsivity in higher visual areas selective for global motion. In a similar vein, the subtle dot size difference in the size-defined stimulus should maximize the signal in V1 and the early visual cortex compared to higher regions. However, if the differential responses were simply due to the visibility of the bar stimulus, we predicted similar responses for the two conditions across different visual areas. We also conducted a third behavioural experiment as a control, presenting each of the above mapping stimuli and requiring observers to localise the bar element. This allowed us to further assess whether variations in the properties of pRFs could be predicted by the visibility of the bar stimuli, or whether these variations can be attributed to the visibility of the neural signals underlying the BOLD response.

2. Experiment 1

Here we examined retinotopic maps and pRF properties across the visual hierarchy using three distinct retinotopic mapping stimuli: the ‘bar-only’ stimulus, similar to standard retinotopic mapping stimuli; the ‘global’ bar stimulus with coherent motion against a background of noise; and the ‘kinetic’ bar stimulus with coherent motion against a background of oppositely-moving dots.

2.1. Materials and methods

2.1.1. Participants

Five participants (one male) took part in Experiment 1, including two of the authors (AH and JG) and three experienced participants naïve to the aims of the study. Participants were aged between 24 and 36 years (mean = 29.6 years) and one participant was left handed. All participants were experienced in an fMRI context and had normal or corrected-to-normal visual acuity. Written consent was acquired from all participants to ensure that they understood the potential risks associated with fMRI. The experiments were approved by the UCL Research Ethics Committee.

2.1.2. Stimuli

Fig. 1 shows a schematic of the experimental set up, and GIF versions of the different experimental stimuli are available as supplementary material.

Fig. 1.

Fig. 1

Schematic diagrams of the experimental set up in Experiment 1. (a) to (j) show one experimental run for the ‘bar-only’ condition, in the trial order presentation used for all stimuli in Experiment 1. Arrows indicate the direction of movement of the bar. Note in (b) that the fixation dot has changed colour, as part of the attentional task used in the experiment. (k) Shows a schematic of the ‘kinetic’ condition. Within the bar (the area within the white lines) the dots moved in one direction (orthogonal to the bar movement); outside this area, they moved in the opposite direction. (l) Shows a schematic of the ‘global’ condition; the movement inside the bar is in one direction, but outside the bar, the dots move in random directions.

Stimuli were created using MATLAB (MathWorks, Natick, MA) and the Psychophysics toolbox (Brainard, 1997; Pelli, 1997). Stimuli were projected onto a screen (resolution 1920 × 1080 pixels, size 36.8 × 20.2 cm) at the rear of the scanner bore, with the screen image reflected off a mirror attached above the head coil. The viewing distance was 67 cm, meaning the screen subtended 30.7 × 17.1 degrees of visual angle on the retina. The refresh rate was 60 Hz.

The experimental stimulus consisted of a field of 2000 dots (diameter = 0.09°) against a uniform median grey field. Half of the dots were randomly selected to be white, and the other half were black. The initial positions of the dots were randomly determined within a rectangular aperture 16 × 16° in size. A mask was then applied to the image, such that only the dots within a circle with 8° radius from the centre of the screen could be seen. A smaller circular mask (diameter = 0.8°) was also applied at the fixation point (diameter = 0.17°) to hide dots in this area and thus aid participants in maintaining their fixation. Further masks were also applied depending upon condition (see below). To further aid fixation, a low contrast “radar screen” pattern was shown behind stimuli (12 radial lines spaced 30 polar degrees apart, extending from just outside the fixation dot to the edge of the screen, (15.5°), along with 11 concentric rings centred on fixation increasing in radius in equal steps of 3°).

A ‘bar’ was defined as a strip within the circular stimulus region that was 2.4° wide and up to 16° in length (the total width of the hidden stimulus rectangle). This area was shifted over the course of each trial in 25 discrete steps of 0.7° and 1 s each, beginning at one edge of the stimulus aperture and ending on the opposite side. The bar could be rotated to start in any one of the four cardinal directions or the four oblique directions, giving eight different trial types. In one run, all eight different trial types were presented, along with two null trials where no bar was present. The null trials were always presented as the 5th and 10th trials. The order of the other directions was fixed for all participants and went anticlockwise from the first direction, which was where the bar was horizontal, starting at the top of the screen and moving downwards.

There were three different stimulus conditions that could be presented in a given run to the participants. In each condition, the dots within the bar always moved in the same way; what differed between conditions was the movement of the background dots outside the bar area.

In the ‘bar-only’ condition, only the dots within the bar were visible, and therefore the bar appeared to be moving across a grey background. In the ‘kinetic’ condition, the dots outside the bar (within the circular stimulus region) were always moving in the opposite direction to the dots within the bar, creating a ‘shearing’ effect which made the bar visible. In the ‘global’ condition, the dots outside the bar moved in random directions, allowing the bar to be detected as a 100% coherent global-motion stimulus against a background of noise in adjacent areas of the stimulus. While this condition therefore also contained kinetic boundaries, they were far less clear than the opposing directions used in the ‘kinetic’ stimulus. When a dot moved into the bar area (either through its normal progression or through a shift of the bar region), it started to move in the same coherent direction as all other bar dots. Similarly, when a dot left the bar area, it began to move randomly again. In all conditions, the null trials (with no bar present) had the same background motion as during the bar trials; a blank screen for the ‘bar-only’ condition, coherent motion in the ‘kinetic’ condition and random motion in the ‘global’ condition.

In all conditions, dots within the bar all moved in the same direction and moved along the length of the bar (so if the bar was moving from the top to the bottom of the screen, the dots moved from left to right or vice versa). All the dots (in both bar and background) changed direction by 180° every 0.5 s, to prevent adaptation to one motion direction. Dots moved at 0.8°/second in all conditions. If any dots moved outside the aperture during the experiment, they were moved back one aperture width in the appropriate direction.

Each trial took 25 s, meaning that a run took 4 min and 10 s (plus a short period at the beginning of the run that was used to ensure that the fMRI signal had reached equilibrium). Each participant completed 4 runs for each condition, giving a total of 12 runs in the entire experiment. The order of the different conditions varied for different participants to control for order effects.

2.1.3. Fixation task

Participants were instructed to focus on a blue fixation dot (diameter = 0.17°) at all times and to press a button on an MRI-compatible button box when they saw it change colour (to a red-purple). The probability of the blue dot changing colour was 0.01 every 200 ms and the colour change periods lasted 200 ms each. The results of this attentional task were unrecorded, and simply served to keep the participant fixated and alert throughout the experiment. An eye tracker (Eyelink 1000, sampling at the screen refresh rate of 60 Hz) was used to monitor eye movements and ensure that participants were fixating correctly. We determined gaze stability using the methods outlined in (Haas and Schwarzkopf, 2018); briefly, this involves calculating the median absolute deviation of the sampled gaze positions along both the horizontal and vertical dimensions for each run, and using these measures to compare the stability of gaze across conditions. Any run where fewer than 10 valid samples were taken was removed from further analysis. One participant was not eye tracked during either experiment, and therefore eye tracking data reflects the average of four participants in both Experiments 1 and 2. Analysis of the difference in eye position between conditions (both in the x and y directions) used general linear mixed models (using condition as a fixed factor, and subject and repeat number as random factors) followed by posthoc pairwise comparisons, with packages lme4 (Bates et al., 2014) and emmeans (Russell, 2018) in R (version 3.5.0).

2.1.4. Data acquisition

Scans were acquired using a Siemens Avanto 1.5T MRI scanner with a 32-channel Siemens head coil located at the Birkbeck-UCL Centre for Neuroimaging. We used a modified version of the head coil without the eye visor to allow an unrestricted view of the screen, leaving 30 effective channels. We used functional T2*-weighted multiband 2D echo planar imaging with a multiband sequence (Breuer et al., 2005) and the following properties: voxel size = 2.3 mm isotropic, field of view = 96 × 96, 36 slices, repetition time (TR) = 1s, echo time (TE) = 55 ms, flip angle = 75°, and acceleration factor = 4. We collected 260–262 vol (depending on stimulus condition) per run, and 4 runs were collected per condition for each participant. We also acquired a T1-weighted anatomical magnetisation-prepared rapid acquisition with gradient echo (MPRAGE) scan for each participant (TR = 2730 ms, TE = 3.57 ms) with a resolution of 1 mm isotropic voxels.

2.1.5. Analysis

The method used for analysing pRFs has been described previously (Alvarez et al., 2015; Dumoulin and Wandell, 2008; Moutsiana et al., 2016; Schwarzkopf et al., 2014; van Dijk et al., 2016). In brief, the SamSrf MATLAB toolbox (available at https://doi.org/10.6084/m9.figshare.1344765) models the pRF of each voxel as a 2D Gaussian in the visual field, incorporating a canonical haemodynamic response function based on the average of 26 participants in a previous study (Haas et al., 2014). For each voxel the model finds the best-fitting visual field location, spread (standard deviation), and overall response amplitude of the pRF function.

Preprocessing of the fMRI data was carried out using SPM12 (Wellcome Centre for Human Neuroimaging, London, http://www.fil.ion.ucl.ac.uk/spm/software/spm12/). The first 10–12 vol (depending on stimulus condition) were removed to allow the signal to reach equilibrium, leaving 250 vol to be used in analysis for all participants and conditions. We then carried out intensity bias correction, realignment, unwarping and coregistration of the functional data to the structural scan, all using the default parameters built into the SPM software. FreeSurfer (https://surfer.nmr.mgh.harvard.edu/fswiki) was used to generate a 3D reconstruction of the grey-white matter surface (Dale et al., 1999; Fischl et al., 1999), and the functional data was then projected to the cortical surface by finding the median position for each vertex of the surface reconstruction between the pial and grey-white matter boundary. Linear detrending was applied to the time series from each vertex in each run, and runs of the same stimulus condition were z-standardised and averaged together.

Population receptive field analysis was carried out on the occipital lobe data in a two-stage procedure. With a binary aperture describing the position of the bar element within each stimulus for each scanning volume (which was identical in each condition), we calculated its overlap with a profile of a pRF to predict the fMRI time series in the experiment. We first carried out a coarse grid search fit on data smoothed with a large kernel on the spherical surface (full width half maximum = 5), allowing calculation of the three pRF parameters that gave the maximal Pearson's correlation between the predicted and observed time series for the full set of search grid parameters and vertices. These parameters were then used to seed an optimisation algorithm (Lagarias et al., 1998; Nelder and Mead, 1965) in a slow fine fit procedure on a vertex by vertex basis using unsmoothed data, allowing refinement of the parameter estimates and the calculation of an estimate of response strength.

Visual areas V1, V2 (dorsal and ventral), V3 (dorsal and ventral), V3A, V3B and MT+ (defined as TO1 and TO2; see supplementary materials for an example) were delineated based on reversals in the polar angle map from the ‘bar-only’ stimulus condition (Sereno et al., 1995). For participants who only completed Experiment 2, the ‘transparent bar-only’ condition was used instead. These regions can be seen in Fig. 2 and Fig. 7. Throughout our main analyses, we used an R2 (goodness-of-fit) threshold of 0.05, which corresponds in our dataset to a p-value of 0.000367 (due to the number of observations per dataset and the number of free parameters in the pRF model). As our experimental conditions often show relatively weak and sparse responses, we chose this relatively liberal threshold to enable us to analyse the residual responses. However, we also carried out all analyses using a more conservative R2 value of 0.1, and results from these analyses can be seen in the supplementary material.

Fig. 2.

Fig. 2

Sphere projection of polar angle data for the left hemispheres of all participants in Experiment 1. The colour of each vertex indicates the polar angle for the corresponding pRF centre (as indicated by the colour wheel). Each person's data forms a column (subject 1 is on the far left, and subject 5 is on the far right), and stimulus condition forms a row. Manual delineations of visual areas V1, V2, V3, V3A, V3B and hMT+ (TO1/2) are shown. (a) Polar angle estimates for the ‘bar-only’ stimulus condition. (b) Polar angle estimates for the ‘kinetic’ stimulus condition. (c) Polar angle estimates for the ‘global’ stimulus condition.

Fig. 7.

Fig. 7

Sphere projection of polar angle data for the left hemispheres of all participants in Experiment 2. The colour of each vertex indicates the polar angle for the corresponding pRF centre (as indicated by the colour wheel). Each person's data forms a column (subject 1 is on the far left and subject 5 is on the far right), and stimulus condition forms a row. Manual delineations of visual areas V1, V2, V3, V3A, V3B, and hMT+ (TO1/2) are shown (if the subject had taken part in Experiment 1, the delineations from this experiment were used). (a) Polar angle estimates for the ‘bar-only’ stimulus condition. (b) Polar angle estimates for the ‘transparent’ stimulus condition. (c) Polar angle estimates for the ‘size-defined’ stimulus condition. Red crosses indicate where participants self-reported low awareness of the stimulus when questioned after the experiment.

2.2. Results and discussion

2.2.1. Relationships between maps

Fig. 2 shows the left hemisphere polar angle maps for each experimental condition and each participant in Experiment 1 (R2 threshold = 0.05). Visual inspection of these images suggests that the ‘bar-only’ condition (Fig. 2A) produces the clearest polar angle maps across participants, while the ‘kinetic’ condition (Fig. 2B) tends to show a similar, if weaker pattern. The ‘global’ stimulus (Fig. 2C) shows a weaker response still, with lower visual areas (e.g. V1) showing very little response. Although we had a relatively small number of participants, these general trends are highly consistent.

In order to quantify the consistency of these maps, we determined the correlation between the values for polar angle obtained for each vertex with each stimulus. As shown in Fig. 3, the correlation between these polar angle estimates was overall clear, particularly for visual areas V2, V3, V3A and V3B where the average correlation was significantly different from zero. The correlation between conditions was less clear in V1 and MT+.

Fig. 3.

Fig. 3

Correlation matrices comparing pRF polar angles between stimulus conditions in Experiment 1. The colour of each cell indicates the strength and sign of each vertex-wide correlation in polar angle. Circular correlations were calculated for each participant, then Z transformed and averaged across participants (as in (Haas and Schwarzkopf, 2018)). The symbols indicate whether the average correlation in individual cells is significantly different from zero (uncorrected). One star = p < 0.05. Two stars = p < 0.001.

We next determined the proportion of vertices within each of these visual areas responding retinotopically in the three experimental conditions (goodness of fit of the pRF model R2 > 0.05, Fig. 4A). The bar stimulus produced the biggest response in areas V1–V3, which then dropped off for the higher visual areas (V3A, V3B and MT+). In contrast, responses to the kinetic stimulus increased across areas V1–V3, levelling off at V3A-V3B and then dropping in MT+. Responses to the global stimulus were even lower in the early visual areas, but again increased, reaching a peak at V3A and V3B. There were therefore large differences in stimulus responsivity in V1–V3, but these differences were much reduced in the higher visual areas. There was a significant interaction between condition and visual area in the final model of the data (interaction: χ2 = 336.06, p < 0.001; main effect of visual area: χ2 = 141.63, p < 0.001; main effect of condition: χ2 = 377.49, p < 0.001). This interaction also helps to explain the lack of significant correlation in polar angle values between conditions in area V1 (Fig. 3) – although the bar stimulus produced clear polar angle estimates in a large number of V1 voxels, this was far less so for the other two stimuli.

Fig. 4.

Fig. 4

(a) Proportion of vertices responding, (b) goodness-of-fit and (c) pRF sizes for each condition and visual area in Experiment 1. The bars show the mean values across all subjects, and the points are individual data for each subject. In (a), this is the mean proportion of vertices responding for each subject, whereas for (b) and (c) these are the median goodness-of-fit values and pRF sizes respectively.

Comparing the goodness of fit across conditions (Fig. 4B) showed a similar pattern, with initially large differences in R2 between conditions in the early visual areas that again decreased in higher regions. This interaction between condition and visual area was again significant (interaction: χ2 = 60.629, p < 0.001; main effect of visual area: χ2 = 62.747, p < 0.001; main effect of condition: χ2 = 242.912, p < 0.001). The area with the highest average R2 value also differed for each condition; the peak was in V3 for the bar stimulus, V3A for the kinetic stimulus and V3B for the global stimulus.

Finally, we analysed pRF size across conditions and visual areas (Fig. 4C). Mean pRF sizes were smallest in the early visual areas, V1–V3, increasing in higher regions. pRFs were also relatively similar in early visual areas for all three conditions, but clear differences emerged in V3A, V3B and MT+. Here, pRFs were largest for the bar condition, smaller for the kinetic condition, and smaller again for the global condition (though in all cases larger than the equivalent condition in earlier regions). Again, there was a significant interaction between condition and visual area (interaction: χ2 = 47.879, p < 0.001; main effect of visual area: χ2 = 174.505, p < 0.001; main effect of condition: χ2 = 24.434, p < 0.001). Similar results were seen when pRF size was examined as a function of eccentricity in the different brain areas and experimental conditions (see Supplementary Fig. 2). pRF size was found to increase as a function of eccentricity in all brain areas, with the lowest rate of increase for the global condition.

2.2.2. Control analyses

As there were clear differences in the proportion of responsive voxels in the three conditions, it is possible that the differences in pRF size between conditions were due to this reduction in the voxels included in each analysis, rather than a specific change in pRF size within each voxel. To examine this possibility, we analysed the data for the ‘bar-only’ condition using just the voxels that survived thresholding for the kinetic and global conditions (see Fig. 5). The pattern of results is similar to Fig. 4C – pRF sizes were again comparable in early visual areas for the three conditions, with clear reductions in pRF size for the kinetic and global conditions in areas V3A, V3B and MT+. In this case however, the reduction can be attributed to the differential selection of voxels responding to the same stimulus. In other words, the observed pRF size differences in Fig. 4C are likely due to changes in the voxels that respond to these stimuli rather than active changes in pRF size across the different conditions. Using a linear model to compare the kinetic and global vertex conditions from the control analysis with the kinetic and global data in the original analysis showed no significant difference between the two data sets (χ2 = 0.758, p = 0.384).

Fig. 5.

Fig. 5

Plot to show pRF sizes for each visual area in Experiment 1 for the bar-only condition, using the responsive vertices for all three conditions. The bars show the mean values across all subjects, and the points are individual data for each subject (median pRF sizes). Any data points with a value of zero (obtained if the vertices for the condition did not overlap with any bar activation) were removed before plotting, leading to unequal numbers of data points in each condition.

It is also possible that differences between conditions could be attributed to differences in fixation stability between conditions. In a second control analysis we therefore examined the median absolute deviations of eye position, which on average were highly consistent and relatively low for both horizontal and vertical eye movements, averaging less than 0.5 degrees of visual angle for every condition (see Fig. 6). General linear mixed models followed by posthoc pairwise comparisons suggested that there were no significant differences in eye position between conditions, either for the X or the Y direction (for X, bar-kinetic: t30.14 = 0.356, p = 0.933, bar-global: t30.26 = 0.457, p = 0.892, kinetic-global: t30.35 = 0.101, p = 0.994. For Y, bar-kinetic: t30.35 = - 0.614, p = 0.814, bar-global: t30.14 = −0.503, p = 0.871, kinetic-global: t30.43 = 0.111, p = 0.993). Differences in fixation are therefore unlikely to have produced the above differences in pRF properties.

Fig. 6.

Fig. 6

Plot showing the mean of the median absolute deviation of eye position across runs, conditions and observers (n = 4) for both the X (horizontal) and Y (vertical) dimensions of Experiment 1, in degrees of visual angle. Error bars are ± 1SD of the mean. Lines connect data points from individual subjects.

2.2.3. Experiment 1 summary

Altogether, the three stimulus types (bar-only, kinetic, and global) produced clear differences in responsivity, goodness-of-fit, and pRF size across the visual hierarchy. Our control analyses reveal that these differences cannot be attributed to differences in gaze stability between the conditions and that the observed differences in pRF size are unlikely to reflect stimulus-driven changes in pRF size within each voxel. Rather, it seems likely that the observed differences in pRF size can be attributed to responses from different sub-populations of voxels in the three different conditions. We next turn our attention to the source of the differences in responsivity that appear to be driving these differences in pRF size. In particular, it is possible that differences in the visibility or salience of the bar between these conditions could drive the differences in responsivity. We explore this possibility in Experiment 2.

3. Experiment 2

As outlined in the introduction, estimates of pRF size and visual field location have been found to vary according to both the properties of the mapping stimulus and the attentional state of the observer. Although it is tempting to attribute the differences observed in Experiment 1 to differences in the effectiveness of these stimuli at driving the selectivity of various stages of the motion-processing hierarchy, the visibility of the bar element in each of our stimuli also varied between the three conditions. Most participants in Experiment 1 informally noted that the bar was less clear in the global condition than in the kinetic or bar-only conditions. It is possible then that the differences observed were driven by the visibility or salience of the bar, rather than any difference in the underlying selectivity of the motion detectors in each brain region.

In Experiment 2 we sought to test this by comparing the bar-only stimulus with two new stimuli. Firstly, we devised a second form of stimulus defined by global motion, a ‘transparent’ stimulus where the bar contained two sets of dots moving in opposite directions to give the appearance of two sheets moving transparently across each other (Snowden and Verstraten, 1999). The global percept of these two directions requires an integration across space that is similar to that for a single direction of global motion amongst noise (Edwards and Greenwood, 2005; Edwards and Nishida, 1999). Transparent motion is not perceived when dots with opposing directions are ‘locally balanced’ within small regions of the visual field (Qian et al., 1994), perhaps due to an intermediate process of ‘local-motion pooling’ prior to the global-motion stage (Edwards et al., 2012; Vidnyánszky et al., 2002). Although the responses of V1 neurons cannot distinguish between transparent and non-transparent stimuli, both fMRI (Muckli et al., 2002) and electrophysiological studies (Qian and Andersen, 1994; Snowden et al., 1991) show a differential response within MT/V5. Transparent-motion stimuli thus offer another stimulus with which we can assess whether pRF parameters differ when stimuli preferentially drive higher levels of the motion-processing hierarchy. We therefore constructed bar stimuli with two opposing directions of transparent motion within the bar, presented against a background of locally-balanced dots that do not appear transparent. If the differences in pRF parameters found in Experiment 1 are due to differences in motion selectivity, we would expect the transparent-motion bar stimulus to produce weaker responses in lower visual areas and a reduction in pRF size in higher visual areas, as with the ‘global’ condition.

Our second comparison stimulus was intended to examine the role of stimulus visibility in these effects. Transparent motion in particular has been found to be less visible in peripheral vision than in the fovea (De Bruyn, 1997), which would likely create issues for the visibility of our transparent bar stimuli as they traverse the visual field, just as it may have been an issue in Experiment 1. We therefore compared these transparent stimuli with a stimulus bar that was not defined by differences in motion, but rather by a subtle difference in stimulus dot size. This bar stimulus would likely differentially drive the responses of early visual areas, given their potential role in the perception of object size (Moutsiana et al., 2016; Murray et al., 2006; Pooresmaeili et al., 2013; Sperandio et al., 2012), but should not differentially drive the responses of higher motion-selective regions as effectively as the motion-defined bars used previously. The size difference in these stimuli does however lead to a substantial reduction in the visibility of the bar stimulus relative to the bar-only condition, particularly in peripheral vision. In particular, we selected a size difference that produced a similar level of subjective visibility to the ‘transparent’ condition (examined during pilot testing). Were this size-defined condition to produce similar responses to the motion-defined condition, this would suggest that visibility or salience is a more likely explanation for the observed differences than the stimulus selectivity of the underlying neural populations.

3.1. Materials and methods

3.1.1. Participants

Five participants (two male) took part in Experiment 2, including all four authors and one non-author participant from the first experiment (age range 28–39 years, mean age: 32.6 years). One participant was left handed. All had normal or corrected-to-normal visual acuity and provided written consent, as in Experiment 1.

3.1.2. Stimuli

The second experiment was set up with the same apparatus and general stimulus properties as the first. Here there were three conditions related to the mapping stimuli: ‘bar-only’, ‘transparent’ and ‘size-defined’. The ‘bar-only’ condition in this experiment was identical to the ‘bar-only’ condition in Experiment 1, except that the dots were paired such that the two dots in each pair travelled in the same direction. However, different pairs travelled in opposite directions to give the impression of transparent motion with two ‘sheets’ of dots travelling over each other in opposite directions. As in the ‘bar-only’ condition of Experiment 1, the null trials for this condition contained only the uniform grey background. The speed of the dots was reduced in this condition to 0.34°/second to increase the impression of transparency.

In the ‘transparent’ condition, the bar was identical to the bar in the ‘bar-only’ condition. In this case however, the background also contained paired dots, with each dot in the pair moving in opposite directions, leading to the perception of non-transparent motion (as with ‘locally paired’ dot stimuli used previously; (Qian et al., 1994). To keep dots within these local regions and avoid them unpairing over time, the motion of the dots periodically reversed. The timing of these reversals was randomised across dot pairs, so that reversals did not occur for all of the dots at the same time, a feature that enhances the percept of transparent sheets moving across one another (Kanai et al., 2004) when dot trajectories are limited (though not when locally paired, as in the background). The consistent pairing of dots between bar and background (differing only in their consistent vs opposing directions) meant that bar and background did not differ in dot density, and that the only feature to distinguish the bar was the percept of transparency against a background of non-transparent flicker.

The ‘size-defined’ condition contained a bar that was defined by a difference in dot size rather than by motion type; the dots in the bar were 0.10° in diameter against a background of dots with 0.09° diameter. All the dots in this condition moved in random directions, though with the same frequency of direction reversal as in the other two conditions (i.e. dots oscillated back-and-forth along a randomly selected axis). Null trials for this condition were the same as for the ‘global’ condition in Experiment 1.

To minimize anticipation effects in this experiment, the starting orientation and direction of the sequence (anticlockwise or clockwise shifts) was randomised for each condition and participant but kept constant across the four runs. Note that this meant that the movement direction of the bar always changed in a sequential fashion. All other presentation and analysis procedures were as in Experiment 1.

3.2. Results and discussion

3.2.1. Relationships between maps

As in Experiment 1, the bar-only stimulus produced clear and consistent polar maps across participants (Fig. 7A). However, the transparent and size-defined stimuli produced much weaker and more variable maps (Fig. 7B and C), particularly in lower visual areas (e.g. V1) where responses were considerably reduced. Interestingly, participants who self-reported that they were frequently unable to detect the transparency or size-defined bar stimuli (shown by red crosses in Fig. 7) also had virtually no discernible map structure for these conditions.

Correlations between these polar angle estimates in the 3 conditions were in general slightly weaker than in Experiment 1 (Fig. 8), but were still generally positive, with significant correlations overall only in areas V2 and V3A. As in Experiment 1, we next determined the proportion of vertices responding retinotopically in the different experimental conditions (see Fig. 9A). Here, the bar-only condition produced similar response levels to the equivalent condition in Experiment 1, with these levels again decreasing in higher regions. In comparison, the transparent and size-defined conditions showed greatly reduced responsivity in early visual areas, similar to the global condition in Experiment 1. Responsivity in these conditions increased in later visual areas, though not quite to the level of the bar-only stimulus (unlike the global condition). There was a significant interaction between condition and visual area in the final model (interaction: χ2 = 52.330, p < 0.001; main effect of visual area: χ2 = 24.933, p < 0.001; main effect of condition: χ2 = 223.071, p < 0.001). This interaction is again likely to explain the lack of significant correlations in polar angle values between conditions, here given the clear drop in responsivity for the latter two stimulus types across the visual hierarchy.

Fig. 8.

Fig. 8

Correlation matrices comparing pRF polar angles between stimulus conditions in Experiment 2. The colour of each cell indicates the strength and sign of each vertex-wide correlation in polar angle. Circular correlations were calculated for each participant, then Z transformed and averaged across participants (as in (Haas and Schwarzkopf, 2018)). The symbols indicate whether the average correlation in individual cells is significantly different from zero (uncorrected). One star = p < 0.05. Two stars = p < 0.001. Note: one participant did not have enough data for valid correlations in the size-defined condition in V3B and MT, and so their data was not used for the average calculation).

Fig. 9.

Fig. 9

(a) Proportion of vertices responding, (b) goodness-of-fit and (c) pRF sizes for each condition (bar-only, transparent, and size-defined) and visual area in Experiment 2. The bars show the mean values across all subjects, and the points are individual data for each subject. In (a), this is the mean proportion of vertices responding for each subject, whereas for (b) and (c) these are the median goodness-of-fit values and pRF sizes respectively.

Goodness of fit also matched the pattern observed in Experiment 1, with the bar-only stimulus having much better goodness of fit compared to the transparent and size-defined stimuli. Goodness of fit was also worst in V1 and MT for all conditions, with R2 values increasing for more intermediate visual areas (Fig. 9B). Modelling of the goodness of fit showed that there was no significant interaction between condition and visual area, (χ2 = 16.505, p = 0.086). However, there were significant main effects of visual area (χ2 = 30.060, p < 0.001) and condition (χ2 = 289.304, p < 0.001).

Finally, we considered differences in pRF size for the different conditions and visual areas (Fig. 9C). Again, as in Experiment 1 there was a highly significant interaction between condition and visual area (interaction: χ2 = 64.806, p < 0.001; main effect of visual area: χ2 = 72.031, p < 0.001; main effect of condition: χ2 = 21.572, p < 0.001). With the bar-only stimulus, pRF size increases dramatically across visual areas, with values comparable to those of Experiment 1. Although pRF values for the transparent and size-defined stimuli are comparable in early areas, the rate of increase is much lower than in the bar-only condition, resulting in considerably smaller pRFs in the highest areas. In MT these values were broadly comparable to the global condition of Experiment 1 for the size-defined stimulus, though generally much smaller for the transparent stimulus. pRF size also increased as a function of eccentricity in all different brain areas and experimental conditions, particularly for the bar-only stimulus (see Supplementary Fig. 3).

3.2.2. Control analyses

As in Experiment 1, a control analysis was run to examine whether the above differences in pRF size between conditions were due to this reduction in the voxels included in each analysis. We again analysed data for the bar-only condition using just the voxels that survived thresholding for the transparent and size-defined conditions. This again produced a clear reduction in pRF size for these two conditions in areas V3A, V3B and MT, suggesting that the observed differences in pRF size may be predominantly explained by differences in the responsivity of voxels (Fig. 10). This is supported by statistical analysis suggesting that there is no significant difference between the transparent and size-defined conditions in the control analysis and in the original experimental analysis (χ2 = 0.111, p = 0.739).

Fig. 10.

Fig. 10

Plot to show pRF sizes for each condition and visual area in Experiment 2 for the transparent bar-only condition, using the responsive vertices for all three conditions. The bars show the mean values across all subjects, and the points are individual data for each subject (median pRF sizes). Any data points with a value of zero (obtained if the vertices for the condition do not overlap with any bar activation) were removed before plotting, leading to unequal numbers of data points in each condition.

As in Experiment 1, the results of Experiment 2 also do not seem to depend on eye movements, as the median absolute deviations of eye position were on average highly consistent and relatively low for both horizontal and vertical eye movements, with averages of less than 0.5 degrees of visual angle for both conditions (see Fig. 11). General linear mixed models followed by posthoc pairwise comparisons suggested that there were no significant differences in eye position between the bar-only and size-defined conditions, either for the X or the Y direction (for X, bar-transparent: t35.20 = −0.055, p = 0.998, bar-size: t35.04 = −1.439, p = 0.332, transparent-size: t35.20 = −1.354, p = 0.375. For Y, bar-transparent: t35.23 = 0.667, p = 0.784, bar-size: t35.05 = −1.513, p = 0.297, transparent-size: t35.23 = −2.149, p = 0.094). As before, this suggests that participants were highly compliant with the fixation instructions and that differences in fixation stability cannot account for our results.

Fig. 11.

Fig. 11

Plot showing the mean of the median absolute deviation of eye position across runs, conditions and observers (n = 4) for both the X (horizontal) and Y (vertical) dimensions of Experiment 2, in degrees of visual angle. Error bars are ± 1SD of the mean. Lines connect data points from individual subjects.

4. Experiment 3

In Experiment 2, we found very similar responses across visual areas for the transparent and size-defined stimuli, both of which differed from the bar-only stimulus. This indicates that the visibility or salience of the bar element could be important in determining responsivity and associated pRF properties. However, there are several possible aspects of visibility that could be involved. One is perceptual visibility, whereby the bar element may be less detectable (or more difficult to attend to) in some stimuli than others, particularly in peripheral vision. Another is neural ‘visibility’, where pRF analyses might be affected by the reduced signal-to-noise ratio in cases where there is a background signal (e.g. from the noise dots in the ‘global’ stimulus) as well as the responses to the bar element. To distinguish between these possibilities, we therefore carried out a control perceptual experiment, quantitatively assessing the visibility of the stimulus types used in Experiments 1 and 2 at different eccentricities, using psychophysical techniques. If perceptual visibility can explain the results found, we would predict that the ability of participants to detect the stimuli should follow the same pattern as the differences in responsiveness seen in Experiments 1 and 2. Specifically, bar stimuli should be highly visible, with a slight reduction in visibility for the kinetic stimuli, and further reductions for the global, transparent, and size-defined stimuli.

4.1. Materials and methods

4.1.1. Participants

Eight participants (three male) took part in the perceptual experiment, including two authors (who participated in both fMRI experiments), one non-author participant who took part in Experiment 1, and five naive participants (age range 21–37 years, mean age: 25.9 years). All had normal or corrected-to-normal visual acuity and provided written consent, as in previous experiments.

4.1.2. Stimuli and procedures

The experiment was carried out in a laboratory setting (i.e. not in the scanner), with stimuli presented on a Display++ monitor (Cambridge Research Systems, UK) with a size of 71 ​× ​39.5 ​cm, a resolution of 2560 ​× ​1440 pixels, and a refresh rate of 120 ​Hz. Viewing distance was 1 ​m, with head movements restricted through the use of a chin and forehead rest. Responses were made via keypad. Stimulus parameters were set to subtend the same visual angle as in the fMRI set up.

On each trial, participants were presented with a single bar location of the stimuli used for the main fMRI experiments, meaning that the bar was presented in one of 22 different locations on the screen for 1 s (the middle three locations were excluded from this experiment to avoid ambiguity regarding their location relative to fixation). The participant's task was to determine whether the bar was above, below, left or right of the fixation point, and press the corresponding button on a keypad to indicate their choice.

Each participant judged the bar location in five different types of stimuli: the three conditions used in Experiment 1 (bar only, kinetic and global) and the transparent and size-defined conditions from Experiment 2. All parameters used for these conditions were unchanged from the fMRI experiments. One block of the experiment used just one of the five stimulus types, with five repeats of all the possible positions (22) presented in two orientations (horizontal and vertical), giving 220 trials per block. The order of trials within a block was randomized. Each participant completed 10 blocks in total (two blocks of each stimulus type) and the order of these blocks was pseudorandomized (each stimulus type appeared once in the first five trials and once in the second five trials).

Testing was completed in two sessions of approximately 40–50 min each. At the beginning of the first session, naïve subjects were given practice trials for all stimuli. This involved showing the stimuli initially for 5s, and then running a practice block for each stimulus type where the bars were presented only at the innermost bar locations. This was done to ensure that naïve participants had a similar amount of experience with the stimuli as those who had participated in the fMRI experiments.

4.2. Results and discussion

For each condition, responses were collated by eccentricity (collapsed across both visual hemifield and bar orientation) and scored as the proportion correct at each point. Fig. 12 shows these responses along with the best-fitting linear function for each condition, a comparison that shows clear differences in visibility for the different stimulus conditions used in our experiments. Both the bar and kinetic stimuli were highly visible at even the furthest eccentricities. In contrast, the global stimulus was highly visible at central eccentricities, with a slight drop in visibility at 10–12 deg, although the difference in the slope of the linear fit was not significantly different from the bar condition (t = −1.473, p = 0.142). The size-defined stimulus was slightly harder to localise, even in central vision, with a steeper decline in visibility with eccentricity that was significantly different from that of the bar stimulus (t = −8.632, p < 0.001). The transparent-motion stimulus was even less visible at central eccentricities (though performance was still well above chance), with a similarly steep decline in visibility with increasing eccentricity that was again significantly different from the bar stimulus (t = −7.857, p < 0.001). Detectability of the size-defined and transparent-motion stimuli was more variable across participants (as indicated by the larger error bars), similar to the variation in visibility reported in Experiment 2.

Fig. 12.

Fig. 12

Plot showing the mean percent correct (n = 8 participants) for each of the five stimulus types in the perceptual experiment, across different eccentricities. All individual data points are represented, along with a linear fit to the full data set for each condition.

Overall however, the above pattern of visibility does not closely match the variations observed in the properties of the pRFs measured with these stimuli. In particular, the global condition was similarly visible to the kinetic and bar conditions but showed a very different pattern of pRF data in Experiment 1. Conversely, the size-defined and transparent conditions showed clear reductions in visibility relative to both the global condition and to each other, and yet the pRF properties measured with these stimuli were both similar to those of the global condition. We conclude that stimulus visibility is unlikely to account for these differences in pRF properties.

5. Discussion

In this study, we show that retinotopic mapping stimuli defined by motion produce clear and highly consistent differences in the properties of population receptive fields (pRFs) measured across the visual hierarchy, including responsivity, goodness of fit, and pRF size. As predicted, we show that a bar mapping stimulus defined by moving dots (against a blank background) produces strong pRF maps in early visual areas, with responsivity decreasing and pRF sizes increasing in areas higher in the motion processing hierarchy. More complex motion stimuli, such as bars defined by kinetic and global motion (against backgrounds of opposing motion or noise, respectively), produce a much lower degree of responsivity in early visual areas, with a reduction in pRF sizes for higher visual areas, and reductions in goodness of fit across the hierarchy as a whole. Control analyses further suggest that the reduction in pRF size can be attributed to the reduction in the voxels included for each stimulus, rather than changes in pRF size within voxels. Although it is tempting to attribute this to differences in the potential for these visual areas to distinguish these higher-order stimuli, a second experiment showed highly similar patterns of responsivity, goodness of fit, and pRF size across visual areas for a stimulus defined by transparent motion (against a non-transparent background) and a size-defined stimulus (with no differences in motion) that were reduced in visibility. This suggests that the observed differences in pRF properties are not specific to mapping stimuli defined by differences in stimulus motion.

As outlined in the introduction, evidence from a variety of experimental approaches suggests that motion is processed hierarchically in the visual system, with local motion processed in early visual areas, such as V1, and global motion processed in higher areas, such as MT+ (Adelson and Movshon, 1982; Braddick et al., 2001; Movshon, 1986; Van Essen and Gallant, 1994; Williams and Sekuler, 1984). These differences in selectivity predict differences in pRF size and responsivity between local motion defined stimuli (such as our bar stimulus) and global motion defined stimuli (such as our global stimulus) for these different areas. Our results indeed show these differences, sometimes quite strikingly; the lack of response in V1 for the global stimulus is highly consistent across observers. Support for differences derived from other forms of motion selectivity is less clear. In particular, it has been suggested that kinetic boundary stimuli are processed preferentially in visual area V3B (Van Oostende et al., 1997). However, our results show that responses were equally as strong in V3A, with comparable pRF sizes and goodness-of-fit values that were, if anything, better than V3B. Our results therefore suggest that although there were clear differences in the pRF properties measured in early visual areas with these stimuli, amongst higher areas these kinetic boundary stimuli produced widespread changes in pRF properties that are difficult to localize to any one area, supporting the notion that motion boundaries may be processed in a wide number of areas in the visual cortex (Larsson et al., 2010; Larsson and Heeger, 2006).

The fact that we see very similar response patterns for both our ‘transparent’ and ‘size-defined’ conditions in Experiment 2 further suggests that we should be cautious about attributing our results to differential motion processing in distinct visual areas and should instead consider alternative explanations. One possibility is that the results could reflect a decrease in visibility because of the presentation of these bar mapping stimuli in peripheral vision. Our sensitivity to global motion (Raymond, 1994), transparent motion (De Bruyn, 1997), and kinetic boundaries or motion-defined form (Regan and Beverley, 1984) are all known to decline in the periphery. These declines can be corrected for by adjusting the contrast, size, and speed of stimuli (Hess and Aaen-Stockdale, 2008; Regan and Hamstra, 1991), at least to some extent. However, were the lack of eccentricity scaling a problem with our stimuli, we should have seen an increase in average pRF size with our motion-defined stimuli due to the loss of responsivity from neurons with small receptive fields (as seen in comparisons between size-invariant and eccentricity-scaled bar stimuli; (Alvarez et al., 2015). Our psychophysical results in Experiment 3 also show that while there are decreases in visibility in peripheral vision, these do not show a consistent relationship with our pRF data. For instance, while the global, transparent and size-defined conditions all showed reductions in pRF responsivity relative to the bar-only stimulus, the bar element in the global condition was in fact clearly visible at all but the most extreme eccentricities. Conversely, the size-defined bars were much harder to localise, with further reductions for the transparent stimulus, and yet pRF responsivity rates for these two conditions were similar. This suggests that perceptual visibility is not the main factor driving the responses we observe.

Previous research has shown that differences in attention to peripheral stimuli can also influence neural responses. One study found decreased responses in V1 (but stronger responses in parietal and frontal areas) when participants distinguished between stimulus and background entirely attentionally (Saygin and Sereno, 2008). It has also recently been shown that it is possible to map retinotopic responses for bar stimuli defined by illusory contours, occluded parts of a bar, or very low luminance contrast (Haas and Schwarzkopf, 2018), suggesting that the mapping reflected spatial attention rather than specific visual properties of these stimuli. In the current experiment, it is likely that the more complex second-order stimuli used were more attentionally demanding than the ‘bar-only’ stimuli. Along these lines, second-order motion has been found to be more difficult to process at multiple locations compared to first-order motion, suggesting that second-order motion is more attentionally demanding (Lu et al., 2000). Direction discrimination thresholds for second-order motion are also influenced more strongly by attention than thresholds for first-order motion (Allen and Ledgeway, 2003). It may be then that the greater attentional demands required to detect second-order stimuli (like our bars defined by differences in global motion) are more important for determining the responsivity and properties than the stimulus parameters. Again, however, the results of our psychophysical experiment argue against this – the global stimulus in particular was highly visible across the visual field (suggesting that observers had no difficulty attending to these bar elements) and yet pRF responses were similar to those for the transparent stimulus that was much less visible (and which therefore may have presented difficulties for attention). In other words, attention does not seem to offer a complete explanation for our results.

Another possibility is that our motion- and size-defined stimuli may have produced an illusory sense of depth for the bar stimulus, which may then have altered our measured pRF properties. It is known that areas such as V3B and V3A are involved in the processing of depth cues (Tyler et al., 2006), and particularly with integration of depth cues with other signals, such as motion (Ban et al., 2012). However, we think this is unlikely to be a complete explanation of our results, as participants did not report strong depth percepts for any of our stimuli. In addition, it is not clear that this hypothesis would explain the patterns observed in our results; we did not see markedly stronger responses in V3A/V3B, as has been observed in previous fMRI studies of depth perception (Anzai and DeAngelis, 2010; Backus et al., 2001; Tsao et al., 2003). A related suggestion is that the responses in higher visual areas may be a consequence of surface segmentation cues; however, again, research has shown that early visual areas such as V1 are also activated by texture detection and surface segregation processes (Scholte et al., 2008).

Finally, rather than psychophysical visibility, it is likely that the visibility of the neural responses to mapping stimuli (relative to background activity and noise levels) may be an issue in the measurement of population receptive fields. pRF analyses rely on a difference in BOLD response when the stimulus bar and the background are presented in a given location of the visual field (though of course the BOLD response is an indirect measure of differences in neural processing that may reflect vascular responses in some situations; (Logothetis, 2008; O'Herron et al., 2016). Changes in the response to the background stimulus may therefore affect our ability to derive these measures. In particular, although unidirectional global motion typically drives BOLD responses in MT to a greater extent than incoherent noise (Braddick et al., 2001), incoherent noise stimuli still produce an increased response within MT relative to stationary stimuli (McKeefry et al., 1997). The same would be true for our kinetic boundary and transparent stimuli. It is possible therefore that the luminance-defined differences produced by the bar-only stimulus produce a clearer difference between bar and background responses than the motion-defined differences of the other conditions, consistent with the observed reductions in goodness of fit for the pRF parameters derived using these motion-based stimuli. This could also explain why size-defined stimuli produced a similar pattern of results, given that the noise dots used in these stimuli would similarly decrease the difference in BOLD response to the stimulus bar relative to the background. Larger pRF estimates are likely to be particularly vulnerable to this issue, given that these voxels tend to show the worst goodness-of-fit. For instance, in Experiment 1 there was a clear negative correlation between R2 and pRF size in V1, even with the bar-only stimuli (ρ = −0.721, p < 0.001). Voxels with large pRF estimates may thus be the first to drop out with our motion-based stimuli, leading to our observed reductions in pRF size. Were this to be the case, our findings would in fact reflect the selectivity of visual brain regions for motion (given the increased responsivity to the stimulus background), though the reduction in pRFs could not be strictly interpreted as a property of the underlying neural populations. Given that our behavioural data suggests that simple psychophysical visibility is not well matched to our pRF results, we would argue that this explanation provides the most parsimonious explanation of our results.

Previous work has suggested that when the stimulus bar is defined by orientation contrast, reductions in pRF size in higher visual areas (such as LO1 and LO2) may be caused by the stimulus mainly driving neurons sensitive to orientation contrast (Yildirim et al., 2018). Here, while we also find reductions in pRF size for our more complex second-order conditions, the similarity in responses between very different conditions (like the size-defined and transparent stimuli) leads us to argue that this reduction can be more parsimoniously explained by reductions in the signal-to-noise ratio of the neural signal, as discussed above. Of course, this does not mean that second-order stimuli are not useful (for example, they may potentially improve the accuracy of pRF estimates by reducing BOLD displacement; (Olman et al., 2007; Yildirim et al., 2018), and it does not mean that there may not be stimulus-specific signals in pRF mapping; for example, recent studies have shown that varying the orientation or direction of motion of the carrier stimulus within the bar apertures used for mapping can lead to differences in pRF parameters (Dumoulin et al., 2014; Harvey and Dumoulin, 2016). We simply urge future researchers to be cautious when interpreting the functional meaning of changes in pRF properties.

In conclusion, we find evidence for variations in the properties of retinotopic maps for different motion-based stimuli. In particular, we find clear retinotopic maps for stimuli defined by a moving bar of dots against a blank background, but much weaker maps when the bar was defined by coherently moving dots against a background of either incoherent or oppositely-moving dots, or by transparent compared to non-transparent motion. However, the similar maps derived from stimuli defined by size differences suggest that these differences do not reflect a change in the responsivity of neurons in different visual areas to different motion properties. We similarly rule out variations in perceptual visibility or attentional selection of the bars with our behavioural data. Rather, we suggest that it is the visibility of the neural signal for retinotopic mapping stimuli, as defined by the signal-to-noise ratio between bar and background responses, that is the most important driver of pRF properties.

Author contributions

All authors conceived and designed the experiments and performed the experiments; A.H. and D.S.S. analysed the data; D.S.S. contributed analysis tools; A.H. wrote the paper; all authors edited the paper and approved the final version.

Conflicts of interest

The authors declare no conflict of interest. The funding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

Acknowledgments

Supported by ERC Starting Grant WMOSPOTWU to DSS, and Career Development Award MR/K024817/1 from the UK Medical Research Council to JAG. Our thanks to Jingxiu Cheng for his help with data collection in the behavioural experiment.

Footnotes

Appendix A

Supplementary data to this article can be found online at https://doi.org/10.1016/j.neuroimage.2019.05.068.

Appendix A. Supplementary data

The following are the Supplementary data to this article:

Supplementary document

containing an example delineation of TO1 and TO2, eccentricity plots and analyses using a R2 threshold of 0.1.

mmc1.docx (4.4MB, docx)
Multimedia component 2
mmc2.xml (947B, xml)
Video S1

Example bar-only stimulus.

mmc3.gif (5.7MB, gif)
Video S2

Example kinetic stimulus.

mmc4.gif (34.3MB, gif)
Video S3

Example global stimulus.

mmc5.gif (34.3MB, gif)
Video S4

Example transparent bar-only stimulus.

mmc6.gif (34.3MB, gif)
Video S5

Example transparent motion stimulus.

mmc7.gif (26.2MB, gif)
Video S6

Example size-defined stimulus.

mmc8.gif (4.5MB, gif)

References

  1. Adelson E.H., Movshon J.A. Phenomenal coherence of moving visual patterns. Nature. 1982;300:523–525. doi: 10.1038/300523a0. [DOI] [PubMed] [Google Scholar]
  2. Allen H.A., Ledgeway T. Attentional modulation of threshold sensitivity to first-order motion and second-order motion patterns. Vis. Res. 2003;43:2927–2936. doi: 10.1016/j.visres.2003.07.005. [DOI] [PubMed] [Google Scholar]
  3. Alvarez I., Haas D.,A.,B., Clark C.A., Rees G., Schwarzkopf D.S. Comparing different stimulus configurations for population receptive field mapping in human fMRI. Front. Hum. Neurosci. 2015;9 doi: 10.3389/fnhum.2015.00096. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Amano K., Wandell B.A., Dumoulin S.O. Visual field maps, population receptive field sizes, and visual field coverage in the human MT+ complex. J. Neurophysiol. 2009;102:2704–2718. doi: 10.1152/jn.00102.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Anzai A., DeAngelis G.C. Neural computations underlying depth perception. Curr. Opin. Neurobiol. 2010;20:367–375. doi: 10.1016/j.conb.2010.04.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Backus B.T., Fleet D.J., Parker A.J., Heeger D.J. Human cortical activity correlates with stereoscopic depth perception. J. Neurophysiol. 2001;86:2054–2068. doi: 10.1152/jn.2001.86.4.2054. [DOI] [PubMed] [Google Scholar]
  7. Ban H., Preston T.J., Meeson A., Welchman A.E. The integration of motion and disparity cues to depth in dorsal visual cortex. Nat. Neurosci. 2012;15:636–643. doi: 10.1038/nn.3046. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Bates D., Mächler M., Bolker B., Walker S. 2014. Fitting Linear Mixed-Effects Models Using Lme4. ArXiv14065823 Stat. [Google Scholar]
  9. Braddick O.J., O'Brien J.M., Wattam-Bell J., Atkinson J., Hartley T., Turner R. Brain areas sensitive to coherent visual motion. Perception. 2001;30:61–72. doi: 10.1068/p3048. [DOI] [PubMed] [Google Scholar]
  10. Brainard D.H. The psychophysics toolbox. Spatial Vis. 1997;10:433–436. [PubMed] [Google Scholar]
  11. Breuer F.A., Blaimer M., Heidemann R.M., Mueller M.F., Griswold M.A., Jakob P.M. Controlled aliasing in parallel imaging results in higher acceleration (CAIPIRINHA) for multi-slice imaging. Magn. Reson. Med. 2005;53:684–691. doi: 10.1002/mrm.20401. [DOI] [PubMed] [Google Scholar]
  12. Britten K.H., Shadlen M.N., Newsome W.T., Movshon J.A. Responses of neurons in macaque MT to stochastic motion signals. Vis. Neurosci. 1993;10:1157–1169. doi: 10.1017/s0952523800010269. [DOI] [PubMed] [Google Scholar]
  13. Britten K.H., Shadlen M.N., Newsome W.T., Movshon J.A. The analysis of visual motion: a comparison of neuronal and psychophysical performance. J. Neurosci. Off. J. Soc. Neurosci. 1992;12:4745–4765. doi: 10.1523/JNEUROSCI.12-12-04745.1992. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Chen S.C., Morley J.W., Solomon S.G. Spatial precision of population activity in primate area MT. J. Neurophysiol. 2015 doi: 10.1152/jn.00152.2015. jn.00152.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Dale A.M., Fischl B., Sereno M.I. Cortical surface-based analysis. I. Segmentation and surface reconstruction. Neuroimage. 1999;9:179–194. doi: 10.1006/nimg.1998.0395. [DOI] [PubMed] [Google Scholar]
  16. De Bruyn B. Blending transparent motion patterns in peripheral vision. Vis. Res. 1997;37:645–648. doi: 10.1016/S0042-6989(96)00117-4. [DOI] [PubMed] [Google Scholar]
  17. Dumoulin S.O., Hess R.F., May K.A., Harvey B.M., Rokers B., Barendregt M. Contour extracting networks in early extrastriate cortex. J. Vis. 2014;14 doi: 10.1167/14.5.18. 18–18. [DOI] [PubMed] [Google Scholar]
  18. Dumoulin S.O., Wandell B.A. Population receptive field estimates in human visual cortex. Neuroimage. 2008;39:647–660. doi: 10.1016/j.neuroimage.2007.09.034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Dupont P., Orban G.A., De Bruyn B., Verbruggen A., Mortelmans L. Many areas in the human brain respond to visual motion. J. Neurophysiol. 1994;72:1420–1424. doi: 10.1152/jn.1994.72.3.1420. [DOI] [PubMed] [Google Scholar]
  20. Edwards M., Cassanello C.R., Kalia K. Adaptation state of the local-motion-pooling units determines the nature of the motion aftereffect to transparent motion. Vis. Res. 2012;64:23–25. doi: 10.1016/j.visres.2012.05.006. [DOI] [PubMed] [Google Scholar]
  21. Edwards M., Greenwood J.A. The perception of motion transparency: a signal-to-noise limit. Vis. Res. 2005;45:1877–1884. doi: 10.1016/j.visres.2005.01.026. [DOI] [PubMed] [Google Scholar]
  22. Edwards M., Nishida S. Global-motion detection with transparent-motion signals. Vis. Res. 1999;39:2239–2249. doi: 10.1016/s0042-6989(98)00325-3. [DOI] [PubMed] [Google Scholar]
  23. Felleman D.J., Van Essen D.C. Distributed hierarchical processing in the primate cerebral cortex. Cerebr. Cortex. 1991;1:1–47. doi: 10.1093/cercor/1.1.1-a. N. Y. N 1991. [DOI] [PubMed] [Google Scholar]
  24. Fischl B., Sereno M.I., Dale A.M. Cortical surface-based analysis. II: inflation, flattening, and a surface-based coordinate system. Neuroimage. 1999;9:195–207. doi: 10.1006/nimg.1998.0396. [DOI] [PubMed] [Google Scholar]
  25. Haas B. de, Schwarzkopf D.S. Spatially selective responses to Kanizsa and occlusion stimuli in human visual cortex. Sci. Rep. 2018;8:611. doi: 10.1038/s41598-017-19121-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Haas B. de, Schwarzkopf D.S., Anderson E.J., Rees G. Perceptual load affects spatial tuning of neuronal populations in human early visual cortex. Curr. Biol. 2014;24:R66–R67. doi: 10.1016/j.cub.2013.11.061. [DOI] [PMC free article] [PubMed] [Google Scholar] [Retracted]
  27. Harvey B.M., Dumoulin S.O. Visual motion transforms visual space representations similarly throughout the human visual hierarchy. Neuroimage. 2016;127:173–185. doi: 10.1016/j.neuroimage.2015.11.070. [DOI] [PubMed] [Google Scholar]
  28. Harvey B.M., Dumoulin S.O. The relationship between cortical magnification factor and population receptive field size in human visual cortex: constancies in cortical architecture. J. Neurosci. 2011;31:13604–13612. doi: 10.1523/JNEUROSCI.2572-11.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Heeger D.J., Simoncelli E.P., Movshon J.A. Computational models of cortical visual processing. Proc. Natl. Acad. Sci. 1996;93:623–627. doi: 10.1073/pnas.93.2.623. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Hess R.F., Aaen-Stockdale C. Global motion processing: the effect of spatial scale and eccentricity. J. Vis. 2008;8 doi: 10.1167/8.4.11. 11.1-11. [DOI] [PubMed] [Google Scholar]
  31. Kanai R., Paffen C.L.E., Gerbino W., Verstraten F.A.J. Blindness to inconsistent local signals in motion transparency from oscillating dots. Vis. Res. 2004;44:2207–2212. doi: 10.1016/j.visres.2004.04.010. [DOI] [PubMed] [Google Scholar]
  32. Kastner S., DeSimone K., Konen C.S., Szczepanski S.M., Weiner K.S., Schneider K.A. Topographic maps in human frontal cortex revealed in memory-guided saccade and spatial working-memory tasks. J. Neurophysiol. 2007;97:3494–3507. doi: 10.1152/jn.00010.2007. [DOI] [PubMed] [Google Scholar]
  33. Kay K.N., Weiner K.S., Grill-Spector K. Attention reduces spatial uncertainty in human ventral temporal cortex. Curr. Biol. 2015;25:595–600. doi: 10.1016/j.cub.2014.12.050. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Klein B.P., Harvey B.M., Dumoulin S.O. Attraction of position preference by spatial attention throughout human visual cortex. Neuron. 2014;84:227–237. doi: 10.1016/j.neuron.2014.08.047. [DOI] [PubMed] [Google Scholar]
  35. Knapen T., Es D. van, Barendregt M. Mapping the dark side: visual selectivity of default network deactivations. 2018. bioRxiv 292524. [DOI]
  36. Lagarias J., Reeds J., Wright M., Wright P. Convergence properties of the nelder--mead simplex method in low dimensions. SIAM J. Optim. 1998;9:112–147. doi: 10.1137/S1052623496303470. [DOI] [Google Scholar]
  37. Larsson J., Heeger D.J. Two retinotopic visual areas in human lateral occipital cortex. J. Neurosci. Off. J. Soc. Neurosci. 2006;26:13128–13142. doi: 10.1523/JNEUROSCI.1657-06.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Larsson J., Heeger D.J., Landy M.S. Orientation selectivity of motion-boundary responses in human visual cortex. J. Neurophysiol. 2010;104:2940–2950. doi: 10.1152/jn.00400.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Leventhal A.G., Wang Y., Schmolesky M.T., Zhou Y. Neural correlates of boundary perception. Vis. Neurosci. 1998;15:1107–1118. doi: 10.1017/s0952523898156110. [DOI] [PubMed] [Google Scholar]
  40. Logothetis N.K. What we can do and what we cannot do with fMRI. Nature. 2008;453:869–878. doi: 10.1038/nature06976. [DOI] [PubMed] [Google Scholar]
  41. Lu Z.L., Liu C.Q., Dosher B.A. Attention mechanisms for multi-location first- and second-order motion perception. Vis. Res. 2000;40:173–186. doi: 10.1016/s0042-6989(99)00172-8. [DOI] [PubMed] [Google Scholar]
  42. Marcar V.L., Raiguel S.E., Xiao D., Orban G.A. Processing of kinetically defined boundaries in areas V1 and V2 of the macaque monkey. J. Neurophysiol. 2000;84:2786–2798. doi: 10.1152/jn.2000.84.6.2786. [DOI] [PubMed] [Google Scholar]
  43. Marr D., Ullman S. Directional selectivity and its use in early visual processing. Proc. R. Soc. Lond. Ser. B Biol. Sci.Proc. R. Soc. Lond. Ser. B Biol. Sci. 1981;211:151–180. doi: 10.1098/rspb.1981.0001. [DOI] [PubMed] [Google Scholar]
  44. McKeefry D.J., Watson J.D., Frackowiak R.S., Fong K., Zeki S. The activity in human areas V1/V2, V3, and V5 during the perception of coherent and incoherent motion. Neuroimage. 1997;5:1–12. doi: 10.1006/nimg.1996.0246. [DOI] [PubMed] [Google Scholar]
  45. Moutsiana C., de Haas B., Papageorgiou A., van Dijk J.A., Balraj A., Greenwood J.A., Schwarzkopf D.S. Cortical idiosyncrasies predict the perception of object size. Nat. Commun. 2016;7:12110. doi: 10.1038/ncomms12110. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Movshon J.A. The analysis of moving patterns. Pattern Recognit. Mech. 1986:117–151. [Google Scholar]
  47. Muckli L., Singer W., Zanella F.E., Goebel R. Integration of multiple motion vectors over space: an fMRI study of transparent motion perception. Neuroimage. 2002;16:843–856. doi: 10.1006/nimg.2002.1085. [DOI] [PubMed] [Google Scholar]
  48. Murray S.O., Boyaci H., Kersten D. The representation of perceived angular size in human primary visual cortex. Nat. Neurosci. 2006;9:429–434. doi: 10.1038/nn1641. [DOI] [PubMed] [Google Scholar]
  49. Nakayama K. Biological image motion processing: a review. Vis. Res. 1985;25:625–660. doi: 10.1016/0042-6989(85)90171-3. [DOI] [PubMed] [Google Scholar]
  50. Nelder J.A., Mead R. A simplex method for function minimization. Comput. J. 1965;7:308–313. doi: 10.1093/comjnl/7.4.308. [DOI] [Google Scholar]
  51. Newsome W.T., Paré E.B. A selective impairment of motion perception following lesions of the middle temporal visual area (MT) J. Neurosci. Off. J. Soc. Neurosci. 1988;8:2201–2211. doi: 10.1523/JNEUROSCI.08-06-02201.1988. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Nishida S. Advancement of motion psychophysics: review 2001–2010. J. Vis. 2011;11 doi: 10.1167/11.5.11. [DOI] [PubMed] [Google Scholar]
  53. O'Herron P., Chhatbar P.Y., Levy M., Shen Z., Schramm A.E., Lu Z., Kara P. Neural correlates of single-vessel haemodynamic responses in vivo. Nature. 2016;534:378–382. doi: 10.1038/nature17965. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Olman C.A., Inati S., Heeger D.J. The effect of large veins on spatial localization with GE BOLD at 3 T: Displacement, not blurring. Neuroimage. 2007;34:1126–1135. doi: 10.1016/j.neuroimage.2006.08.045. [DOI] [PubMed] [Google Scholar]
  55. Pelli D.G. The VideoToolbox software for visual psychophysics: transforming numbers into movies. Spatial Vis. 1997;10:437–442. [PubMed] [Google Scholar]
  56. Pitzalis S., Sereno M.I., Committeri G., Fattori P., Galati G., Patria F., Galletti C. Human V6: the medial motion area. Cerebr. Cortex. 2010;20:411–424. doi: 10.1093/cercor/bhp112. N. Y. NY. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Pooresmaeili A., Arrighi R., Biagi L., Morrone M.C. Blood oxygen level-dependent activation of the primary visual cortex predicts size adaptation illusion. J. Neurosci. 2013;33:15999–16008. doi: 10.1523/JNEUROSCI.1770-13.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Qian N., Andersen R.A. Transparent motion perception as detection of unbalanced motion signals. II. Physiology. 1994;14:7367–7380. doi: 10.1523/JNEUROSCI.14-12-07367.1994. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Qian N., Andersen R.A., Adelson E.H. Transparent motion perception as detection of unbalanced motion signals. I. Psychophysics. J. Neurosci. 1994;14:7357–7366. doi: 10.1523/JNEUROSCI.14-12-07357.1994. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Raymond J.E. Directional anisotropy of motion sensitivity across the visual field. Vis. Res. 1994;34:1029–1037. doi: 10.1016/0042-6989(94)90007-8. [DOI] [PubMed] [Google Scholar]
  61. Regan D., Beverley K.I. Figure–ground segregation by motion contrast and by luminance contrast. J. Opt. Soc. Am. A. 1984;1:433–442. doi: 10.1364/JOSAA.1.000433. [DOI] [PubMed] [Google Scholar]
  62. Regan D., Hamstra S. Shape discrimination for motion-defined and contrast-defined form: squareness in special. Perception. 1991;20:315–336. doi: 10.1068/p200315. [DOI] [PubMed] [Google Scholar]
  63. Russell L. 2018. Emmeans: Estimated Marginal Means, Aka Least-Squares Means. R Package Version 1.2.1. [Google Scholar]
  64. Saygin A.P., Sereno M.I. Retinotopy and attention in human occipital, temporal, parietal, and frontal cortex. Cerebr. Cortex. 2008;18:2158–2168. doi: 10.1093/cercor/bhm242. N. Y. N 1991. [DOI] [PubMed] [Google Scholar]
  65. Scase M.O., Braddick O.J., Raymond J.E. What is noise for the motion system? Vis. Res. 1996;36:2579–2586. doi: 10.1016/0042-6989(95)00325-8. [DOI] [PubMed] [Google Scholar]
  66. Scholte H.S., Jolij J., Fahrenfort J.J., Lamme V.A.F. Feedforward and recurrent processing in scene segmentation: electroencephalography and functional magnetic resonance imaging. J. Cogn. Neurosci. 2008;20:2097–2109. doi: 10.1162/jocn.2008.20142. [DOI] [PubMed] [Google Scholar]
  67. Schwarzkopf D.S., Anderson E.J., Haas B. de, White S.J., Rees G. Larger extrastriate population receptive fields in autism spectrum disorders. J. Neurosci. 2014;34:2713–2724. doi: 10.1523/JNEUROSCI.4416-13.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Sereno M.I., Dale A.M., Reppas J.B., Kwong K.K., Belliveau J.W., Brady T.J., Rosen B.R., Tootell R.B. Borders of multiple visual areas in humans revealed by functional magnetic resonance imaging. Science. 1995;268:889–893. doi: 10.1126/science.7754376. [DOI] [PubMed] [Google Scholar]
  69. Simoncelli E.P., Heeger D.J. A model of neuronal responses in visual area MT. Vis. Res. 1998;38:743–761. doi: 10.1016/S0042-6989(97)00183-1. [DOI] [PubMed] [Google Scholar]
  70. Snowden R.J., Treue S., Erickson R.G., Andersen R.A. The response of area MT and V1 neurons to transparent motion. J. Neurosci. 1991;11:2768–2785. doi: 10.1523/JNEUROSCI.11-09-02768.1991. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Snowden R.J., Verstraten F.A.J. Motion transparency: making models of motion perception transparent. Trends Cognit. Sci. 1999;3:369–377. doi: 10.1016/S1364-6613(99)01381-9. [DOI] [PubMed] [Google Scholar]
  72. Sperandio I., Chouinard P.A., Goodale M.A. Retinotopic activity in V1 reflects the perceived and not the retinal size of an afterimage. Nat. Neurosci. 2012;15:540–542. doi: 10.1038/nn.3069. [DOI] [PubMed] [Google Scholar]
  73. Sunaert S., Van Hecke P., Marchal G., Orban G.A. Motion-responsive regions of the human brain. Exp. Brain Res. 1999;127:355–370. doi: 10.1007/s002210050804. [DOI] [PubMed] [Google Scholar]
  74. Tootell R.B.H., Hadjikhani N., Hall E.K., Marrett S., Vanduffel W., Vaughan J.T., Dale A.M. The retinotopy of visual spatial attention. Neuron. 1998;21:1409–1422. doi: 10.1016/S0896-6273(00)80659-5. [DOI] [PubMed] [Google Scholar]
  75. Tootell R.B.H., Mendola J.D., Hadjikhani N.K., Ledden P.J., Liu A.K., Reppas J.B., Sereno M.I., Dale A.M. Functional analysis of V3A and related areas in human visual cortex. J. Neurosci. 1997;17:7060–7078. doi: 10.1523/JNEUROSCI.17-18-07060.1997. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Tsao D.Y., Vanduffel W., Sasaki Y., Fize D., Knutsen T.A., Mandeville J.B., Wald L.L., Dale A.M., Rosen B.R., Essen D.C.V., Livingstone M.S., Orban G.A., Tootell R.B.H. Stereopsis activates V3A and caudal intraparietal areas in macaques and humans. Neuron. 2003;39:555–568. doi: 10.1016/S0896-6273(03)00459-8. [DOI] [PubMed] [Google Scholar]
  77. Tyler C.W., Likova L.T., Kontsevich L.L., Wade A.R. The specificity of cortical region KO to depth structure. Neuroimage. 2006;30:228–238. doi: 10.1016/j.neuroimage.2005.09.067. [DOI] [PubMed] [Google Scholar]
  78. van Dijk J.A., de Haas B., Moutsiana C., Schwarzkopf D.S. Intersession reliability of population receptive field estimates. Neuroimage. 2016;143:293–303. doi: 10.1016/j.neuroimage.2016.09.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Van Essen D.C., Gallant J.L. Neural mechanisms of form and motion processing in the primate visual system. Neuron. 1994;13:1–10. doi: 10.1016/0896-6273(94)90455-3. [DOI] [PubMed] [Google Scholar]
  80. Van Oostende S., Sunaert S., Van Hecke P., Marchal G., Orban G.A. The kinetic occipital (KO) region in man: an fMRI study. Cerebr. Cortex. 1997;7:690–701. doi: 10.1093/cercor/7.7.690. Cortex N. Y. N 1991. [DOI] [PubMed] [Google Scholar]
  81. Vidnyánszky Z., Blaser E., Papathomas T.V. Motion integration during motion aftereffects. Trends Cognit. Sci. 2002;6:157–161. doi: 10.1016/S1364-6613(02)01871-5. [DOI] [PubMed] [Google Scholar]
  82. Vo V.A., Sprague T.C., Serences J.T. Spatial tuning shifts increase the discriminability and fidelity of population codes in visual cortex. J. Neurosci. Off. J. Soc. Neurosci. 2017;37:3386–3401. doi: 10.1523/JNEUROSCI.3484-16.2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Wallach H. Über visuell wahrgenommene Bewegungsrichtung. Psychol. Forsch. 1935;20:325–380. doi: 10.1007/BF02409790. [DOI] [Google Scholar]
  84. Watamaniuk S.N. Ideal observer for discrimination of the global direction of dynamic random-dot stimuli. J. Opt. Soc. Am. A. 1993;10:16–28. doi: 10.1364/josaa.10.000016. [DOI] [PubMed] [Google Scholar]
  85. Williams D.W., Sekuler R. Coherent global motion percepts from stochastic local motions. Vis. Res. 1984;24:55–62. doi: 10.1016/0042-6989(84)90144-5. [DOI] [PubMed] [Google Scholar]
  86. Yildirim F., Carvalho J., Cornelissen F.W. A second-order orientation-contrast stimulus for population-receptive-field-based retinotopic mapping. NeuroImage. 2018;164:183–193. doi: 10.1016/j.neuroimage.2017.06.073. [DOI] [PubMed] [Google Scholar]
  87. Zeki S., Perry R.J., Bartels A. The processing of kinetic contours in the brain. Cerebr. Cortex. 2003;13:189–202. doi: 10.1093/cercor/13.2.189. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary document

containing an example delineation of TO1 and TO2, eccentricity plots and analyses using a R2 threshold of 0.1.

mmc1.docx (4.4MB, docx)
Multimedia component 2
mmc2.xml (947B, xml)
Video S1

Example bar-only stimulus.

mmc3.gif (5.7MB, gif)
Video S2

Example kinetic stimulus.

mmc4.gif (34.3MB, gif)
Video S3

Example global stimulus.

mmc5.gif (34.3MB, gif)
Video S4

Example transparent bar-only stimulus.

mmc6.gif (34.3MB, gif)
Video S5

Example transparent motion stimulus.

mmc7.gif (26.2MB, gif)
Video S6

Example size-defined stimulus.

mmc8.gif (4.5MB, gif)

RESOURCES