Abstract
Eye-centered (egocentric) and landmark-centered (allocentric) visual signals influence spatial cognition, navigation, and goal-directed action, but the neural mechanisms that integrate these signals for motor control are poorly understood. A likely candidate for egocentric/allocentric integration in the gaze control system is the supplementary eye fields (SEF), a mediofrontal structure with high-level “executive” functions, spatially tuned visual/motor response fields, and reciprocal projections with the frontal eye fields (FEF). To test this hypothesis, we trained two head-unrestrained monkeys (Macaca mulatta) to saccade toward a remembered visual target in the presence of a visual landmark that shifted during the delay, causing gaze end points to shift partially in the same direction. A total of 256 SEF neurons were recorded, including 68 with spatially tuned response fields. Model fits to the latter established that, like the FEF and superior colliculus (SC), spatially tuned SEF responses primarily showed an egocentric (eye-centered) target-to-gaze position transformation. However, the landmark shift influenced this default egocentric transformation: during the delay, motor neurons (with no visual response) showed a transient but unintegrated shift (i.e., not correlated with the target-to-gaze transformation), whereas during the saccade-related burst visuomotor (VM) neurons showed an integrated shift (i.e., correlated with the target-to-gaze transformation). This differed from our simultaneous FEF recordings (Bharmauria et al., 2020), which showed a transient shift in VM neurons, followed by an integrated response in all motor responses. Based on these findings and past literature, we propose that prefrontal cortex incorporates landmark-centered information into a distributed, eye-centered target-to-gaze transformation through a reciprocal prefrontal circuit.
Keywords: frontal cortex, head-unrestrained, landmarks, macaques, neural response fields, reference frames
Significance Statement
It is thought that the brain integrates egocentric (self-centered) and allocentric (landmark-centered) visual signals to generate accurate goal-directed movements, but the neural mechanism is not known. Here, by shifting a visual landmark while recording frontal cortex activity in awake behaving monkeys, we show that the supplementary eye fields (SEF) incorporates landmark-centered information (in memory and motor activity) when it transforms target location into future gaze position commands. We propose a circuit model in which the SEF provides control signals to implement an integrated gaze command in the frontal eye fields (FEF; Bharmauria et al., 2020). Taken together, these experiments explain normal egocentric/allocentric integration and might suggest rehabilitation strategies for neurologic patients who have lost one of these visual mechanisms.
Introduction
The brain integrates egocentric (eye-centered) and allocentric (landmark-centered) visual cues to guide goal-directed behavior (Goodale and Haffenden, 1998; Ball et al., 2009; Chen et al., 2011; Karimpur et al., 2020). For example, to score a goal, a soccer forward must derive allocentric relationships (e.g., where is the goaltender relative to the posts?) from eye-centered visual inputs to predict an opening, and then transform this into body-centered motor commands. The neural mechanisms for egocentric/allocentric coding for vision, cognition, and navigation have been extensively studied (O'Keefe and Dostrovsky, 1971; O'Keefe, 1976; Rosenbaum et al., 2004; Milner and Goodale, 2006; Schenk, 2006; Ekstrom et al., 2014), but the mechanisms for goal-directed behavior are poorly understood. One clue is that humans aim reaches toward some intermediate point between conflicting egocentric/allocentric cues, suggesting Bayesian integration (Bridgeman et al., 1997; Lemay et al., 2004; Neely et al., 2008; Byrne and Crawford, 2010; Fiehler et al., 2014; Klinghammer et al., 2017). Neuroimaging studies suggest this may occur in parietofrontal cortex (Chen et al., 2018) but could not reveal the cellular mechanisms. However, similar behavior has been observed in the primate gaze system (Li et al., 2017), suggesting this system can be also used to study egocentric/allocentric integration.
It is thought that higher level gaze structures, lateral intraparietal cortex (LIP), frontal eye fields (FEF), superior colliculus (SC), primarily employ eye-centered codes (Russo and Bruce, 1993; Tehovnik et al., 2000; Klier et al., 2001; Paré and Wurtz, 2001; Goldberg et al., 2002), and transform target location (T) into future gaze position (G; Schall et al., 1995; Everling et al., 1999; Constantin et al., 2007). We recently confirmed this by fitting various spatial models against FEF and SC response field activity (Sadeh et al., 2015, 2020; Sajad et al., 2015, 2016). Visual responses coded for target position in eye coordinates (Te), whereas motor responses (separated from vision by a delay) coded for future gaze position in eye coordinates (Ge), and a progressive target-to-gaze (T-G) transformation (where G includes errors relative to T) along visual-memory-motor activity. Further, when we introduced a visual landmark, and then shifted it during the delay period, response field coordinate systems (in cells with visual, delay, and motor activity) shifted partially in the same direction (Bharmauria et al., 2020). Eventually this allocentric shift became integrated into the egocentric (T-G) transformation of all cells that produced a motor burst during the saccade. However, it is unclear whether this occurs independently through a direct visuomotor (VM) path to FEF, or in concert with higher control mechanisms.
One likely executive control mechanism could be the supplementary eye fields (SEF), located in the dorsomedial frontal cortex (Schlag and Schlag-Rey, 1987) and reciprocally connected to the FEF (Huerta and Kaas, 1990; Stuphorn, 2015). The SEF has visual and gaze motor response fields (Schlag and Schlag-Rey, 1987; Schall, 1991), but its role in gaze control is controversial (Abzug and Sommer, 2017). The SEF is involved in various high-level oculomotor functions (Olson and Gettner, 1995; Stuphorn et al., 2000; Tremblay et al., 2002; Sajad et al., 2019) and reference frame transformations, both egocentric (Schlag and Schlag-Rey, 1987; Schall et al., 1993; Martinez-Trujillo et al., 2004), and object-centered, i.e., one part of an object relative to another (Olson and Gettner, 1995; Tremblay et al., 2002). However, the role of the SEF in coding egocentric VM signals for head-unrestrained gaze shifts is untested, and its role in the implicit coding of gaze relative to independent visual landmarks is unknown.
Here, we recorded SEF neurons in two head-unrestrained monkeys while they performed gaze shifts in the presence of implicitly conflicting egocentric and allocentric cues (Fig. 1A). As reported previously, gaze shifted away from the target, in the direction of a shifted landmark (Byrne and Crawford, 2010; Li et al., 2017). We employed our model-fitting procedure (Keith et al., 2009; DeSouza et al., 2011; Sadeh et al., 2015; Sajad et al., 2015) to analyze the neural data. First, we tested all possible egocentric and allocentric models. Second, we performed a spatial continuum analysis (between both egocentric and allocentric models) through time to see if, when, and how egocentric and allocentric transformations are integrated in the SEF. We find that (1) SEF neurons predominantly possess an eye-centered transformation from target-to-gaze coding and (2) landmark-centered information is integrated into this transformation, but through somewhat different cellular mechanisms than the FEF (Bharmauria et al., 2020). We thus propose a reciprocal SEF-FEF model for allocentric/egocentric integration in the gaze system.
Materials and Methods
Surgical procedures and recordings of 3D gaze, eye, and head
The experimental protocols complied the guidelines of Canadian Council on Animal Care on the use of laboratory animals and were approved by the York University Animal Care Committee. Neural recordings were performed on two female Macaca mulatta monkeys (Monkeys V and L) and they were implanted with 2D and 3D sclera search coils in left eyes for eye-movement and electrophysiological recordings (Crawford et al., 1999; Klier et al., 2003). The eye coils allowed us to register 3D eye movements (i.e., gaze) and orientation (horizontal, vertical, and torsional components of eye orientation relative to space). During the experiment, two head coils (orthogonal to each other) were also mounted that allowed similar recordings of the head orientation in space. We then implanted the recording chamber centered in stereotaxic coordinates at 25 mm anterior and 0 mm lateral for both animals. A craniotomy of 19 mm (diameter) that covered the chamber base (adhered over the trephination with dental acrylic) allowed access to the right SEF. Animals were seated within a custom-made primate chair during experiments, and this allowed free head movements at the center of three mutually orthogonal magnetic fields (Crawford et al., 1999). The values acquired from 2D and 3D eye and head coils allowed us to compute other variables such as the orientation of the eye relative to the head, eye-velocity and head-velocity, and accelerations (Crawford et al., 1999).
Basic behavioral paradigm
The visual stimuli were presented on a flat screen (placed 80 cm away from the animal) using laser projections (Fig. 1A). The monkeys were trained on a standard memory-guided gaze task to remember a target relative to a visual allocentric landmark (two intersecting lines acted as an allocentric landmark) thus leading to a temporal delay between the target presentation and initiation of the eye movement. This allowed us to separately analyze visual (aligned to target) and eye-movement-related (saccade onset) responses in the SEF. To not provide any additional allocentric cues, the experiment was conducted in a dark room. In a single trial, the animal began by fixating on a red dot (placed centrally) for 500 ms while the landmark was present on the screen. This was followed by a brief flash of visual target (T, white dot) for 100 ms, and then a brief delay (300 ms), a grid-like mask (200 ms, this hides the past visual traces, and also the current and future landmark), and a second memory delay (200–600 ms, i.e., from the onset of the landmark until the go signal). As the red fixation dot extinguished, the animal was signaled to saccade head-unrestrained (indicated by the solid green arrow) toward the memorized location of the target either in the presence of a shifted (indicated by broken green arrow) landmark (90% of trials) or in absence of it (10%, no-shift/zero-shift condition, i.e., landmark was present at the same location as before mask). These trials with zero-shift were used to compute data at the “origin” of the coordinate system for the T-T’ spatial model fits. The saccade targets were flashed one-by-one randomly throughout the response field of a neuron. Note: green color highlights the items that were not presented on the screen (they are only for representational purposes).
The spatial details of the task are in Figure 1B. Figure 1B1 shows an illustration of a gaze shift (blue curve) to an example target (T) in presence of a shifted landmark (blue cross). G refers to the final gaze endpoint, and T’ stands for the virtual target (fixed to the shifted landmark). The landmark vertex could initially appear at one of four spots located 11° obliquely relative to the target and then shift in any one of eight directions (Fig. 1B2). Importantly, the timing and amplitude (8°) of this shift was fixed. Since these animals had been trained, tested behaviorally (Li et al., 2017) and then retrained for this study over a period exceeding two years, it is reasonable to expect that they may have learned to anticipate the timing and the amount of influence of the landmark shift. However, we were careful not to bias this influence: animals were rewarded with a water-drop if gaze was placed (G) within 8–12° radius around the original target (i.e., they were rewarded if they looked at T, toward or away from T’, or anywhere in between). Based on our previous behavioral result in these animals (Li et al., 2017), we expected this paradigm to cause gaze to shift partially toward the virtually shifted target in landmark coordinates (T’).
Note that this paradigm was optimized for our method for fitting spatial models to neural activity (see below in the section ‘Fitting neural response fields against spatial models’), which is based on variable dissociations between measurable parameters such as target location and effectors (gaze, eye, head), and various egocentric/allocentric reference frames (Keith et al., 2009; Sajad et al., 2015). This was optimized by providing variable landmark locations and shift directions, and the use of a large reward window to allow these shifts (and other endogenous factors) to influence gaze errors relative to T. We also jittered the initial fixation locations within a 7–12° window to dissociate gaze-centered and space-centered frames of reference (note that no correlation was observed between the initial gaze location and final gaze errors). Further dissociations between effectors and egocentric frames were provided by the animals themselves, i.e., in the naturally variable contributions of the eye and head to initial gaze position and the amplitude/direction of gaze shifts. Details of such behavior have been described in detail in our previous papers (Sadeh et al., 2015; Sajad et al., 2015).
Behavioral recordings and analysis
During experiments, we recorded the movement of eye and head orientations (in space) with a sampling rate of 1000 Hz. For the analysis of eye movement, the saccade onset (eye movement in space) was marked at the point in time when the gaze velocity exceeded 50°/s, and the gaze offset was marked as the point in time when the velocity declined below 30°/s. The head movement was marked from the saccade onset till the time point at which the head velocity declined below 15°/s.
When the landmark shifted (90% of trials), its influence on measured future gaze position (G) was called allocentric weight (AW), computed as follows:
(1) |
where d is the component of T-G (error space between the actual target location and the final measured gaze position) that projects onto the vector direction of the landmark shift, and D is the magnitude of the landmark shift (Byrne and Crawford, 2010; Li et al., 2017). This was done for each trial, and then averaged to find the representative landmark influence on behavior in a large number of trials. A mean AW score of zero signifies no landmark influence, i.e., gaze shifts headed on average toward T. A mean AW score of 1.0 means that on average, gaze headed toward a virtual target position (T’) that remained fixed to the shifted landmark position. As we shall see (Fig. 1C), AW scores for individual trials often fell between 0 and 1 but varied considerably, possibly because of trial-to-trial variations in landmark influence and/or other sources of variable gaze error that are present without a landmark shift (Sajad et al., 2015, 2016, 2020).
Electrophysiological recordings and response field mapping
Tungsten electrodes (0.2- to 2.0-mΩ impedance, FHC Inc.) were lowered into the SEF [using Narishige (MO-90) hydraulic micromanipulator] to record extracellular activity. Then the recorded activity was digitized, amplified, filtered, and saved for offline spike sorting using template matching and applying principal component analysis on the isolated clusters with Plexon MAP System. The recorded sites of SEF (in head-restrained conditions) were further confirmed by injecting a low-threshold electrical microstimulation (50 μA) as previously used (Bruce et al., 1985). A superimposed pictorial of the recorded sites from both animals is presented in Figure 2A,B (Monkey L in blue and Monkey V in red).
We mostly searched for neurons while the monkey freely (head-unrestrained) scanned the environment. Once we noticed that a neuron had reliable spiking, the experiment started. The response field (visual and/or motor) of neuron was characterized while the animal performed the memory-guided gaze shift paradigm as described above in the section ‘Basic behavioral paradigm’. After an initial sampling period to determine the horizontal and vertical extent of the response field, targets were presented (one per trial) in a 4 × 4 to 7 × 7 array (5–10° from each other) spanning 30–80°. This allowed the mapping of visual and motor response fields such as those shown in Results. For analysis we aimed at ∼10 trials per target, so the bigger the response field (and thus the more targets), the more the number of recorded trials were required and vice versa. On average, 343 ± 166 (mean ± SD) trials/neuron were recorded, again depending on the size of the response field. We did such recordings from >200 SEF sites, often in conjunction with simultaneous FEF recordings, as reported previously (Bharmauria et al., 2020).
Data inclusion criteria, sampling window, and neuronal classification
In total, we isolated 256 SEF neurons: 102 and 154 neurons were recorded from Monkeys V and L, respectively. Of these, we only analyzed task-modulated neurons with clear visual burst and/or with perisaccadic movement response (Fig. 2C,D). Neurons that only had postsaccadic activity (activity after the saccade onset) were excluded. Moreover, neurons that lacked significant spatial tuning were also eliminated (see below, Testing for spatial tuning). In the end, after applying our exclusion criteria, we were left with 37 and 31 spatially tuned neurons in Monkeys V and L, respectively. Only those trials were included where monkeys landed their gaze within the reward acceptance window, however, we eliminated gaze end points beyond ±2° of the mean distribution from our analysis. For the analysis of the neural activity, the “visual epoch” sampling window was chosen as a fixed 100-ms window of 80–180 ms aligned to the target onset, and the “movement period” was characterized as a high-frequency perisaccadic 100-ms (−50 to +50 ms relative to saccade onset) period (Sajad et al., 2015). This allowed us to get a good signal/noise ratio for neuronal activity analysis, and most likely corresponded to the epoch in which gaze shifts were influenced by SEF activity. After rigorously employing our exclusion criteria and data analyses (Table 1), we further dissociated the spatially tuned neurons into pure visual (V; n = 6), pure motor (M; n = 26) and VM neurons (which possessed visual, maximal delay, and motor activity, n = 36) based on the common dissociation procedure (Bruce and Goldberg, 1985; Sajad et al., 2015; Schall, 2015; Khanna et al., 2019). Note: the motor population also includes neurons which have delayed memory activity (Sajad et al., 2016).
Table 1.
Analysis | Statistical test | Power |
---|---|---|
Distribution significantly shifted from a point | One-sampled Wilcoxon signed-rank test | p < 0.05 |
Comparison between models | Brown–Forsythe test | p = 100 = 1; the best model |
Comparison between distributions | Mann–Whitney U test | p < 0.05 |
Transformation between the visual and the motor responses for VM neurons |
Wilcoxon matched-pairs signed-rank test | p < 0.05 |
Spatial tuning test for a neuron | CI (see above in the section ‘Testing for spatial tuning’) | p < 0.05 |
Comparison between VM and motor neurons in SEF and FEF |
Three-factor ANOVA analysis | p < 0.05 |
Correlation between T-G and T-T’ | Spearman correlation | p < 0.05 |
Spatial models analyzed (egocentric and allocentric)
Figure 3 graphically shows how we derived the 11 egocentric “canonical” models tested in this study and then chose the best egocentric model to compare with the allocentric models. Briefly, based on these models, this figure illustrates the formal means for comparing between target versus gaze position coding and gaze versus eye versus head displacement/position (Fig. 3A–C), with each plotted in different possible frames of reference. We then use the landmark-centered information to test between the egocentric versus allocentric modes of coding (Fig. 3D). Note that in our analysis, these models were based on actual target, gaze, eye, and head data, either derived from geometric calculations (in the case of T) or eye/head coil measurements. As described previously, the variability required to distinguish between these models was either provided by ourselves (i.e., stimulus placement) or the monkeys’ own behavior (i.e., variable gaze errors and variable combinations of eye-head position; Sajad et al., 2015).
We first tested for all the egocentric models as reported in our previous studies (Sajad et al., 2015, 2016) and then tested the best egocentric model with all “pure” allocentric models. Based on different spatial parameters [most importantly target (T) and final gaze position (G)], we fitted visual and motor response fields of SEF against previously tested 11 (Fig. 3A–C) egocentric canonical models in FEF (Sajad et al., 2015). Note that Te and Ge codes were obtained by mathematically rotating the experimental measures of T and G in space coordinates by the inverse of experimentally measured initial eye orientation to obtain the measures in eye coordinates (Klier et al., 2001). In other words, Te is based on the actual target location relative to initial eye orientation, whereas Ge is based on the final gaze position relative to initial eye orientation.
The following models were also tested: dH, difference between the initial and final head orientation in space coordinates; dE: difference between the initial and final eye orientation in head coordinates; dG, difference between the initial and final gaze position in space coordinates; Hs, final position of the head in space coordinates; Eh, final position of the eye in head coordinates; Gs, Ge, Gh: gaze in space, eye, and coordinates, respectively; Ts, Te, Th: target in space, eye, and head coordinates, respectively. The final position refers to orientation of eye/head after the gaze saccade. These models are detailed in the previous study (Sajad et al., 2015). Note that some of these models (like dG and Ge) might not be distinguishable within the range of initial gaze positions and saccade amplitudes described here (Crawford and Guitton, 1997).
Since an allocentric landmark was involved in this task, we analyzed eight additional (allocentric) models of target coding based on the original and shifted landmark location (Fig. 3D). In the allocentric analysis, we retained the best egocentric model (Te for visual neurons and Ge for motor neurons) for comparison. The tested allocentric models are: Ls, landmark within the space coordinates; L’s, shifted landmark within the space coordinates; Le, landmark within eye coordinates; L’e, shifted landmark within eye coordinates; T’s, shifted target in space coordinates; T’e, shifted target in eye coordinates; TL, target relative to landmark; TL’, target relative to shifted landmark. Note: prime (’) stands for the positions that are related to the shifted location of the landmark.
Intermediate spatial models used in main analysis
Previous reports on FEF responses from our lab have reported that responses do not fit exactly against spatial models like Te or Ge, but actually may fit best against intermediate models between the canonical ones (Fig. 4A, lower left; Sajad et al., 2020). As in our previous studies (Sajad et al., 2016; Sadeh et al., 2020), we found that a T-G continuum (specifically, steps along the “error line” between Te and Ge) best quantified the SEF egocentric transformation (Fig. 4A, lower left). This continuum is similar to the concept of an intermediate frame of reference (e.g., between the eye and head) but is instead intermediate between target and gaze position within the same frame of reference. Hereafter, we will sometimes refer to T-G as “the egocentric code.”
Further, to quantify the influence of the landmark shift, we created (Fig. 4A, lower right) another continuum (T-T’, the line between the original target and the target if it were fixed to the landmark) between Te (target fixed in eye coordinates) and T’e (virtual target fixed in landmark coordinates), computed in eye coordinates with 10 intermediary steps between, and additional steps on the flanks (10 beyond Te and 10 beyond T’e). These additional 10 steps beyond the canonical models were included (1) to quantify if neurons can carry abstract spatial codes around the canonical models and (2) to eliminate the misleading edge effects (else the best spatial models would cluster around the canonical models). The T-T’ continuum will allow us to test whether SEF code is purely egocentric (based on T), allocentric (based on T’), or contains the integrated allocentric + egocentric information that eventually actuates the behavior. Note that AW and T-T’ are geometrically similar, but the first describes behavioral data, whereas the second describes neural data. Hereafter, we will refer to T-T’ as the “allocentric shift.”
Fitting neural response fields against spatial models
In order to test different spatial models, they should be spatially separable (Keith et al., 2009; Sajad et al., 2015). The natural variation in monkeys’ behavior allowed this spatial separation (for details, see Results). For instance, the variability induced by memory-guided gaze shifts permitted us to discriminate target coding from the gaze coding; the initial eye and head locations allowed us to distinguish between different egocentric reference frames and variable eye and head movements for a gaze shift permitted the separation of different effectors. As opposed to the decoding methods that generally test whether a spatial characteristic is implicitly coded in patterns of population activity (Bremmer et al., 2016; Brandman et al., 2017), the method employed in this study method directly examines which model best predicts the activity in spatially tuned neural responses. The logic of our response field fitting in different reference frames is schematized in Figure 4B. Precisely, if the response field data are plotted in the correct best reference frame, this will yield the lowest residuals (errors between the fit and data points) compared with other models, i.e., if a fit computed to its response field matches the data, then this will yield low residuals (Fig. 4B, left). On the other hand, if the fit does not match the data well, this will lead to higher residuals (Fig. 4B, right). For example, if an eye-fixed response field is computed in eye-coordinates this will lead to lower residuals and if it is plotted in any other inferior/incorrect coordinate, such as space coordinate, this will produce higher residuals (Sajad et al., 2015).
In reality, we employed a non-parametric fitting method to characterize the neural activity with reference to a spatial location and also varied the spatial bandwidth of the fit to plot any response field size, shape, or contour (Keith et al., 2009). We tested between various spatial models using predicted residual error some of squares (PRESS)-statistics. To independently compute the residual for a trial, the actual activity related to it was subtracted from the corresponding point on the fit computed over all the other trials (like cross-validation). Notably, if the spatial physical shift between two models results in a systematic shift (direction and amount), this will appear as a shifted response field or expanded response field and our model fitting approach would not be able to distinguish two models as they would produce virtually indistinguishable residuals. Because in our study, the distribution of relative positions across different models also possesses a non-systematic variable component (e.g., variability in gaze endpoint errors, or unpredictable landmark shifts), the response fields invariably stayed at the same location, but the dissociation between spatial models was based on the residual analysis.
We plotted response fields (visual and movement) of neurons in the coordinates of all the canonical (and intermediate) models. To map the visual response in egocentric coordinates, we took eye and head orientations at the time of target presentation, and for movement response fields, we used behavioral measurements at the time when the gaze shift started (Keith et al., 2009; DeSouza et al., 2011; Sajad et al., 2015). Likewise, for the allocentric models, we used the initial and the shifted landmark location to plot our data. Since we did not know the size and shape of a response field a priori and since the spatial distribution of data were different for every spatial model (e.g., the models would have a smaller range for head than the eye models), we computed the non-parametric fits with different kernel bandwidth (2–25°), thus making sure that we did not bias the fits toward a particular size and spatial distribution. For all the tested models with different non-parametric fits, we computed the PRESS residuals to reveal the best model for the neural activity (that yielded the least PRESS residuals). We then statistically (Keith et al., 2009) compared the mean PRESS residuals of the best model with the mean PRESS residuals of other models at the same kernel bandwidth (two-tailed Brown–Forsythe test). Finally, we performed the same statistical analysis (Brown–Forsythe) at the population level (Keith et al., 2009; DeSouza et al., 2011). For the models in intermediate continua, a similar procedure was used to compute the best fits.
Testing for spatial tuning
The above-described method in the section ‘Fitting neural response fields against spatial models’ assumes that neuronal activity is structured as spatially tuned response fields. This does not imply that other neurons do not subserve the overall population code (Goris et al., 2014; Bharmauria et al., 2016; Leavitt et al., 2017; Chaplin et al., 2018; Zylberberg, 2018; Pruszynski and Zylberberg, 2019), but with our method, only tuned neurons can be explicitly tested. We tested for the neuronal spatial tuning as follows. We randomly (100 times to obtain random 100 response fields) shuffled the firing rate data points across the position data that we obtained from the best model. The mean PRESS residual distribution (PRESSrandom) of the 100 randomly generated response fields was then statistically compared with the mean PRESS residual (PRESSbest-fit) distribution of the best-fit model (unshuffled, original data). If the best-fit mean PRESS fell outside of the 95% confidence interval of the distribution of the shuffled mean PRESS, then the neuron’s activity was deemed spatially selective. At the population level, some neurons displayed spatial tuning at certain time-steps and others did not because of low signal/noise ratio. Thus, we removed the time steps where the populational mean spatial coherence (goodness of fit) was statistically indiscriminable from the baseline (before target onset) because there was no task-related information at this time and thus neural activity exhibited no spatial tuning. We defined an index (coherence index; CI) for spatial tuning. CI for a single neuron, which was calculated as (Sajad et al., 2016):
(2) |
If the PRESSbest-fit was similar to PRESSrandom then the CI would be roughly 0, whereas if the best-fit model is a perfect fit (i.e., PRESSbest-fit = 0), then the CI would be 1. We only included those neurons in our analysis that showed significant spatial tuning.
Time normalization for spatiotemporal analysis
A major aim of this study was to track the progression of the T-G and T-T’ codes in spatially tuned neuron populations, from the landmark-shift until the saccade onset. The challenge here was that this period was variable, so that averaging data across trials for each neuron would result in the mixing of very different signals (see Extended Data Fig. 11-1). To compensate for this, we time-normalized this data (Sajad et al., 2016; Bharmauria et al., 2020). To do this, the neural firing rate (in spikes per second; the number of spikes divided by the sampling interval for each trial) was sampled at seven half-overlapping time windows (with a range of 80.5–180.5 ms depending on trial duration). The rationale behind the bin number choice was to make sure that the sampling time window was wide enough, and therefore robust enough to account for the stochastic nature of spiking activity (thus ensuring that there were sufficient neuronal spikes in the sampling window to do effective spatial analysis; Sajad et al., 2016). Once the firing rate was estimated for each trial at a given time-step, they were pooled together for the spatial modeling. Note that the final (seventh) time step also contained some part of the perisaccadic sampling window. Finally, we performed our T-G/T-T’ fits on the data for each of these time bins. In short, this procedure allowed us to treat the whole sequence of memory-motor responses from the landmark shift until saccade onset as a continuum.
Results
Influence of landmark shift on behavior
To investigate how the landmark influences behavior toward an object of interest, we used a cue-conflict task, where a visual landmark shifted during the memory delay, i.e., between the target presentation and the gaze shift (Fig. 1A). The task is further schematized in Figure 1B with possible target, original landmark, virtual target and shifted landmark locations. We then computed the influence of the landmark shift as described previously (Li et al., 2017) and further confirmed it in the current dataset. This is computed as AW, i.e., the component of gaze end points along the axis between the target and the cue shift direction. If AW equals 0, then it implies no influence of the shifted landmark, whereas if AW equals 1 then it indicates a complete influence of landmark shift on the gaze (see Materials and Methods). Note that we performed this analysis on the same trials as used for the analysis of spatially tuned neural activity in the results section.
In both animals (Monkeys V and L), the gaze endpoints scattered along the one-dimensional axis of the landmark shift (Fig. 1C), producing a bell-like distribution of gaze errors. However, both distributions showed a highly significant (p < 0.001, one sampled Wilcoxon signed-rank test) shift from the original target (0) in the direction of the landmark shift (with a mean AW of 0.36 in Monkey V and 0.29 in Monkey L). There was considerable trial-to-trial variance around these means, but note that such variance was present or even larger in the absence of a landmark (Sajad et al., 2015; Li et al., 2017). The overall average error (distance) between T and final gaze, i.e., including errors in all directions, was 8.19 ± 5.27 (mean ± SD) and 10.35 ± 5.89 for Monkeys V and L, respectively. No correlation was found between the AW as a function of saccade latency (Bharmauria et al., 2020). These results generally agree with previous cue-conflict landmark studies (Neggers et al., 2005; Byrne and Crawford, 2010; Fiehler et al., 2014) and specifically with our use of this paradigm in the same animals, with only slight differences owing to collection of data on different days (Li et al., 2017; Bharmauria et al., 2020).
In order to determine when the landmark first influenced the SEF spatial code, and when that influence became fully integrated with the default SEF codes, we asked the following questions in their logical order: (1) what are the fundamental egocentric and/or allocentric codes in the SEF; (2) how does the landmark shift influence these neural codes; (3) what is the contribution of different SEF cell types (visual, VM, and motor); and finally, (4) is there any correlation between the SEF allocentric and egocentric codes, as we found in the FEF?
Neural recordings: general observation and classification of neurons
To understand the neural underpinnings of the behavior (as revealed above in the section ‘Influence of landmark shift on behavior’) in the SEF activity, we recorded visual and motor responses from over 200 SEF sites using tungsten microelectrodes, while the monkey performed the cue-conflict task (Fig. 2A,B). We analyzed a total of 256 neurons and after applying our rigorous exclusion criteria (see Materials and Methods and Table 1), we were finally left with 68 significantly spatially tuned neurons (see Materials and Methods): 32 significantly spatially tuned visual responses (including V and VM) and 53 (including M and VM) significantly spatially tuned motor responses. Many other neurons (n = 188) were not spatially tuned or did not respond to the event and were thus excluded from further analyses. Typically, neurons that were not spatially tuned were responsive at some time during the task throughout both visual fields and did not show preference for any of the spatial models.
The mean spike density graphs (with the 100-ms sampling window) for the visual (red) and motor (black) neurons are shown in Figure 2C,D, respectively. We then performed the non-parametric analysis (Fig. 4A,B) while using all trials throughout the response field of each neuron (dark black and red curves), but we also display the top 10% from every neuron during the sampling window (light red and gray for visual and motor activity, respectively), roughly corresponding to the “hot-spot” of the response field. We first employed our model-fitting approach to investigate SEF spatial codes, starting (since this has not been done with SEF before) with a test of the most fundamental models.
Visual activity fitting in egocentric and allocentric spatial models
We began by testing all the egocentric and allocentric models (Fig. 3) in the sampling window (80–180 ms relative to target onset) of the visual response (including pure visual and VM neurons). Figure 5 shows a typical example of analysis of a visual response field. Figure 5A displays the raster and spike density (pink curve) plot of a visually responsive neuron aligned to the target onset (blue line as indicated by downward blue arrow). The shaded pink area corresponds to the sampling window for response field analysis. Figure 5B shows the closed (spatially restricted) response field for this neuron in the original target-in-space coordinates (Ts) corresponding to actual stimulus locations. Each circle corresponds to the magnitude of response in a single trial. The heat map represents the non-paramedic fit to the neural data, with residuals (difference between data and fit) plotted to the right. This neuron has a hot-spot (red) near the center of the response field.
Figure 5C provides a statistical summary (p values) comparing the goodness of fits for each of our canonical egocentric models (Fig. 3A, x-axis) relative to the model with the lowest residuals. For this neuron, the response field was best fit across target in eye coordinates (Te) with a Gaussian kernel bandwidth of 7° (Fig. 5C, inset, the red vertical line indicates the best kernel bandwidth). Thus, Te is the preferred egocentric model for this neuron, although most other models were not significantly eliminated. Figure 5D shows the response field plot of the neuron in Te coordinates. Comparing Ts and Te, one can see how larger circles (larger bursts of activity) tend to cluster more together at the response field hot spot in Te than Ts, and the residuals look (and were) smaller, implying that Te is the best coordinate system for this neuron. The bottom row of Figure 5 shows a similar analysis of the same neuron with respect to our allocentric models (from Fig. 3B). Figure 5E provides a statistical comparison (p values) of the residuals of these models relative to Te, which we kept as a reference from the egocentric analysis. Te still provided the lowest residuals, whereas nearly every allocentric model was statistically eliminated. The exception was T’e (target-in-eye coordinates but shifted with the landmark) perhaps because it was most similar to Te. We have used this coordinate system for the example response field plot in Figure 5F. An example of a spatially untuned neuron is shown in Extended Data Figure 5-1.
Figure 6 shows the pooled fits for visual (V + VM) responses across all spatially tuned neurons (n = 32) for both egocentric (left column) and allocentric (right column) analyses. In each column, the top row provides the distributions of mean residuals, the middle the p value comparisons, and the bottom row the fraction of spatially tuned neurons that preferred each model. In the egocentric analysis (Fig. 6A,B), Te was still the best fit overall (and was preferred in ∼40% of the neurons). The head-centered models and most of the effector-specific models were statistically eliminated but dE, dG, Ge, and Ts were not eliminated. Compared with the allocentric models (Fig. 6C,D), Te performed even better, statistically eliminating all other models. Overall, these analyses of egocentric and allocentric models suggest that Te was the predominant code in the SEF visual responses, just as we found for the FEF in the same task (Bharmauria et al., 2020) and both the FEF and SC in a purely egocentric gaze task (Sadeh et al., 2015; Sajad et al., 2015).
Motor activity fitting in egocentric and allocentric spatial models
We then proceeded with the analysis of neural activity for motor responses (including the M and VM neurons), i.e., neurons firing in relation to the saccade onset. Despite the landmark-induced shifts and variable errors in gaze, overall there was still a strong correlation between target direction and final gaze position (Bharmauria et al., 2020), and thus motor response fields tended to align with sensory response fields. Figure 7 shows a typical analysis of a motor neuron. Figure 7A displays the raster and spike density plot for a motor response (the shaded area corresponds to the sampling window and Fig. 7B shows the corresponding response field, plotted in Ts coordinates). Figure 7C provides a statistical analysis of this neuron’s activity against all egocentric spatial models (same convention as Fig. 5). Here, Ge (future gaze direction relative to initial eye orientation) was the preferred model for this neuron, although several other models were also retained. Figure 7D shows the corresponding data in Ge coordinates, where again, residuals are lower and similar firing rates cluster. Figure 7E shows the statistical testing of different allocentric models using Ge as the egocentric reference. Once again, the egocentric model “wins,” eliminating all of the allocentric models except for T’e (used in the example response field plot; Fig. 7F).
Figure 8 summarizes our population analysis of motor responses (n = 53, M + VM) against our egocentric and allocentric models, using the same conventions as Figure 6. Overall, of the egocentric models, dE yielded the lowest residuals, but dE was statistically indistinguishable from dG and Ge, which also yielded similar amounts of spatial tuning (Fig. 8A). Importantly, Te was now eliminated, along with dH, Hs, Ts, and Th (Fig. 8B). We retained Ge as our egocentric reference for comparison with the allocentric fits, because it is mathematically similar to Te (used for the visual analysis), and has usually outperformed most motor models in previous studies (Klier et al., 2001; Martinez-Trujillo et al., 2004; Sadeh et al., 2015; Sajad et al., 2015; Bharmauria et al., 2020). This time, comparisons with Ge statistically eliminated all of our allocentric models at the population level. In summary, these analyses suggest that Ge (or something similar), and not Te, was the preferred model for motor responses, as reported in our studies on FEF and SC (Sadeh et al., 2015; Sajad et al., 2015; Bharmauria et al., 2020).
Visual-motor transformation along the intermediate frames: T-G and T-T’ continua
Thus far, we found that SEF continued to be dominated by eye-centered target (T) and gaze (G) codes, like other saccade-related areas (FEF and SC). However, it is possible that the actual codes fall within some intermediate code, as we have found previously in the FEF (Bharmauria et al., 2020). Therefore, as described in Materials and Methods (Fig. 4) and Figure 9, we constructed the same spatial continua to quantify the detailed sensorimotor transformations in the SEF: a T-G (specifically Te-Ge) continuum to quantify the amount of transition from target to future gaze coding, and a T-T’ continuum (specifically Te-T’e; similar to our behavioral AW score) to quantify the influence of the landmark on the target code. Following analysis shows an example of a visual and a motor response.
Figure 9A shows the raster and the spike density plot for a visually responsive neuron aligned to the target onset (same neuron as Fig. 5). Figure 9B displays the best-fit response field plot of the neuron along the T-G continuum, where the circle represents the magnitude of the response, the heat map represents the non-parametric fit to the data, and the residuals are plotted to the right. The converging broken lines pointing to bar at the top represent the corresponding point of best fit along the 10 equal steps between T and G. Here, the response field of neuron fits best exactly at T as indicated by the broken lines. Figure 9C shows the response field plot of the same data along the T-T’ continuum. Here, the best fit for the response field was located only one step (10% beyond T, in the direction away from T’) demonstrating no influence of the future landmark shift on the initial visual response.
What then happens after the landmark shift? Figure 9D depicts the raster and spike density plot of a motor neuron, aligned to the saccade onset. Along the T-G continuum (Fig. 9E), the best response field fit was at the ninth step, i.e., 90% toward G (suggesting a near-complete transformation to gaze coordinates), whereas the best fit along the T-T’ continuum (Fig. 9F), was at fourth step from T, i.e., 40% toward T’ (suggesting landmark influence similar to that seen in our behavioral measure).
Population analysis along the T-G and T-T’ continuum
How representative were the examples shown above (Fig. 9) of our entire population of data? To answer that question, we performed the same analysis on our entire population of spatially tuned visual (both visual and VM) and motor (both VM and motor) responses. Figure 10 shows the distribution of best fits for visual responses (top row) and motor responses (bottom row) neurons. T-G distribution of visually responding neurons (Fig. 10A) showed a primary cluster and peak around T, but overall was shifted slightly toward G (mean = 0.3; median = 0.15), because of a smaller secondary peak at G. This suggests that most visual responses encoded the target, but some already predicated the future gaze location. This is similar to what has been reported in FEF too (Sajad et al., 2015; Bharmauria et al., 2020). The motor distribution (Fig. 10B, n = 53) showed the opposite trend: a smaller cluster remained at T, but the major cluster and peaks were near G. Overall, this resulted in a significant (p < 0.0001, one sampled Wilcoxon signed-rank test) shift toward G (mean = 0.72; median= 0.8). Notably, the motor and visual distributions were significantly different from each other (p < 0.0001; Mann–Whitney U test).
Along the T-T’ continuum (Fig. 10C), the best fits for the visual population peaked mainly around T, but overall showed a small (mean = 0.26; median = 0.15) but non-significant shift toward T’ (p = 0.07, one sampled Wilcoxon signed-rank test). The motor population (Fig. 10D) shifted further toward T’ (mean = 0.32; median= 0.3). This overall motor shift was not significantly different from the overall visual population (p = 0.53; Mann–Whitney U test), but it was significantly shifted from T (p = 0.0002, one sampled Wilcoxon signed-rank test). In general, this T-T’ shift resembled the landmark influence on actual gaze behavior. Notably, at the single cell level there was a significant T-G transition (Extended Data Fig. 10-1) between the visual and the motor responses within VM neurons (n = 16; p = 0.04, Wilcoxon matched-pairs signed-rank test), but not along the TT’ continuum (n = 16; p = 0.32, Wilcoxon matched-pairs signed-rank test).
Overall, this demonstrates a target-to-gaze transformation similar to the SC and FEF (Sadeh et al., 2015; Sajad et al., 2015) and a similar significant landmark influence in the motor response as we found in FEF (Bharmauria et al., 2020). However, we do not yet know how different cell types contribute to this shift and if this landmark influence has some relationship to the egocentric (T-G) transformation as revealed in FEF (Bharmauria et al., 2020).
Contribution of different cell types and the allocentric shift
We next examined how different neuronal classes are implicated in the landmark influence as noticed above (Fig. 11). To this goal, as in our FEF study on the same task (Bharmauria et al., 2020), we focused our analysis on a seven-step time-normalized analysis aligned to onset of the landmark-shift until the saccade onset (Extended Data Fig. 11-1). Since the second delay was variable, the time-normalization procedure allowed us to treat the corresponding neural activity as a single temporal continuum (Sajad et al., 2016; Bharmauria et al., 2020). By employing this procedure on the neural activity, we tracked the progression of T-G and T-T’ continua to quantify the gaze and landmark-shift influence, respectively.
Figure 11A displays the mean activity of the entire spatially tuned population (n = 68) of neurons divided into seven time-normalized bins from landmark-shift onset to the saccade onset (for details, see Materials and Methods). The mean spike density plots are shown for (1) all trials (bottom trace), (2) top 10% activity corresponding to each time step (top trace), and (3) top 10% activity from 80 to 180 ms aligned to the landmark-shift (middle trace). Note (the red histograms below the spike density plots) that the delay period possessed substantial spatially tuned neural activity along the T-T’ continuum, ∼50% of the neurons were tuned. A similar trend was noticed along the T-G continuum (data not shown). Figure 11B shows the data (mean ± SEM) in the corresponding time steps for the population along the T-G (blue) and the T-T’ (red) continua. The solid circles indicate a significant shift from T (p < 0.005, one sampled Wilcoxon signed-rank test), whereas the empty circle indicates a non-significant shift. The T-G code showed a significant shift at all the steps as reported previously (Sajad et al., 2015; Bharmauria et al., 2020). The T-T’ fits were slightly shifted from T at the first step, but this shift was significantly embedded only at the fourth step (p = 0.02), then it shifted back at the fifth step before significantly shifting toward T’ at the sixth (p = 0.001) and seventh (p = 0.003) steps when the gaze was just imminent.
To further tease apart the contribution of different cell types to the embedding of landmark influence, we divided the population into visual only (V), VM, and motor (M) cell types. The V neurons (n = 6; data not shown) did not display any significant shift at any of the time steps, therefore they were eliminated from further analyses. Figure 11C shows the spike density plots as shown for the population in Figure 11A. We did not notice any significant shift in the delay period along the T-T’ continuum for the VM neurons (Fig. 11D). For the delay activity (Fig. 11E) in M neurons, along the T-T’ continuum, a significant shift was observed at the third step (p = 0.04), then the code shifted back before significantly shifting toward T’ at the seventh step (p = 0.02) with impending gaze.
Integration of the allocentric shift with the egocentric code
Until this point we observed that the landmark shift influences the motor code along the T-T’ coordinates, but we still need to address whether there is any relation between the T-G and T-T’ transformations. To address this, post landmark-shift until the perisaccadic burst, we plotted the T-T’ score as a function of the corresponding T-G score for each neuron that exhibited spatial tuning for both (Fig. 12A). As for FEF (Bharmauria et al., 2020), we made following predictions for the embedding of the allocentric influence with the egocentric coordinates (Fig. 12B): (1) no influence, i.e., the coding was purely egocentric but as we have shown above (Figs. 10 and 11) that is not the case; (2) independent, the egocentric and allocentric codes are completely independent of each other; (3) fully integrated, the allocentric influence varies as a function of G; and (4) partial integration, a mix of (2) and (3).
We first did this analysis for early postshift visual response in the 80- to 180-ms window of analysis for the VM and M neurons (Fig. 12C,D). No significant correlation was noticed for both the VM (Spearman R = 0.14; slope = 0.09 ± 0.18, intercept = 0.03 ± 0.16, p = 0.60; Fig. 12C) and the M (Spearman R = –0.06; slope = –0.06 ± 0.37, intercept = 0.52 ± 0.32, p = 0.87; Fig. 12D) neurons, suggesting no integration in this period. We further plotted the correlation for the delay activity from early postshift response until the saccade onset (roughly corresponding to steps 2–7 from the previous figure) for the whole population (M + VM) and the individual M and VM populations. We found no significant correlation at each of these steps either for the entire population and the subpopulations, implying that yet integration had not occurred, although a shift was noticed in the delay and the impending saccadic activity of M neurons (Fig. 11F). Finally, a significant correlation between the T-T’ and T-G was noticed for the VM (n = 27; Spearman R = 0.46, slope = 0.36 ± 0.15, intercept = 0.05 ± 0.13, p = 0.02; Fig. 12E) neurons in the perisaccadic burst (–50 to 50 ms), but not for the M (n = 26; Spearman R = 0.06, slope = 0.09 ± 0.20, intercept = 0.27 ± 0.18, p = 0.64; Fig. 12E) neurons. After combining the M and the VM neurons, a significant correlation still existed (Spearman R = 0.28, slope = 0.26 ± 0.12, intercept = 0.13 ± 0.11, p = 0.03).
Comparison with spatiotemporal integration in FEF
Notably, we performed SEF and FEF recordings concurrently, providing an opportunity to compare the current dataset with the FEF dataset published previously published (Bharmauria et al., 2020); focusing on the spatiotemporal progression of egocentric/allocentric integration after the landmark shift. To do this, we performed a three-factor ANOVA analysis on the VM and motor populations from both areas (F1 = FEF/SEF, F2 = M/VM, F3 = time step). We found a significant difference between the VM and M neurons along the T-G (p = 0.009), T-T’ (p = 0.04) continua. We found no significant difference between the SEF and FEF along the T-G continuum (p = 0.10, suggesting similar egocentric transformations), but we found a significant difference along the T-T’ continuum (p = 0.009), implying a difference in allocentric processing. Moreover, a significant interaction was also noticed between the VM/M neurons of FEF/SEF (p = 0.04) along the T-T’ continuum. Finally, the M neurons of SEF displayed a significant shift in their delay activity (fourth step, Bonferroni corrected Mann–Whitney U test, p = 0.006) toward T’ compared with the M neurons of FEF. These statistics support the observation that both areas showed transient T-T’ shifts in delay activity, but this primarily occurred in VM neurons in the FEF (Bharmauria et al., 2020), as opposed to motor neurons in SEF. Finally, we note that whereas only SEF VM neurons showed T-G/T-T’ correlation during the saccade burst (Fig. 12), both VM and M neurons showed this correlation in the FEF (Bharmauria et al., 2020).
Discussion
This study addressed a fundamental question in cognitive neuroscience: how does the brain represent and integrate allocentric and egocentric spatial information? We used a cue-conflict memory-guided saccade task, in which a visual landmark shifted after a mask, to establish the basic egocentric coding mechanisms used by the SEF during head-unrestrained gaze shifts, and investigate how allocentric information is incorporated into these transformations. We found the following. (1) Despite the presence of a visual landmark, spatially tuned SEF neurons predominantly show the same eye-centered codes as the FEF and SC (Sadeh et al., 2015; Sajad et al., 2015), i.e., target coding (T) in the visual burst and gaze position coding (G) in the motor burst. (2) After the landmark shift, motor neuron delay activity showed a transient shift in the same direction (T-T’). (3) A second perisaccadic shift was observed in VM neurons. (4) Only the latter shift was correlated with T-G. Overall, the SEF showed similar egocentric visual-motor transformations, however, it integrated the landmark information into this transformation in a manner complementary to the FEF. Briefly, the novel results of this investigation implicate the SEF (and thus the frontal cortex) in the integration of allocentric and egocentric visual cues.
General SEF function: spatial or non-spatial?
The FEF, LIP, and SC show (primarily) contralateral visual and motor response fields involved in various spatial functions for gaze control (Andersen et al., 1985; Schlag and Schlag-Rey, 1987; Schall, 1991; Munoz, 2002), but the role of SEF is less clear (Purcell et al., 2012; Abzug and Sommer, 2018). Only 27% of our SEF neurons were spatially tuned, lower than our FEF recordings (50%) in the same sessions (Bharmauria et al., 2020). This is consistent with previous studies (Schall, 1991; Purcell et al., 2012) and the notion that the SEF also has non-spatial functions, such as, learning (Chen and Wise, 1995) prediction error encoding (Schlag-Rey et al., 1997; Amador et al., 2000; So and Stuphorn, 2012), performance monitoring (Sajad et al., 2019) and decision-making (Abzug and Sommer, 2018). The general consensus is that SEF subserves various cognitive functions (Stuphorn et al., 2000; Tremblay et al., 2002) while also representing multiple spatial frames (Martinez-Trujillo et al., 2004; Stuphorn, 2015; Abzug and Sommer, 2017). It should be noted that these diverse signals (So and Stuphorn, 2012; Abzug and Sommer, 2018; Sajad et al., 2019) may be prominent in many of the spatially untuned neurons that were rejected in our analysis. However, the possibility of these signals influencing the spatially tuned response fields cannot be eliminated here (Purcell et al., 2012; Abzug and Sommer, 2017; Sajad et al., 2019). It can be reasonably hypothesized that spatially tuned neurons integrate non-spatial signals from untuned neurons (Pruszynski and Zylberberg, 2019) with spatial signals and forward these integrated signals to FEF neurons, thereby influencing gaze behavior in real space, thus providing a mechanism for SEF to implement executive function as behavior.
Egocentric transformations in the gaze system
In the gaze control system, the consensus is that eye-centered visual and motor codes predominate (Russo and Bruce, 1993; Tehovnik et al., 2000; Klier et al., 2001; Paré and Wurtz, 2001; Goldberg et al., 2002), but alternative views persist (Mueller and Fiehler, 2017; Caruso et al., 2018). Visual-motor dissociation tasks (e.g., antisaccades) found that visual and motor activities coded target and saccade direction, respectively (Everling and Munoz, 2000; Sato and Schall, 2003; Takeda and Funahashi, 2004). However, this requires additional training and signals that would not be present during ordinary visually guided saccades (Munoz and Everling, 2004; Medendorp et al., 2005; Amemori and Sawaguchi, 2006), and the head was fixed in most such studies. We have previously extended these results to natural head-unrestrained gaze shifts in the SC and FEF (Sadeh et al., 2015, 2020; Sajad et al., 2015) and here in the SEF.
Consistent with previous reports we found that SEF response fields are primarily organized in eye-centered coordinates (Russo and Bruce, 1993; Park et al., 2006), and participate in progressive target-to-gaze transition like FEF and SEF (Sajad et al., 2016; Sadeh et al., 2020). Note that in the current study, deviations of gaze from the target (used to fit response fields against G) were produced in part by the landmark shift. However, this alone does not likely explain the T-G transition in our cells, because it happened continuously through the task, it was spatially separable and often uncorrelated with the neural response to the landmark shift (discussed below in the section ‘A circuit model for allocentric/egocentric integration’), and much of the gaze errors used to calculate this transition were not because of the landmark shift, but appeared to be because of general internal “noise,” as in our previous studies (Sajad et al., 2016, 2020; Sadeh et al., 2020) which was even larger without a landmark (Li et al., 2017). In short, the landmark shift clearly contributed to gaze errors, but cannot alone explain the T-G transition we observed here in SEF cells.
Besides the general resemblance of the T-G transition in SC, FEF, and SEF, we were able to compare the latter two directly in the current experiments, and found no significant difference. This level of spatial simplicity and homogeneity across these gaze-related areas likely results from shared inputs and extensive interconnectivity, and perhaps serves as a common baseline signal to carry more subtle cognitive modulations (Munoz, 2002; Munoz and Everling, 2004; Schall, 2015). This does not preclude different influences on spatial behavior, because (1) these structures may carry other more subtle spatial signals that might be decoded downstream (Bulkin and Groh, 2006; Gandhi and Katnani, 2011), and (2) the influence of these signals on behavior depends on how they project to downstream motor structures, and through further modulations such as “gain fields” (Andersen and Buneo, 2002; Smith and Crawford, 2005; Blohm and Crawford, 2009). For example, microstimulation of the SC induces eye-fixed gaze shifts (Wurtz and Albano, 1980; Klier et al., 2001; Gandhi and Katnani, 2011), whereas microstimulation of frontal cortex produces gaze shifts toward a spectrum of eye, head, and body-centered goals (Schlag and Schlag-Rey, 1987; Martinez-Trujillo et al., 2004; Monteon et al., 2013; Sato and Schall, 2003; Abzug and Sommer, 2017), suggesting more complex motor transformations.
Integration of landmark-centered codes with egocentric codes
Previous humans studies suggest that egocentric and allocentric codes are initially separated in the visual system (Milner and Goodale, 2006; Schenk, 2006; Chen et al., 2014, 2018) and then reintegrated in parietofrontal cortex (Chen et al., 2018). Relatively few neurophysiological studies have aimed at discriminating egocentric/allocentric codes in visual and VM systems (Dean and Platt, 2006; Uchimura et al., 2017). Our recent study confirmed the role of the FEFs in integrating these codes for gaze (Bharmauria et al., 2020). Specifically, the landmark shift paradigm used here induced an initial transient T-T’ shift in delay activity, followed by a shift that was integrated with the egocentric T-G code in the motor burst.
In some ways our SEF and FEF findings were similar, i.e., we found an influence of the landmark shift on the more basic egocentric codes, first as a multiplexed but independent code, and finally as an integrated influence within the egocentric motor code. However, there were mechanistic differences between these structures: the landmark shift was coded initially in SEF M neurons (vs VM neurons in the FEF) and integration only occurred in the VM burst (vs all motor responses in FEF), suggesting a complementary mechanism between these areas.
In both the FEF and SEF, there was good agreement between the degree of landmark influence on the neural signals (measured either as a T-T’ shift or the final T-T’/T-G slope) and the actual gaze behavior, suggesting that these structures participate in the optimal integration described in previous studies, where “optimal” is defined as the best estimate of target direction based on statistical weighing of signal uncertainty (Körding and Wolpert, 2004; Byrne and Crawford, 2010; Karimpur et al., 2020). This is consistent with the general notion that the brain employs Bayesian methods of statistical learning, using probabilistic strategy from the task/target distribution and feedback uncertainty, to optimize performance (Körding and Wolpert, 2004; Byrne and Crawford, 2010; Mutluturk and Boduroglu, 2014; Aagten-Murphy and Bays, 2019; Karimpur et al., 2020).
Finally, at first glance, our findings contradict the classic findings of object-centered coding in the SEF (Olson and Gettner, 1995; Tremblay et al., 2002). However, there are important differences: besides being head-restrained, those studies involved explicit training and coding of one part of an object relative to other part, whereas ours involved untrained and implicit coding of a target relative to a background landmark. Taken together, the SEF plays a role in both the implicit and explicit use of allocentric cues for gaze coding.
A circuit model for allocentric/egocentric integration
Figure 13 provides a hypothetical circuit model for allocentric/egocentric integration in frontal cortex (Fig. 13), based on our current results, FEF recordings obtained in the same recording sessions (Bharmauria et al., 2020), and previous literature. First, our data show that the FEF and SEF are initially driven by low latency (∼80 ms) egocentric visual inputs (target-in-eye coordinates), whereas the landmark influence has a higher latency (∼200–300 ms), consistent with more complex visual processing pathway. Such signals are thought to arise in the ventral visual stream (Milner and Goodale, 2006; Schenk, 2006; Chen et al., 2014), which projects to parietal cortex (Milner, 2017; Budisavljevic et al., 2018). Consistent with this, saccade-related activity in human parietal cortex is modulated by landmarks (Chen and Crawford 2017). Direct inputs to the prefrontal gaze system include LIP (Andersen et al., 1990; Schall et al., 1993; Stuphorn, 2015), and nuclei in central thalamus that relay inputs from the SC and substantia nigra pars reticulata (SnPR; Lynch et al., 1994; Parent and Hazrati, 1995).
Second, we adopt the general convention that the SEF is involved more in executive control, whereas the FEF is more closely linked with eye control (Stuphorn et al., 2010; Stuphorn, 2015; Abzug and Sommer, 2017, 2018; Sajad et al., 2019). As discussed above in the section ‘General SEF function: spatial or non-spatial?’, the non-spatial aspects could be subserved by a progression from non-spatial SEF neurons to spatial SEF and hence FEF neurons. This directional flow seems to hold for egocentric/allocentric integration, because in our data integration was more complete in the FEF motor burst than in the SEF motor burst. In this context, an executive control mechanism might explain why different contexts influence allocentric weighting (Neggers et al., 2005; Byrne and Crawford, 2010; Fiehler et al., 2014; Klinghammer et al., 2017).
Based on the preceding assumptions, we speculate that the SEF provides control signals to the FEF for context-appropriate reference frame integration. In this scheme, the landmark-shift is first assessed in SEF preparatory activity (Fig. 11F) and then relayed (red arrow) to FEF delay activity (Bharmauria et al., 2020). In the FEF, the egocentric-allocentric conflict was multiplexed, but not integrated (no correlation between T-G and T-T’) in the earlier delay activity of VM neurons, however, an integration was observed in both the VM and M neurons in the final motor burst (Bharmauria et al., 2020). On the contrary, in the SEF, the multiplexed (non-integrated) signal first appeared in the delay activity of M neurons and later (motor burst) integration only occurs in the VM activity. Thus, in both areas, allocentric and egocentric signals are initially multiplexed in a decorrelated state within the reciprocal FEF-SEF loop (yellow arrows) but perhaps other working memory circuits are also involved (Christophel et al., 2017; Pinotsis et al., 2019). Finally, these signals become integrated in a path relayed from SEF VM neurons to a more complete integration in the FEF motor response (red arrow), consistent with the latter being closer to motor output (Isa and Sasaki, 2002). We also speculate that the FEF and SEF neuronal classes have opposing roles with strongly and closely wired memory-related inter-class circuitries for a final gaze command. Briefly, we speculate that the SEF influences the final FEF burst in a learned, task-related capacity, although we cannot show that here. Thus, the SEF relatively has more an “executive” control over the “motor” role of the FEF for goal-directed behavior.
This speculative model makes specific predictions for SEF-FEF spike correlation, i.e., (1) spatially tuned neurons should correlate across structures; (2) output neurons would correlate best with work-related activity in the opposite structure; (3) during the delay, SEF motor neurons should correlate best with FEF VM neurons; (4) during saccades, SEF VM neurons would correlate best with the FEF motor burst; and (5) neuron pairs that show temporal spike correlations will also show correlated spatial codes along the T-G/T-T’ continua. Since ∼80% of our SEF neurons were recorded in conjunction with FEF neurons during our experiments, this is a feasible goal for a future study.
General implications and conclusion
Previous studies have suggested that egocentric and allocentric visual cues are optimally integrated for goal directed action (Byrne and Crawford, 2010; Karimpur et al., 2020). Here, we emulated this behavior in the gaze system, and found that the SEF (like the FEF) is involved in an eye-centered transformation of target signals into gaze signals while incorporating landmark-centered information (Bharmauria et al., 2020). Taken together, these results suggest a neurophysiological model for optimal egocentric-allocentric integration in animals and humans. This is relevant for understanding normal function in daily (normal) and abnormal behavior, where brain damage preferentially affects egocentric versus allocentric mechanisms (Milner and Goodale, 2006; Schenk, 2006). Knowledge of their integration circuits, combined with neuroplasticity, might provide access to preserved visual functions through targeted rehabilitation strategies.
Synthesis
Reviewing Editor: Satu Palva, University of Helsinki
Decisions are customarily a result of the Reviewing Editor and the peer reviewers coming together and discussing their recommendations until a consensus is reached. When revisions are invited, a fact-based synthesis statement explaining their decision and outlining what is needed to prepare a revision will be listed below. Note: If this manuscript was transferred from JNeurosci and a decision was made to accept the manuscript without peer review, a brief statement to this effect will instead be what is listed below.
All concerns have been addressed
References
- Aagten-Murphy D, Bays PM (2019) Independent working memory resources for egocentric and allocentric spatial information. PLoS Comput Biol 15:e1006563. 10.1371/journal.pcbi.1006563 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Abzug ZM, Sommer MA (2017) Supplementary eye fields. In: Reference module in neuroscience and biobehavioral psychology. Amsterdam: Elsevier. [Google Scholar]
- Abzug ZM, Sommer MA (2018) Neuronal correlates of serial decision-making in the supplementary eye field. J Neurosci 38:7280–7292. 10.1523/JNEUROSCI.3643-17.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Amador N, Schlag-Rey M, Schlag J (2000) Reward-predicting and reward-detecting neuronal activity in the primate supplementary eye field. J Neurophysiol 84:2166–2170. 10.1152/jn.2000.84.4.2166 [DOI] [PubMed] [Google Scholar]
- Amemori KI, Sawaguchi T (2006) Rule-dependent shifting of sensorimotor representation in the primate prefrontal cortex. Eur J Neurosci 23:1895–1909. 10.1111/j.1460-9568.2006.04702.x [DOI] [PubMed] [Google Scholar]
- Anderse RA, Asanuma C, Essick G, Siegel RM (1990) Corticocortical connections of anatomically and physiologically defined subdivisions within the inferior parietal lobule. J Comp Neurol 296:65–113. [DOI] [PubMed] [Google Scholar]
- Andersen RA, Buneo CA (2002) Intentional Maps in Posterior Parietal Cortex. Annu Rev Neurosci 25:189–220. 10.1146/annurev.neuro.25.112701.142922 [DOI] [PubMed] [Google Scholar]
- Andersen RA, Essick GK, Siegel RM (1985) Encoding of spatial location by posterior parietal neurons. Science 230:456–458. 10.1126/science.4048942 [DOI] [PubMed] [Google Scholar]
- Ball K, Smith D, Ellison A, Schenk T (2009) Both egocentric and allocentric cues support spatial priming in visual search. Neuropsychologia 47:1585–1591. 10.1016/j.neuropsychologia.2008.11.017 [DOI] [PubMed] [Google Scholar]
- Bharmauria V, Bachatene L, Cattan S, Brodeur S, Chanauria N, Rouat J, Molotchnikoff S (2016) Network-selectivity and stimulus-discrimination in the primary visual cortex: cell-assembly dynamics. Eur J Neurosci 43:204–219. 10.1111/ejn.13101 [DOI] [PubMed] [Google Scholar]
- Bharmauria V, Sajad A, Li J, Yan X, Wang H, Crawford JD (2020) Integration of eye-centered and landmark-centered codes in frontal eye field gaze responses. Cereb Cortex 30:4995–5013. [DOI] [PubMed] [Google Scholar]
- Blohm G, Crawford JD (2009) Fields of gain in the brain. Neuron 64:598–600. 10.1016/j.neuron.2009.11.022 [DOI] [PubMed] [Google Scholar]
- Brandman DM, Cash SS, Hochberg LR (2017) Review: human intracortical recording and neural decoding for brain computer interfaces. IEEE Trans Neural Syst Rehabil Eng 25:1687–1696. 10.1109/TNSRE.2017.2677443 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bremmer F, Kaminiarz A, Klingenhoefer S, Churan J (2016) Decoding target distance and saccade amplitude from population activity in the macaque lateral intraparietal area (LIP). Front Integr Neurosci 10:30. 10.3389/fnint.2016.00030 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bridgeman B, Peery S, Anand S (1997) Interaction of cognitive and sensorimotor maps of visual space. Percept Psychophys 59:456–469. 10.3758/bf03211912 [DOI] [PubMed] [Google Scholar]
- Bruce CJ, Goldberg ME (1985) Primate frontal eye fields. I. Single neurons discharging before saccades. J Neurophysiol 53:603–635. 10.1152/jn.1985.53.3.603 [DOI] [PubMed] [Google Scholar]
- Bruce CJ, Goldberg ME, Bushnell MC, Stanton GB (1985) Primate frontal eye fields. II. Physiological and anatomical correlates of electrically evoked eye movements. J Neurophysiol 54:714–734. 10.1152/jn.1985.54.3.714 [DOI] [PubMed] [Google Scholar]
- Budisavljevic S, Dell’Acqua F, Castiello U (2018) Cross-talk connections underlying dorsal and ventral stream integration during hand actions. Cortex 103:224–239. [DOI] [PubMed] [Google Scholar]
- Bulkin DA, Groh JM (2006) Seeing sounds: visual and auditory interactions in the brain. Curr Opin Neurobiol 16:415–419. [DOI] [PubMed] [Google Scholar]
- Byrne PA, Crawford JD (2010) Cue reliability and a landmark stability heuristic determine relative weighting between egocentric and allocentric visual information in memory-guided reach. J Neurophysiol 103:3054–3069. 10.1152/jn.01008.2009 [DOI] [PubMed] [Google Scholar]
- Caruso VC, Pages DS, Sommer MA, Groh JM (2018) Beyond the labeled line: variation in visual reference frames from intraparietal cortex to frontal eye fields and the superior colliculus. J Neurophysiol 119:1411–1421. 10.1152/jn.00584.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chaplin TA, Hagan MA, Allitt BJ, Lui LL (2018) Neuronal correlations in MT and MST impair population decoding of opposite directions of random dot motion. eNeuro 5. 10.1523/ENEURO.0336-18.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chatham CH, Badre D (2015) Multiple gates on working memory. Curr Opin Behav Sci 1:23–31. 10.1016/j.cobeha.2014.08.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen LL, Wise SP (1995) Neuronal activity in the supplementary eye field during acquisition of conditional oculomotor associations. J Neurophysiol 73:1101–1121. 10.1152/jn.1995.73.3.1101 [DOI] [PubMed] [Google Scholar]
- Chen Y, Byrne P, Crawford JD (2011) Time course of allocentric decay, egocentric decay, and allocentric-to-egocentric conversion in memory-guided reach. Neuropsychologia 49:49–60. 10.1016/j.neuropsychologia.2010.10.031 [DOI] [PubMed] [Google Scholar]
- Chen Y, Monaco S, Byrne P, Yan X, Henriques DYP, Crawford JD (2014) Allocentric versus egocentric representation of remembered reach targets in human cortex. J Neurosci 34:12515–12526. 10.1523/JNEUROSCI.1445-14.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen Y, Crawford JD (2017) Landmark-Centered vs. Gaze-Centered Memory of Saccade Targets in the Human: An FMRI Study. Front Syst Neurosci 11:44. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen Y, Monaco S, Crawford JD (2018) Neural substrates for allocentric-to-egocentric conversion of remembered reach targets in humans. Eur J Neurosci 47:901–917. 10.1111/ejn.13885 [DOI] [PubMed] [Google Scholar]
- Christophel TB, Klink PC, Spitzer B, Roelfsema PR, Haynes J-D (2017) The distributed nature of working memory. Trends Cogn Sci 21:111–124. 10.1016/j.tics.2016.12.007 [DOI] [PubMed] [Google Scholar]
- Constantin AG, Wang H, Martinez-Trujillo JC, Crawford JD (2007) Frames of reference for gaze saccades evoked during stimulation of lateral intraparietal cortex. J Neurophysiol 98:696–709. 10.1152/jn.00206.2007 [DOI] [PubMed] [Google Scholar]
- Crawford JD, Guitton D (1997) Visual-motor transformations required for accurate and kinematically correct saccades. J Neurophysiol 78:1447–1467. 10.1152/jn.1997.78.3.1447 [DOI] [PubMed] [Google Scholar]
- Crawford JD, Ceylan MZ, Klier EM, Guitton D (1999) Three-dimensional eye-head coordination during gaze saccades in the primate. J Neurophysiol 81:1760–1782. 10.1152/jn.1999.81.4.1760 [DOI] [PubMed] [Google Scholar]
- Dean HL, Platt ML (2006) Allocentric spatial referencing of neuronal activity in macaque posterior cingulate cortex. J Neurosci 26:1117–1127. 10.1523/JNEUROSCI.2497-05.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- DeSouza JFX, Keith GP, Yan X, Blohm G, Wang H, Crawford JD (2011) Intrinsic reference frames of superior colliculus visuomotor receptive fields during head-unrestrained gaze shifts. J Neurosci 31:18313–18326. 10.1523/JNEUROSCI.0990-11.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ekstrom AD, Arnold AEGF, Iaria G (2014) A critical review of the allocentric spatial representation and its neural underpinnings: toward a network-based perspective. Front Hum Neurosci 8:803. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Everling S, Munoz DP (2000) Neuronal correlates for preparatory set associated with pro-saccades and anti-saccades in the primate frontal eye field. J Neurosci 20:387–400. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Everling S, Dorris MC, Klein RM, Munoz DP (1999) Role of primate superior colliculus in preparation and execution of anti-saccades and pro-saccades. J Neurosci 19:2740–2754. 10.1523/JNEUROSCI.19-07-02740.1999 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fiehler K, Wolf C, Klinghammer M, Blohm G (2014) Integration of egocentric and allocentric information during memory-guided reaching to images of a natural environment. Front Hum Neurosci 8:636. 10.3389/fnhum.2014.00636 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gandhi NJ, Katnani HA (2011) Motor functions of the superior colliculus. Annu Rev Neurosci 34:205–231. 10.1146/annurev-neuro-061010-113728 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Goldberg ME, Bisley J, Powell KD, Gottlieb J, Kusunoki M (2002) The role of the lateral intraparietal area of the monkey in the generation of saccades and visuospatial attention. Ann NY Acad Sci 956:205–215. 10.1111/j.1749-6632.2002.tb02820.x [DOI] [PubMed] [Google Scholar]
- Goodale MA, Haffenden A (1998) Frames of reference for perception and action in the human visual system. Neurosci Biobehav Rev 22:161–172. 10.1016/s0149-7634(97)00007-9 [DOI] [PubMed] [Google Scholar]
- Goris RLT, Movshon JA, Simoncelli EP (2014) Partitioning neuronal variability. Nat Neurosci 17:858–865. 10.1038/nn.3711 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Huerta MF, Kaas JH (1990) Supplementary eye field as defined by intracortical microstimulation: connections in macaques. J Comp Neurol 293:299–330. 10.1002/cne.902930211 [DOI] [PubMed] [Google Scholar]
- Isa T, Sasaki S (2002) Brainstem control of head movements during orienting; organization of the premotor circuits. Prog Neurobiol 66:205–241. 10.1016/s0301-0082(02)00006-0 [DOI] [PubMed] [Google Scholar]
- Karimpur H, Kurz J, Fiehler K (2020) The role of perception and action on the use of allocentric information in a large-scale virtual environment. Exp Brain Res 238:1813–1814. 10.1007/s00221-020-05839-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Keith GP, DeSouza JFX, Yan X, Wang H, Crawford JD (2009) A method for mapping response fields and determining intrinsic reference frames of single-unit activity: applied to 3D head-unrestrained gaze shifts. J Neurosci Methods 180:171–184. 10.1016/j.jneumeth.2009.03.004 [DOI] [PubMed] [Google Scholar]
- Khanna SB, Snyder AC, Smith MA (2019) Distinct sources of variability affect eye movement preparation. J Neurosci 39:4511–4526. 10.1523/JNEUROSCI.2329-18.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Klier EM, Wang H, Crawford JD (2001) The superior colliculus encodes gaze commands in retinal coordinates. Nat Neurosci 4:627–632. 10.1038/88450 [DOI] [PubMed] [Google Scholar]
- Klier EM, Wang H, Crawford JD (2003) Three-dimensional eye-head coordination is implemented downstream from the superior colliculus. J Neurophysiol 89:2839–2853. 10.1152/jn.00763.2002 [DOI] [PubMed] [Google Scholar]
- Klinghammer M, Blohm G, Fiehler K (2017) Scene configuration and object reliability affect the use of allocentric information for memory-guided reaching. Front Neurosci 11:204. 10.3389/fnins.2017.00204 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Körding KP, Wolpert DM (2004) Bayesian integration in sensorimotor learning. Nature 427:244–247. 10.1038/nature02169 [DOI] [PubMed] [Google Scholar]
- Leavitt ML, Pieper F, Sachs AJ, Martinez-Trujillo JC (2017) Correlated variability modifies working memory fidelity in primate prefrontal neuronal ensembles. Proc Natl Acad Sci USA 114:E2494–E2503. 10.1073/pnas.1619949114 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lemay M, Bertram CP, Stelmach GE (2004) Pointing to an allocentric and egocentric remembered target in younger and older adults. Exp Aging Res 30:391–406. 10.1080/03610730490484443 [DOI] [PubMed] [Google Scholar]
- Li J, Sajad A, Marino R, Yan X, Sun S, Wang H, Crawford JD (2017) Effect of allocentric landmarks on primate gaze behavior in a cue conflict task. J Vis 17:20. 10.1167/17.5.20 [DOI] [PubMed] [Google Scholar]
- Lynch JC, Hoover JE, Strick PL (1994) Input to the primate frontal eye field from the substantia nigra, superior colliculus, and dentate nucleus demonstrated by transneuronal transport. Exp Brain Res 100:181–186. 10.1167/17.5.20 [DOI] [PubMed] [Google Scholar]
- Martinez-Trujillo JC, Medendorp WP, Wang H, Crawford JD (2004) Frames of reference for eye-head gaze commands in primate supplementary eye fields. Neuron 44:1057–1066. 10.1016/j.neuron.2004.12.004 [DOI] [PubMed] [Google Scholar]
- Medendorp WP, Goltz HC, Vilis T (2005) Remapping the remembered target location for anti-saccades in human posterior parietal cortex. J Neurophysiol 94:734–740. 10.1152/jn.01331.2004 [DOI] [PubMed] [Google Scholar]
- Milner AD (2017) How do the two visual streams interact with each other? Exp Brain Res 235:1297–1308. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Milner D, Goodale M (2006) The visual brain in action. Oxford: Oxford University Press. [Google Scholar]
- Monteon JA, Wang H, Martinez-Trujillo J, Crawford JD (2013) Frames of reference for eye-head gaze shifts evoked during frontal eye field stimulation. Eur J Neurosci 37:1754–1765. 10.1111/ejn.12175 [DOI] [PubMed] [Google Scholar]
- Mueller S, Fiehler K (2017) Gaze-centered coding of proprioceptive reach targets after effector movement: testing the impact of online information, time of movement, and target distance. PLoS One 12:e0180782. 10.1371/journal.pone.0180782 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Munoz DP (2002) Commentary: saccadic eye movements: overview of neural circuitry. Prog Brain Res 140:89–96. 10.1016/S0079-6123(02)40044-1 [DOI] [PubMed] [Google Scholar]
- Munoz DP, Everling S (2004) Look away: the anti-saccade task and the voluntary control of eye movement. Nat Rev Neurosci 5:218–228. 10.1038/nrn1345 [DOI] [PubMed] [Google Scholar]
- Mutluturk A, Boduroglu A (2014) Effects of spatial configurations on the resolution of spatial working memory. Atten Percept Psychophys 76:2276–2285. 10.3758/s13414-014-0713-4 [DOI] [PubMed] [Google Scholar]
- Neely KA, Tessmer A, Binsted G, Heath M (2008) Goal-directed reaching: movement strategies influence the weighting of allocentric and egocentric visual cues. Exp Brain Res 186:375–384. 10.1007/s00221-007-1238-z [DOI] [PubMed] [Google Scholar]
- Neggers SFW, Schölvinck ML, van der Lubbe RHJ, Postma A (2005) Quantifying the interactions between allo- and egocentric representations of space. Acta Psychol (Amst) 118:25–45. 10.1016/j.actpsy.2004.10.002 [DOI] [PubMed] [Google Scholar]
- O'Keefe J (1976) Place units in the hippocampus of the freely moving rat. Exp Neurol 51:78–109. [DOI] [PubMed] [Google Scholar]
- O'Keefe J, Dostrovsky J (1971) The hippocampus as a spatial map. Preliminary evidence from unit activity in the freely-moving rat. Brain Res 34:171–175. 10.1016/0006-8993(71)90358-1 [DOI] [PubMed] [Google Scholar]
- Olson CR, Gettner SN (1995) Object-centered direction selectivity in the macaque supplementary eye field. Science 269:985–988. 10.1126/science.7638625 [DOI] [PubMed] [Google Scholar]
- Paré M, Wurtz RH (2001) Progression in neuronal processing for saccadic eye movements from parietal cortex area LIP to superior colliculus. J Neurophysiol 85:2545–2562. 10.1152/jn.2001.85.6.2545 [DOI] [PubMed] [Google Scholar]
- Parent A, Hazrati LJ (1995) Functional anatomy of the basal ganglia. I. The cortico-basal ganglia-thalamo-cortical loop. Brain Res Rev 20:91–127. [DOI] [PubMed] [Google Scholar]
- Park J, Schlag-Rey M, Schlag J (2006) Frames of reference for saccadic command tested by saccade collision in the supplementary eye field. J Neurophysiol 95:159–170. 10.1152/jn.00268.2005 [DOI] [PubMed] [Google Scholar]
- Pinotsis DA, Buschman TJ, Miller EK (2019) Working memory load modulates neuronal coupling. Cereb Cortex 29:1670–1681. 10.1093/cercor/bhy065 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pruszynski JA, Zylberberg J (2019) The language of the brain: real-world neural population codes. Curr Opin Neurobiol 58:30–36. 10.1016/j.conb.2019.06.005 [DOI] [PubMed] [Google Scholar]
- Purcell BA, Weigand PK, Schall JD (2012) Supplementary eye field during visual search: salience, cognitive control, and performance monitoring. J Neurosci 32:10273–10285. 10.1523/JNEUROSCI.6386-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rosenbaum RS, Ziegler M, Winocur G, Grady CL, Moscovitch M (2004) “I have often walked down this street before”: fMRI studies on the hippocampus and other structures during mental navigation of an old environment. Hippocampus 14:826–835. 10.1002/hipo.10218 [DOI] [PubMed] [Google Scholar]
- Russo GS, Bruce CJ (1993) Effect of eye position within the orbit on electrically elicited saccadic eye movements: a comparison of the macaque monkey’s frontal and supplementary eye fields. J Neurophysiol 69:800–818. 10.1152/jn.1993.69.3.800 [DOI] [PubMed] [Google Scholar]
- Sadeh M, Sajad A, Wang H, Yan X, Crawford JD (2015) Spatial transformations between superior colliculus visual and motor response fields during head-unrestrained gaze shifts. Eur J Neurosci 42:2934–2951. 10.1111/ejn.13093 [DOI] [PubMed] [Google Scholar]
- Sadeh M, Sajad A, Wang H, Yan X, Crawford JD (2020) Timing determines tuning: a rapid spatial transformation in superior colliculus neurons during reactive gaze shifts. eNeuro 7. 10.1523/ENEURO.0359-18.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sajad A, Sadeh M, Keith GP, Yan X, Wang H, Crawford JD (2015) Visual-motor transformations within frontal eye fields during head-unrestrained gaze shifts in the monkey. Cereb Cortex 25:3932–3952. 10.1093/cercor/bhu279 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sajad A, Sadeh M, Yan X, Wang H, Crawford JD (2016) Transition from target to gaze coding in primate frontal eye field during memory delay and memory-motor transformation. eNeuro 3. 10.1523/ENEURO.0040-16.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sajad A, Godlove DC, Schall JD (2019) Cortical microcircuitry of performance monitoring. Nat Neurosci 22:265–274. 10.1038/s41593-018-0309-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sajad A, Sadeh M, Crawford JD (2020) Spatiotemporal transformations for gaze control. Physiol Rep 8. 10.14814/phy2.14533 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sato TR, Schall JD (2003) Effects of stimulus-response compatibility on neural selection in frontal eye field. Neuron 38:637–648. 10.1016/s0896-6273(03)00237-x [DOI] [PubMed] [Google Scholar]
- Schall JD (1991) Neuronal activity related to visually guided saccades in the frontal eye fields of rhesus monkeys: comparison with supplementary eye fields. J Neurophysiol 66:559–579. 10.1152/jn.1991.66.2.559 [DOI] [PubMed] [Google Scholar]
- Schall JD (2015) Visuomotor functions in the frontal lobe. Annu Rev Vis Sci 1:469–498. 10.1146/annurev-vision-082114-035317 [DOI] [PubMed] [Google Scholar]
- Schall JD, Morel A, Kaas JH (1993) Topography of supplementary eye field afferents to frontal eye field in macaque: implications for mapping between saccade coordinate systems. Vis Neurosci 10:385–393. 10.1017/s0952523800003771 [DOI] [PubMed] [Google Scholar]
- Schall JD, Hanes DP, Thompson KG, King DJ (1995) Saccade target selection in frontal eye field of macaque. I. Visual and premovement activation. J Neurosci 15:6905–6918. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schenk T (2006) An allocentric rather than perceptual deficit in patient D.F. Nat Neurosci 9:1369–1370. 10.1038/nn1784 [DOI] [PubMed] [Google Scholar]
- Schlag J, Schlag-Rey M (1987) Evidence for a supplementary eye field. J Neurophysiol 57:179–200. 10.1152/jn.1987.57.1.179 [DOI] [PubMed] [Google Scholar]
- Schlag-Rey M, Amador N, Sanchez H, Schlag J (1997) Antisaccade performance predicted by neuronal activity in the supplementary eye field. Nature 390:398–401. 10.1038/37114 [DOI] [PubMed] [Google Scholar]
- Smith MA, Crawford JD (2005) Distributed population mechanism for the 3-D oculomotor reference frame transformation. J Neurophysiol 93:1742–1761. 10.1152/jn.00306.2004 [DOI] [PubMed] [Google Scholar]
- So N, Stuphorn V (2012) Supplementary eye field encodes reward prediction error. J Neurosci 32:2950–2963. 10.1523/JNEUROSCI.4419-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stuphorn V (2015) The role of supplementary eye field in goal-directed behavior. J Physiol Paris 109:118–128. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stuphorn V, Taylor TL, Schall JD (2000) Performance monitoring by the supplementary eye field. Nature 408:857–860. 10.1038/35048576 [DOI] [PubMed] [Google Scholar]
- Stuphorn V, Brown JW, Schall JD (2010) Role of supplementary eye field in saccade initiation: executive, not direct, control. J Neurophysiol 103:801–816. 10.1152/jn.00221.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Takeda K, Funahashi S (2004) Population vector analysis of primate prefrontal activity during spatial working memory. Cereb Cortex 14:1328–1339. 10.1093/cercor/bhh093 [DOI] [PubMed] [Google Scholar]
- Tehovnik EJ, Sommer MA, Chou IH, Slocum WM, Schiller PH (2000) Eye fields in the frontal lobes of primates. Brain Res Rev 32:413–448. [DOI] [PubMed] [Google Scholar]
- Tremblay L, Gettner SN, Olson CR (2002) Neurons with object-centered spatial selectivity in macaque SEF: do they represent locations or rules? J Neurophysiol 87:333–350. 10.1152/jn.00356.2001 [DOI] [PubMed] [Google Scholar]
- Abstr.Uchimura M, Kumano H, Kitazawa S (2017) Rapid allocentric coding in the monkey precuneus. 2017 Annual Meeting. Soc Neurosci 589.24/GG19. [Google Scholar]
- Wurtz RH, Albano JE (1980) Visual-motor function of the primate superior colliculus. Annu Rev Neurosci 3:189–226. 10.1146/annurev.ne.03.030180.001201 [DOI] [PubMed] [Google Scholar]
- Zylberberg J (2018) The role of untuned neurons in sensory information coding. bioRxiv. doi: https://doi.org/10.1101/134379. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.