Significance
When making decisions, humans and other primates move their eyes freely to gather information about their environment. A large literature has explored the factors that determine where the eyes fall during natural scene perception and visual search, concluding that deviant or surprising perceptual information attracts attention and gaze. Here, we describe a sampling policy that paradoxically shows that in a decision-making task, the eyes are attracted to expected rather than unexpected information: When classifying the average of an array of numbers, human observers looked by preference at the numbers that were closest to the mean. This policy drove a behavioral tendency to discount the influence of outliers when making choices, leading to “robust” choices about the stimulus array.
Keywords: decision-making, categorization, eye movements, numerical averaging, information sampling
Abstract
Humans move their eyes to gather information about the visual world. However, saccadic sampling has largely been explored in paradigms that involve searching for a lone target in a cluttered array or natural scene. Here, we investigated the policy that humans use to overtly sample information in a perceptual decision task that required information from across multiple spatial locations to be combined. Participants viewed a spatial array of numbers and judged whether the average was greater or smaller than a reference value. Participants preferentially sampled items that were less diagnostic of the correct answer (“inlying” elements; that is, elements closer to the reference value). This preference to sample inlying items was linked to decisions, enhancing the tendency to give more weight to inlying elements in the final choice (“robust averaging”). These findings contrast with a large body of evidence indicating that gaze is directed preferentially to deviant information during natural scene viewing and visual search, and suggest that humans may sample information “robustly” with their eyes during perceptual decision-making.
A well-validated computational framework argues that decisions are made by integrating information toward a threshold, at which point a response is made (1, 2). Within this integration-to-bound framework, decisions are optimized by sequential sampling, at the cost of delaying responding (3). Studies using this framework in conjunction with psychophysical tasks have suggested that humans and monkeys make approximately optimal perceptual category judgments, with choice accuracy limited mainly by sensory noise or uncertainty in the external world (2, 4, 5). These psychophysical studies have been argued to offer a window on the general mechanisms by which humans make decisions (6).
However, paradigms in which stimulation is continuously available and controlled exclusively by the experimenter differ from everyday choices in real-world environments. For example, in the random dot motion (RDM) paradigm, subjects passively sample from a single, continuous, centrally presented stream of information and decide whether average dot motion tends in one direction or another (7). By contrast, decisions about the average feature in a natural scene are made by actively saccading (or covertly attending) to successive image locations and sampling the subset of information that is available there. For example, a foraging monkey might move its eyes freely to several locations in the forest canopy and average the available fruit yield to decide where to move. Real-world category decisions, thus, are affected not only by the quality of the sensory information and the dynamics of integration, but also by the policy that dictates how information is sampled and weighted when forming a decision. However, the nature of this policy remains largely unknown.
The determinants of saccadic eye movements have been extensively investigated during visual search (8) and natural scene perception (9). In visual search tasks, eye movements are attracted to image locations that contain perceptually salient or surprising items (10, 11) or to those points at which task-relevant information is maximal (12–15). In value-based choice tasks, in which for example participants judge which of two or more food items is preferred, choices and eye movements are interlinked, in that items that are fixated more frequently are chosen more often (16, 17). Other models have suggested that more uncertain sources are sampled more thoroughly with the eyes (18) or that decisions and information-gathering mutually drive one another in a reciprocal cascade (19). However, little is known about the policy by which participants sample information during choice tasks that involve integrating decision information from multiple locations across the visual field. Various studies have investigated the sampling policy in differing domains, such as category learning (20) and classification tasks (21). Recently, researchers have highlighted the importance of investigating evidence accumulation as an active sampling process (22). We investigated this sampling process by measuring eye movements while human participants performed a categorization task that involved averaging numbers in a spatial array. We used fully visible symbolic numbers as stimuli, allowing us to assess how gaze placement was determined by the statistics of decision information, rather than the perceptual arrangement of the stimulus array.
When averaging perceptual features across multiple locations under central fixation, participants tend to downweight or ignore extreme values, such as a deviant expression in a crowd of faces (23) or outlying color values in an array of shapes (24). This has led to the proposal that “inlying” items (those that are closer to the category boundary) are processed with higher gain than outliers (24, 25). However, a related possibility is that observers tend to sample by preference those items that are closer to the category boundary. Testing this theory, we found humans prefer to fixate the inlying items in a stimulus array during a perceptual decision-making task—a policy that contradicts the well-established preference for fixating deviant items in visual search tasks. This “robust” sampling policy drove decisions to ignore or discount outlying information during perceptual decisions.
Results
Participants (n = 54) viewed eight two-digit numbers distributed quasirandomly around the screen (Fig. 1A). Numbers were drawn from a distribution with mean µ and SD σ, and participants were asked to compare their average to a reference value R that varied across blocks (µ = R ± 2 or R ± 4 in low and high µ trials, respectively, and σ = 6 or 12 in low and high σ trials, respectively). Half of the participants indicated whether the average was higher (or not) than the reference; the others indicated whether the average was lower (or not). Numbers were interspersed among task-irrelevant items (letters), which introduced a gentle crowding effect and encouraged shifts of gaze during decision formation. By tracking gaze during task performance, we were able to ask how the information sampled with the eyes related to decisions, with a focus on whether participants sampled by preference those numbers that were inlying (close to reference) or outlying (far from reference).
We report data collapsed over two experiments that differed only in terms of the response deadline (3,000 ms and 5,000 ms in experiment 1 and experiment 2, respectively). Overall mean accuracy was 70.5% (n = 54). Performance depended on both the mean and variance of the distribution from which the numbers were drawn (Fig. 1B for accuracy and Table S1 for response time). Accuracy increased on high µ trials compared with low µ trials [F(1, 53) = 76, P < 0.001] and decreased on high σ (more variable evidence) trials compared with low σ trials [F(1, 53) = 47, P < 0.001], as previously reported (24).
Table S1.
Experiment | σ | High µ | Low µ | Effect on RT | Significance |
1 | High σ | 1.57 ± 0.27 s | 1.61 ± 0.27 s | µ ↑ → RT ↓ | F(1, 25) = 31, P < 0.001 |
Low σ | 1.62 ± 0.27 s | 1.67 ± 0.28 s | σ↑ → RT ↓ | F(1, 25) = 33, P < 0.001 | |
2 | High σ | 2.05 ± 0.55 s | 2.1 ± 0.60 s | µ ↑ → RT ↓ | F(1, 27) = 19, P < 0.001 |
Low σ | 2.11 ± 0.58 s | 2.18 ± 0.62 s | σ↑ → RT ↓ | F(1, 27) = 25, P < 0.001 |
Mean reaction times (RTs) and SD are shown per condition. RTs decreased on high µ trials compared with low µ trials and RTs decreased on high σ trials compared with low σ trials, which is in contrast with previous studies of feature-averaging using color and shape (25). No significant interaction effects were found for either accuracy or RT.
Gaze Trajectories.
We developed an approach to analyzing eye-tracking data (“landscape analysis”) that measured the decision information sampled by the eyes as a continuous, time-varying quantity (the “evidence trajectory,” or ET) across each trial (Fig. S1 and SI Methods for a detailed description of the analysis method). We first converted the stimulus array on each trial to a smooth topography (or landscape) of decision values, with peaks or troughs centered on the spatial position of each element and a height and sign that was determined by the difference between each element Vk and the reference R (positive peaks favoring the decision “higher than reference,” negative troughs favoring “lower”; Fig. 1C). The momentary evidence at each point in time was defined as the landscape value under the gaze location, and the ET constituted the time-series of values obtained as the eyes moved across the stimulus array on each trial (Fig. 1D and SI Methods). This is equivalent to conceiving of the gaze as a “soft” (or smooth) information-harvesting spotlight moving across the screen, allowing evidence to be accumulated up to a decision threshold. The width of the spotlight was estimated independently of further analyses under the assumption that participants’ gaze maximized the acquisition of total sensory evidence. We fit this value by searching for values that maximized “quality trajectory” (QT)—that is, the extent to which items were viewed irrespective of their task-relevant values (SI Methods). We also validated all analyses using a “hard” spotlight—that is, an aperture of fixed width with no spatial smoothing, obtaining qualitatively identical results (Fig. S2).
Decisions Depend on Gaze Location.
Using this approach, we first compared cumulative ET on trials where participants responded “higher” and those where they responded lower, correcting for differences in available information on “high” and “low” trials (SI Methods). As expected, choices depended strongly on the decision information sampled, with average cumulative ET building up steadily until the response: participants responded lower when the sampled evidence was lower than the reference and higher when the sampled evidence was higher than the reference (Fig. 1D, Bottom). In other words, consistent with previous findings, choices were strongly affected by where participants looked (16).
Gaze Is Directed Preferentially to Inliers Over Outliers.
Next, we used a similar landscape approach to determine the extent to which participants’ gaze favored items that were inlying or outlying (numerically closer or further from the reference). We generated a new set of landscapes in which the topography on each trial was determined by the rank of the absolute value of each element (number) within the stimulus array, so that more inlying elements had higher values and more outlying elements had lower values (Fig. 2A; similar results were obtained when this analysis was carried out using the absolute values instead of the ranks). The integral of the eye movement trajectory under these landscapes, which we call the “prototypicality trajectory” (PT), thus encoded participants’ time-varying preference to fixate those elements that were numerically similar (inliers; denoted by positive values) or dissimilar (outliers; negative values) to the reference. We found that inlying items were sampled preferentially, from 400 to 800 ms poststimulus presentation, indicated by a significantly positive PT value (Fig. 2B, black line and shading). This result survived correction for multiple comparisons across time points using a well-validated cluster-based approach (26).
Next, we evaluated this preference for inliers using a different approach. On each fixation, we defined the item being viewed as that which was closest to the current gaze location, allowing us to calculate the probability of gaze moving from an item of one rank (i.e., ordinal position in the array) to another. To avoid any biases that may arise from nonindependence of nearby items, we constructed a null distribution specific to our experiment by randomly permuting item-location assignments on each trial and repeating the analysis 1,000 times. Next, we computed the difference between gaze-transition probabilities in our dataset and under the null distribution. Finally, we calculated the probability for a single fixation to go to an inlier/outlier, split for short, medium, and long Euclidean distance in pixels on the screen between the currently and subsequently fixated items (Fig. 3). The main effect of preference to sample inliers (sampling policy) was significant [F(1, 53) = 8.58, P < 0.01]. This effect was largely driven by fixations of medium distance [t(53) = 2.74, P < 0.01] in comparison with short and long distances (both P > 0.05). The main effect of distance [F(1.9, 100.67) = 1.71, P > 0.05] and its interaction with sampling policy [F(1.89, 100.38) = 1.168, P > 0.05] were not significant.
These findings obtained using discrete fixations rather than the landscape approach demonstrate that observers exhibited a tendency to shift their gaze toward inliers rather than merely dwelling for longer while viewing these items. Unsurprisingly, these effects diminished for long Euclidean distances between previous and current fixations, as the decision information associated with distant items cannot be evaluated through parafoveal processing. Thus, the data suggest that participants’ parafoveal processing is not unbounded but occurs within a fixed range, and the sampling bias (moving to inlying information) is only possible for items falling within this range.
Humans Show Robust Averaging of Number.
Next, we investigated whether participants’ decisions show evidence of robust averaging—that is, a decision policy in which outliers carry less weight than inliers. We used probit regression to predict human choices as a function of (i) the numbers associated with each item (ranked from lowest to highest), (ii) the time spent sampling those numbers, and (iii) the interaction between these two factors, all entered as competitive predictors in the predictor matrix (Fig. 4 A–C and Methods). The weighting function plotted in Fig. 4A shows the main effect of stimulus—that is, the weight given to each number as a function of its rank within the numerical array. The inverted-u shape indicates that outliers are downweighted. This observation is qualified by a statistical comparison between the weight given to inlying ranks [3, 4, 5, 6] and outlying ranks [1, 2, 7, 8], t(53) = 3.72, P < 0.001. This finding of robust averaging with numbers complements earlier reports with color, shape, and facial expression, confirming that outliers carry less influence during perceptual averaging across domains and task settings (24, 25).
Linking Sampling Policy and Human Decisions.
Next, we show how the main effect of sampled information influenced choices, finding that the resulting estimates grew with rank: sampling low numbers was associated with a propensity to respond low, and sampling high numbers was associated with responding high (Fig. 4B). This is expected from the ET analyses above (Fig. 1D), which show that the information sampled was strongly predictive of the response (high vs. low). Critically, the third predictor (interaction term) in the regression allowed us to assess how the weight given to each stimulus rank was modulated by gaze duration, over and above the influence of stimulus value. Ranks where this term differs positively from zero are those where directing gaze to the relevant numbers amplifies their weight multiplicatively. For example, if gaze were merely a filter that acted equally for all numbers irrespective of their rank, we would see weights that are equal for all ranks. However, this is not what is observed (Fig. 4C). The time spent sampling inlying items enhances the effect that they will have on choice (compared with outlying items) above and beyond what would be predicted simply by their values alone [inlying vs. outlying: t(53) = 5.10, P < 0.001]. We note that this finding is not simply due to a logarithmic (Weber-like) compression of the number line, because it occurred both for those participants that were asked whether the average was higher than the reference [t(28) = 4.99, P < 0.001] and those that reported whether the average was lower than the reference [t(24) = 2.18, P = 0.039], a framing variable that we counterbalanced over participants (Methods). This result also persisted when we used the average (rather than total) ET calculated over the trial, suggesting that it is not simply due to different sampling policies being deployed on trials with short and long RTs.
Given that the prototypicality analysis (above) showed that inlying items were preferentially viewed 400–800 ms poststimulus, we examined whether this downweighting, due to preferential sampling, is also most apparent in this time period. To do this, we conducted the previous regression for each sampling time point individually and tested which of these resulting coefficients (for the Gaze × Stimulus interaction) were significantly different from zero. In Fig. 4D, these are plotted (t values > 3) on a heat map, showing that this downweighting of outliers effect is also maximally depended on gaze during this 400–800 ms timeframe. These results suggest that robust averaging is due at least in part to the sampling policy with which participants actively sample inlying rather than “outlying” information.
Together, Fig. 4 A and C show that (i) inlying stimuli (i.e., those close to the reference) have a stronger influence on choice than outlying stimuli (i.e., those that are much larger or much smaller than the reference) and (ii) that this effect is multiplicatively enhanced by gaze. Thus, any tendency that inlying stimuli might have had to drive choices more strongly is exaggerated when they are fixated, but this is not true for outlying stimuli. In other words, human observers are not only “robust averagers” but also “robust samplers”—selectively viewing inlying category information and thereby ensuring that this information is a more potent driver of choice. In further control analyses, we ruled out the possibility that this effect was due to a “counting” strategy, in which all items above or below the reference contributed equally to decisions (Fig. S3).
Computational Simulations.
Next, to validate our approach, we used computational simulations to estimate how the ET and PT should vary under differing sampling policies. Using distributions of decision information (i.e., numbers) identical to those viewed by human participants (i.e., same µ and σ), we constructed simulated ET and PT vectors consisting of alternating periods of fixation and eye movements, prefixed by a delay period. The duration of all of these periods (initial delay, fixation, eye movement) was sampled from the distributions exhibited by human participants. During simulated fixations, simulated ET was set to the decision values Vk – R (where Vk is the numerical value of item k), and during delay and eye movements, ET values were set to zero. Simulated PT was constructed in the same manner by setting PT during a fixation of stimulus k to its value ranked within the array, as for the human analyses. Model choices occurred when the cumulative simulated ET reached a fixed threshold, as in standard integration-to-bound models (2). The sign of the cumulative ET at this fixed threshold determined model choice and thus accuracy. To ensure that human PT and model PT values were on the same scale, we multiplied the model PT with the average QT for the corresponding participant.
This framework allowed us to manipulate the model sampling policy—that is, to simulate the order in which the array items were sampled. We implemented this via a single parameter π that dictated whether the observer preferred to view inliers before outliers (π > 0), outliers before inliers (π < 0), or sampled the elements in random order (π ∼ 0). The parameter π encodes the (positive or negative) correlation between the observed sampling order and the most “prototypical” ordering—that is, from most inlying to most outlying. In Fig. 2C, we plot simulated PT for different values of π. The best fitting parameterization was determined per subject by calculating the mean squared error (MSE) between human and simulated PT for each value of π. Fig. 2B shows the simulated values of PT for the best fitting average π value over subjects (magenta line) plotted against human data. Human PT curves were best accounted for by positive π values, consistent with a policy that involved sampling inliers before outliers and thus with the analyses described above. Fig. S4 shows the distribution of best fitting values for π over subjects. Next, using the best fitting π values for each subject, we returned to the analysis of ET and compared the human and simulated ET values for the different conditions—that is, for different means µ and SDs σ of the decision values with respect to the reference. We observed a striking convergence between the human and model ET values (Fig. S5), providing an independent validation of our best fitting model.
Determinants of Human and Model Performance.
Finally, we tested how model performance (accuracy) varied with π. Interestingly, performance for the model linearly decreased as the sampling policy π changed from a policy to sample outlying information first to a policy to sample inlying information (Fig. 2D). Thus, for an “ideal” model—in which decisions are only limited by the quality of information that is sampled from the screen—a policy that preferentially samples inlying information incurs a model performance cost relative to random sampling. To test whether this was the case for human observers, we conducted an additional analysis that evaluated how estimated human values of π predicted accuracy at the single-trial level. We computed single-trial estimates of π by correlating the order in which items were actually sampled with their ranked equivalent and then used logistic regression to estimate how these values predicted choice accuracy (Fig. S6). Both the sampling policy π [t(53) = 2.37, P < 0.05] and the average prototypicality [t(53) = 11.3, P < 0.001] were positive predictors of accuracy, indicating that unlike the model, participants performed better as they exhibited more robust sampling.
SI Methods
We exhaustively computed the QT across a space of possible parameter values (range 0–100) for s and identified the value of s that maximized QT. In other words, we made the assumption that participants move their eyes to locations that maximize available information (QT), irrespective of any decision values (ET or PT). Note that an excessively narrow spotlight would lead to items even close to the point of gaze being neglected, whereas (because landscapes were normalized to have a mean of zero) an excessively broad spotlight would lead to large portions of uninformative space being sampled, reducing the overall level of information (QT) obtained. The aperture size s that maximizes QT is the one that maximizes this trade-off. The resulting function was found to be concave with a peak at 50 pixels.
The QT-maximizing SD of the Gaussian for the entire cohort (50 pixels) was used for subsequent analyses. This approach thus defines the width of the information-processing spotlight within the same task but in a way that is independent of all key analyses described in the paper, which related to the decision values (numbers relative to the reference) rather than the position of the items on the screen. We opted to develop this specific ad hoc computation of the width of the information-harvesting spotlight for two reasons. First, this approach assumes that information falling further from the point of gaze has a weaker influence on choices than that falling directly under a fixation point rather than assuming that all information within a fixed window contributes equally to choices. Second, recent work has suggested that sampling policies are not always well predicted by mapping retinal sensitivity based on independent data from a task with different constraints (38).
In a subsequent step, we used a variant of this method to obtain real-time measures of the decision information available at each point in the trial, a quantity we call ET. To this end, we set the height of the Gaussian gk associated with each item k equal to Vk – R, where Vk is the value (number) of each item and R is the reference value. As a consequence, peaks in the landscape are located at numbers higher than the reference value, and troughs are located at numbers lower than the reference value (Fig. 1C). The amplitude of peaks (red) and troughs (blue) depended linearly on the difference between each number and the reference point. These values were inverted for participants in the lower framing group, such that positive deflections in each landscape were always in the frame of reference of the choice, not the absolute number. Time point-by-time point measures of decision information are then obtained by defining gaze trajectory as the temporal sequence of landscape values falling under the observer’s gaze at each time point t. We averaged the resulting ET curves across trials and conditions (for each chosen response) to assess how decision information is sampled across time during trials and how it relates to behavior.
The approach we use here has some drawbacks. Notably, we assume that all information falling under the gaze trajectory contributes to the decision, in direct proportion to its viewing time, an assumption that is probably incorrect; it is known that participants are blind to information arriving at the retina during saccades (39). However, using a standard analysis approach, in which we defined fixations using Eyelink preset criteria (velocity threshold of 30, acceleration threshold of 8,000, and a saccade motion threshold of 0.1), saccadic trajectories accounted for only 18.75% of time points, and as these were largely made across the space separating stimuli, they contributed only 0.00004% to the final decision value, so any bias that arises by accounting for inputs from between fixations is likely to be negligible. However, our approach has the major benefit that it allows us to characterize, in real time on each trial, information that is available to the participant and thus to explore how stimulation and gaze interact in determining choices. It is also immune to otherwise arbitrary choices that might have to be made during a more standard fixation analysis, such as the size and shape of the window that surrounds each stimulus or the criteria for determining dwell times. Moreover, we provide multiple validation analyses that show that our approach offers sensible information about how people move their eyes and how this determines choices.
Nevertheless, to verify that our findings were not in some way specific to our use of this “landscape” approach, we repeated the analysis using a more conventional method of defining an aperture with fixed width. In this “hard” filter approach, the landscape is composed of a series of circular “cylinders” placed at the x, y position of each item k, with height ck corresponding to the quantityVk – R and width s, where s is a free parameter fit to the QT as for standard landscape analysis. The QT values are computed in the same way as for the “soft” filter approach—that is, as the integral under the eye movement trace. The two methods thus differ in that here, any time the gaze trajectory falls within the space defined by a cylinder, the ET value is incremented by ck, rather than depending on its distance to the x, y position. All of the results were replicated using a hard information-harvesting spotlight (i.e., a spotlight of fixed width with no spatial smoothing), and the ETs under both approaches were qualitatively identical (Fig. S2).
Having developed and tested the landscape approach, we then used it to assess our main question of interest: Do participants sample preferentially from those numbers that are inlying or outlying with respect to the reference? To this end, we defined a new quantity, protypicality or p, that encoded how inlying or outlying each sample was:
where Pk is calculated by computing the ranks for each item k (tied ranks are set to their average rank) and subtracting 4.5 from this value. Thus, Pk is more positive for inlying items and more negative for outlying items.
Next, we set the height of the Gaussian gk associated with each item k equal to this prototypicality Pk. As a consequence, peaks in the landscape are located at more inlying numbers, and troughs are located at more outlying numbers. Time point-by-time point measures of PT are then obtained as above, by defining gaze trajectory as the temporal sequence of landscape values falling under the observer’s gaze at each time point t. For the results shown in Fig. 2, we averaged the resulting PT curves across trials and conditions (up to each chosen response) to assess the prototypicality of the sampling process across the trial and how this relates to behavior.
To assess statistical significance (i.e., deviation from zero of PT) in a manner that corrected for multiple comparisons across time points, we used a nonparametric cluster-correction method. First, we constructed a null distribution of cluster sizes by sampling randomly from the t distribution at each time point and logging the maximum cluster size (i.e., greatest number of contiguous significant time points) for each of 1,000 random samples. We then report only significant (P < 0.05) clusters with a length that fell within the top 50 (5%) of this null distribution (26).
Discussion
In real-world settings, information arises in crowded natural scenes, replete with multiple features and objects that compete to drive decisions. Studies of saccadic control have revealed that attention is automatically captured by perceptually salient information, such as a target defined by a unique color in a visual search experiment (27), and may also be captured by semantically deviant information, such as an object in an unusual context (e.g., a sofa on the beach), although the latter claim remains controversial (11). However, the task at hand is a key determinant of saccadic control, and the eyes are drawn to those scene elements that are task-relevant, or valuable, or to those locations where the most information can be gleaned (10, 12–15, 28). Nevertheless, most decision-making studies either use a single, centrally presented source of information (such as a dot motion stimulus) precluding the detailed investigation of the policy by which participants sample decision information or assume that gaze allocation is a stochastic process that drives a preference for fixated alternatives, rather than vice versa (16, 17). Recently, researchers have highlighted the importance of investigating evidence accumulation as an active sampling process (22).
Here, we reveal an aspect of the sampling policy that determines where gaze is allocated during decision-making: faced with multiple spatially distinct sources of decision information, observers tend to ignore those sources that carry the most extreme information and focus instead on the inlying decision information—that is, that falling closest to the boundary that segregates one choice from another (or close to the mode of the overall information presented on one trial; our approach was not designed to distinguish these possibilities). This is surprising given that participants exhibit a well-characterized policy to saccade toward perceptually deviant information during natural scene viewing, or visual search (10, 11). The focus on inlying elements influenced human choices, with gaze driving a multiplicative effect to weight inlying elements more heavily, which in turn drove “robust averaging” in perceptual choices made during the experiment.
Robust averaging of perceptual information has been described before, for faces (23) and for colors and shapes (24, 25). Here, we add to these findings, revealing that humans engage in robust averaging of symbolic number. Importantly, we assumed that participants used an approximate number sense, meaning that numbers are treated as noisy estimates of magnitude, which can be averaged like continuous sensory signals (29). This assumption follows from the short duration for which numerical information was available and the low average dwell times observed for each number, which would preclude overt arithmetic approaches. Previous studies have left unresolved the question of whether robust averaging occurs because inlying items are sampled (i.e., overtly attended) by preference or whether the influence of outliers is attenuated at a later decision stage. For example, one model proposes that observers simply compute a posterior likelihood ratio conditional on the past distribution of features and their associated responses. This is equivalent to using a nonlinear decoder that “squashes” outlying information, muting its influence in the average and prolonging response latencies when decision information is heterogeneous (24). However, here we show that the robust averaging is driven at least in part by the policy with which gaze is directed during information integration. In other words, humans are not just robust averagers but robust samplers of decision information.
Why do humans sample and average robustly? We describe a computational simulation of the sampling policy adopted during our spatial averaging task that was able to faithfully recreate the relationship between the evidence sampled and the choices made by humans. Exploring model performance under a range of policies (spanning a tendency to sample inliers or outliers by preference) revealed that for an ideal model limited only by the information available on the screen, sampling outliers rather than inliers by preference boosted performance. This is consistent with the observation that when assigning samples to one of two Gaussian-distributed classes, eccentric or “off-channel” features are most diagnostic of the correct response (30, 31). It is very different, however, from the optimal policy that emerges as participants gradually learn a category boundary through repeated sampling (32) (Fig. S7 for discussion).
However, in the human data, a tendency to sample inliers was a positive predictor of accuracy at the single-trial level (Fig. S6). This implies that a robust sampling policy has evolved to counter additional sources of loss that arise during the decision process, such as “late” noise that corrupts the information of integration, perhaps due to capacity limits on attention and working memory processes. For example, robust policies may lead to more stable estimates of decision information, just as nonparametric statistical inference is robust to the inclusion of aberrant data points. Other recent work has suggested that apparent “suboptimalities” in human judgment may in fact maximize performance when this late noise is taken into account (33, 34). The effect of the interplay between the decision environment (i.e., task characteristics) and cognitive constraints on the relative (sub)optimality of various sampling policies is an interesting avenue for future research.
In summary, the current results highlight that perceptual judgments are not only constrained by the quality of the sensory information and the dynamics of integration but also by the sampling policy, which dictates how the information is selected for entry into the decision process. However, other questions remain open. In particular, the information processing steps by which fixations are guided to subsequent inliers remain unclear in our work, although this presumably relies on parafoveal processing of nonfixated locations. One possibility is that covert attention is first shifted to candidate items and the eyes are then “pulled” to this location (35, 36). Nor does our study directly address how saccade placement varies according to the spatial arrangement of items on the screen. For example, participants may prefer to saccade toward the center of mass of clusters of objects (37). Thus, gaze can indeed be driven by the spatial arrangement of items on the screen, but gaze is additionally driven by the decision values of the available evidence, exhibited in this work via a preference to sample inlying samples.
Methods
Participants.
Sixty healthy human participants with normal or corrected-to-normal vision were recruited in two experiments (experiment 1, n = 30; experiment 2, n = 30). Participants provided written consent beforehand, and the study was approved by the Oxford University Medical Sciences Division Ethics Committee (approval no. MSD-IDREC-C1-2009-1). There were only minor differences between the two experiments (highlighted below), and so we collapse across their results. Six participants whose performance failed to differ from chance were excluded from all further analyses, leaving n = 54.
Stimuli.
On each trial, participants viewed an array composed of eight two-digit numbers. Stimuli were generated using PsychToolbox (psychtoolbox.org/) for MATLAB (Mathworks) and presented on a 17-inch screen (resolution, 1024 × 768) viewed from a distance of 60 cm. Numbers (∼1.15° horizontal and ∼0.95° vertical visual angle) were presented at randomly selected locations within an invisible 8 × 8 grid. Eight additional randomly selected grid locations contained letter pairs, which were distracters. No letter or number pairs were presented within the central four grid locations. An example stimulus array can be seen in Fig. 1A.
At the start of each block of 96 trials, participants were informed of the reference value, against which they compared the average number for all trials across that block. Subsequently, each trial began with the onset of a central fixation cross (500 ms), followed by the stimulus array, which remained on the screen for either up to 3,000 ms (experiment 1) or up to 5,000 ms (experiment 2). During array presentation, participants freely moved their eyes across the screen and chose which was higher (lower): the reference number or the average of the number array. The decision framing (decide whether numbers higher vs. lower than reference) was counterbalanced across participants within each experiment. Participants could respond at any point during the array presentation by pressing one of two keys (“F” or “J”), and they received auditory feedback in the form of a high or low tone lasting 100 ms (800/400 Hz, correct/incorrect). In experiment 2 only, participants were additionally paid in proportion to their accuracy. In the high frame, participants received a bonus corresponding to their choice (reference vs. numbers, in pence) on 20 randomly selected trials. In the low frame, the reward of participants started at £20, from which their choice was deducted on 20 randomly selected trials. Participants earned £10 on average for their participation. Participants performed eight blocks of 96 trials (768 trials in total) in an experiment lasting ∼50 min.
Design.
The reference number R on each block was drawn randomly from a uniform distribution between 25 and 75. The sample numbers were drawn from one of eight Gaussian distributions whose mean µ differed from R by either ±2 or ±4 and whose SD σ was either 6 or 12. Sample values were resampled to ensure that the sample mean and variance fell within the desired value for that trial with a tolerance of 0.7 and 0.1, respectively. Each of the eight trial types occurred equally often, and their presentation order within a block was random.
Eye-Tracking.
All eye movement data were recorded monocularly using an SR Research EyeLink 1000 eye-tracking system at a sampling rate of 1,000 Hz. Participants placed their heads on a chin rest that was mounted on the table at a fixed distance of 68 cm from the screen. Before the start of each new block, a standard calibration/validation procedure was executed, in which participants tracked a dot that appeared at 13 evenly spaced locations across the screen. Data were converted to vectors of x, y position in the frame of reference of the screen exported to MATLAB using the EyeLink toolbox and analyzed using in-house scripts. Missing data (due to blinks or eye movements away from the screen) were ignored (i.e., set to NaN) in all analyses. Eye movement traces were cut off at reaction time for all analyses described above. All of our analyses assume that the eye-tracker faithfully estimates the pixel position of the gaze on the screen.
Analysis.
We analyzed the eye-tracking data using both a combination of conventional methods (e.g., identifying the location and dwell time of individual fixations) and a technique that treated the stimulus array as a topography of decision information that was sampled continuously with the gaze (landscape analysis). We describe this method in detail in SI Methods (Fig. S1). The methods associated with other analyses are detailed in accompanying supplementary sections that are referenced in the main text (Figs. S3, S6, and S7).
Weighting of Decision Information on Choice.
Probit regression was used to predict human choices (higher or lower than reference) as a function of three sets of predictors: (i) the decision values of each number ranked from lowest to highest, (ii) QT for each number QTk (once again, with k ranked from lowest to highest), and (iii) the interaction between these two quantities. The QT for each number QTk was computed as the sum of the integral under an element-specific QT landscape—that is, a landscape constructed using each element in isolation (with the height of the Gaussian set to 1)—up to the response time (Fig. S1). QTk is then proportional to the time spent sampling element k.
The regression model we used was as follows, where in each case:
where ϕ denotes the cumulative normal distribution, k the sample number, and t the time point of the QT vector. For the data shown in Fig. 4 A–C, we dropped the indexing by t and used the sum of QT for all time points up to the response.
Acknowledgments
This work was supported by European Research Council Starter Grant 281628 (to C.S.), Economic and Social Research Council Funding Award 1369566 (to H.V.), and Wellcome Trust Grant 0099741/Z/12/Z (to S.H.C.).
Footnotes
The authors declare no conflict of interest.
This article is a PNAS Direct Submission. J.G. is a Guest Editor invited by the Editorial Board.
This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1613950114/-/DCSupplemental.
References
- 1.Gold JI, Shadlen MN. The neural basis of decision making. Annu Rev Neurosci. 2007;30:535–574. doi: 10.1146/annurev.neuro.29.051605.113038. [DOI] [PubMed] [Google Scholar]
- 2.Ratcliff R, McKoon G. The diffusion decision model: Theory and data for two-choice decision tasks. Neural Comput. 2008;20(4):873–922. doi: 10.1162/neco.2008.12-06-420. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Bogacz R, Brown E, Moehlis J, Holmes P, Cohen JD. The physics of optimal decision making: A formal analysis of models of performance in two-alternative forced-choice tasks. Psychol Rev. 2006;113(4):700–765. doi: 10.1037/0033-295X.113.4.700. [DOI] [PubMed] [Google Scholar]
- 4.Gold JI, Shadlen MN. Neural computations that underlie decisions about sensory stimuli. Trends Cogn Sci. 2001;5(1):10–16. doi: 10.1016/s1364-6613(00)01567-9. [DOI] [PubMed] [Google Scholar]
- 5.Beck JM, et al. Probabilistic population codes for Bayesian decision making. Neuron. 2008;60(6):1142–1152. doi: 10.1016/j.neuron.2008.09.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Shadlen MN, Kiani R. Decision making as a window on cognition. Neuron. 2013;80(3):791–806. doi: 10.1016/j.neuron.2013.10.047. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Newsome WT, Paré EB. A selective impairment of motion perception following lesions of the middle temporal visual area (MT) J Neurosci. 1988;8(6):2201–2211. doi: 10.1523/JNEUROSCI.08-06-02201.1988. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Eckstein MP. Visual search: A retrospective. J Vis. 2011;11(5):14. doi: 10.1167/11.5.14. [DOI] [PubMed] [Google Scholar]
- 9.Torralba A, Oliva A, Castelhano MS, Henderson JM. Contextual guidance of eye movements and attention in real-world scenes: The role of global features in object search. Psychol Rev. 2006;113(4):766–786. doi: 10.1037/0033-295X.113.4.766. [DOI] [PubMed] [Google Scholar]
- 10.Towal RB, Mormann M, Koch C. Simultaneous modeling of visual saliency and value computation improves predictions of economic choice. Proc Natl Acad Sci USA. 2013;110(40):E3858–E3867. doi: 10.1073/pnas.1304429110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Võ ML-H, Henderson JM. The time course of initial scene processing for eye movement guidance in natural scene search. J Vis. 2010;10(3):1–13. doi: 10.1167/10.3.14. [DOI] [PubMed] [Google Scholar]
- 12.Itti L, Baldi P. Bayesian surprise attracts human attention. Vision Res. 2009;49(10):1295–1306. doi: 10.1016/j.visres.2008.09.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Eckstein MP, Schoonveld W, Zhang S, Mack SC, Akbas E. Optimal and human eye movements to clustered low value cues to increase decision rewards during search. Vision Res. 2015;113(Pt B):137–154. doi: 10.1016/j.visres.2015.05.016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Navalpakkam V, Koch C, Rangel A, Perona P. Optimal reward harvesting in complex perceptual environments. Proc Natl Acad Sci USA. 2010;107(11):5232–5237. doi: 10.1073/pnas.0911972107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Manohar SG, Husain M. Attention as foraging for information and value. Front Hum Neurosci. 2013;7:711. doi: 10.3389/fnhum.2013.00711. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Krajbich I, Armel C, Rangel A. Visual fixations and the computation and comparison of value in simple choice. Nat Neurosci. 2010;13(10):1292–1298. doi: 10.1038/nn.2635. [DOI] [PubMed] [Google Scholar]
- 17.Krajbich I, Rangel A. Multialternative drift-diffusion model predicts the relationship between visual fixations and choice in value-based decisions. Proc Natl Acad Sci USA. 2011;108(33):13852–13857. doi: 10.1073/pnas.1101328108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Cassey TC, Evens DR, Bogacz R, Marshall JA, Ludwig CJ. Adaptive sampling of information in perceptual decision-making. PLoS One. 2013;8(11):e78993. doi: 10.1371/journal.pone.0078993. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Shimojo S, Simion C, Shimojo E, Scheier C. Gaze bias both reflects and influences preference. Nat Neurosci. 2003;6(12):1317–1322. doi: 10.1038/nn1150. [DOI] [PubMed] [Google Scholar]
- 20.Rehder B, Hoffman AB. Eyetracking and selective attention in category learning. Cognit Psychol. 2005;51(1):1–41. doi: 10.1016/j.cogpsych.2004.11.001. [DOI] [PubMed] [Google Scholar]
- 21.Meier KM, Blair MR. Waiting and weighting: Information sampling is a balance between efficiency and error-reduction. Cognition. 2013;126(2):319–325. doi: 10.1016/j.cognition.2012.09.014. [DOI] [PubMed] [Google Scholar]
- 22.Gottlieb J, Hayhoe M, Hikosaka O, Rangel A. Attention, reward, and information seeking. J Neurosci. 2014;34(46):15497–15504. doi: 10.1523/JNEUROSCI.3270-14.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Haberman J, Whitney D. The visual system discounts emotional deviants when extracting average expression. Atten Percept Psychophys. 2010;72(7):1825–1838. doi: 10.3758/APP.72.7.1825. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.de Gardelle V, Summerfield C. Robust averaging during perceptual judgment. Proc Natl Acad Sci USA. 2011;108(32):13341–13346. doi: 10.1073/pnas.1104517108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Michael E, de Gardelle V, Nevado-Holgado A, Summerfield C. Unreliable evidence: 2 sources of uncertainty during perceptual choice. Cereb Cortex. 2015;25(4):937–947. doi: 10.1093/cercor/bht287. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Maris E, Oostenveld R. Nonparametric statistical testing of EEG- and MEG-data. J Neurosci Methods. 2007;164(1):177–190. doi: 10.1016/j.jneumeth.2007.03.024. [DOI] [PubMed] [Google Scholar]
- 27.Yarbus AL. Eye Movements During Perception of Complex Objects. Springer; New York: 1967. [Google Scholar]
- 28.Najemnik J, Geisler WS. Eye movement statistics in humans are consistent with an optimal search strategy. J Vis. 2008;8(3):1–14. doi: 10.1167/8.3.4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Brezis N, Bronfman ZZ, Usher M. Adaptive spontaneous transitions between two mechanisms of numerical averaging. Sci Rep. 2015;5:10415. doi: 10.1038/srep10415. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Navalpakkam V, Itti L. Search goal tunes visual features optimally. Neuron. 2007;53(4):605–617. doi: 10.1016/j.neuron.2007.01.018. [DOI] [PubMed] [Google Scholar]
- 31.Scolari M, Serences JT. Adaptive allocation of attentional gain. J Neurosci. 2009;29(38):11933–11942. doi: 10.1523/JNEUROSCI.5642-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Markant DB, Settles B, Gureckis TM. Self-directed learning favors local, rather than global, uncertainty. Cogn Sci. 2016;40(1):100–120. doi: 10.1111/cogs.12220. [DOI] [PubMed] [Google Scholar]
- 33.Tsetsos K, et al. Economic irrationality is optimal during noisy decision making. Proc Natl Acad Sci USA. 2016;113(11):3102–3107. doi: 10.1073/pnas.1519157113. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Howes A, Warren PA, Farmer G, El-Deredy W, Lewis RL. Why contextual preference reversals maximize expected value. Psychol Rev. 2016;123(4):368–391. doi: 10.1037/a0039996. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Inhoff AW, Rayner K. Parafoveal word processing during eye fixations in reading: Effects of word frequency. Percept Psychophys. 1986;40(6):431–439. doi: 10.3758/bf03208203. [DOI] [PubMed] [Google Scholar]
- 36.Henderson JM, Ferreira F. Effects of foveal processing difficulty on the perceptual span in reading: Implications for attention and eye movement control. J Exp Psychol Learn Mem Cogn. 1990;16(3):417–429. doi: 10.1037//0278-7393.16.3.417. [DOI] [PubMed] [Google Scholar]
- 37.He PY, Kowler E. The role of location probability in the programming of saccades: Implications for “center-of-gravity” tendencies. Vision Res. 1989;29(9):1165–1181. doi: 10.1016/0042-6989(89)90063-1. [DOI] [PubMed] [Google Scholar]
- 38.Morvan C, Maloney LT. Human visual search does not maximize the post-saccadic probability of identifying targets. PLOS Comput Biol. 2012;8(2):e1002342. doi: 10.1371/journal.pcbi.1002342. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Dodge R. Visual perception during eye movement. Psychol Rev. 1900;7(5):454. [Google Scholar]
- 40.Cheadle S, et al. Adaptive gain control during human perceptual choice. Neuron. 2014;81(6):1429–1441. doi: 10.1016/j.neuron.2014.01.020. [DOI] [PMC free article] [PubMed] [Google Scholar]