Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2021 Aug 16;17(8):e1009190. doi: 10.1371/journal.pcbi.1009190

An uncertainty-based model of the effects of fixation on choice

Zhi-Wei Li 1, Wei Ji Ma 1,*
Editor: Samuel J Gershman2
PMCID: PMC8389845  PMID: 34398884

Abstract

When people view a consumable item for a longer amount of time, they choose it more frequently; this also seems to be the direction of causality. The leading model of this effect is a drift-diffusion model with a fixation-based attentional bias. Here, we propose an explicitly Bayesian account for the same data. This account is based on the notion that the brain builds a posterior belief over the value of an item in the same way it would over a sensory variable. As the agent gathers evidence about the item from sensory observations and from retrieved memories, the posterior distribution narrows. We further postulate that the utility of an item is a weighted sum of the posterior mean and the negative posterior standard deviation, with the latter accounting for risk aversion. Fixating for longer can increase or decrease the posterior mean, but will inevitably lower the posterior standard deviation. This model fits the data better than the original attentional drift-diffusion model but worse than a variant with a collapsing bound. We discuss the often overlooked technical challenges in fitting models simultaneously to choice and response time data in the absence of an analytical expression. Our results hopefully contribute to emerging accounts of valuation as an inference process.

Author summary

When people look longer at a food item, they tend to like it more. We propose a new theory in which this occurs because looking gathers information that reduces uncertainty, and people are uncertainty averse. We turn the theory into a mathematical model and fit it to previously published data. It fits better than the leading model, although we also find that the leading model can be improved.

Introduction

Eye fixations affect choices in value-based decision-making. This was originally demonstrated in a consumer decision-making task by Krajbich and colleagues [1]. Their experiment consisted of two phases. In the rating phase, subjects rated 70 snack food items on a scale from -10 to 10. In the choice phase, subjects chose between two previously rated items. The authors found that subjects more often chose the item that they fixated on for longer, even when the subjective ratings of the two items were equal. There is also evidence that the fixations directly cause the choice bias, instead of an underlying preference causing both [2, 3] (also see [4] for moral decisions and [5] for perceptual decisions; [6] for a review on causality between attention and choice).

To quantitatively describe the decision-making process, Krajbich and colleagues introduced the attentional drift-diffusion model (aDDM). This model predicts choices and reaction times based on the subjective ratings of the two items and the sequence of fixation times. Like the traditional DDM [7], the aDDM assumes a hypothetical decision variable that drifts noisily towards a bound corresponding to the item with higher subjective rating. Here, however, the drift is accelerated for the fixated item. A choice is made when the decision variable reaches one of two bounds (one for each item). The aDDM accounts for choice and response time data not only in binary choice [1], but also in ternary choice [8] and in purchasing choices [9].

The aDDM does not express choice behavior as the result of the maximization of a utility function. As such, it is somewhat disjoint from many other models in behavioral economics. To bridge this gap, we build on recent work in psychology that has formulated utility functions in value-based decision-making in terms of subjective beliefs [1012]. In computer science, this notion is already much older, appearing for example in Bayesian Q-learning [13]. In these models, the agent maintains a continuum of hypotheses about value (or item attractiveness) and computes a Bayesian posterior distribution over value, which reflects the degree of belief in each value on the continuum; the studies differ in how the posterior is subsequently used. Intriguingly, dopamine neurons also seem to encode a distribution over future rewards [14].

These previous studies introduced the notion of probabilistic inference of value, but did not compare the resulting models to the aDDM in terms of their ability to fit data. Here, we propose a new value inference model that uses the posterior distribution as the basis of a utility function with an “uncertainty aversion” component. We fitted this model to the joint choice and response time data from Krajbich et al. (2010) [1], and formally compared our model against the aDDM. To anticipate our results: we cannot confidently conclude that our model fits better than aDDM, but we show that it is at least competitive, while arguably being more principled. We also point out overlooked technical intricacies in fitting both models, which might be of interest for future studies.

Results

Posterior-utility-choice (PUC) model

Background: Posterior uncertainty about value

We conceptualize the value of an item as the amount of future satisfaction resulting from consuming the item. An agent typically has incomplete or imperfect information about future satisfaction. For example, when choosing a food item, the saltiness, fat content, texture, etc. of the item are only approximately known, and so is the extent to which these properties will match the agent’s personal preferences.

In the perception literature, incomplete and imperfect knowledge is most often modeled using the framework of Bayesian inference [1518]. In this framework, the observer uses the sensory observations to build a subjective belief about a state of the world, as captured by a posterior probability distribution p(world state|observations). Recent work has proposed that value, despite being highly subjective and individualistic, may be inferred in much the same way as a world state in perception [1012]; we use this notion as the basis of our model. In the Krajbich task, the agent can try to infer value from physical cues, such as product brand and information on the packaging, as well as from memories of previous experiences with the category [19]. Combining sensory and memory cues with knowledge of one’s internal state will allow the agent to form a belief about the value of the item. This belief can be expressed as a posterior probability distribution p(value|cues).

Based on this posterior distribution, posterior uncertainty can be defined as a summary statistic, for example as the standard deviation. Intuitively, when information is accumulated for a longer time, posterior uncertainty should generally decrease. Posterior uncertainty is different from classic risk, such as one when an agent chooses between two lotteries, e.g. $5 for sure, or a 50% chance of receiving $10 [20]. In such an experiment, the outcome is uncertain solely due to the stochastic step that follows the agent’s choice. By contrast, posterior uncertainty can exist in a fully deterministic world. The distinction between posterior uncertainty and risk is an instance of the distinction between epistemic and aleatoric uncertainty, respectively [21]. We illustrate this distinction and how we view it in the full decision-making process in Fig 1.

Fig 1. Types of uncertainty in risky choice.

Fig 1

(A) In lotteries, uncertainty arises due to stochasticity in the mapping from decision to outcome. This is a form of aleatoric uncertainty. (B) Uncertainty can also arise from information about a choice item being incomplete or imperfect. In a probabilistic framework, the result of inferring the unknown value of a future outcome is captured by a posterior distribution, the width of which is a measure of uncertainty. This is a form of epistemic uncertainty.

Model overview

The basic premises of our model are as follows: a) fixating on an item will reduce posterior uncertainty about future satisfaction, either through the acquisition of visual information, or by triggering memory recall; b) lower uncertainty leads to higher utility in an uncertainty-averse agent. By combining these mechanisms, the model can in principle account for the effect of fixation on choice: longer fixation on an item leads to lower uncertainty, which leads to higher utility, which leads to a higher probability of choosing the item. We will now turn these premises into a concrete mathematical model, comment on the relations with other models, fit the model to Krajbich’ data, and compare the fit with the fit of the aDDM.

We first specify the model mathematically. In the model, when viewing a food item, the agent is trying to predict the value (future satisfaction) derived from consuming the item. On a given trial, the decision process consists of three steps:

  1. the computation of a posterior distribution over value (with associated posterior uncertainty);

  2. converting a given posterior distribution over value into a scalar utility;

  3. converting utility to choice.

Accordingly, we will refer to the model as the Posterior-Utility-Choice (PUC) model. The PUC model is partially normative because it computes a Bayesian posterior in Step 1 and a utility function in Step 2. However, the form of the utility function is ad-hoc, and unlike in recent work by Tajima et al. [10], Step 3 is not normative.

Generative model

Before we can describe the decision model, we need to specify the generative model, which describes the statistical origins of the agent’s data. In PUC, step 1 (posterior computation) is directly based on the generative model.

We denote the state resulting from consuming a food item by s. This high-dimensional state could comprise many factors, including satiety, nutritional state, and health status. The agent, however, does not have direct access to s, but has to form a beliefs about s from the available data D, which include observations of the properties of the item (size, ingredients, brand, etc.), interoceptive data about one’s own homeostatic state [22], and memories of consuming similar items [19]. We assume that in a limited time, only a limited amount of imperfect data can be collected. The mapping from s to D is stochastic and can be described by a probability distribution p(D|s).

Next, we assume that the state s will produce—in a deterministic or stochastic fashion—a value v, representing an amount of future satisfaction. We model this mapping as a probability distribution p(v|s). The graphical representation of the generative model is then Dsv, where s generates both D and v.

Decision model

We now use the generative model to specify the agent’s decision-making process. This process consists of three steps (P-U-C): from the sensory and memory data to a posterior distribution (Step 1), from a posterior distribution to utility (Step 2), and from utility to choice (Step 3). We now discuss each step.

  • Step 1: From data to posterior distribution over value

    Given specific observed data Dobs, the agent entertains a range of hypotheses about value. The likelihood of hypothesized value v, denoted by L(v; Dobs), is now the probability that Dobs were produced by a state of value v:
    L(v;Dobs)=p(Dobs|v)=p(Dobs|s)p(s|v)ds.
    Here, Dobs and s are both high-dimensional, whereas v is one-dimensional. We assume that the effect of viewing longer is that more data are gathered. Unfortunately, we know neither p(Dobs|s) nor p(s|v) since the researchers cannot observe the complete interoceptive state s. Therefore, we make a simplification, where we assume that the likelihood over v is Gaussian with mean x and standard deviation σ, which we will assume to be the same for all items:
    L(v;Dobs)N(v;x(Dobs),σ(Dobs)2).
    Here, x is the maximum-likelihood estimate of v, the best guess one could make about v based on Dobs. By analogy with perception, we will refer to x as a measurement, and a consistent generative model for x would be p(x|v)=N(x;v,σ2), where v is the true value. This simplification is largely analogous to the mapping from a neural population model to a behavioral model [23]. As a proxy for the true value of consuming an item, we take the rating given by the subject for that item in the rating phase of the experiment.
    We next assume a Gaussian prior p(v)=N(v;μp,σp2). Including the prior, the posterior over v becomes
    p(v|Dobs)L(v;Dobs)p(v), (1)
    which we approximate by
    p(v|x)p(x|v)p(v). (2)
    The accumulation of evidence is gated by fixations. Moreover, the longer the agent fixates on an item, the more measurements are made. Within a trial, the number of measurements can differ between the two items due to unequal fixation times. For each item, the posterior distribution starts as the prior distribution. As the agent gathers measurements by fixating on an item, the posterior distribution for that item updates iteratively:
    p(v|x1,,xT+1)p(xT+1|v)p(v|x1,,xT). (3)
    Qualitatively, this updating has two effects: first, the posterior mean E[v|x1,,xT] will move away from the mean of the prior, μp, toward the true value, v; second, the variance of v under the posterior distribution will decrease (Fig 2, step 1).
Fig 2. Posterior-Utility-Choice model.

Fig 2

The PUC model describes how an agent maps noisy measurements of value to a decision variable. Top: Flow diagram of the model. Bottom: Components of the model. Step 1: The agent computes a posterior distribution over hypothesized value. As viewing time increases (darker colors), the posterior distribution shifts from the prior towards the true value, and becomes narrower. Step 2: Utility incorporates both the mean and the standard deviation of the posterior over value. Both higher mean and lower standard deviation are preferred. Utility is evaluated separately for each item. Step 3: Evolution of the decision variable on an example trial. L and R denote fixations on the left or the right item. The decision variable, DV, is the utility difference of the two items. A decision is made when DV crosses the collapsing bound (dashed).

  • Step 2: From posterior distribution over value to utility

    Now that the agent has a belief over value as expressed in the posterior distribution, they have to turn this belief into a utility. The simplest way is simply to take a central tendency of the posterior (mean, median, or mode of value under the posterior). But this might discard too much information. In principle, any mapping from the posterior to a number can serve to compute the utility of an item. Since the posterior is a function, utility is then a function of a function, and we will denote it by U[p(v|x)].

    Furthermore, we define utility as a weighted average of the mean and the standard deviation of value under its posterior distribution based on a sequence of measurements x = (x1, …, xT) (Fig 2B):
    U[p(v|x)]=μposterior-A·σposterior, (4)
    where A is a constant that we will call the “uncertainty aversion parameter”, and
    μposterior=E[v|x] (5)
    σposterior=SD[v|x] (6)
    The standard deviation term can be motivated in at least three, not mutually exclusive ways. First, it is possible that the proper definition of utility is U=E[v|Dobs], i.e. the expected value of the posterior based on Dobs. We made the approximation that the likelihood over v is Gaussian, but if it is not, an error will be introduced by instead using E[v|x]. The standard deviation term could partially compensate for this error, in a way similar to [24]. Second, even when the likelihood is exactly Gaussian, the agent might be uncertainty-averse. The uncertainty arises from incomplete and imperfect data (i.e. stochasticity in the measurements), rather than from post-decision stochasticity, as in classic risk aversion. Standard models of risky choice such as in Fig 1A are a specific case of our model. In the classic risk models, the future state s is typically the state of having received a specific monetary amount, and no data need to be collected to know the distribution of s. Moreover, the mapping from s to v is typically assumed deterministic and monotonic, say v = F(s). Then Eq (1) for the posterior over v reduces to p(v) = ∫δ(vF(s))p(s)ds (all of this can be conditioned on an action). Thus, our model generalizes standard models of risky choice by replacing the distribution p(v) by a posterior distribution p(v|Dobs). A term similar to the standard deviation term has been used in portfolio theory [25]. Third, we are inspired by the literature on the so-called mere exposure effect [26], the phenomenon that mere exposure to a stimulus (for example a consumer item or a work of art) increases the observer’s preference for this stimulus. A leading explanation of the mere exposure effect is uncertainty reduction [27, 28], the idea that people prefer stimuli that are familiar.
    Under our Gaussian assumptions for likelihood and prior, the mean and standard deviation of the posterior become
    μposterior=μpσp2+Tx¯σ21σp2+Tσ2 (7)
    σposterior=11σp2+Tσ2, (8)
    where x¯ is the mean of the measurements (x1, …, xT). Increasing viewing time will change utility in two ways: by moving the posterior mean from μp to the true value v, and by decreasing the posterior standard deviation. The former of these changes can be negative or positive depending on the prior estimation of value and the specific value of the current item, the latter is always positive. In Fig 2, step 2 we show two examples of how one item can have higher utility than the other.
  • Step 3: From utility to choice

    The final step is to map utility to choice. We posit that the agent’s decision variable, denoted by DV, is the difference between the utilities of the two items:
    DV=Uleft-Uright. (9)
    Finally, the agent terminates the decision process when DV crosses a decision bound. The agent will then choose the item with the higher utility (Fig 2, step 3). We choose the decision bound to be decreasing over time (“collapsing”). A general motivation for using a collapsing bound instead of a fixed one is to prevent the model from predicting unrealistically long reaction times when deciding between two very similar items [2932]. For the specific form of collapsing bound, we use a special case of the Weibull function suggested by [30]:
    Bt=B0e-(tλ)k. (10)
    An example of the evolution of the decision variable and the decision bound is shown in Fig 2, step 3).

    In addition, we allow for non-decision (or residual) time τ. This means that the decision process may end before the fixation series has ended, and the remaining time would then not contribute to the decision, regardless of how many fixations occur in that time; in practice, the estimated non-decision time is so short that fixation rarely switches in this time. Our non-decision time is conceptually different from the one in [8, 9]: theirs can be interpreted as accounting for the transitions between fixations, while ours is part of the total fixation time and interpreted as the time spent looking at items after the decision has been made.

    Finally, we add a guessing rate parameter g. For each trial, there is a probability g of making a random decision at a random moment (based on the empirical decision time distribution fitted by a Weibull function, details see S2 Text), then the rest probability of 1 − l making decision based on the model prediction. This is necessary to avoid a zero probability for a smooth likelihood function landscape that allows parameter fitting.

Attentional drift-diffusion model

Krajbich and colleagues [1, 8, 9, 33] have proposed the attentional drift diffusion model (aDDM), which conceptualizes the consequences of fixation as biasing the drift velocity of the fixated item. The decision variable is defined as:

DVt=DVt-1+d(rleft-θrright)+ϵt (11)

when fixating on the left item, and

DVt=DVt-1+d(θrleft-rright)+ϵt (12)

when fixating on the right item, where d is a scaling constant, rleft and rright are the ratings for the two items, and ϵt is diffusion noise, drawn independently across time points from a normal distribution N(0,σ2). aDDM differs from the standard drift-diffusion model [7] in the attentional bias factor θ, which takes values between 0 and 1. Diffusion continues until the DV hits one of two boundaries, which are assumed symmetric with respect to 0. In the original work by Krajbich and colleagues [1], the boundaries were fixed over time. As usual, there is an arbitrary scaling in aDDM which allows us to set B0 = 1; in PUC, this is not possible because the scale is already set by v, which we approximate by the subject’s rating of the item.

aDDM with collapsing bounds

We also consider a more flexible variant of the aDDM, namely one that has the same parametric family of collapsing bounds as the PUC model (Eq (10)). We call the resulting model the attentional collapsing-bound drift-diffusion model (acbDDM), by analogy to the non-attentional version, which has been called cbDDM, with “cb” standing for “collapsing bound”. Milosavlevic et al. [31] previously considered a collapsing bound that is a special case of ours, namely with k = 1. In principle, we could also allow for an increasing bound by changing the parameter ranges of the boundary function. We limit ourselves to a collapsing bound here, following previous work and because it is psychologically easier to interpret.

Differences between PUC model and aDDM

The PUC Model and the a(cb)DDM have mechanistic similarities: in both models, the agent decides when the decision variable crosses a bound, allowing the model to make predictions for the relation between choice and total fixation time. However, the PUC model differs conceptually from the a(cb)DDM in the following aspects: (a) the PUC agent chooses the item with the highest utility, whereas the a(cb)DDM does not have an immediate interpretation in terms of utility. (This stands in contrast to the basic DDM model, in which the decision variable can be interpreted as the difference between the values of the two items [34]); (b) in the PUC, noise is specifically associated with the agent’s observed sensory information and retrieved memories, whereas the origin of noise in the a(cb)DDM is not well specified; (c) In the PUC model, later measurements have a smaller effect on the decision variable than earlier ones, because all measurements are generated from the same distribution and there are diminishing returns to information collection as the estimated value approaches the underlying true value. By contrast, in the a(cb)DDM, the variance of the noise added at each time point stays the same across time; (d) in the PUC model, two main mechanisms influence the preference: the uncertainty reduction term is independent of the item rating, indicating an additive effect of attention, whereas the posterior mean update depends on the difference between the prior mean value and the specific item value, thus indicating a multiplicative effect of attention. In contrast, a(cb)DDM will always boost the item with higher value more, thus the influence of attention is always multiplicative. For more studies regarding the additive versus multiplicative nature of attention, see [3537].

In addition, on the surface, it seems that in the PUC model, the agent keeps two distinct value distributions, one for each item, whereas in the a(cb)DDM, the information of the two items are combined into a single “relative decision variable”. However, in the later extension of aDDM model where more than two options are being compared ([8]), separate accumulators for each item were used and and it could be reduced to the aDDM with two items. Thus, the a(cb)DDM can also be conceptualized as two distinct accumulators, followed by another step of deriving the relative value difference, which is similar to the PUC model.

To model the effects of fixation on choice, we introduced the Posterior-Utility-Choice (PUC) model, in which the agent judges the value of an item by (1) accumulating evidence, gated by fixations; (2) computing and updating two posterior probability distributions over value, one for each item; (3) calculating the utility of each item not only from its posterior mean but also from its posterior standard deviation, with the latter accounting for uncertainty aversion. We compared the PUC model to the established attentional Drift Diffusion Model (aDDM). We fitted model parameters to individual-subject data using maximum-likelihood estimation.

PUC versus aDDM

To compare the goodness of fit of the PUC model with that of the aDDM, we first inspected model fits to several summary statistics plotted by [1] (Fig 3): the proportion of choices of an item as a function of the fixation time advantage for that item, the same but conditioned on the item’s rating, and the proportion of choices of an item as a function of the rating difference between the two items and which item was fixated last. To obtain the fits of a model to a summary statistic, we ran each model in generative mode using the fitted parameters for an individual subject to obtain a model prediction for each trial, then aggregated these predictions to compute the summary statistic.

Fig 3. Fits of the aDDM, the acbDDM, and the PUC model to summary statistics of the data.

Fig 3

(A) When the total fixation time advantage of an item increases, that item is chosen more often. (B) Same as A but conditioned on item rating. Both models predicted that when the absolute values of both items are higher (i.e. both are more preferred items), the fixation modulation effect is larger, which trend is less significant in the empirical data. (C) Besides total fixation duration, the last fixation also biases the choice. (D) Distribution of total fixation time. The aDDM fits we obtained differ from those in [1] not only because of differences in parameter estimation methods (see Model fitting and model comparison—Differences from Krajbich et al. (2010)), but also because of a difference in trial aggregation (see Results—PUC versus aDDM).

Both the PUC model and the aDDM qualitatively capture the behavioral phenomena that an item viewed for a longer time (Fig 3A and 3B) or viewed last (Fig 3C) is more likely to be chosen. However, both models show modest deviations from the data. In addition to choice data, we also plotted the distribution of the total fixation time (Fig 3D), which shows a somewhat worse fit for the aDDM than for the PUC model.

Note that our fits of the aDDM to the summary statistics look different from those given by [1]. This is mainly because of a difference in trial aggregation. Krajbich et al. aggregated the model predictions across rating pairs without taking into account the frequencies of these pairs in the experiment. We instead aggregated individual-trial predictions; this is necessary in our case because we used the individual-trial fixation series, but it also ensures a proportional representation of each rating pair in the summary statistics.

To quantitatively compare the models, we performed formal model comparison (Table 1). The PUC model has a lower AICc than the aDDM (thus a better fit), with a summed difference across subjects of −461. The bootstrapped 95% confidence interval was (−887, −136). BIC penalizes the number of parameters more, causing the 95% confidence interval to include 0. Overall, the PUC model performs better than the aDDM.

Table 1. Comparing the main PUC model to alternative models according to negative log likelihood (not corrected for the number of free parameters), AICc, and BIC.

Lower values are better for the first-mentioned model. All values are summed across subjects; bootstrapped 95% confidence intervals are given in parentheses.

PUC-aDDM PUC-acbDDM acbDDM-aDDM
neg LL -321 (-533, -160) 127 (55, 197) -448 (-613, -318)
AICc -461 (-887, -136) 253 (110, 394) -714 (-1045, -456)
BIC -288 (-708, 35) 253 (110, 394) -541 (-869, -284)

Collapsing bound

Next, we included the aDDM with collapsing bound, acbDDM, into the comparison. The fits of the acbDDM to the summary statistics are similar to those of the aDDM (Fig 3), but the acbDDM performs substantially better than the aDDM in model comparison (summed AICc difference: -714; summed BIC difference: -541). The acbDDM model also clearly outperforms the PUC model (summed AICc and BIC difference: 253; these two metrics are same because both models have the same number of parameters).

Hierarchical Bayesian model selection

So far, we assumed that every subject followed the same model. However, there might be heterogeneity in the population. To account for this possibility, we performed hierarchical Bayesian model selection [38] using the VBA package [39]. According to this analysis, the proportions of the population following the PUC, the aDDM, and the acbDDM are 20.5%, 25.6%, and 53.9% based on AICc (and 18.0%, 35.9%, and 46.1% based on BIC). However, more work would be needed to conclusively establish heterogeneity in the population.

PUC model variants

Finally, we examined three variants of the PUC model: first, a variant in which the prior variance is a free parameter instead of a fixed constant; second, a variant in which both the prior variance and the prior mean are free parameters; third, a variant without the uncertainty term in Eq (4) (i.e. A = 0). We found that all three variants fit worse than the main model we presented in the paper, in terms of both AICc and BIC (see S1 Text).

Discussion

In this work, we use the recent idea that valuation is a form of Bayesian inference to explain the effect of fixation on choice. In the Posterior-Utility-Choice (PUC) model, the agent continuously updates a posterior distribution over the value of an item based on a sequence of noisy measurements and computes the utility of that item as a weighted difference between the posterior mean and posterior uncertainty, the latter reflecting uncertainty aversion. The decision is made when the utility difference between the two items reaches a bound. We found that the PUC model accounts better for the effects of fixation on choice than the original model, the attentional drift-diffusion model (aDDM), but not better than its generalization with a collapsing bound, the acbDDM; thus we provided some supports for a flexible bound as a decision model component (although, evidence is mixed in the literature [30, 31]).

Setting aside goodness of fit, the PUC model is a different type of model than the Attentional Drift-Diffusion models. The PUC model postulates what the agent cares about at a behavioral level, through a utility function derived from a posterior distribution over value. By contrast, the a(cb)DDM is neither stated in terms of utility nor involves computing a belief over value. In addition, the PUC model makes a more explicit commitment to the origin of behavioral variability: ultimately, it stems from the noise in sensory measurements or retrieved memories. The a(cb)DDM does not make such a commitment.

These conceptual differences between the PUC model and the a(cb)DDM allow for novel predictions. First, the role of the uncertainty and therefore the utility of an item can be studied more explicitly in a new experimental design. Other than explicitly measure the uncertainty in the rating phase like in [40], future experiments could even manipulate the uncertainty by compromising the quality of the data that the subject receives about the value of an item. This could be done through a simple visual manipulation, such as lowering contrast or blurring the image, or through a memory manipulation, such as presenting items that differ in the time elapsed since the subject last interacted with them. We predict that lower quality of data will lead to lower utility and in turn to the item being less likely to be chosen. Second, a similar prediction could be tested in a timed rating experiment, where speeded judgments should lead to lower ratings. Third, we predict that changing the prior distribution in Eq (2) will affect choices in a specific manner. For example, consider a choice between two items with a true value of 3, yet the participant has a prior estimate of 0. Fixating longer on one item then has two effects on the utility of that item: it increases the posterior mean (since the likelihood mean is greater than the prior) and it decreases posterior uncertainty. Now consider the same agent but with a prior mean of 6. Then, longer fixation has two counteracting effects on utility: it will still decrease uncertainty, but now it will decrease the posterior mean. Comparing to the prior mean of 0, we expect a weaker or even reversed effect of fixation on choice. Thus, it might be interesting to experimentally manipulate the prior distribution.

Our work has several limitations:

  • All extant models exhibit noticeable deviations from the data, which leaves a challenge for future modelers.

  • We assumed a specific direction of causality: that fixation increases preference, instead of the other way round. Both the PUC model and the aDDM assume this causal direction, but the present data do not speak directly to this potential confound. Earlier work did to some extent: when the presentation times of items are controlled by the experimenter, the subjects will prefer the item with longer exposure duration [2]. However, it is not clear whether the magnitude of the effect is comparable between self-directed and passive fixation. This issue needs to be addressed experimentally.

  • We assumed the prior to be Gaussian. Instead, one could allow for a richer parametrization of the prior or use an empirically grounded distribution as the prior. In addition, the agent might update the prior over the course of the experiment.

  • Although Step 1 (computing the posterior) and Step 2 (computing the utility difference) of the PUC model are normative, Step 3 (the collapsing bound) is not. One could make this step normative for example by postulating that the agent maximizes expected reward rate [32], but this would make the model quite complicated without clear prospects for additional insight, in part because expected reward rate is only one way to take into account the cost of time.

  • One could explore alternative forms of the uncertainty aversion term in Eq (4). We subtracted the standard deviation, but this is somewhat arbitrary. Instead, we could have subtracted a power of the standard deviation (e.g. variance), or added the inverse standard deviation.

  • It is not clear how to generalize our model to the loss domain. For aversive items, the fixation bias seems to be in the opposite direction than for attractive items [3]; in other words, looking longer at an aversive item makes the item less likely to be chosen. An account for this effect could start from the finding that people tend to be risk-seeking in the loss domain [41]. Replacing risk attitude by uncertainty attitude, it is tempting to simply use A < 0 for aversive items in Eq (4) of the model. However, this would not produce a good process model, since it would be ill-specified how the agent sets the sign of A. Instead, we see greater promise in taking a step back and designing an alternative utility function, to replace Eq (4), that is the probability that the item under consideration has a value higher than a criterion vcrit:
    U=Pr(v>vcrit|x). (13)
    If the posteriors are Gaussian, as in our model, this becomes
    U=Φ(μposterior;vcrit,σposterior2), (14)
    where Φ(⋅; ⋅, ⋅) is the cumulative normal distribution with mean and variance parameters. The decision variable would still be the difference of the utilities of the left and right item, as in Eq (9). As an example, we now consider the case of vcrit = 0 and σp → ∞, use the properties of the cumulative normal distribution, and substitute Eqs (7) and (8). This yields
    U=Φ(μposteriorσposterior;0,1)=Φ(x¯Tσ;0,1) (15)
    For positive v, this U tends to increase with more observations, but for negative v, U tends to decrease. Thus, an aversive item will become less preferred with longer looking time. We conclude that Eq (13) might provide a starting point for future models that generalize better to the loss domain (and that are also less arbitrary in the sense of the previous point).

Finally, we briefly address recent studies that also apply a value inference framework to understand attention-modulated decision-making [11, 42, 43]. These studies interpret the switching of attention as an active sampling process and derive the switching strategy from a optimal policy. The optimal policies are derived in different ways, some with an explicit decreasing threshold like in the PUC model [11], while others assume that the sampling and switching costs need to be balanced with accurate posterior estimation [42, 43]. In addition, these models differ in how the behavioral signature of “more fixated item being more preferred” is reproduced. One assumption that they shared is that the prior mean of the item value is either zero or lower than the true value (“prior bias” in [43]). As a result, the less sampled item will have the posterior value closer to prior mean rather than the true value which is higher than the prior. Jang et al. [42] also assumed that attention changes the precision of observations, so that the unattended item will incur samples with bigger variance, thus the expected mean will approach to the true mean even slower; this is similar to the PUC model. Song et al., on the other hand, assumed a distorted value perception for the unattended item by assuming a lower sample mean for that item [11].

A difference between these approaches and our model is that our reproduction of the qualitative effect does not require an assumption that the prior mean is zero or lower than the true mean; instead, we fix the prior mean to the empirical mean of the ratings on an individual basis. Instead, we introduced the subjective utility with an uncertainty aversion component, the underlying consideration being that consumer decisions may not merely amount to a value comparison similar to perceptual tasks, but also involve choice bias mechanisms. A shortcoming of our approach, however, is that we are not able to predict fixation times. We believe it would be worthwhile to examine whether the PUC model can be equipped with an active sampling mechanism.

At a high level, our work fits in a broader set of recent attempts to appreciate the role of evidence accumulation and inference in value-based decision-making (other examples include [10, 12, 44]). We expect the probabilistic inference to become increasingly central in the study of valuation and decision-making.

Materials and methods

Data

The data from the experiment in [1] were made available to us by the authors. The data set contained 39 participants with an average of 95 trials per participant (25 participants completed the maximum number of 100 trials). On each trial, the data of interest consisted of the previously collected ratings of the presented items and the eye fixation series summarized as a binary sequence with values “left” and “right”, with the corresponding fixation times. We are not fitting reaction times, but instead total fixation time.

Likelihood

The PUC model, as introduced in “Decision models” above, has 4 parameters for the value estimation (σ, σp, μp,A), three bound parameters (B0, k, and λ), one guessing rate parameter (g) and a non-decision time parameter τ. To simplify, we fixed the prior mean σp and variance μp to be the empirical mean and variance extracted from the rating data, thus leaving 5 parameters to be fitted. We tested the more flexible versions of PUC too, as well as another reduced PUC model(see summary in the “Result” section and details in S1 Text). The aDDM has 3 parameters for the drift process: σ, d, θ, two bound parameters (k and λ, since in their model, B0 can be set to 1 without loss of generality). To match with PUC for a fair model comparison, we added a guessing rate parameter (g) and a non-decision time parameter τ for aDDM.

We fitted the parameters in each model on an individual-subject basis using maximum-likelihood estimation. The inputs to the model on the ith trial consist of the ratings of the left and right items, rleft,i and rright,i. The model predicts the joint probability of the choice Ci and the total fixation time Ti. The log likelihood of the parameters is then

logL(parameters)=i=1ntrialslogp(Ci,Ti|rleft,i,rleft,i,fixationseriesi,parameters). (16)

We did not constrain Ti to be later than all of the observed fixations. Thus, the model is allowed to wrongly predict a total fixation time that falls somewhere in the middle of the fixation series.

Fitting procedure

We used maximum-likelihood estimation to fit the parameters, separately for each participant. This involves maximizing Eq (16), which in turn involves calculating for a given parameter combination and on each trial the probability that the model observer produces the participant’s choice Ci with the participants’ total fixation time Ti. To calculate this probability, we numerically propagated the probability distribution of the decision variable (See Eqs 9, 11 and 12) across time. At each time step, we used the boundaries to truncate the distribution, with the truncated probability being our estimate of the probability of the corresponding response. The (non-normalized) remaining part of the distribution is propagated further.

To maximize the log likelihood, we used Bayesian Adaptive Direct Search [45], a global optimization algorithm that uses Bayesian methods to approximate the shape of the likelihood function. To minimize the risk of getting stuck in local optima, we ran the optimization algorithm with multiple initial conditions (see S2 Text).

Differences from Krajbich et al. (2010)

Our fitting procedure differs from the one used in the original paper [1] in the following aspects:

  • Instead of fitting parameters at the group level, we fitted parameters to individual subjects. This is more accurate if individuals differ from each other.

  • Krajbich et al. binned the data in 100 ms bins, but simulated the time course of their model evolution in 1 ms steps. Instead, we used steps of 100 ms for the latter as well. Apart from saving computational time, this choice is more consistent with the assumption that measurements are independent across time points (in view of neural autocorrelation functions). Moreover, 100 ms is still reasonably fine compared to the median total fixation time (which was about 1.4 seconds; mean being 1.9 seconds).

  • Instead of fitting the reaction time, we fitted total fixation time, in an attempt to avoid epochs in which the observer did not fixate on either item.

  • Krajbich et al. used randomly sampled fixation durations from the empirical distribution to perform the simulation both in this and later work [8, 9]. Instead, we used the actual fixation data for each trial, simulating only until the time when the actual fixation has ended and calculating the probability that the choice was made at the end of the empirical fixation series. All remaining probability went into a single bin representing later decision times. Note one exception is when we plot the summary statistics of probability of making choice against the total fixation time. We used the distribution of fixation times split out by subject, then independently and sequentially drew from this distribution to create a synthetic fixation series for each trial after the empirical fixation series has ended; we repeated this 10 times for each trial which is enough to achieve a stable result. This allows us to obtain an unrestricted model prediction for total fixation times (which we use as a proxy for reaction times throughout the paper).

  • Krajbich et al. performed 1000 simulations to calculate the log likelihood for each rating pair (3000 in their later work [9]). However, this number is low relative to the number of possible responses (in Krajbich et al., 2 choices times 52 reaction time bins; for us, 2 choices times 50 total fixation time quantiles). In such situations, estimating the log likelihood through brute-force simulation not only causes variance to be high, but is also biased due to the nonlinearity of the logarithm [46]. This problem is particularly stark when the simulation assigns zero samples to an observed response. Therefore, instead, we used the numerical approximation mentioned above.

  • Instead of using a grid search in parameter space, we used Bayesian Adaptive Direct Search [45], which is a more precise and more reliable optimization method.

Parameter recovery and model checking

To confirm the validity of our model fitting choices, we fitted synthetic data using the same fitting procedures as for the real data. To generate synthetic data, we used the exact same rating distribution as the real data by matching each synthetic trial with a real trial. Each synthetic subject was given the parameters that best fitted one real subject. Then we performed fitting for individual synthetic subjects using methods introduced above. Results are presented in S3 Text. The summary statistics are recovered very well. Parameter recovery is generally good but somewhat worse for the more complex models (acbDDM and PUC). This is likely due to soft trade-offs between parameters. As a result, the parameter estimates in the real data should be taken with a grain of salt. However, the results of our paper do not rely on parameter estimates but only on log likelihoods and summary statistics. Therefore, the results are not affected by issues with parameter recovery.

Model comparison

To compare models, we used the corrected Akaike Information Criterion (AICc; [47, 48]) and the Bayesian Information Criterion (BIC; [49]).

Supporting information

S1 Text. PUC model variants.

(PDF)

S2 Text. Details of model fitting.

(PDF)

S3 Text. Parameter recovery.

(PDF)

Data Availability

The data and code code used for generating synthesized data, model fitting, and plotting is available on osf (DOI 10.17605/OSF.IO/8YAC9).

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Krajbich I, Armel C, Rangel A. Visual fixations and the computation and comparison of value in simple choice. Nature Neuroscience. 2010;13(10):1292–1298. doi: 10.1038/nn.2635 [DOI] [PubMed] [Google Scholar]
  • 2.Armel KC, Beaumel A, Rangel A. Biasing simple choices by manipulating relative visual attention. Judgment and Decision Making. 2008;3(5):396. [Google Scholar]
  • 3.Shimojo S, Simion C, Shimojo E, Scheier C. Gaze bias both reflects and influences preference. Nature Neuroscience. 2003;6(12):1317–1322. doi: 10.1038/nn1150 [DOI] [PubMed] [Google Scholar]
  • 4.Pärnamets P, Johansson P, Hall L, Balkenius C, Spivey MJ, Richardson DC. Biasing moral decisions by exploiting the dynamics of eye gaze. Proceedings of the National Academy of Sciences. 2015;112(13):4170–4175. doi: 10.1073/pnas.1415250112 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Tavares G, Perona P, Rangel A. The attentional drift diffusion model of simple perceptual decision-making. Frontiers in neuroscience. 2017;11:468. doi: 10.3389/fnins.2017.00468 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Krajbich I. Accounting for attention in sequential sampling models of decision making. Current opinion in psychology. 2019;29:6–11. doi: 10.1016/j.copsyc.2018.10.008 [DOI] [PubMed] [Google Scholar]
  • 7.Ratcliff R, McKoon G. The diffusion decision model: theory and data for two-choice decision tasks. Neural computation. 2008;20(4):873–922. doi: 10.1162/neco.2008.12-06-420 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Krajbich I, Rangel A. Multialternative drift-diffusion model predicts the relationship between visual fixations and choice in value-based decisions. Proceedings of the National Academy of Sciences. 2011;108(33):13852–13857. doi: 10.1073/pnas.1101328108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Krajbich I, Lu D, Camerer C, Rangel A. The Attentional Drift-Diffusion Model Extends to Simple Purchasing Decisions. Frontiers in Psychology. 2012;3. doi: 10.3389/fpsyg.2012.00193 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Tajima S, Drugowitsch J, Pouget A. Optimal policy for value-based decision-making. Nature communications. 2016;7:12400. doi: 10.1038/ncomms12400 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Song M, Wang X, Zhang H, Li J. Proactive information sampling in value-based decision-making: deciding when and where to saccade. Frontiers in human neuroscience. 2019;13:35. doi: 10.3389/fnhum.2019.00035 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Polania R, Woodford M, Ruff CC. Efficient coding of subjective value. Nature neuroscience. 2019;22(1):134. doi: 10.1038/s41593-018-0292-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Dearden R, Friedman N, Russell S. Bayesian Q-learning. In: Aaai/iaai; 1998. p. 761–768.
  • 14.Dabney W, Kurth-Nelson Z, Uchida N, Starkweather CK, Hassabis D, Munos R, et al. A distributional code for value in dopamine-based reinforcement learning. Nature. 2020; p. 1–5. doi: 10.1038/s41586-019-1924-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.de Laplace PS. Théorie analytique des probabilités. Courcier; 1820.
  • 16.Knill DC, Richards W. Perception as Bayesian inference. Cambridge University Press; 1996. [Google Scholar]
  • 17.Pouget A, Beck JM, Ma WJ, Latham PE. Probabilistic brains: knowns and unknowns. Nature Neuroscience. 2013;16(9):1170–1178. doi: 10.1038/nn.3495 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Ma WJ, Jazayeri M. Neural Coding of Uncertainty and Probability. Annual Review of Neuroscience. 2014;37(1):205–220. doi: 10.1146/annurev-neuro-071013-014017 [DOI] [PubMed] [Google Scholar]
  • 19.Shadlen MN, Shohamy D. Decision making and sequential sampling from memory. Neuron. 2016;90(5):927–939. doi: 10.1016/j.neuron.2016.04.036 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Harrison GW, Elisabet Rutström E. Risk aversion in the laboratory. In: Risk aversion in experiments. Emerald Group Publishing Limited; 2008. p. 41–196. [Google Scholar]
  • 21.Chowdhary K, Dupuis P. Distinguishing and integrating aleatoric and epistemic variation in uncertainty quantification*. ESAIM: Mathematical Modelling and Numerical Analysis. 2013;47(3):635–662. doi: 10.1051/m2an/2012038 [DOI] [Google Scholar]
  • 22.Craig A. Interoception: the sense of the physiological condition of the body. Current opinion in neurobiology. 2003;13(4):500–505. doi: 10.1016/S0959-4388(03)00090-4 [DOI] [PubMed] [Google Scholar]
  • 23.Ma WJ. Signal detection theory, uncertainty, and Poisson-like population codes. Vision Research. 2010;50(22):2308–2319. doi: 10.1016/j.visres.2010.08.035 [DOI] [PubMed] [Google Scholar]
  • 24.Levy H, Markowitz HM. Approximating expected utility by a function of mean and variance. The American Economic Review. 1979;69(3):308–317. [Google Scholar]
  • 25.Markowitz H. Portfolio selection. The journal of finance. 1952;7(1):77–91. doi: 10.2307/2975974 [DOI] [Google Scholar]
  • 26.Zajonc RB. Mere exposure: A gateway to the subliminal. Current directions in psychological science. 2001;10(6):224–228. doi: 10.1111/1467-8721.00154 [DOI] [Google Scholar]
  • 27.Bornstein RF. Exposure and affect: overview and meta-analysis of research, 1968–1987. Psychological bulletin. 1989;106(2):265. doi: 10.1037/0033-2909.106.2.265 [DOI] [Google Scholar]
  • 28.Lee AY. The mere exposure effect: An uncertainty reduction explanation revisited. Personality and Social Psychology Bulletin. 2001;27(10):1255–1266. doi: 10.1177/01461672012710002 [DOI] [Google Scholar]
  • 29.Churchland AK, Kiani R, Shadlen MN. Decision-making with multiple alternatives. Nature neuroscience. 2008;11(6):693–702. doi: 10.1038/nn.2123 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Hawkins GE, Forstmann BU, Wagenmakers EJ, Ratcliff R, Brown SD. Revisiting the Evidence for Collapsing Boundaries and Urgency Signals in Perceptual Decision-Making. Journal of Neuroscience. 2015;35(6):2476–2484. doi: 10.1523/JNEUROSCI.2410-14.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Milosavljevic M, Malmaud J, Huth A, Koch C, Rangel A. The Drift Diffusion Model can account for the accuracy and reaction time of value-based choices under high and low time pressure. Judgment and Decision Making. 2010;5(6):437–449. [Google Scholar]
  • 32.Drugowitsch J, Moreno-Bote R, Churchland AK, Shadlen MN, Pouget A. The cost of accumulating evidence in perceptual decision making. Journal of Neuroscience. 2012;32(11):3612–3628. doi: 10.1523/JNEUROSCI.4010-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Krajbich I, Smith SM. Modeling Eye Movements and Response Times in Consumer Choice. Journal of Agricultural & Food Industrial Organization. 2015;13(1). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Webb R. The (neural) dynamics of stochastic choice. Management Science. 2019;65(1):230–255. doi: 10.1287/mnsc.2017.2931 [DOI] [Google Scholar]
  • 35.Cavanagh JF, Wiecki TV, Kochar A, Frank MJ. Eye tracking and pupillometry are indicators of dissociable latent decision processes. Journal of Experimental Psychology: General. 2014;143(4):1476. doi: 10.1037/a0035813 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Smith SM, Krajbich I. Gaze amplifies value in decision making. Psychological science. 2019;30(1):116–128. doi: 10.1177/0956797618810521 [DOI] [PubMed] [Google Scholar]
  • 37.Westbrook A, van den Bosch R, Määttä J, Hofmans L, Papadopetraki D, Cools R, et al. Dopamine promotes cognitive effort by biasing the benefits versus costs of cognitive work. Science. 2020;367(6484):1362–1366. doi: 10.1126/science.aaz5891 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Stephan KE, Penny WD, Daunizeau J, Moran RJ, Friston KJ. Bayesian model selection for group studies. NeuroImage. 2009;46(4):1004–1017. doi: 10.1016/j.neuroimage.2009.03.025 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Daunizeau J, Adam V, Rigoux L. VBA: a probabilistic treatment of nonlinear models for neurobiological and behavioural data. PLoS computational biology. 2014;10(1). doi: 10.1371/journal.pcbi.1003441 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Gwinn R, Krajbich I. Attitudes and attention. Journal of Experimental Social Psychology. 2020;86:103892. doi: 10.1016/j.jesp.2019.103892 [DOI] [Google Scholar]
  • 41.Hershey JC, Schoemaker PJ. Risk taking and problem context in the domain of losses: An expected utility analysis. Journal of Risk and Insurance. 1980; p. 111–132. doi: 10.2307/252685 [DOI] [Google Scholar]
  • 42.Jang AI, Sharma R, Drugowitsch J. Optimal policy for attention-modulated decisions explains human fixation behavior. Elife. 2021;10:e63436. doi: 10.7554/eLife.63436 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Callaway F, Rangel A, Griffiths TL. Fixation patterns in simple choice reflect optimal information sampling. PLOS Computational Biology;17(3):e1008863. doi: 10.1371/journal.pcbi.1008863 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Gabaix X, Laibson D. Myopia and Discounting [Working Paper]; 2017. Available from: https://scholar.harvard.edu/xgabaix/publications/myopia-and-discounting-working-paper.
  • 45.Acerbi L, Ma WJ. Practical Bayesian Optimization for Model Fitting with Bayesian Adaptive Direct Search. arXiv:170504405 [q-bio, stat]. 2017.
  • 46.van Opheusden B, Acerbi L, Ma WJ. Unbiased and Efficient Log-Likelihood Estimation with Inverse Binomial Sampling. arXiv preprint arXiv:200103985. 2020. [DOI] [PMC free article] [PubMed]
  • 47.Akaike H. A new look at the statistical model identification. IEEE transactions on automatic control. 1974;19(6):716–723. doi: 10.1109/TAC.1974.1100705 [DOI] [Google Scholar]
  • 48.Burnham KP, Anderson DR. Multimodel inference: understanding AIC and BIC in model selection. Sociological methods & research. 2004;33(2):261–304. doi: 10.1177/0049124104268644 [DOI] [Google Scholar]
  • 49.Schwarz G, et al. Estimating the dimension of a model. The annals of statistics. 1978;6(2):461–464. doi: 10.1214/aos/1176344136 [DOI] [Google Scholar]
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009190.r001

Decision Letter 0

Samuel J Gershman

9 Oct 2020

Dear Dr. Ma,

Thank you very much for submitting your manuscript "A posterior-based model of the effects of fixation on choice" for consideration at PLOS Computational Biology.

As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments. I thought the comments are reasonable and clear, so I don't need to expand on them.

My only meta-comment concerns R2's first point: it's not clear to me that a Gaussian rating distribution is necessary to validate the model. If a subject's internal model and the true distribution over values are both Gaussian, then it is true you would expect a Gaussian rating distribution, but it's possible for the Gaussian assumptions to be a good description of the subject's internal model even if the true value distribution is non-Gaussian, in which case you would expect a non-Gaussian rating distribution. Another way of saying this is that Bayesian models of cognition need not be well-calibrated. It's certainly nice when they are, but finding poor calibration is not by itself evidence against a Bayesian model. It's especially tricky in this context even just to measure calibration because we don't have independent access to the true values. This means that finding a non-Gaussian rating distribution could mean the prior is non-Gaussian, or the observation noise is non-Gaussian, or the true distribution over values is non-Gaussian, or any combination of these. In any case, it's worth checking --- you might get lucky!

We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.

Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Samuel J. Gershman

Deputy Editor

PLOS Computational Biology

***********************

Reviewer's Responses to Questions

Comments to the Authors:

Reviewer #1: This is a really nice paper investigating a Bayesian approach to value-based decision-making. Drift-diffusion models (DDM) assume that evidence is accumulated over time for each alternative. The attentional DDM (aDDM) assumes that the rate of evidence accumulation increases for the alternative that is attended to. This leads to a choice bias in favor of attended items. This paper proposes an interesting alternative model for this phenomenon. The model assumes that subjects have a prior expected value (and associated variance) for each alternative. When the subject attends to an alternative, they update its expected value and reduce the variance around that expectation. Subjects are risk-averse and so they prefer alternatives that they are more certain about. The preference for higher values and higher certainty leads subjects to choose alternatives that are higher valued but also attended to longer.

The model presented in this paper is very sensible and has a lot of nice features. It blends economic intuition about decision-making under uncertainty with cognitive knowledge about attention and information gathering. While the model does not substantially outperform the existing model (aDDM) in goodness-of-fit, it is still an important step, for at least a couple of reasons. First, the aDDM is silent about how/why decision-makers sample information. The authors’ PUC model makes it clear that the point of attending to an alternative is to get a more precise and accurate estimate of the value of that alternative. This intuition is critical to advance our understanding of how people sample information. While this paper does not attempt to model the fixation process, one can see how you would use this framework to tackle that question. Second, the authors’ model seems to (marginally) outperform the standard aDDM. Only when the authors give the aDDM collapsing bounds do the models perform similarly well. Thus, another takeaway point of the paper is that collapsing bounds are an important feature to consider in these model-comparison exercises, and they may be an important addition to the aDDM (though visually the improvements were imperceptible, see Fig. 3).

In summary, I think this is an important and well written paper. I have a number of mostly minor comments. If the authors can address those comments, I would be happy to endorse publication.

(1) Regarding the comparison between the aDDM and cbaDDM: Figure 3 shows little difference between the models. These are all choice plots, where collapsing bounds will have little impact. It might be useful to add a fourth column comparing the reaction time curves with each model compared to the data.

(2) In comparing the advantages of the PUC to the aDDM, the authors repeatedly state that in the aDDM the “meaning of the decision variable is unclear”. I’m not really sure what the authors mean by this. In the (a)DDM, the goal is to choose the better item. That is why drift rate is driven primarily by the alternatives’ values; this leads to the decision-maker choosing the higher-value item in most cases. As outlined by Webb (2019) in Management Science, the DDM is essentially just a choice rule that takes utilities as inputs and outputs logit choices. In other words, the authors’ model first updates values then calculates utilities, while the DDM first generates utilities then updates the decision variable. It isn’t that one model is more normative than the other, its that the models do things in a different order.

(3) More generally, I think that the authors claims about what is normative (or what isn’t) should be set aside. What is normative depends on lots of assumptions. The aDDM could certainly be normative if attention switching is costly and attention is limited; we know that the standard DDM is normative under certain conditions (Wald,1945). The authors’ model is already admittedly semi-normative, and that doesn’t even consider the fact that risk-aversion is not normative per se (though it is standard) and that there is no explanation in this model for the fixation series themselves, which seemingly lead to non-normative information gathering.

I would instead focus on the fact that this model is explicitly Bayesian and so is set up to tackle important questions of how attention should be allocated, how to incorporate prior information about the alternatives, what confidence ratings the model should produce, etc.

(4) The authors mention at some point that, unlike in the aDDM, later evidence has a smaller effect on the decision variable. Could the authors elaborate on this point? In Bayesian models the order of data is not supposed to matter. Is that violated here? Or do the authors simply mean that there are diminishing returns to information collection from the same alternative?

(5) In simulating their model, the authors use 100ms time steps, which is much larger than other papers in the literature. Did the authors verify that such large time steps do not lead to distortions in the simulated data? In particular, did the authors do any parameter recovery exercises to ensure that this method was effective?

(6) On a related note, in their simulations the authors use the empirical fixation series. What happened when the observed fixation series terminated (in the data) and a decision hadn’t yet been reached in the simulation? Did the decision suddenly end or were additional fixations generated in some way?

(7) I found the simplified PUCp to be a strange model to consider. Why would subjects have a prior of 0 when only appetitive items were presented in the experiment? Even if subjects for some reason started the experiment with this prior, one would think it would change after the rating task, or at least after a handful of choice trials. Why not instead try a model where the prior mean is the mean of each subject’s rating distribution? With a prior of 0, you are essentially guaranteeing that more attention will lead to a higher choice probability. The whole thing seems quite anti-Bayesian. To be clear, I’m not demanding that the authors test such a model, I would just like more explanation why they are even considering this prior=0 model. What would be worth doing is looking at the relation between the recovered mu_p parameters and subjects’ actual mean ratings. On that note, it would be useful to see the distributions of recovered parameters for each of the models (in the appendix would be fine).

(8) There has been a debate about the multiplicative vs. additive nature of attention in these choice processes, with Cavanagh et al. 2014 (JEP: General), Smith & Krajbich 2019 (Psych Sci), and Westbrook et al. 2020 (Science). I wonder if the authors could comment on how their findings might inform this debate. Their model would seem to fall somewhere in between, with the uncertainty reduction being “additive” (independent of value), while the benefit of updating from the prior would be “multiplicative”, in the sense that it is more beneficial the farther the true value is from the prior.

Minor comments

(9) There are some references the authors might find useful (in addition to the ones mentioned above).

Gwinn & Krajbich 2020 (JESP) - they discuss how certainty (as well as accessibility) about item ratings affects choice outcomes.

Parnamets et al. 2015 (PNAS) - they provide evidence for how exogenous manipulations of attention affect choice outcomes

Tavares et al. 2017 (Frontiers) - they essentially replicate Parnamets, but with perceptual stimuli

There are also a number of papers on these attention effects that the authors might want to consider including in their references. See Krajbich 2019 (Current Opinion in Psychology) for a review.

(10) I’m not sure, but I’m concerned there may be an error in Equation 7. It would seem that as T goes to infinity (as deliberation time increases), mu_posterior goes to zero. But perhaps I am missing something.

(11) Non-decision time is also thought to account for early orienting to the stimuli, in addition to motor latency at the end of the decision.

(12) In the aDDM, d is a scaling constant, not the drift rate per se. The drift is d multiplied by the attention-weighted values.

(13) The authors say that in the data the median RT was 4 seconds. Can they elaborate on this, given that in Krajbich et al. 2010, mean RT appears to be about 2 seconds?

(14) The authors claim that Krajbich et al. 2010 used a “fixation model” in their simulations, but in the paper they state that they “sampl[ed] fixation lengths from the empirical fixation data….”

(15) I think it would be more accurate to state that there is “some” evidence for collapsing bounds. Milosvaljevic et al. 2010 argue that collapsing bounds do not help once across trial drift variability is accounted for. Hawkins et al. 2015 show mixed evidence for collapsing bounds. And here, while collapsing bounds do seem to help in the goodness-of-fits statistics, qualitatively the improvement is imperceptible, at least in the plots the authors have shown.

(16) Prospect theory states that risk-seeking occurs in the loss domain because of diminishing marginal sensitivity, not “so that they will still have some probability to obtain a good outcome.” So while this is an interesting way to think about the loss domain, it is not based on the standard interpretation.

(17) Table 2 - please briefly describe the models in the table caption.

Reviewer #2: In this work, Li and Ma propose a partial normative model to study the effects of fixation on decision-making. In the model they propose, the utility is computed based on weighted sum of the posterior mean and standard deviation of the posterior distribution. These values depend on both the alternative input values and the amount of time fixating at each alternative. I think that this is an interesting attempt to formalize the originally proposed instantiation of the aDDM. There are however a couple of critical aspects that the authors may want to clarify related key assumptions adopted in their models and parameter fitting.

Major comments

1) The authors adopt a Bayesian approach to estimate a posterior distribution, which ultimately leads to the computation of an utility that is then used to compute the decision variable (by comparing the utilities of each alternative). The authors assume a Gaussian prior and a Gaussian likelihood function with constant noise (irrespective of value input). The authors use the choice data to fit both the prior (mean and variance), and the variance in the likelihood function. Here there are two critical issues. First, given that the authors fit both prior and likelihood using the choice data, there must be at least some evidence that the use of this prior shape (and the resulting fit) is valid. Do the rating data follow a Gaussian distribution? Is the estimated prior distribution comparable to the rating data distribution? It is essential that the authors show and report this information in the article. I appreciate that the authors briefly mention in the discussion section the fact that the Gaussian distribution assumption is a limitation. However, in my opinion, such strong assumptions should be backed up by a principled argument otherwise the Bayesian model ceases to be “normative”. Also, the motivation for such assumption should be stated already in the description of the model. This is not to say that I am against Bayesian approaches (actually quite the opposite), but this is precisely the issue that one sector cognitive psychology has used as ammunition against Bayesian approaches in our field. The authors could for instance obtain the best possible fit of a Gaussian distribution to the rating data and then use this as prior distribution to fit the choice data. Alternatively, out of sample fits could be adopted to estimate out of sample priors (however, this might not sully solve the problem).

2) The model proposed by the authors has 8 free parameters that are used to fit about 95 choice trials per participant. It has been previously argued that the reliability of fitting that many parameters to DDMs (in particular with dynamic bounds) can be problematic and might be not identifiable, in particular for low number of trials as is the case here. A more critical problem is that the authors are also freely fitting the prior distribution. As it has been shown in previous work, some DDM parameters and in particular boundary parameters are very much related to prior distributions (see related work by Drugowitsch). This in addition to the prior fitting issue highlighted in my previous point 1. Here it is essential to demonstrate via simulations that the Bayesian inference parameters and the boundary parameters are identifiable using the same number of trials used to fit the real data (i.e., about 100 trials). Without this evidence, it is difficult to evaluate whether all free model parameters assumed by the authors are jointly identifiable.

Minor comment

3) It would be informative to study further qualitative predictions that could uniquely emerge in the model proposed by the authors. At present it is not clear (at least qualitatively) what the key differences are between the PUC and the aDDM. For instance, it is possible that the model makes different predictions for long versus short trials as in the PUC the model is directly influenced by viewing time?

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: No: Data are available from the original authors.

Reviewer #2: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, PLOS recommends that you deposit laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions, please see http://journals.plos.org/compbiol/s/submission-guidelines#loc-materials-and-methods

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009190.r003

Decision Letter 1

Samuel J Gershman

14 May 2021

Dear Dr. Ma,

Thank you very much for submitting your manuscript "An uncertainty-based model of the effects of fixation on choice" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations.

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Samuel J. Gershman

Deputy Editor

PLOS Computational Biology

***********************

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

[LINK]

Reviewer's Responses to Questions

Comments to the Authors:

Reviewer #1: The authors have done a quite thorough job addressing my comments from the previous version. There are just a few minor points that I think should be corrected before publication.

(1) One of the takeaway points of the paper is that collapsing bounds are important, as they improve the aDDM fits substantially. However, one issue with this claim is that the boundary parameters k and lambda were constrained to be positive in the fits, and there appears to be bunching of the estimates close to or at zero. This constraint essentially guarantees that you will find support for collapsing bounds as long as there is variability across participants. I'm not asking the authors to refit all the models, but if they don't, I think they need to qualify that they did not allow for increasing bounds over time. What they have shown is that allowing bounds to vary improves model fit, but not that bounds are necessarily collapsing.

(2) I think the authors need to be clearer about the parameter recovery results in the main text. I wouldn't say that parameter recovery is "somewhat worse" for the more complex models; it is non-existent. To my eye the correlation between true and fitted parameters looks to be about 0 for several of the parameters, for example lambda and k in the PUC model. While this might not impact model comparison, it does mean that these parameters can't really be interpreted, so readers should not try to fit these models to data and compare parameter values (as readers are wont to do).

Side note - where do the "original" parameter values come from anyways? They don't appear to be from the participants' best fitting values.

(3) In the "Differences between PUC and aDDM" section, point (c) says that in the aDDM the information about the two items is combined into a single variable while in the PUC there are two distinct value distributions. However, Krajbich & Rangel 2011 showed that the multi-alternative version of aDDM has separate accumulators for each item and that it reduces to the aDDM with two items. In other words, the aDDM can be thought of as two distinct accumulators with a downstream difference calculation to compute the relative variable. In other words, both models allow the agent to keep two distinct value representations, and in the multi-alternative case this is required. I would suggest removing point (c).

Minor comments

What are the units on d? I guess this is in "evidence per 100ms"?

aDDM B set to 0? I think you mean 1?

The PUC has 9 parameters, you fix 2, but that leaves 7 not 5…

The "DDM2 parameter recovery" figure has one panel that says “lps” with unreadable units - should that be k?

The PUC parameter recovery figure has a panel labeled “ObsVar” - should that be sigma?

Across the parameter recovery and parameter distribution figures, you are inconsistent in whether you use the greek symbol itself or spell it out.

Reviewer #2: I commend authors for their efforts to carefully address my concerns and questions.

The authors did an excellent job responding to my concerns, and I do not have any further comments.

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Rafael Polania

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

References:

Review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript.

If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009190.r005

Decision Letter 2

Samuel J Gershman

17 Jun 2021

Dear Dr. Ma,

We are pleased to inform you that your manuscript 'An uncertainty-based model of the effects of fixation on choice' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Samuel J. Gershman

Deputy Editor

PLOS Computational Biology

***********************************************************

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009190.r006

Acceptance letter

Samuel J Gershman

11 Aug 2021

PCOMPBIOL-D-20-01716R2

An uncertainty-based model of the effects of fixation on choice

Dear Dr Ma,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Andrea Szabo

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Text. PUC model variants.

    (PDF)

    S2 Text. Details of model fitting.

    (PDF)

    S3 Text. Parameter recovery.

    (PDF)

    Attachment

    Submitted filename: response_to_review_fixation.pdf

    Attachment

    Submitted filename: second_response_to_review_fixation.pdf

    Data Availability Statement

    The data and code code used for generating synthesized data, model fitting, and plotting is available on osf (DOI 10.17605/OSF.IO/8YAC9).


    Articles from PLoS Computational Biology are provided here courtesy of PLOS

    RESOURCES