Abstract
Time is an extremely valuable resource but little is known about the efficiency of time allocation in decision-making. Empirical evidence suggests that in many ecologically relevant situations, decision difficulty and the relative reward from making a correct choice, compared to an incorrect one, are inversely linked, implying that it is optimal to use relatively less time for difficult choice problems. This applies, in particular, to value-based choices, in which the relative reward from choosing the higher valued item shrinks as the values of the other options get closer to the best option and are thus more difficult to discriminate. Here, we experimentally show that people behave sub-optimally in such contexts. They do not respond to incentives that favour the allocation of time to choice problems in which the relative reward for choosing the best option is high; instead they spend too much time on problems in which the reward difference between the options is low. We demonstrate this by showing that it is possible to improve subjects' time allocation with a simple intervention that cuts them off when their decisions take too long. Thus, we provide a novel form of evidence that organisms systematically spend their valuable time in an inefficient way, and simultaneously offer a potential solution to the problem.
Keywords: decision-making, speed-accuracy trade-off, neuroeconomics, sequential-sampling model, evidence accumulation, optimality
1. Introduction
We all know the phrase ‘time is money’, and yet at some point or another many of us have caught ourselves agonizing too long even where it makes little difference what we choose, such as what to order for dinner at a restaurant or what movie to watch. Far from being a uniquely human problem, many species exhibit such behaviour. Naturally, the question arises whether this phenomenon is simply an unlucky outcome of an optimal decision-making process, or whether the process itself is sub-optimal. Much work in decision science has focused on whether organisms achieve optimal decision outcomes (e.g. [1–4] and much of the experimental economics literature) but relatively little attention has been paid to how they allocate their time while making decisions.
The problem arises due to the well-known speed-accuracy trade-off, where more time invested into a decision yields a more accurate response [5,6]. One explanation for this phenomenon in many choice contexts is due to the way the brain gradually accumulates noisy evidence for the different choice options, up to predetermined thresholds. Theoretical work has shown how speed-accuracy trade-offs can optimally be resolved [7–9]. For example, when choice difficulty and the benefit of a correct response are held constant, the drift-diffusion model (DDM) is known to be optimal [10,11]. By optimal we mean that for a desired accuracy rate, the DDM minimizes the expected response time (RT) [12]. Recent years have seen much research showing that organisms including flies [13], ants [14], bees [15–18], rats [19,20], primates [7,21–28] and humans [29–34] use sequential sampling model (SSM) processes (like the DDM) to make many decisions, and that they do respond to speed or accuracy constraints. Moreover, these models apply not only to many perceptual decisions, where there is an objectively correct response, but also to several value-based decisions, where the correct answer is based on subjective preference [35–49].
For instance, the house-hunting behaviour of ants and bees is an example of collective value-based decision-making where individual organisms evaluate the suitability of potential nest sites and then recruit and compete with other members of the group in order to guide the final choice of the whole colony. Ants recruit by physically leading, and eventually carrying each other to attractive nest sites, while bees use a ‘waggle dance’ and head butting to communicate the location of attractive hive sites and inhibit bees favouring other sites, respectively. In both cases, these scouts recruit other colony members more readily for higher quality sites, and so support builds more quickly for better options. Once there is enough support for a particular location, the decision is made and the entire colony picks up and moves. This collective behaviour is governed by the speed-accuracy trade-off; colonies may emphasize either speed or accuracy depending on the urgency of the situation.
More typically, SSMs are applied to individual decision-making behaviour. For instance, an animal may have to quickly evaluate the attractiveness of various potential foraging sites based on the likely quantity and quality of food available, exposure to predators, distance away, etc. [50,51]. An overemphasis on accuracy may demand an unreasonable amount of time to evaluate the potential foraging sites, while an overemphasis on speed may lead the animal to a poor site.
This literature on the optimality of time allocation has mainly focused on the simple case where difficulty and the relative reward for a correct response (compared to an incorrect response) are both held constant, but in the real world these can vary [16,17]. One point that has not been widely acknowledged is that in many ecologically important situations, difficulty and relative reward are in fact linked. In particular, in value-based (economic) decisions where the individual receives the item that he chooses, the benefit of making the correct decision decreases as the options get closer together in subjective value. Simultaneously, as this occurs, the items become harder to distinguish, and we know from the SSM literature that mean decision time increases. As a result, more difficult choices generally take longer, even though the correct choice yields only a minor increase in benefit over the incorrect choice. For example, a foraging animal might find itself torn between two equally attractive patches, wasting time that could better be spent quickly sorting the edible items from the rest.
In settings like these, the optimization problem becomes more complex and there is no clear way to determine whether decision-makers are behaving optimally. In the case of fixed difficulty, it is optimal for the decision-maker to accumulate evidence until the total net evidence reaches a constant threshold [52]. However, when the relative reward for making the correct decision is tied to the difficulty level, the decision-maker can update his/her prior about the subjective-value difference between the options in this particular trial based on how long the decision has taken so far. The decision-maker should realize that as time goes on, the expected relative benefit of making the correct decision is decreasing. When there is limited time to make many decisions, time spent on a low-benefit decision represents an ‘opportunity cost’ [53]. Thus, when the benefit of making the correct choice differs across choice problems, the decision-maker should re-allocate time away from the problems where the relative rewards are small and more time towards the problems where the relative rewards are large. Prior work, for instance in similar settings where information acquisition is increasingly costly over time [8,54], indicates that these situations call for the decision thresholds to decrease over time within a trial [9,52,55–58]. In the neuroscience literature, collapsing-threshold models (and similar urgency models [55,59]) have been gaining popularity, though the behavioural evidence for them is mixed [60,61].
It is critical to note that all these models nevertheless predict that decisions between similar options (‘hard’ choices) will on average take more time than between dissimilar options (‘easy’ choices). Conceivably, it may be that hard choices are unavoidably slower than easy choices because easy choices can be made more quickly than they can be distinguished from hard choices. That is, it may not be possible to quickly identify the difficulty of a choice problem. Therefore, the observation of RT differences between easy versus hard decisions is not by itself sufficient to establish the sub-optimality of time allocation. Here, we tackle this issue by developing a novel empirical method for testing the optimality of behaviour.
We begin with an economic task and then also investigate a perceptual decision-making task that incorporates the difficulty-relative-reward connection that one finds in economic choice. In each task, subjects made a series of choices where time was both scarce and valuable. The first uses naturalistic stimuli and relies on subjects' own valuations, whereas the second uses an approach that affords external control of relative value. While the two studies appear quite different on the surface, they share a very important feature. At the beginning of both studies, subjects had a ‘baseline’ expected outcome that they would earn if they did nothing. By making good choices, subjects could increase their expected earnings from this baseline level. The amount of this increase varied from trial to trial, along with the difficulty of the decision.
The results clearly demonstrate that subjects misallocated their time. We established this by introducing a simple intervention that improved subjects' performance on both decision tasks, using only information that they themselves had available. Importantly, subjects seemed to learn from our intervention and so some of the benefits remained even in the subsequent absence of the intervention. Thus we not only show that decision-makers sometimes wasted their valuable time, but that it was possible to use simple training to help them improve.
2. Material and methods
(a). Study 1
Forty-nine subjects provided informed consent and were paid a flat fee of CHF 30 for their participation, plus possible additional cash of up to CHF 2.50 from the first part of the study. Subjects first indicated their willingness to pay (WTP) for 100 different snack foods, using a Becker–deGroot–Marshak (BDM) mechanism [62], which has the property that it is in subjects' best interest to reveal their true WTP (see details below). For each trial, subjects saw a colour photograph of the item and a slider bar (with a random starting location) that they could use to select a WTP from CHF 0 to 2.50, in steps of CHF 0.25. Subjects used the ‘left’ and ‘right’ arrow keys to move the slider and the ‘up’ arrow key to confirm their choice.
Subjects then proceeded through five blocks of binary decisions between pairs of these items. Using these WTPs, we constructed choice pairs with known valuation differences between the two items. Each block contained 100 trials, half of which were constructed as ‘easy/high-stakes' choices (large valuation difference) and half ‘hard/low-stakes' (small valuation difference). It was impossible to reach every trial in any given block, since each block's duration was 150 s, with 1.5 s inter-trial interval (ITI). Subjects indicated their decision by pressing the ‘left’ or ‘right’ arrow keys on the keyboard. Critically, subjects were informed that the computer would randomly make any uncompleted choices at the end of the 150 s.
At the end of the experiment, subjects were rewarded for one random trial. This trial could be a BDM trial (p = 1/6) or a binary choice trial (p = 5/6). For a choice trial, subjects simply received the item that they, or the computer, chose on that trial. For a BDM trial, the computer generated a random price between CHF 0 and 2.50. If the random price was equal to or less than the subject's WTP for that item, then the subject received the food and paid the random price (out of an endowment of CHF 2.50). If the random price was above the subject's WTP for that item, then the subject did not receive the food and kept the endowment of CHF 2.50. In addition to these earnings, all subjects earned CHF 30 for their participation in the study.
The first of the five blocks (T) was used to obtain an individual empirical distribution function for the RTs in the task. Four more blocks followed. Two of these were non-intervention blocks (N), which were constructed identically to the first block, but with different choice pairs. The other two blocks were intervention blocks (I), in which subjects were reminded on screen to ‘choose now’ after a pre-specified amount of time had passed. If they did not make a choice within 0.5 s of the message, the choice was randomly made for them, and the next trial commenced (after the ITI). The mean deadline was defined for each subject separately, such that it would have cut off the slowest 30% of their decisions in the T block. For each subject, there was a 50% chance that they would experience the sequence T-I-N-I-N, and a 50% chance that they would experience the sequence T-N-I-N-I.
To assess performance on the task, we created a measure of surplus that captures the subjective value generated through making choices and is analogous to the points earned in study 2. To do this, we used each individual subject's WTPs to create the following measure:
Here, is the WTP for the chosen item, vL is the lower of the two WTPs and vH is the higher of the two WTPs. Thus, choice surplus represents the degree to which the surplus from actual human choices outperforms chance. Computer choices were treated as performing at chance level (zero by construction), regardless of their actual random realization, to reduce artificial noise in the measure.
(b). Study 2
Forty-two subjects were recruited through a Princeton University online subject recruitment system and provided informed consent to participate in this study. Two subjects were excluded from analysis due to outlier behaviour. These two subjects scored more than 3 s.d. away from the mean for one of the two conditions, leaving us with an analysed sample size of 40 subjects.
The instructions were provided on screen. Subjects were informed that they would be paid $1 for every 1000 points they earned during the study, rounded down to the nearest dollar. For example, a subject with a score of 17 232 points would receive $17.00. The minimum payment was set at $12 and the average payment was $18.29 (all results still hold if we exclude the subset of subjects who scored under 12 000 points).
The task was to indicate, using the keyboard, which side of the computer screen contained more flickering dots. The difference in the number of stars between the two sides of the screen was either 10 (hard trials) or 80 (easy trials), with the mean number of stars equal to 100 (e.g. a ‘hard’ trial had 95 versus 105 stars, and an ‘easy’ trial had 60 versus 140 stars).
The study was divided into five blocks. The first of the five blocks was a 5 min unpaid trial block (T) to familiarize subjects with the task, with an ITI of 0.5 s separating trials. The four remaining blocks each took 10 min, with subjects earning points that were later converted to cash. On each trial in these blocks, participants could either gain or lose a specific number of points, which we refer to as the ‘stake’ for that trial. Subjects were self-paced and continued to make decisions until the block time was up. The ITI in the paid blocks was 2 s, plus the time needed to prepare the next trial, resulting in an average empirical ITI of approximately 2.2 s.
Subjects were informed that the stakes corresponded to half the difference in the number of stars between the two sides of the screen. For example, if there were 105 stars on the left and 95 stars on the right, the stakes were 5 (since (105–95)/2 = 5). Since subjects did not know the number of dots in advance, on every trial they had to infer the stakes based on the on-screen stimuli.
As in study 1, there were two within-subject experimental conditions: intervention blocks and non-intervention blocks. Participants were introduced to the two experimental conditions in the following way:
On some runs [blocks], there will be a deadline. If you do not respond by the deadline, the trial will be aborted, and you will earn no points. A short time before the deadline, the stars will disappear - respond quickly when this happens!
The mean deadline was determined in the same way as in study 1. Each trial, the actual deadline was drawn uniformly from within 50 ms of the mean deadline.
(c). Analysis
The mixed-effects regressions reported for studies 1 and 2 use the following model: where β is a vector of coefficients, xij is the vector of regressors in trial j of individual i, vi is an individual-specific noise term and ɛij is a general noise term. For study 2 (electronic supplementary material, table S2, columns 4–6), the dependent variable y is cumulative surplus per block, in points, since every trial was paid in full (1000 points = 1 USD). For study 1 (electronic supplementary material, table S2, columns 1–3), the dependent variable y is the blockwise mean surplus, in CHF per trial. Since there were 100 trials per block, conversion to the block level requires multiplying by 100. Before regressing, the data were first collapsed to obtain blockwise mean surplus for each participant, resulting in four data points per participant (representing blocks 2–5). The mixed-effects regression models for both studies were estimated using maximum likelihood. Standard errors were clustered at the individual level. The first (trial) block was excluded from all analyses.
3. Results
Here, we report the results of two separate decision-making studies, one using an economic value-based food-choice task, and one using a perceptual choice task. Further (consistent) results from a related third study are reported in the electronic supplementary material. In each study, subjects faced a fixed amount of time to make as many decisions as possible.
In the value-based food-choice task (study 1), subjects had to decide which of two food items they would prefer to eat at the end of the experiment (figure 1a). Using the elicited values from a separate valuation task (see Material and methods), we constructed, for each participant, trials in which the valuation difference was large (easy trials), and trials in which the valuation difference was small (hard trials). In each block, there were more decisions (100) than could be made in the time available. Any remaining decisions at the end of the time-limit were made randomly by the computer. At the end of the experiment, subjects received the chosen food item from one randomly chosen trial. Crucially, the trial randomly chosen for payment could be a self-made or computer-made decision. Thus the subjects had an incentive to make as many of their own choices as possible, to reduce the chance of receiving a randomly chosen food item.
In the twinkling-stars task (study 2), subjects had to decide which side of the computer screen contained more dots. The dots disappeared and reappeared at random, giving the appearance of twinkling stars (figure 1b), but the underlying number of stars on each side of the screen remained constant throughout a trial. In each trial we varied the difficulty of the task by changing the difference in the number of stars between the two sides of the screen. Analogous to study 1, a negative link between choice difficulty and relative reward was also induced in study 2. Here, participants received points according to the difference in the number of stars between the two sides of the screen.
Crucially, in both studies 1 and 2, there is more to be gained from making the correct choice in easy trials than in hard trials. By construction, trials with a high relative reward are both easier to answer accurately and have a larger impact on the expected earnings, thus constituting a better investment of time than the trials with a low relative reward.
Despite this reasoning, we found that subjects spent significantly more time on trials with a low relative reward than they did on high relative reward trials in both the food-choice task (figure 2a) and the twinkling-stars task (figure 2b). In the twinkling-stars task, the average, across individuals, of median RT in hard trials was 1.02 s, compared with 0.58 s in the easy trials. A non-parametric Wilcoxon matched-pairs signed-ranks test rejects the hypothesis that these two values are equal at p < 0.0001. In the food-choice task, there is a clear association between the absolute difference in willingness to pay and RT. We see that a CHF 1 increase in corresponded, on average, to a decrease in log(RT) by 0.218 units (p < 0.001, mixed-effects regression, see the electronic supplementary material, table S1).
The RT pattern displayed in figure 2a,b is suggestive, but to truly identify whether behaviour is sub-optimal, more direct evidence is needed. To convincingly demonstrate the sub-optimality of this pattern, it is sufficient to demonstrate that unrestricted performance can be improved upon without using any additional information. Specifically, we hypothesized that it might be possible to improve subjects' performance in these tasks by imposing a per-trial deadline on decision-making.
In both studies, subjects were informed that there would be five blocks of decisions and that each block was time-limited. They were also told that in some blocks there would be within-trial deadlines. Being cut off only occurred in 1.60% and 2.14% of intervention block trials, in studies 1 and 2, respectively. In what follows, we will refer to the blocks with deadlines as intervention blocks (I), and those without as non-intervention blocks (N).
Comparing the performance on the intervention blocks to the non-intervention blocks (excluding the first blocks), we found that the effect of the intervention was beneficial for 80% of participants in study 1 and 60% of participants in Study 2. The intervention helped subjects earn significantly more points in the twinkling-stars task (t39 = −2.2215, p = 0.032, paired) and significantly more value (see Material and methods) in the food-choice task (t48 = −4.1973, p < 0.0001, paired).
Since we have repeated measures, and each subject went through a total of five blocks, it is possible that gaining experience with the task improved performance. To rule out this possible confound, blocks 2–5 were run either in sequence N-I-N-I or in sequence I-N-I-N. Figure 3 shows mean task performance block by block for studies 1 (panel (a)) and 2 (panel (b)). To analyse the benefit of the intervention while statistically controlling for the sequence of blocks, we regressed blockwise performance from blocks 2–5 on both a dummy variable for the intervention blocks, as well as an integer variable (2–5) that encoded the block number. The mixed-effects regression results (see the electronic supplementary material, table S2) show that, all else being equal, with each additional block, earnings increased by 61.86 points (z = 5.90, p < 0.0001) in the twinkling-stars task. In the food-choice task, value per block increased by CHF 1.33 (z = 6.69, p < 0.0001) (if all the trials had been realized; in reality only one trial was realized since we could not give a subject 500 food items to eat). However, the positive effect of the intervention remains significant when we control for experience, with subjects earning an average of 76.28 points more in an intervention block of the twinkling-stars task (z = −2.11, p = 0.035) and a value of CHF 1.69 more in an intervention block of the food-choice task (z = −3.75, p = 0.0002).
Finally, we investigated whether subjects learned from the intervention and so improved in subsequent non-intervention blocks. In order to test for this, while controlling for the effects of experience, we ran a mixed-effects regression (see the electronic supplementary material, table S2, specifications 2 and 5) that included a block-number regressor, as well as dummy variables for intervention blocks and for pre-intervention blocks. The regression results show that performance in post-intervention non-intervention blocks was higher than in pre-intervention blocks, though the effect was only significant in the food-choice task. In the twinkling-stars task, pre-intervention blocks fared 26.32 points worse than the non-intervention blocks that followed (z = −0.40, p = 0.68), while in the food-choice task, pre-intervention blocks fared the equivalent of CHF 3.55 worse (if every trial had been realized) than the post-intervention ones (z = −4.20, p < 0.0001). While intervention blocks continued to outperform post-intervention non-intervention blocks (see the electronic supplementary material, table S2, specifications 3 and 6), this effect was only marginal with a remaining performance increase of 69.58 points per block in the twinkling-stars task (z = 1.64, p = 0.1) and CHF 0.75 per block in the food-choice task (z = 1.33, p = 0.18).
4. Discussion
Here, we have shown that decision-makers are consistently sub-optimal at investing scarce decision time, but that this can be mitigated using a simple intervention where we impose choice deadlines. We observed behaviour that failed to maximize earnings when subjects had to decide how to allocate their time across many binary choices. This finding replicated across two separate studies, one involving a food-choice task, the other involving a perceptual decision-making task. These two studies tell a consistent story, in which people apparently misallocate their time, spending too much on those choice problems in which the relative reward is low.
These findings are economically counterintuitive because we find that imposing an additional constraint (a deadline) onto individual decisions actually improves the overall outcome. Theoretically, the same constraints could have been self-imposed by the subjects. The fact that our intervention improved performance thus means that behaviour was not optimized to maximize earnings in these settings.
This work highlights the fact that the classic speed-accuracy trade-off is an oversimplification of the typical trade-offs faced by organisms in their natural environments [16,17,51]. Each choice is not equally important and so rather than trying to maximize accuracy, organisms should be looking to maximize the relative benefits from their decisions. Optimally behaving organisms should know the relationship between strength of preference and RT, and so infer over time that the current decision is less and less worth making correctly. This idea is captured by the well-known paradox of Buridan's ass, where an ass that is equally hungry and thirsty is placed halfway between a stack of hay and a pail of water, and unable to choose between them, dies. In the SSM framework, collapsing decision thresholds (along with noise in the decision process) allow the organism to avert this deadlock. For example, Seeley et al. [15] describe how bee colonies are able to avoid deadlock when deciding between two equally attractive new hive sites. Others have used SSMs to argue that rats, monkeys and humans use collapsing thresholds, urgency signals or nonlinear dynamics to avoid indecision [20,55,57–59,63–67]. On the other hand, Hawkins et al. [61] have argued that the evidence for such behaviour is not quite so clear and that it may be present only in extensively trained animals.
In any case, even when such mechanisms are present, it remains unclear whether they simply serve to break deadlock or whether they produce optimal time allocation. There has indeed been some suggestion that time allocation is not optimal. For example, in the domain of bee colour discrimination, researchers have found that in some laboratory settings, bees will overemphasize accuracy and improve their flower colour discrimination by an order of magnitude compared to what is typically observed in the field. This enhanced accuracy comes at a substantial time cost; one that is probably higher than the cost of visiting poorly rewarded flowers [51]. Indeed, in an analysis of bee heterogeneity in the flower colour discrimination task, Burns found that the fast, inaccurate bees performed better (in terms of nectar collection rate) than the slow, accurate bees [68]. In another example involving a mouse odour discrimination task, the researchers found that their mice exhibited evidence-sampling times that were independent of the difficulty of the decisions, when instead they might have done better skipping through the difficult trials [69]. However, this evidence is limited both in quantity and in the conclusions that one can draw regarding sub-optimal behaviour.
Our results provide a new source of evidence supporting the view that whatever deadlock-breaking mechanisms exist, they are not maximizing reward rate. While we cannot say whether decision thresholds are collapsing or not, we can say that they are clearly not collapsing optimally. Thus our findings support the view that organisms may be less capable of dealing with these trade-offs than we might have expected. Future research could use a similar choice-deadline procedure to test for sub-optimal time allocation in other species or in highly trained decision-makers.
It is worth noting that we purposefully implemented our intervention soft-handedly. We did this to enable subjects to retain agency and avoid being cut off by the computer. While this procedure clearly improved subjects' outcomes, it is unclear whether the intervention worked primarily by serving as a cue to terminate the decision process, or by altering subjects' thresholds. The comparison between pre- and post-non-intervention blocks suggests that in fact the intervention may have had different effects in the two studies. In study 1, the intervention led to an improvement in non-intervention blocks, suggesting a change in subjects' thresholds, while in study 2, this effect was not significant. The lingering effect of the intervention even in its aftermath, somewhat weakens the normative case for sustaining the intervention.
Also, while the data show that the intervention enhanced the material benefit of the subjects, it is beyond the scope of the current research to evaluate its subjective benefits, all things considered. For example, it is possible that some organisms assign an intrinsic value to being correct [32]. Indeed, research in signal detection theory has demonstrated that people tend to overemphasize accuracy [3,4]. If this intrinsic value was higher for hard problems, as research on achievement motivation indeed suggests (e.g. [70]), this could explain why subjects might allocate more time to them. However, note that this cannot explain why subjects' behaviour improved post-intervention (nor can it explain the lack of a difference between payment schemes documented in study 3, which is reported in the electronic supplementary material). Additionally, a common explanation for the overemphasis on accuracy is that throughout their lives people are reinforced for making correct decisions. However, in our study 1, there is technically no ‘accuracy’, only consistency with the prior WTP for the different items. In similar ‘real world’ settings, organisms are not typically given feedback about the accuracy of such preference-based decisions, since preferences are subjective.
One might wonder if it matters whether the opportunity costs are certain or uncertain. In study 1, subjects were compensated for one randomly selected trial and so it is possible that their motivation to optimize performance was reduced. However in study 2, subjects were compensated for all of their decisions and yet sub-optimal behaviour remained. So while probabilistic outcomes may reduce motivations to optimize, they do not seem to be necessary to observe sub-optimal behaviour.
Taken together, our research here demonstrates a new, simple way to test for sub-optimal behaviour. Rather than taking the traditional modelling approach to derive what optimal behaviour might look like, we instead used experimental manipulations and behavioural interventions to show that subjects' unrestricted behaviour is not pay-off maximizing. Of course, it was the modelling literature that first suggested to us that such inefficiency might exist. We thus close with the hope that this work highlights the important complementarities between theory, modelling and experiments.
Ethics
Study 1 was approved by the University of Zurich ethics committee and all subjects gave informed written consent before participating. Study 2 was approved by the Princeton University IRB and all subjects gave informed written consent before participating. Both experiments were conducted using Matlab Psychophysics Toolbox [71].
Data accessibility
All data are available on Dryad (http://datadryad.org).
Authors' contributions
B.O., I.K. and E.F. designed study 1. B.O. and I.K. conducted study 1. B.O. analysed data of studies 1, 2 and 3. K.M., J.C., M.B. and I.K. designed study 2. K.M. and M.B. designed study 3. K.M. and J.C conducted studies 2 and 3. B.O., I.K., E.F., K.M., J.C. and M.B. wrote the article.
Competing interests
We have no competing interests.
Funding
The research undertaken at the University of Zurich is part of the advanced ERC grant ‘Foundations of Economic Preferences’ and the authors gratefully acknowledge financial support from the European Research Council. The research undertaken at Princeton University was funded by the National Science Foundation (CRCNS 1207833) and the John Templeton Foundation. The opinions expressed in this publication are those of the authors and do not necessarily reflect the views of the funding agencies.
References
- 1.Herrnstein RJ. 1961. Relative and absolute strength of response as a function of frequency of reinforcement. J. Exp. Anal. Behav. 4, 267–272. ( 10.1901/jeab.1961.4-267) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Tversky A, Kahneman D. 1974. Judgment under uncertainty: heuristics and biases. Science 185, 1124–1131. ( 10.1126/science.185.4157.1124) [DOI] [PubMed] [Google Scholar]
- 3.Maddox WT. 2002. Toward a unified theory of decision criterion learning in perceptual categorization. J. Exp. Anal. Behav. 78, 567–595. ( 10.1901/jeab.2002.78-567) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Bohil CJ, Maddox WT. 2003. On the generality of optimal versus objective classifier feedback effects on decision criterion learning in perceptual categorization. Mem. Cogn. 31, 181–198. ( 10.3758/BF03194378) [DOI] [PubMed] [Google Scholar]
- 5.Pachella RG. 1974. The interpretation of reaction time in information processing research. In Human information processing: tutorials in performance and cognition (ed. Kantowitz B.), pp. 41–82. Hillsdale, NJ: Halstead Press. [Google Scholar]
- 6.Wickelgren WA. 1977. Speed-accuracy tradeoff and information processing dynamics. Acta Psychol. 41, 67–85. ( 10.1016/0001-6918(77)90012-9) [DOI] [Google Scholar]
- 7.Bogacz R. 2007. Optimal decision-making theories: linking neurobiology with behaviour. Trends Cogn. Sci. 11, 118–125. ( 10.1016/j.tics.2006.12.006) [DOI] [PubMed] [Google Scholar]
- 8.Busemeyer JR, Rapoport A. 1988. Psychological models of deferred decision making. J. Math. Psychol. 32, 91–134. ( 10.1016/0022-2496(88)90042-9) [DOI] [Google Scholar]
- 9.Frazier P, Yu AJ. 2008. Sequential hypothesis testing under stochastic deadlines. In Proc. of Neural Inf. Process. Syst. 2007 (eds JC Platt, Y Singer, ST Roweis), pp. 465–472.
- 10.Ratcliff R. 1978. A theory of memory retrieval. Psychol. Rev. 85, 59 ( 10.1037/0033-295X.85.2.59) [DOI] [PubMed] [Google Scholar]
- 11.Ratcliff R, McKoon G. 2008. The diffusion decision model: theory and data for two-choice decision tasks. Neural Comput. 20, 873–922. ( 10.1162/neco.2008.12-06-420) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Wald A, Wolfowitz J. 1948. Optimum character of the sequential probability ratio test. Ann. Math. Stat. 19, 326–339. ( 10.1214/aoms/1177730197) [DOI] [Google Scholar]
- 13.DasGupta S, Ferreira CH, Miesenböck G. 2014. FoxP influences the speed and accuracy of a perceptual decision in Drosophila. Science 344, 901–904. ( 10.1126/science.1252114) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Stroeymeyt N, Giurfa M, Franks NR. 2010. Improving decision speed, accuracy and group cohesion through early information gathering in house-hunting ants. PLoS ONE 5, e13059 ( 10.1371/journal.pone.0013059) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Seeley TD, Visscher PK, Schlegel T, Hogan PM, Franks NR, Marshall JAR. 2012. Stop signals provide cross inhibition in collective decision-making by honeybee swarms. Science 335, 108–111. ( 10.1126/science.1210361) [DOI] [PubMed] [Google Scholar]
- 16.Pais D, Hogan PM, Schlegel T, Franks NR, Leonard NE, Marshall JAR. 2013. A mechanism for value-sensitive decision-making. PLoS ONE 8, e73216 ( 10.1371/journal.pone.0073216) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Pirrone A, Stafford T, Marshall JA. 2014. When natural selection should optimize speed-accuracy trade-offs. Front. Neurosci. 8, 73 ( 10.3389/fnins.2014.00073) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Chittka L, Dyer AG, Bock F, Dornhaus A. 2003. Psychophysics: bees trade off foraging speed for accuracy. Nature 424, 388 ( 10.1038/424388a) [DOI] [PubMed] [Google Scholar]
- 19.Kaneko H, Tamura H, Kawashima T, Suzuki SS. 2006. A choice reaction-time task in the rat: a new model using air-puff stimuli and lever-release responses. Behav. Brain Res. 174, 151–159. ( 10.1016/j.bbr.2006.07.020) [DOI] [PubMed] [Google Scholar]
- 20.Brunton BW, Botvinick MM, Brody CD. 2013. Rats and humans can optimally accumulate evidence for decision-making. Science 340, 95–98. ( 10.1126/science.1233912) [DOI] [PubMed] [Google Scholar]
- 21.Roitman JD, Shadlen MN. 2002. Response of neurons in the lateral intraparietal area during a combined visual discrimination reaction time task. J. Neurosci. 22, 9475–9489. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Hanes DP, Schall JD. 1996. Neural control of voluntary movement initiation. Science 274, 427–430. ( 10.1126/science.274.5286.427) [DOI] [PubMed] [Google Scholar]
- 23.Wong K-F, Huk AC, Shadlen MN, Wang X-J. 2007. Neural circuit dynamics underlying accumulation of time-varying evidence during perceptual decision making. Front. Comput. Neurosci. 1, 6 ( 10.3389/neuro.10.006.2007) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Ratcliff R, Cherian A, Segraves M. 2003. A comparison of macaque behavior and superior colliculus neuronal activity to predictions from models of two-choice decisions. J. Neurophysiol. 90, 1392–1407. ( 10.1152/jn.01049.2002) [DOI] [PubMed] [Google Scholar]
- 25.Ding L, Gold JI. 2012. Separate, causal roles of the caudate in saccadic choice and execution in a perceptual decision task. Neuron 75, 865–874. ( 10.1016/j.neuron.2012.07.021) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Purcell BA, Schall JD, Logan GD, Palmeri TJ. 2012. From salience to saccades: multiple-alternative gated stochastic accumulator model of visual search. J. Neurosci. 32, 3433–3446. ( 10.1523/JNEUROSCI.4622-11.2012) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Bogacz R, Gurney K. 2007. The basal ganglia and cortex implement optimal decision making between alternative actions. Neural Comput. 19, 442–477. ( 10.1162/neco.2007.19.2.442) [DOI] [PubMed] [Google Scholar]
- 28.Gold JI, Shadlen MN. 2007. The neural basis of decision making. Annu. Rev. Neurosci. 30, 535–574. ( 10.1146/annurev.neuro.29.051605.113038) [DOI] [PubMed] [Google Scholar]
- 29.Balci F, Simen P, Niyogi R, Saxe A, Hughes JA, Holmes P, Cohen JD. 2010. Acquisition of decision making criteria: reward rate ultimately beats accuracy. Atten. Percept. Psychophys. 73, 640–657. ( 10.3758/s13414-010-0049-7) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Forstmann BU, Anwander A, Schafer A, Neumann J, Brown S, Wagenmakers E-J, Bogacz R, Turner R. 2010. Cortico-striatal connections predict control over speed and accuracy in perceptual decision-making. Proc. Natl Acad. USA 107, 15 916–15 920. ( 10.1073/pnas.1004932107) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Starns JJ, Ratcliff R. 2010. The effects of aging on the speed–accuracy compromise: boundary optimality in the diffusion model. Psychol. Aging 25, 377–390. ( 10.1037/a0018022) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Starns JJ, Ratcliff R. 2011. Age-related differences in diffusion model boundary optimality with both trial-limited and time-limited tasks. Psychon Bull. Rev. 19, 139–145. ( 10.3758/s13423-011-0189-3) [DOI] [PubMed] [Google Scholar]
- 33.Usher M, McClelland JL. 2001. The time course of perceptual choice: the leaky, competing accumulator model. Psychol. Rev. 108, 550–592. ( 10.1037//0033-295X.108.3.550) [DOI] [PubMed] [Google Scholar]
- 34.Wenzlaff H, Bauer M, Maess B, Heekeren HR. 2011. Neural characterization of the speed-accuracy tradeoff in a perceptual decision-making task. J. Neurosci. 31, 1254–1266. ( 10.1523/JNEUROSCI.4000-10.2011) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Busemeyer JR. 1985. Decision making under uncertainty: a comparison of simple scalability, fixed-sample, and sequential-sampling models. J. Exp. Psychol. Learn. Mem. Cogn. 11, 538–564. ( 10.1037/0278-7393.11.3.538) [DOI] [PubMed] [Google Scholar]
- 36.Cavanagh JF, Wiecki TV, Kochar A, Frank MJ. 2014. Eye tracking and pupillometry are indicators of dissociable latent decision processes. J. Exp. Psychol. Gen. 143: 1476–1488. ( 10.1037/a0035813) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.De Martino B, Fleming SM, Garrett N, Dolan RJ. 2013. Confidence in value-based choice. Nat. Neurosci. 16, 105–110. ( 10.1038/nn.3279) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Dickhaut J, Rustichini A, Smith V. 2009. A neuroeconomic theory of the decision process. Proc. Natl Acad. Sci. USA 106, 22 145–22 150. ( 10.1073/pnas.0912500106) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Diederich A. 1997. Dynamic stochastic models for decision making under time constraints. J. Math. Psychol. 41, 260–274. ( 10.1006/jmps.1997.1167) [DOI] [PubMed] [Google Scholar]
- 40.Hunt LT, Kolling N, Soltani A, Woolrich MW, Rushworth MFS, Behrens TEJ. 2012. Mechanisms underlying cortical activity during value-guided choice. Nat. Neurosci. 15, 470–476. ( 10.1038/nn.3017) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Johnson JG, Busemeyer JR. 2005. A dynamic, stochastic, computational model of preference reversal phenomena. Psychol. Rev. 112, 841–861. ( 10.1037/0033-295X.112.4.841) [DOI] [PubMed] [Google Scholar]
- 42.Krajbich I, Oud B, Fehr E. 2014. Benefits of neuroeconomic modeling: new policy interventions and predictors of preference. Am. Econ. Rev. Papers Proc. 104, 501–506. ( 10.1257/aer.104.5.501) [DOI] [Google Scholar]
- 43.Krajbich I, Armel C, Rangel A. 2010. Visual fixations and the computation and comparison of value in simple choice. Nat. Neurosci. 13, 1292–1298. ( 10.1038/nn.2635) [DOI] [PubMed] [Google Scholar]
- 44.Krajbich I, Lu D, Camerer C, Rangel A. 2012. The attentional drift-diffusion model extends to simple purchasing decisions. Front. Psychol. 3, 193 ( 10.3389/fpsyg.2012.00193) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Krajbich I, Rangel A. 2011. Multialternative drift-diffusion model predicts the relationship between visual fixations and choice in value-based decisions. Proc Natl Acad. Sci. USA 108, 13 852–13 857. ( 10.1073/pnas.1101328108) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Milosavljevic M, Koch C, Rangel A. 2011. Consumers can make decisions in as little as a third of a second. Judgment Decis. Mak. 6, 520–530. [Google Scholar]
- 47.Rodriguez CA, Turner BM, McClure SM. 2014. Intertemporal choice as discounted value accumulation. PLoS ONE 9, e90138 ( 10.1371/journal.pone.0090138) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Roe RM, Busemeyer JR, Townsend JT. 2001. Multialternative decision field theory: a dynamic connectionst model of decision making. Psychol. Rev. 108, 370–392. ( 10.1037/0033-295X.108.2.370) [DOI] [PubMed] [Google Scholar]
- 49.Towal RB, Mormann M, Koch C. 2013. Simultaneous modeling of visual saliency and value computation improves predictions of economic choice. Proc. Natl Acad. Sci. USA 110, E3858–E3867. ( 10.1073/pnas.1304429110) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Hayden BY, Pearson JM, Platt ML. 2011. Neuronal basis of sequential foraging decisions in a patchy environment. Nat. Neurosci. 14, 933–939. ( 10.1038/nn.2856) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Chittka L, Skorupski P, Raine NE. 2009. Speed–accuracy tradeoffs in animal decision making. Trends Ecol. Evol. 24, 400–407. ( 10.1016/j.tree.2009.02.010) [DOI] [PubMed] [Google Scholar]
- 52.Bogacz R, Brown E, Moehlis J, Holmes P, Cohen JD. 2006. The physics of optimal decision making: a formal analysis of models of performance in two-alternative forced choice tasks. Psychol. Rev. 113, 700–765. ( 10.1037/0033-295X.113.4.700) [DOI] [PubMed] [Google Scholar]
- 53.Rustichini A. 2009. Neuroeconomics: formal models of decision making and cognitive neuroscience. In Neuroeconomics: decision making and the brain (eds Glimcher PW, Camerer CF, Fehr E), pp. 33–46. London, UK: Academic Press. [Google Scholar]
- 54.Rapoport A, Burkheimer GJ. 1971. Models for deferred decision making. J. Math. Psychol. 8, 508–538. ( 10.1016/0022-2496(71)90005-8) [DOI] [Google Scholar]
- 55.Cisek P, Puskas GA, El-Murr S. 2009. Decisions in changing conditions: the urgency-gating model. J. Neurosci. 29, 11 560–11 571. ( 10.1523/JNEUROSCI.1844-09.2009) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Edwards W. 1965. Optimal strategies for seeking information: models for statistics, choice reaction times and human information processing. J. Math. Psychol. 2, 312–329. ( 10.1016/0022-2496(65)90007-6) [DOI] [Google Scholar]
- 57.Hanks TD, Mazurek ME, Kiani R, Hopp E, Shadlen MN. 2011. Elapsed decision time affects the weighting of prior probability in a perceptual decision task. J. Neurosci. 31, 6339–6352. ( 10.1523/JNEUROSCI.5613-10.2011) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Drugowitsch J, Moreno-Bote R, Churchland AK, Shadlen MN, Pouget A. 2012. The cost of accumulating evidence in perceptual decision making. J. Neurosci. 32, 3612–3628. ( 10.1523/JNEUROSCI.4010-11.2012) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Polania R, Krajbich I, Grueschow M, Ruff CC. 2014. Neural oscillations and synchronization differentially support evidence accumulation in perceptual and value-based decision making. Neuron 82, 709–720. ( 10.1016/j.neuron.2014.03.014) [DOI] [PubMed] [Google Scholar]
- 60.Milosavljevic M, Malmaud J, Huth A, Koch C, Rangel A. 2010. The drift diffusion model can account for the accuracy and reaction time of value-based choices under high and low time pressure. Judgment Decis. Mak. 5, 437–449. ( 10.2139/ssrn.1901533) [DOI] [Google Scholar]
- 61.Hawkins GE, Forstmann BU, Wagenmakers E-J, Ratcliff R, Brown SD. 2015. Revisiting the evidence for collapsing boundaries and urgency signals in perceptual decision-making. J. Neurosci. 35, 2476–2484. ( 10.1523/JNEUROSCI.2410-14.2015) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Becker GM, Degroot MH, Marschak J. 1964. Measuring utility by a single-response sequential method. Behav. Sci. 9, 226–232. ( 10.1002/bs.3830090304) [DOI] [PubMed] [Google Scholar]
- 63.Mormann MM, Malmaud J, Huth A, Koch C, Rangel A. 2010. The drift diffusion model can account for the accuracy and reaction time of value-based choices under high and low time pressure. Judgm. Decis. Mak. 5, 437–449. [Google Scholar]
- 64.Bowman NE, Kording KP, Gottfried JA. 2012. Temporal integration of olfactory perceptual evidence in human orbitofrontal cortex. Neuron 75, 916–927. ( 10.1016/j.neuron.2012.06.035) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Ditterich J. 2006. Stochastic models of decisions about motion direction: behavior and physiology. Neural Netw. 19, 981–1012. ( 10.1016/j.neunet.2006.05.042) [DOI] [PubMed] [Google Scholar]
- 66.Thura D, Beauregard-Racine J, Fradet C-W, Cisek P. 2012. Decision making by urgency gating: theory and experimental support. J. Neurophysiol. 108, 2912–2930. ( 10.1152/jn.01071.2011) [DOI] [PubMed] [Google Scholar]
- 67.Shadlen MN, Kiani R. 2013. Decision making as a window on cognition. Neuron 80, 791–806. ( 10.1016/j.neuron.2013.10.047) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Burns JG. 2005. Impulsive bees forage better: the advantage of quick, sometimes inaccurate foraging decisions. Anim. Behav. 70, e1–e5. ( 10.1016/j.anbehav.2005.06.002) [DOI] [Google Scholar]
- 69.Rinberg D, Koulakov A, Gelperin A. 2006. Speed-accuracy tradeoff in olfaction. Neuron 51, 351–358. ( 10.1016/j.neuron.2006.07.013) [DOI] [PubMed] [Google Scholar]
- 70.Matsui T, Okada A, Mizuguchi R. 1981. Expectancy theory prediction of the goal theory postulate,’ The harder the goals, the higher the performance’. J. Appl. Psychol. 66, 54 ( 10.1037/0021-9010.66.1.54) [DOI] [Google Scholar]
- 71.Brainard DH. 1997. The psychophysics toolbox. Spatial Vis. 10, 433–436. ( 10.1163/156856897X00357) [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
All data are available on Dryad (http://datadryad.org).