Abstract
In order to discover the most rewarding actions, agents must collect information about their environment, potentially foregoing reward. The optimal solution to this “explore-exploit” dilemma is often computationally challenging, but principled algorithmic approximations exist. These approximations utilize uncertainty about action values in different ways. Some random exploration algorithms scale the level of choice stochasticity with the level of uncertainty. Other directed exploration algorithms add a “bonus” to action values with high uncertainty. Random exploration algorithms are sensitive to total uncertainty across actions, whereas directed exploration algorithms are sensitive to relative uncertainty. This paper reports a multi-armed bandit experiment in which total and relative uncertainty were orthogonally manipulated. We found that humans employ both exploration strategies, and that these strategies are independently controlled by different uncertainty computations.
Keywords: explore-exploit dilemma, reinforcement learning, Bayesian inference
Introduction
Uncertainty lies at the heart of decision making in the real world. A bee in search of nectar and a venture capitalist in search of an investment both need to explore their options in order to reduce their uncertainty, at the expense of exploiting the currently best option. In the absence of uncertainty, no explore-exploit dilemma would exist. The question addressed here is how humans use uncertainty to guide exploration.
The optimal solution to the explore-exploit dilemma is, except for some special cases (e.g., in foraging theory; Charnov, 1976; Stephens & Krebs, 1986), computationally intractable, but computer scientists have developed algorithmic approximations that can provably approach optimal behavior (Sutton & Barto, 1998). Psychologists have also studied algorithmic hypotheses about how humans balance exploration and exploitation (Cohen, McClure, & Yu, 2007; Hills et al., 2015; Mehlhorn et al., 2015), but only recently have the links between modern machine learning algorithms and psychological processes been systematically investigated (Gershman, 2018; Schulz, Konstantinidis, & Speekenbrink, 2017; Speekenbrink & Konstantinidis, 2015). Key to this synthesis is the idea that uncertainty can guide exploration in two qualitatively different ways: by adding randomness into choice behavior, or by directing choice towards uncertain options.
The pioneering work of Wilson, Geana, White, Ludvig, and Cohen (2014) demonstrated that humans use both random and directed exploration in a carefully designed two-armed bandit task. When the subject had more experience with one option (and hence less uncertainty), she favored the more uncertain option, indicating a form of directed exploration. In addition, subjects increased the randomness in their choices when they were more uncertain, spreading their choices across both high- and low-value options. Directed and random exploration strategies are dissociable, developing on different timescales across the lifespan (Somerville et al., 2017) and relying on different neural substrates (Zajkowski, Kossut, & Wilson, 2017).
The directed/random distinction is closely related to the distinction between two families of exploration algorithms that have been studied extensively in machine learning. Directed exploration can be realized by adding uncertainty bonuses to estimated values (Auer, Cesa-Bianchi, & Fischer, 2002; Brafman & Tennenholtz, 2002; Dayan & Sejnowski, 1996; Kolter & Ng, 2009; Srinivas, Krause, Seeger, & Kakade, 2010). One of the most prominent versions of this approach is known as the upper confidence bound (UCB) algorithm (Auer et al., 2002), which chooses action at on trial t according to:
(1) |
where k indexes actions and Ut(k) is the upper confidence bound that plays the role of an uncertainty bonus. Under a Bayesian analysis (Srinivas et al., 2010), Qt(k) corresponds to the posterior mean and the uncertainty bonus is proportional to the posterior standard deviation σt(k).
Random exploration has an even older pedigree in machine learning, dating back to Thompson’s work in the 1930s (Thompson, 1933). In what is now known as Thompson sampling, a random value function from the posterior is drawn and then the agent chooses greedily with respect to the random draw. Like UCB, Thompson sampling uses uncertainty to promote exploration, but does so by encouraging stochasticity (i.e., a form of probability matching) rather than via a response bias.1
In psychophysical terms, uncertainty in Thompson sampling changes the slope of the function relating action values to choice probabilities, whereas uncertainty in UCB changes the intercept (indifference point).2 The role of uncertainty in Thompson sampling can be understood by recognizing that sampling from a distribution over values implies that variability in this distribution directly translates to variability in choices. By contrast, the role of uncertainty in UCB can be understood by recognizing that the uncertainty bonus acts in the same way as a boost in reward, shifting choice probabilities towards the more uncertain option without altering the level of choice stochasticity.
Recognizing the dissociable (and possibly complementary) nature of directed and random exploration, computer scientists have also constructed hybrids of UCB and Thompson sampling (Chapelle & Li, 2011; May, Korda, Lee, & Leslie, 2012). A recent report (Gershman, 2018) provided the most direct evidence for such hybrids in human decision making, demonstrating that uncertainty influences both the intercept and slope of the choice probability function. Critically, the intercept and slope effects derive from different uncertainty computations: the intercept is governed by relative uncertainty (the difference in posterior uncertainty between the two options, defined formally below), whereas the slope is governed by total uncertainty (the sum of posterior uncertainty across the options). This suggests that experimental manipulations of these two factors should produce dissociable effects.
Here we pursue this line of reasoning using a two-armed bandit task, with the additional twist that we inform subjects about the riskiness of each arm (Figure 1). On each trial, participants were given a choice between two arms, labeled either as “safe” (S) or “risky” (R). The safe arms always delivered the same reward, whereas the risky arms delivered stochastic rewards (with Gaussian noise). Denoting trial types by compound labels (e.g., “SR” denote trials in which the left arm is safe and the right arm is risky), we used the comparison between preference for arm 1 on SR and RS trials to isolate the effects of relative uncertainty, holding total uncertainty fixed, and the comparison between preference for arm 1 on SS and RR trials to isolate the effects of total uncertainty, holding relative uncertainty fixed (Figure 2). By independently manipulating relative uncertainty (SR vs. RS) and total uncertainty (SS vs. RR), we can go beyond the correlational findings of Gershman (2018) to causally test the predictions of UCB and Thompson sampling. In particular, we predicted that the SS condition would increase the slope of the choice probability function (affecting random but not directed exploration) relative to the RR condition, whereas the RS condition would shift the intercept of the choice probability function (affecting directed but not random exploration) relative to the SR condition.
Materials and Methods
Code and data for reproducing all analyses reported in this paper, as well as Javascript code for re-running the experiments, are available at https://github.com/sjgershm/exploration_uncertainty.
Subjects
Forty-six subjects were recruited via the Amazon Mechanical Turk web service and paid $2.00. The sample size was chosen to be comparable to previous studies using a similar experimental paradigm (Gershman, 2018). The experiments were approved by the Harvard Institutional Review Board.
Stimuli and Procedure
Participants played 30 two-armed bandits, each for one block of 10 trials. On each trial, participants chose one of the arms and received reward feedback (points). They were instructed to choose the “slot machine” (corresponding to an arm) that maximizes their total points. On each block, the mean reward μ(k) for each arm was drawn from a Gaussian with mean 0 and variance . The arms were randomly designated “safe” or “risky,” indicated by an S or R (respectively), and these designations were randomly resampled after a block transition. When participants chose the risky arm, they received stochastic rewards drawn from a Gaussian with mean μ(R) and variance τ2(R) = 16. When participants chose the safe arm, they received a reward of μ(S).
The exact instructions for participants were as follows:
In this task, you have a choice between two slot machines, represented by colored buttons. When you click one of the buttons, you will win or lose points. Choosing the same slot machine will not always give you the same points, but one slot machine is always better than the other. Your goal is to choose the slot machine that will give you the most points. After making your choice, you will receive feedback about the outcome. Sometimes the machines are “safe” (always delivering the same feedback), and sometimes the machines are “risky” (delivering variable feedback). Before you make a choice, you will get information about each machine: “S” indicates SAFE, “R” indicates RISKY. Note that safe/risky is independent of how rewarding a machine is: a risky machine may deliver more points on average than a safe machine, and vice versa. You cannot predict how good a machine is based on whether it is safe or risky. You will play 30 games, each with a different pair of slot machines. Each game will consist of 10 trials.
Belief updating model
To derive estimates of expected value and uncertainty, we assume that subjects approximate an ideal Bayesian learner. Given the Gaussian distributional structure underlying our task, the posterior over the value of arm k is Gaussian with mean Qt(k) and variance . These suffcient statistics can be recursively updated using the Kalman filtering equations:
(2) |
(3) |
where at is the chosen arm, rt is the received reward, and the learning rate αt is given by:
(4) |
Note that only the chosen option’s mean and variance are updated after each trial. The initial values were set to the prior means, Q1(k) = 0 for all k, and the initial variances were set to the prior variance, . The value of τ2 was set to 16 (its true value) for risky options, and to 0.00001 for safe options (to avoid numerical issues, the variance was not set exactly to 0). Although the Kalman filter is an idealization of human learning, it has been shown to account well for human behavior in bandit tasks (Daw, O’doherty, Dayan, Seymour, & Dolan, 2006; Gershman, 2018; Schulz, Konstantinidis, & Speekenbrink, 2015; Speekenbrink & Konstantinidis, 2015).
Choice probability analysis
Gershman (2018) showed that Thompson sampling, UCB, and a particular hybrid of the two imply a probit regression model of choice:
(5) |
where Φ(·) is the cumulative distribution function of the standard Gaussian distribution (mean 0 and variance 1), and the regressors are defined as follows:
Estimated value difference, Vt = Qt(1) − Qt(2).
Relative uncertainty, RUt = σt(1) – σt(2).
Total uncertainty, .
As shown in Gershman (2018), Thompson sampling predicts a significant positive effect of V/TU on choice probability, but not of RU or V, whereas UCB predicts a significant positive effect of both V and RU, but not of V/TU (Figure 2, top left). We used maximum likelihood estimation to fit the coe cients (w) in the probit regression model.
In addition to this model-based analysis, we analyzed choices as a function of experimental condition (i.e., RS, SR, RR, SS).
(6) |
where πtj = 1 if trial t is assigned to condition j, and 0 otherwise. We refer to the w1 terms as intercepts and the w2 terms as slopes.3
Response time analysis
We examined response times as an additional source of evidence about exploration strategies. Our hypotheses are motivated by a sequential sampling framework, according to which the value difference between two options drives a stochastic accumulator until it reaches a decision threshold (Busemeyer & Townsend, 1993; Krajbich, Armel, & Rangel, 2010; Milosavljevic, Malmaud, Huth, Koch, & Rangel, 2010; Ratcliff & Frank, 2012; Summerfield & Tsetsos, 2012). Formally, the decision variable μt(τ) evolves over time (within a trial, index by τ) as follows:
(7) |
where Vt (the value difference between options 1 and 2 defined above) is a deterministic “drift” term, and ϵ(τ) is a stochastic “diffusion” term. Without loss of generality, it is conventional to assume that ϵ(τ) is drawn from a standard Gaussian distribution (mean of 0 and variance of 1). The decision variable evolves until it hits one of two thresholds ±B (corresponding to the two options), at which point a decision is made. If we assume, following the logic of the previous section, that both UCB and Thompson sampling are implemented by a linear transformation of the value difference (see Gershman, 2018, for more details), then the decision variable will evolve according to:
(8) |
where , and β and γ are coe cients controlling the relative contribution of random and directed exploration, respectively. Under this model, the expected response time is given analytically by:
(9) |
where tanh is the hyperbolic tangent function. Since the expected response time is a monotonically decreasing function of , it decreases with RU and increases with TU (see also Busemeyer & Townsend, 1993).
Results
The hybrid random/directed exploration model hypothesizes that all three computational regressors (V, RU, V/TU) should be predictors of choice. We therefore confirmed that maximum likelihood estimates of the corresponding coe cients were significantly greater than 0 [Figure 3; V: t(13797) = 16.48, p < 0.0001; RU: t(13797) = 4.9, p < 0.0001; V/TU: t(13797) = 6.04, p < 0.0001]. Model comparison using the Bayesian and Akaike information criteria strongly favored this three-parameter model over a one-parameter model with only 12 the V regressor, a one-parameter model with V/TU, and a two-parameter model with V and RU (Table 1). This supports previous results showing that humans use both random and directed exploration strategies (Gershman, 2018; Somerville et al., 2017; Wilson et al., 2014).
Table 1: Model comparison results.
Model | BIC | AIC |
---|---|---|
V | 13493 | 13478 |
V/TU | 17262 | 17247 |
V + RU | 12874 | 12836 |
V + RU + V/TU | 12689 | 12621 |
We next addressed the central question of the paper: can random and directed exploration be independently manipulated? As predicted, changing relative uncertainty (RS vs. SR) alter the intercept of the choice probability function (Figure 3): there was a significant difference between the intercepts for the RS and SR conditions [F (1, 13792) = 13.92, p < 0.001]. Furthermore, the RS intercept was significantly greater than 0 [t(13792) = 2.33, p < 0.02] and the SR intercept was significantly less than 0 [t(13792) = 4.47, p < 0.001], indicating an uncertainty-directed choice bias, as predicted by the theory. In other words, uncertainty boosted the value of an arm, shifting choice probability towards that arm. Critically, there was no significant effect of total uncertainty [RR vs. SS, p = 0.40], consistent with the hypothesis that uncertainty-directed biases only emerge when there is a difference in relative uncertainty.
The pattern is flipped when we inspect the slope parameter estimates (Figure 4): increasing total uncertainty (RR vs. SS) reduced the slope [F (1, 13792) = 20.12, p < 0.0001], but relative uncertainty (RS vs. SR) did not have a significant effect (p = 0.72). This finding is consistent with our hypothesis that the random component of exploration would be specifically sensitive to changes in total uncertainty. In other words, the decreased slope for RR indicates greater choice stochasticity when total uncertainty was higher. An alternative possibility is that the increased level of payoff variance in RR compared to SS causes value differences to be more strongly driven by payoff noise. In other words, even if subjects were not using random exploration, they might still show a difference in choice behavior between the RR and SS conditions. However, this hypothesis cannot account completely for the RR vs. SS effect, because our analysis quantifies how choice probability differs between conditions for the same estimated value difference. Thus, even if the conditions differed in their effects on value learning, this analysis controls for such differences.
Sequential sampling models have also been shown to jointly predict choice and response time in reinforcement learning tasks, where the value is dynamically updated based on experience (Frank et al., 2015; Millner, Gershman, Nock, & den Ouden, 2018; Pedersen, Frank, & Biele, 2017). Within the sequential sampling framework, a directed exploration strategy predicts that response times should be faster for risky choices than for safe choices on RS and SR trials, because uncertainty acts as a bonus added to the values (see Materials and Methods). This prediction was confirmed in our data [Figure 5; t(45) = 2.01, p < 0.05].
When uncertainty is equated across the arms, directed exploration does not predict any difference in response time as a function of total uncertainty. Random exploration strategies, in contrast, correctly predict that total uncertainty will act divisively on values, thereby slowing response times when both arms are risky compared to when both arms are safe [Figure 5; t(45) = 2.28, p < 0.05].
Discussion
By separately manipulating relative and total uncertainty, we were able to independently influence directed and random exploration, lending support to the contention that these strategies coexist and jointly determine exploratory behavior (Gershman, 2018; Wilson et al., 2014). In particular, increasing relative uncertainty by making one option riskier than the other caused participants to shift their preference towards the risky option in a value-independent manner, consistent with a change in the intercept (indifference point) of the choice probability function. This manipulation had no effect on the slope of the choice probability function. In contrast, increasing total uncertainty by making both options risky decreased the slope relative to when both options were safe, with no effect on the intercept. In other words, increasing total uncertainty caused choices to become more stochastic.
This dissociation between strategies guiding choice behavior was mirrored by a dissociation in response times. When one option was riskier than the other, subjects were faster in choosing the risky option. However, the same subjects were slower when both options were risky compared to when both options were safe. Taken together, these findings demonstrate that risk can have qualitatively different effects on both choice and response time, depending on the underlying uncertainty computation (i.e., total vs. relative). Consistent with our findings, previous work has shown that payoff variance increases exploration (Lejarraga, Hertwig, & Gonzalez, 2012; Wulff, Mergenthaler-Canseco, & Hertwig, 2018). However, this work did not distinguish between random and directed exploration.
We note here that preference for the risky option appears to be in opposition to what would be predicted by theories of risk aversion, as pointed out by previous authors (Gershman, 2018; Payzan-LeNestour & Bossaerts, 2012; Wilson et al., 2014). This does not mean that our subjects were not risk averse; rather, we suspect that the imperative to explore in a multi-armed bandit task overwhelmed the tendency to avoid risks seen in typical gambling paradigms.
The role of uncertainty-guided exploration has come to occupy an increasingly important place in theories of reinforcement learning (Gershman & Niv, 2015; Knox, Otto, Stone, & Love, 2011; Navarro, Newell, & Schulze, 2016; Payzan-LeNestour & Bossaerts, 2011; Pearson, Hayden, Raghavachari, & Platt, 2009; Schulz et al., 2015; Speekenbrink & Konstantinidis, 2015; Zhang & Yu, 2013), superseding earlier models of exploratory choice based on a fixed source of decision noise, as in ϵ-greedy and softmax policies (e.g., Daw et al., 2006). This shift has been accompanied by a deeper understanding of how reinforcement learning circuits in the basal ganglia compute, represent, and transmit uncertainty to downstream decision making circuits (Gershman, 2017; Lak, Nomoto, Keramati, Sakagami, & Kepecs, 2017; Starkweather, Babayan, Uchida, & Gershman, 2017). Despite this progress, we are only beginning to understand how these circuits implement dissociable channels regulating directed and random exploration (Warren et al., 2017; Zajkowski et al., 2017). Furthermore, a number of inconsistencies exist in the literature; for example, some studies do not find evidence for uncertainty bonuses in exploration (Daw et al., 2006; Riefer, Prior, Blair, Pavey, & Love, 2017). Resolving these inconsistencies will require a more complete picture of the different factors governing directed and random exploration, including individual differences, internal states (e.g., stress, cognitive load), and task structure.
Acknowledgments
I am grateful to Eric Schulz for helpful discussions. This research was supported by the NSF Collaborative Research in Computational Neuroscience (CRCNS) Program Grant R01–1207833 and the Office of Naval Research under grant N000141712984.
Footnotes
The idea that uncertainty promotes choice stochasticity is present in some theories of decision making, notably Decision Field Theory (Busemeyer & Townsend, 1993). Thompson sampling di_ers formally in that stochasticity is driven by posterior uncertainty (see the formal description below), whereas in Decision Field Theory it is driven by payoff variance. Typically uncertainty and payoff variance are correlated, a fact that we exploit in our experimental design (see also Leuker, Pachur, Hertwig, & Pleskac, 2018).
We have assumed here, following Gershman (2018), that some noise is added to the values that enter into the UCB computation. Without this assumption (or some other sources of stochasticity), we would not be able to capture variability in choices.
Note that although this analysis is “model-free” in the sense that it does not use the computationally derived uncertainty regressors, it is still dependent on the model-based value estimates.
References
- Auer P, Cesa-Bianchi N, & Fischer P (2002). Finite-time analysis of the multiarmed bandit problem. Machine Learning, 47, 235–256. [Google Scholar]
- Brafman RI, & Tennenholtz M (2002). R-max-a general polynomial time algorithm for near-optimal reinforcement learning. Journal of Machine Learning Research, 3, 213–231. [Google Scholar]
- Busemeyer JR, & Townsend JT (1993). Decision field theory: a dynamic-cognitive approach to decision making in an uncertain environment. Psychological Review, 100, 432–459. [DOI] [PubMed] [Google Scholar]
- Chapelle O, & Li L (2011). An empirical evaluation of Thompson sampling. In Advances in neural information processing systems (pp. 2249–2257). [Google Scholar]
- Charnov EL (1976). Optimal foraging, the marginal value theorem. Theoretical Population Biology, 9, 129–136. [DOI] [PubMed] [Google Scholar]
- Cohen JD, McClure SM, & Yu AJ (2007). Should i stay or should i go? how the human brain manages the trade-off between exploitation and exploration. Philosophical Transactions of the Royal Society of London B: Biological Sciences, 362, 933–942. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Daw ND, O’doherty JP, Dayan P, Seymour B, & Dolan RJ (2006). Cortical substrates for exploratory decisions in humans. Nature, 441, 876–879. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dayan P, & Sejnowski TJ (1996). Exploration bonuses and dual control. Machine Learning, 25, 5–22. [Google Scholar]
- Frank MJ, Gagne C, Nyhus E, Masters S, Wiecki TV, Cavanagh JF, & Badre D (2015). fmri and eeg predictors of dynamic decision parameters during human reinforcement learning. Journal of Neuroscience, 35, 485–494. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gershman SJ (2017). Dopamine, inference, and uncertainty. Neural Computation, 29, 3311–3326. [DOI] [PubMed] [Google Scholar]
- Gershman SJ (2018). Deconstructing the human algorithms for exploration. Cognition, 173, 34–42. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gershman SJ, & Niv Y (2015). Novelty and inductive generalization in human reinforcement learning. Topics in Cognitive Science, 7, 391–415. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hills TT, Todd PM, Lazer D, Redish AD, Couzin ID, & Group CSR (2015). Exploration versus exploitation in space, mind, and society. Trends in Cognitive Sciences, 19, 46–54. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Knox WB, Otto AR, Stone P, & Love BC (2011). The nature of belief-directed exploratory choice in human decision-making. Frontiers in Psychology, 2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kolter JZ, & Ng AY (2009). Near-Bayesian exploration in polynomial time. In Proceedings of the 26th annual international conference on machine learning (pp. 513–520). [Google Scholar]
- Krajbich I, Armel C, & Rangel A (2010). Visual fixations and the computation and comparison of value in simple choice. Nature Neuroscience, 13, 1292. [DOI] [PubMed] [Google Scholar]
- Lak A, Nomoto K, Keramati M, Sakagami M, & Kepecs A (2017). Midbrain dopamine neurons signal belief in choice accuracy during a perceptual decision. Current Biology, 27, 821–832. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lejarraga T, Hertwig R, & Gonzalez C (2012). How choice ecology influences search in decisions from experience. Cognition, 124, 334–342. [DOI] [PubMed] [Google Scholar]
- Leuker C, Pachur T, Hertwig R, & Pleskac TJ (2018). Exploiting risk–reward structures in decision making under uncertainty. Cognition, 175, 186–200. [DOI] [PubMed] [Google Scholar]
- May BC, Korda N, Lee A, & Leslie DS (2012). Optimistic Bayesian sampling in contextual-bandit problems. Journal of Machine Learning Research, 13, 2069–2106. [Google Scholar]
- Mehlhorn K, Newell BR, Todd PM, Lee MD, Morgan K, Braithwaite VA, … Gonzalez C (2015). Unpacking the exploration–exploitation tradeoff: A synthesis of human and animal literatures. Decision, 2, 191–215. [Google Scholar]
- Millner AJ, Gershman SJ, Nock MK, & den Ouden HE (2018). Pavlovian control of escape and avoidance. Journal of Cognitive Neuroscience, 1–12. [DOI] [PubMed] [Google Scholar]
- Milosavljevic M, Malmaud J, Huth A, Koch C, & Rangel A (2010). The drift diffusion model can account for value-based choice response times under high and low time pressure. Judgment and Decision Making, 5, 437–449. [Google Scholar]
- Navarro DJ, Newell BR, & Schulze C (2016). Learning and choosing in an uncertain world: An investigation of the explore-exploit dilemma in static and dynamic environments. Cognitive Psychology, 85, 43–77. [DOI] [PubMed] [Google Scholar]
- Payzan-LeNestour E, & Bossaerts P (2011). Risk, unexpected uncertainty, and estimation uncertainty: Bayesian learning in unstable settings. PLoS Computational Biology, 7, e1001048. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Payzan-LeNestour E, & Bossaerts P (2012). Do not bet on the unknown versus try to find out more: estimation uncertainty and “unexpected uncertainty” both modulate exploration. Frontiers in Neuroscience, 6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pearson JM, Hayden BY, Raghavachari S, & Platt ML (2009). Neurons in posterior cingulate cortex signal exploratory decisions in a dynamic multioption choice task. Current Biology, 19, 1532–1537. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pedersen ML, Frank MJ, & Biele G (2017). The drift diffusion model as the choice rule in reinforcement learning. Psychonomic Bulletin & Review, 24, 1234–1251. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ratclif R, & Frank MJ (2012). Reinforcement-based decision making in corticostriatal circuits: mutual constraints by neurocomputational and diffusion models. Neural Computation, 24, 1186–1229. [DOI] [PubMed] [Google Scholar]
- Riefer PS, Prior R, Blair N, Pavey G, & Love BC (2017). Coherency-maximizing exploration in the supermarket. Nature Human Behaviour, 1, 0017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schulz E, Konstantinidis E, & Speekenbrink M (2015). Learning and decisions in contextual multi-armed bandit tasks. In Proceedings of the 37th annual conference of the cognitive science society (pp. 2122–2127). [Google Scholar]
- Schulz E, Konstantinidis E, & Speekenbrink M (2017). Putting bandits into context: How function learning supports decision making. Journal of Experimental Psychology: Learning, Memory, and Cognition. [DOI] [PubMed] [Google Scholar]
- Somerville LH, Sasse SF, Garrad MC, Drysdale AT, Abi Akar N, Insel C, & Wilson RC (2017). Charting the expansion of strategic exploratory behavior during adolescence. Journal of Experimental Psychology: General, 146, 155–164. [DOI] [PubMed] [Google Scholar]
- Speekenbrink M, & Konstantinidis E (2015). Uncertainty and exploration in a restless bandit problem. Topics in Cognitive Science, 7, 351–367. [DOI] [PubMed] [Google Scholar]
- Srinivas N, Krause A, Seeger M, & Kakade SM (2010). Gaussian process optimization in the bandit setting: No regret and experimental design. In Proceedings of the 27th international conference on machine learning (pp. 1015–1022). [Google Scholar]
- Starkweather CK, Babayan BM, Uchida N, & Gershman SJ (2017). Dopamine reward prediction errors reflect hidden-state inference across time. Nature Neuroscience, 20, 581. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stephens DW, & Krebs JR (1986). Foraging theory. Princeton University Press. [Google Scholar]
- Summerfield C, & Tsetsos K (2012). Building bridges between perceptual and economic decision-making: neural and computational mechanisms. Frontiers in Neuroscience, 6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sutton RS, & Barto AG (1998). Reinforcement learning: An introduction. MIT Press. [Google Scholar]
- Thompson WR (1933). On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika, 25, 285–294. [Google Scholar]
- Warren CM, Wilson RC, van der Wee NJ, Giltay EJ, van Noorden MS, Cohen JD, & Nieuwenhuis S (2017). The effect of atomoxetine on random and directed exploration in humans. PloS One, 12, e0176034. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wilson RC, Geana A, White JM, Ludvig EA, & Cohen JD (2014). Humans use directed and random exploration to solve the explore-exploit dilemma. Journal of Experimental Psychology: General, 143, 2074–2081. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wulff DU, Mergenthaler-Canseco M, & Hertwig R (2018). A meta-analytic review of two modes of learning and the description-experience gap. Psychological Bulletin, 144, 140–176. [DOI] [PubMed] [Google Scholar]
- Zajkowski WK, Kossut M, & Wilson RC (2017). A causal role for right frontopolar cortex in directed, but not random, exploration. eLife, 6, e27430. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang S, & Yu AJ (2013). Forgetful Bayes and myopic planning: Human learning and decision-making in a bandit setting. In Advances in neural information processing systems (pp. 2607–2615). [Google Scholar]