Abstract
Agents living in volatile environments must be able to detect changes in contingencies while refraining to adapt to unexpected events that are caused by noise. In Reinforcement Learning (RL) frameworks, this requires learning rates that adapt to past reliability of the model. The observation that behavioural flexibility in animals tends to decrease following prolonged training in stable environment provides experimental evidence for such adaptive learning rates. However, in classical RL models, learning rate is either fixed or scheduled and can thus not adapt dynamically to environmental changes. Here, we propose a new Bayesian learning model, using variational inference, that achieves adaptive change detection by the use of Stabilized Forgetting, updating its current belief based on a mixture of fixed, initial priors and previous posterior beliefs. The weight given to these two sources is optimized alongside the other parameters, allowing the model to adapt dynamically to changes in environmental volatility and to unexpected observations. This approach is used to implement the “critic” of an actor-critic RL model, while the actor samples the resulting value distributions to choose which action to undertake. We show that our model can emulate different adaptation strategies to contingency changes, depending on its prior assumptions of environmental stability, and that model parameters can be fit to real data with high accuracy. The model also exhibits trade-offs between flexibility and computational costs that mirror those observed in real data. Overall, the proposed method provides a general framework to study learning flexibility and decision making in RL contexts.
Author summary
In stable contexts, animals and humans exhibit automatic behaviour that allows them to make fast decisions. However, these automatic processes exhibit a lack of flexibility when environmental contingencies change. In the present paper, we propose a model of behavioural automatization that is based on adaptive forgetting and that emulates these properties. The model builds an estimate of the stability of the environment and uses this estimate to adjust its learning rate and the balance between exploration and exploitation policies. The model performs Bayesian inference on latent variables that represent relevant environmental properties, such as reward functions, optimal policies or environment stability. From there, the model makes decisions in order to maximize long-term rewards, with a noise proportional to environmental uncertainty. This rich model encompasses many aspects of Reinforcement Learning (RL), such as Temporal Difference RL and counterfactual learning, and accounts for the reduced computational cost of automatic behaviour. Using simulations, we show that this model leads to interesting predictions about the efficiency with which subjects adapt to sudden change of contingencies after prolonged training.
Introduction
Learning agents must be able to deal efficiently with surprising events when trying to represent the current state of the environment. Ideally, agents’ response to such events should depend on their belief about how likely the environment is to change. When expecting a steady environment, a surprising event should be considered as an accident and should not lead to updating previous beliefs. Conversely, if the agent assumes the environment is volatile, then a single unexpected event should trigger forgetting of past beliefs and relearning of the (presumably) new contingency. Importantly, assumptions about environmental volatility can also be learned from experience.
Here, we propose a general model that implements this adaptive behaviour using Bayesian inference. This model is divided in two parts: the critic which learns the environment and the actor that makes decision on the basis of the learned model of the environment.
The critic side of the model (called Hierarchical Adaptive Forgetting Variational Filter, HAFVF [1]) discriminates contingency changes from accidents on the basis of past environmental volatility, and adapts its learning accordingly. This learner is a special case of Stabilized Forgetting (SF) [2]: practically, learning is modulated by a forgetting factor that controls the relative influence of past data with respect to a fixed prior distribution reflecting the naive knowledge of the agent. At each time step, the goal of the learner is to infer whether the environment has changed or not. In the former case, she erases her memory of past events and resets her prior belief to her initial prior knowledge. In the latter, she can learn a new posterior belief of the environment structure based on her previous belief. The value of the forgetting factor encodes these two opposite behaviours: small values tend to bring parameters back to their original prior, whereas large values tend to keep previous posteriors in memory. The first novel contribution of our work lies in the fact that the posterior distribution of the forgetting factor depends on the estimated stability of past observations. The second and most crucial contribution lies in the hierarchical structure of this forgetting scheme: indeed, the posterior distribution of the forgetting factor is itself subject to a certain forgetting, learned in a similar manner. This leads to a 3-level hierarchical organization in which the bottom level learns to predict the environment, the intermediate level represents its volatility and the top level learns how likely the environment is to change its volatility. We show that this model implements a generalization of classical Q-learning algorithms.
The actor side of the model is framed as a full Drift-Diffusion Model of decision making [3] (Normal-Inverse-Gamma Diffusion Process; NIGDM) that samples from the value distributions inferred from the critic in order to select actions in proportion to their probability of being the most valued. We show that this approach predicts plausible results in terms of exploration-exploitation policy balance, reward rate, reaction times (RT) and cognitive cost of decision. Using simulated data, we also show that the model can uncover specific features of human behaviour in single and multi-stage environments. The whole model is outlined in Algorithm 8: the agent first selects an action given an (approximate) Q-sampling policy, which is temporally implemented as a Full DDM [3] with variable drift rate and accumulation noise, then learns based on the return of the action executed (reward r(sj, aj) and transition s′ = T(sj, aj)). Then, the critic updates its approximate posterior belief about the state of the environment qj ≈ p.
Algorithm 1: AC-HAFVF sketch a represents actions, r stands for rewards, s stands for state and x stands for observations. μr is the expected value of the reward. q represents the approximate posterior of the latent variable z and θ0 stands for the prior parameter values of the distribution of z
1 for j = 1 to J do
2 Actor: NIGDM Line 8;
3 select ;
4 Observe xj = {r(sj, aj), sj+1};
5
6 Critic: HAFVF Line 8;
7 update qj(z(sj, aj)) ≈ p(z(sj, aj)|xj;x<j, θ0);
8 end
We apply the proposed approach to Model-Free RL contexts (i.e. to agents limiting their knowledge of the environment to a set of reward functions) in an extensive manner. We explore in detail the application of our algorithm to Temporal Discounting RL, in which we study the possibility of learning the discounting factor as a latent variable of the model. We also highlight various methods for accounting for unobserved events in a changing environment. Finally, we show that the way our algorithm optimizes the exploration-exploitation balance is similar to Q-Value Sampling when using large DDM thresholds.
Importantly, the proposed approach is very general, and even though we apply it here only to Model-Free Reinforcement Learning, it could be also extended to Model-Based RL [4], where the agent models a state-action-state transition table together with the reward functions. Additionally, other machine-learning algorithms can also benefit from this approach [1].
The paper is structured as follows: first (Related work section) we review briefly the state of the art and place our work within the context of current literature. in the Methods section, we present the mathematical details of the model. We derive the analytical expressions of the learning rule, and frame them in a biological context. We then show how this learning scheme directly translates into a decision rule that constitutes a special case of the Sequential Sampling family of algorithms. In the Results section, we show various predictions of our model in terms of learning and decision making. More importantly, we show that despite its complexity, the model can be fitted to behavioural data. We conclude by reviewing the contributions of the present work, highlighting its limitations and putting it in a broader perspective.
Related work
The adaptation of learning to contingency changes and noise has numerous connections to various scientific fields from cognitive psychology to machine learning. A classical finding in behavioural neuroscience is that instrumental behaviours tend to be less and less flexible as subjects repeatedly receive positive reinforcement after selecting a certain action in a certain context, both in animals [5–8] and humans [9–13]. This suggests that biological agents indeed adapt their learning rate to inferred environmental stability: when the environment appears stable (e.g. after prolonged experience of a rewarded stimulus-response association), they show increased tendency to maintain their model of the environment unchanged despite reception of unexpected data.
Most studies on such automatization of behaviour have focused on action selection. However, weighting new evidence against previous belief is also a fundamental problem for perception and cognition [14–16]. Predictive coding [17–22] provides a rich, global, framework that has the potential to tackle this problem, but an explicit formulation of cognitive flexibility is still lacking. For example, whereas [22] provides an elegant Kalman-like Bayesian filter that learns the current state of the environment based on its past observations and predicts the effect of its actions, it assumes a stable environment and cannot, therefore, adapt dynamically to contingency changes. The Hierarchical Gaussian Filter (HGF) proposed by Mathys and colleagues [23, 24] provides a mathematical framework that implements learning of a sensory input in a hierarchical manner, and that can account for the emergence of inflexibility in various situations. This model deals with the problem of flexibility (framed as expected “volatility”) by building a hierarchy of random variables: each of these variables is distributed with a Gaussian distribution with a mean equal to this variable at the trial before and the variance equal to a non-linear transform of the variable at superior level. Each level encodes the distribution of the volatility of the level below. Although it has shown its efficiency in numerous applications [25–30], a major limitation of this model, within the context of our present concern, is that While the HGF accommodates a dynamically varying volatility, it assumes that the precision of the likelihood at the lowest level is static. To understand why it is the case, one should first observe that in the HGF the variance at each level is the product of two factors: a first “tonic” component, which is constant throughout the experiment, and a “phasic” component that is time-varying and controlled by the level above. These terms recall the concepts of “expected” and “unexpected” uncertainty [31, 32], and in the present paper, we will refer to these as variance (of the observation) and volatility (of the contingency). Now consider an experiment with two distinct successive signals, one with a low variance and one with a high variance. When fitted to this dataset, the HGF will consider the lower variance as the first tonic component, and all the extra variance in the second part of the signal will be assigned to the “phasic” part of the volatility, thus wrongfully considering noise of the signal as a change of contingency (see Fig 1). In summary, the HGF will have difficulties accounting for changes in the variance of the observations. Moreover, the HGF model cannot forget past experience after changes of contingency, but can only adapt its learning to the current contingency. This contrasts with the approach we propose, where the assessment of a change of contingency is made with the use of a reference, naive prior that plays the role of a “null hypothesis”. This way of making the learning process gravitate around a naive prior allows the model to actively forget past events and to eventually come back to a stable learning state even after very surprising events. These caveats limit the applicability of the HGF to a certain class of datasets in which contingency changes affect the mean rather than the variance of observations and in which the training set contains all possible future changes that the model may encounter at testing.
As will be shown in detail below, in the model proposed in the present paper, volatility is not only a function of the variance of the observations: if a new observation falls close enough to previous estimates then the agent will refine its posterior estimate of the variance and will decrease its forgetting factor (i.e. will move its prior away from the fixed initial prior and closer to the learned posterior from the previous trial), but if the new observation is not likely given this posterior estimate, the forgetting factor will increase (i.e. will move closer to the fixed initial prior) and the model will tend to update to a novel state (because of the low precision of the initial prior). In the results of this manuscript, we show that our model outperforms the HGF in such situations.
In Machine Learning and in Statistics, too, the question of whether new unexpected data should be classified as outlier or environmental change is important [33]. This problem of “denoising” or “filtering” the data is ubiquitous in science, and usually relies on arbitrary assumptions about environmental stability. In signal processing and system identification, adaptive forgetting is a broad field where optimality is highly context (and prior)-dependant [2]. Bayesian Filtering (BF) [34], and in particular the Kalman Filter [35] often lack the necessary flexibility to model real-life signals that are, by nature, changing. One can discriminate two approaches to deal with this problem: whereas Particle Filtering (PF) [36–38] is computationally expensive, the SF family of algorithms [2, 39], from which our model is a special case, usually has greater accuracy for a given amount of resources [36] (for more information, we refer to [35] where SF is reviewed). Most previous approaches in SF have used a truncated exponential prior [40, 41] or a fixed, linear mixture prior to account for the stability of the process [37]. Our approach is innovative in this field in two ways: first, we use a Beta prior on the mixing coefficient (unusual but not unique [42]), and we adapt the posterior of this forgetting factor on the basis of past observations, the prior of this parameter and its own adaptive forgetting factor. Second, we introduce a hierarchy of forgetting that stabilizes the learning when the training length is long.
We therefore intend to focus our present research on the very question of flexibility. We will show how flexibility can be implemented in a Bayesian framework using an adaptive forgetting factor, and what prediction this framework makes when applied to learning and decision making in Model-Free paradigms.
Methods
Bayesian Q-learning and the problem of flexibility
Classical RL [43], or Bayesian RL [44, 45] cannot discriminate learners that are more prone to believe in a contingency change from those who tend to disregard unexpected events and consider them as noise. To show this, we take the following example: let p(ρ|r≤j) = Beta(α, β) be the posterior probability at trial j of a binary reward rj ∼ Bern(ρ) with prior probability ρ ∼ Beta(α0, β0). It can be shown that, at the trial j = vj + uj, where vj is the number of successes and uj the number of failures, the posterior probability parameters read {αj = α0 + vj, βj = β0 + uj}. This can be easily mapped to a classical RL algorithm if one considers that, at each update of v and u, the posterior expectation of ρ is updated by
(1) |
(2) |
(3) |
(4) |
which has the form of a classical myopic Q-learning algorithm with a decreasing learning rate.
The drawback of this fixed-schedule learning rate is that, if the number of observed successes outnumbers greatly the number of failures (v ≫ u) at the time of a contingency change in which failures become suddenly more frequent, the agent will need v − u + 1 failures to start considering that p(rj = 0|r≤j) > p(rj = 1|r≤j). This behaviour is obviously sub-optimal in a changing environment, and Dearden [44] suggests adding a constant forgetting factor to the updates of the posterior, making therefore the agent progressively blind to past outcomes. Consider the case in which
with w ∈ [0; 1] being the forgetting factor. We can easily see that, in the limit case of α0 = 0 and β0 = 0, as j → ∞. We can define as the efficient memory of the agent, which provides a bound on the effective memory, represented by the total amount of trials taken into account so far (e.g. αj+ βj in the previous example). This produces an upper and lower bound to the variance of the posterior estimate of p(ρ|r≤j). This can be seen from the variance of the beta distribution which is maximized when αj = βj, and minimized when either αj = α0 or βj = β0. In a steady environment, agents with larger memory are advantaged since they can better estimate the variance of the observations. But when the environment changes, large memory becomes disadvantageous because it requires longer time to adapt to the new contingency. Here, we propose a natural solution to this problem by having the agent erase its memory when a new observation (or a series of observations) is unlikely given the past experience.
General framework
Our framework is based on the following assumptions:
Assumption 1 The environment is fully Markovian: the probability of the current observation given all the past history is equal to the probability of this observation given the previous observation.
Assumption 2 At a given time point, all the observations (rewards, state transitions, etc.) are i.i.d. and follow a distribution p(x|z) that is issued from the exponential family and has a conjugate prior that also belongs to the exponential family p(z|θ0).
For conciseness, the latent variables z (i.e. action value, transition probability etc.) and their prior θ will represent the natural parameters of the corresponding distributions in what follows.
Assumption 3 The agent builds a hierarchical model of the environment, where each of the distributions at the lower level (reward and state transitions) are independent, i.e. the reward distribution for one state-action cannot be predicted from the distribution of the other state-actions.
Assumption 4 The agent can only observe the effects of the action she performs.
Finally, an important assumption that will guide the development of the model is that the evolution of the environment is unpredictable (i.e. transition probabilities are uniformly distributed for all states of the environment) with the notable exception that it is more or less likely to stay in the same state than to switch to another state. Formally:
Assumption 5 Let be a set of environment states, with A ≫ 0 and a, b, c ∈ {1, 2, …, A}, a ∉ {b, c}. We assume that the transition probabilities are uniformly distributed for b, c ∈ {1: A}¬a, which reads:
Assumption 5 implies that any attempt to learn new transitions from state to state based on a uniform prior over these transitions will harm the performance of the predictive model, and the best strategy one could adopt is to learn the probability of staying in the same state and group the probabilities of changing to any other state together. Then, the only two transition probabilities to learn are:
This is what the “critic” part of the AC-HAFVF we propose achieves. Of course, the model could be improved by learning the other transition probabilities, if needed, but we leave this for future work (see for instance [35]).
Model specifications
We are interested in deriving the posterior probability of some datapoint-specific measure pj(z ∣ x≤j, θ0), where j indicates the point in time, given the past and current observations x≤j and some prior belief θ0. Bayes theorem states that this is equal to
(5) |
We now consider the case of a subject that observes the stream of data and updates her posterior belief on-line as data are gathered. According to Assumption 2, one can express the posterior of z given the current observation xj
(6) |
It appears immediately that the prior p(z|x<j) has the same form as the previous posterior, so that our posterior probability function can be easily estimated recursively using the last posterior estimate as a prior (yesterday’s posterior is today’s prior) until p(z|θ0) is reached. Assumption 2 implies that the posterior p(z|xj, x<j) will be tractable: since p(x|z) is from the exponential family and has a conjugate prior p(z|θ0), the posterior probability has the same form as the prior, and has a convenient expression:
(7) |
where T(xi) is the sufficient statistics of the ith sample, and where we have made explicit the fact that can be partitioned in two parts, from which represents the prior number of observations. Consequently, is equivalent to the effective memory introduced above.
This simple form of recursive posterior estimate suffers from the drawbacks we want to avoid, i.e. it does not forget past experience. Let us therefore assume that zj can be different from zj−1 with a given probability, which we first assume to be known. We introduce a two-component mixture prior where the previous posterior is weighted against the original prior belief:
(8) |
The exponential weights on this prior mixture allow us to easily write its logarithmic form, but it still demands that we compute the normalizing constant Z(w, x<j, θ0) = ∫ p(z|x<j)w p(z|θ0)1−w d z. This constant has, however, a closed-form if both the prior and the previous posterior are from the same distribution, issued from the exponential family.
In Eq 8, we assumed that the forgetting factor was known. However, it is more likely that the learner will need to infer it from the data at hand. Putting a beta prior on this parameter, and under the assumption that the posterior probability factorizes (Mean-Field assumption), the joint probability at time j reads:
(9) |
where ϕ0 is the vector of the parameters of the beta prior of w. The model in Eq 9 is not conjugate, and the posterior is therefore not guaranteed to be tractable anymore.
Hierarchical filter
Let us now analyze the expected behaviour of an agent using a model similar to the one just described, in a steady environment: if all belong to the same, unknown distribution p(x|z), the value of will progressively converge to its true value as the prior (or the previous posterior) over w will eventually put a lot of weight on the past experience (i.e. it favours high values of w), since the distribution from which x is drawn is stationary. We have shown that such models rapidly tend to an overconfident posterior over w [1]. In practice, when the previous posterior of w is confident on the value that w should take (i.e. has low variance), it tends to favor updates that reduce variance further, corresponding to values of pj(w|x≤j) that match pj−1(w|x<j), even if this means ignoring an observed mismatch between pj−1(z|x<j) and pj(z|x≤j). In order to deal with this issue, we enrich our model by introducing a third level in the hierarchy.
We re-define the prior over w as a two-component mixture of priors:
and the full joint probability has the form
(10) |
This additional hierarchical level allows the model to forget w as a function of observed data (i.e. not at a fixed rate) providing it with the capacity to adapt the approximate posterior distribution over w with greater flexibility [1]. The latent variable b can be seen as a regulizer for pj(w ∣ x≤j).
The prior parameters of the HAFVF and their interpretation is outlined in Table 1.
Table 1. HAFVF prior parameters in the case of normally distributed variables.
General identifier | Parameter | Domain | Name | Interpretation |
---|---|---|---|---|
θ0 | Prior mean | Expected value of the observations | ||
Prior number of observations (over μ) | Importance of the prior belief of μ | |||
Gamma shape parameter | Importance of the prior belief of σ | |||
Gamma rate parameter | Sum of squared residuals | |||
ϕ0 | Beta shape parameter | Stability belief of {μ, σ} | ||
Beta shape parameter | Volatility belief of {μ, σ} | |||
β0 | Beta shape parameter | Stability belief of w | ||
Beta shape parameter | Volatility belief of w |
Variational Inference
Eqs 6–10 involve the posterior probability distributions of the parameters given the previous observations. When these quantities have no closed-form formula, two classes of methods can be used to estimate them. Simulation-based algorithms [46] such as importance sampling, particle filtering or Markov Chain Monte Carlo, are asymptotically exact but computationally expensive, especially in the present case where the estimate has to be refined at each time step. The other class of methods, approximate inference [47, 48], consists in formulating, for a model with parameters y and data x, an approximate posterior q(y), that we will use as a proxy to the true posterior p(y|x). Roughly, approximate inference can be partitioned into Expectation Propagation and Variational Bayes (VB) methods. Let us consider in more detail VB, as it is the core engine of our learning model. In VB, optimizing the approximate posterior amounts to computing a lower-bound to the log model evidence (ELBO) , whose distance from the true log model evidence can be reduced by gradient descent [49]. Hybrid methods, that combine sampling methods with approximate inference, also exist (e.g. Stochastic Gradient Variational Bayes [50] or Markov Chain Variational Inference [51]). With the use of refined approximate posterior distributions [52–54], they allow for highly accurate estimates of the true posterior with possibly complex, non-conjugate models.
We define a variational distribution over y with parameters υ: q(y|υ), which we will use as a proxy to the real, but unknown, posterior distribution p(y|x). The two distribution match exactly when their Kullback-Leibler divergences are equal to zero, i.e.
where we have omitted the approximate posterior parameters υ for sparsity of the expressions. Given some arbitrary constraints on q(y), we can choose (for mathematical convenience) to reduce DKL[q(y)||p(y|x)] wrt q(y). Formally, this can be written as
We can now substitute log p(y|x) by its rhs in the log-Bayes formula
(11) |
Because log p(x) does not depend on the model parameters, it is fixed for a given dataset. Therefore, as we maximize in Eq 11, we decrease the divergence DKL[q(y)||p(y|x)] between the approximate and the true posterior. When a maximum is reached, we can consider that (1) we have obtained the most accurate approximate posterior given our initial assumptions about q(y) and (2) provides a lower bound to log p(x). It should be noted here that the more q(y) is flexible, the closer we can hope to get from the true posterior, but this is generally at the expense of tractability and/or computational resources.
The ELBO in Eq 11 is the sum of the expected log joint probability and the entropy of the approximate posterior. In order for the former to be tractable, one must carefully choose the form of the approximate posterior. The Mean-field assumption we have made allows us to select, for each factor of the approximate posteriors, a distribution with the same form as their conjugate prior, which is the best possible configuration in this context [55].
Applying now this approach to Eq 10, our spherical approximate posterior looks like:
In addition, in order to recursively estimate the current posterior probability of the model parameters given the past, we make the natural approximation that the true previous posterior can be substituted by its variational approximation:
(12) |
and similarly for p(w|x<j) and p(b|x<j). The use of this distribution as a proxy to the posterior greatly simplifies the optimization of qj(z, w, b).
The full, approximate joint probability distribution at time j therefore looks like
where θj−1, ϕj−1 and βj−1 are the variational parameters at the last trial for z, w and b respectively. A further advantage of the approximation made in Eq 12 is that the prior of z and w simplifies elegantly:
(13) |
(see Appendix A for the full derivation).
A conjugate distribution for p(w) is hard to find. Šmìdl and Quinn [56] propose a uniform prior and a truncated exponential approximate posterior over w. They interpolate the normalizing constant between two fixed value of w, which allows them to perform closed-form updates of this parameter. Here, we chose p(w|ϕ0) and q(w|ϕj) to be both beta distributions, a choice that does not impair our ability to perform closed-form updates of the variational parameters as we will see in the Update equation section.
In this model (see Fig 2), named the Hierarchical Adaptive Forgetting Variational Filter [1], specific prior configurations will bend the learning process to categorize surprising events either as contingency changes, or as accidents. In contrast with other models [57], w and b are represented with a rich probability distribution where both the expected values and variances have an impact on the model’s behaviour. For a given prior belief on z, a confident prior over w, centered on high values of this parameter, will lead to a lack of flexibility that would not be observed with a less confident prior, even if they have the same expectation.
The critic: HAFVF as a reinforcement learning algorithm
Application of this scheme of learning to the RL case is straightforward, if one considers as being the observed rewards and z as the parameters of the distribution of these rewards. In the following, we will assume that the agent models a normally distributed state-action reward function xj = r(s, a), from which she tries to estimate the posterior distribution natural parameters z ≜ η(μ(s, a), σ(s, a)) where η(⋅) is the natural parameter vector of the normal distribution. In this context, an intuitive choice for the prior (and the approximate posterior) of these parameters is a Normal Inverse-Gamma distribution (): for the prior, we have
and the approximate posterior can be defined similarly with a normal component and a gamma component .
Update equation
Even though the model is not formally conjugate, the use of a mixture of priors with exponential weights makes the variational update equations easy to implement for the first level. Let us first assert a few basic principles from the Mean-Field Variational Inference framework: it can be shown that, under the assumption that the approximate posterior factorizes in q(y1 y2) = q(y1)q(y2), then the optimal distribution q*(y1) given our current estimate of q(y2) is given by
(14) |
where log Z is some log-normalizer that does not depend on y. Eq 14 states that each set of variational parameters can be updated independently given the current value of the others: this usually lead to an approach similar to EM [46], where one iterates through the updates of variational posterior successively until convergence.
Fortunately, thanks to the conjugate form of the lower level of the HAFVF, Eq 14 can be unpacked to a form that recalls Eq 7 where the update of the variational parameters of z reads:
(15) |
where
is the weighted prior of z (see Appendix A). This update scheme can be mapped onto and be interpreted in terms of Q-learning [43] (see Appendix B). Again, is the updated effective memory of the subject, which is bounded on the long term by the (approximate) efficient memory . One can indeed see that the actual efficient memory, which reads
is undefined when b ≤ 1. To solve this issue, we used the first order Taylor approximation
which is a biased but consistent estimator of the efficient memory, as it approaches its true value for large values of ϕ.
More specifically, for the approximate posterior of a single observation stream with corresponding parameters , we can apply this principle easily, as the resulting distribution has the form of a with parameters:
(16) |
where we have used .
Deriving updates for the approximate posterior over the mixture weights w and b is more challenging, as the optimal approximate posterior in Eq 14 does not have the same form as the beta prior due to the non-conjugacy of the model. Fortunately, non-conjugate variational message passing (NCVMP) [58] can be used in this context. In short, NCVMP minimizes an approximate KL divergence in order to find the value of the approximate posterior parameters that maximize the ELBO. Although NCVMP convergence is not guaranteed, this issue can be somehow alleviated by damping of the updates (i.e. updating the variational parameters to a value lying in between the previous value they endorsed and the value computed using NCVMP, see [58] for more details). The need for a closed-form formula of the expected log-joint probability constitutes another obstacle for the naive implementation of NCVMP to the present problem: indeed, computing the expected value of the log-partition functions log Z(w) and log Z(b) involves a weighted sum of the past variational parameters θj−1 and the prior θ0, which are known, with a weight w, which is unknown. Expectation of this expression given q(w) does not, in general, have an analytical expression. To solve this problem, we used the second order Taylor expansion around (see Appendix A).
The derivation of the update equations of ϕ and β can be found in [1].
Counterfactual learning
As an agent performs a series of choices in an environment, she must also keep track of the actions not chosen and update her belief accordingly: ideally, the variance of the approximate posterior of the reward function associated with a given action should increase when that action is not selected, to reflect the increased uncertainty about its outcome during the period when no outcome was observed. This requirement implies counterfactual learning capability [59–61].
Two options will be considered here: the first option consists in updating the approximate posterior parameters of the non-selected action at each time step with an update scheme that pulls the approximate posterior progressively towards the prior θ0, with a speed that depends on w, i.e. as a function of the belief the agent has about environment stability. The second approach will consist in updating the approximate posterior of the actions only when they are actually selected, but accounting for past trials during which that action was not selected. The mathematical details of these approaches are detailed in Appendix C.
Delayed updating
Even though the agent learns only actions that are selected, it can adapt its learning rate as a function of how distant in the past was the last time the action was selected. Formally, this approach considers that if the posterior probability had been updated at each time step and the forgetting factor w had been stable, then the impact of the observations n trials back in time would currently have an influence that would decrease geometrically with a rate ω ≜ wn. We can then substitute the prior over z by:
(17) |
which is identical to Eq 8 except that w has been substituted by ω.
We name this strategy Delayed Approximate Posterior Updating.
Continuous updating
When an action is not selected, the agent can infer what value it would have had given the observed stability of the environment. In practice, this is done by updating the variational parameters of the selected and non-selected action using the observed reward for the former, and the expected reward and variance of this reward for the latter.
This approach can be beneficial for the agent in order to optimize her exploration/exploitation balance. In Appendix C.1, it is shown that if the agent has the prior belief that the reward variance is high, then the probability of exploring the non-chosen option will increase as the lag between the current trial and the last observation of the reward associated with this option increases.
This feature makes this approach intuitively more suited for exploration among multiple alternatives in changing environments, and we therefore selected it for the simulations achieved in this paper.
Temporal difference learning
An important feature required for an efficient Model-Free RL algorithm is to be able to account for future rewards in order to make choices that might seem suboptimal to a myopic agent, but that make sense on the long run. This is especially useful when large rewards (or the avoidance of large punishments) can be expected in a near future.
In order to do this, one can simply sum the expected value of the next state to the current reward in order to perform the update of the reward distribution parameters. However, because the evolution of the environment is somehow chaotic, it is usually considered wiser to decay slightly the future rewards by a discount rate γ. This mechanism is in accordance with many empirical observations of animal behaviours [62–64], neurophysiological processes [65–67] and theories [68–70].
As the optimal value of γ is unknown to the agent, we can assume that she will try to estimate its posterior distribution from the data as she does for the mean and variance of the reward function. Appendix D shows how this can be implemented in the current context. An example of TD learning in a changing environment is given in the TD learning with the HAFVF section.
We now focus on the problem of decision making under the HAFVF.
The actor: Decision making under the HAFVF
Bayesian policy
In a stable environment where the distribution of the action values are known precisely, the optimal choice (i.e. the choice that will maximize reward on the long run) is the choice with the maximum expected value: indeed, it is easy to see that if , then (here and for the next few paragraphs, we will restrict our analysis to the case of single stage tasks, and omit the s input in the reward function). However, in the context of a volatile environment, the agent has no certainty that the reward function has not changed since the last time she visited this state, and she has no precise estimate of the reward distribution. This should motivate her to devote part of her choices to exploration rather than exploitation. In a randomly changing environment, there is no general, optimal balance between the two, as there is no way to know how similar is the environment wrt the last trials. The best thing an agent can do is therefore to update her current policy wrt her current estimate of the uncertainty of the latent state of the environment.
Various policies have been proposed in order to use the Bayesian belief the agent has about its environment to make a decision that maximizes expected rewards in the long run. Here we will focus more particularly on Q-value sampling, or Q-sampling (QS) [71]. Note that we use the terminology Q-value sampling in accordance with Dearden [44, 71], but one should recall that QS is virtually indistinguishable from Thompson sampling [72, 73]. Our framework can also be connected to another algorithm used in the study of animal RL [74, 75], based on the Value of Perfect Information (VPI) [44, 45, 76], which we describe in Appendix E.
QS [71] is an exploration policy based on the posterior predictive probability that an action value exceeds all the other actions available:
(18) |
The expectation of Eq 18 provides a clear policy to the agent. The QS approach is compelling in our case: in general, the learning algorithm we propose will produce a trial-wise posterior probability that should, most of the time, be easy to sample from.
In bandit tasks, the policy dictated by QS is optimal provided that the subject has an equal knowledge of all the options she has. If the environment is only partly and unequally explored, the value of some actions may be overestimated (or underestimated), in which case QS will fail to detect that exploration might be beneficial. QS can lead to the same policy in a context where two actions (a1 and a2) have similar uncertainty associated with their reward distributions (σ1 = σ2) but different means (μ1 > μ2), and in a context where one action has a much larger expected reward (μ1 ≫ μ2) but also larger uncertainty (σ1 ≫ σ2) (see [44] for an example). This can be sub-optimal, as the action with the larger uncertainty could lead to a higher (or lower) reward than expected: in this specific case, choosing the action with the largest expected reward should be even more encouraged due to the lack of of knowledge about its true reward distribution, which might be much higher than expected. A strategy to solve this problem is to give to each action value a bonus, the Value of Perfect Information, that reflects the expected information gain that will follow the selection of an action. This approach, and its relationship to our algorithm, is discussed in Appendix E.
Q-sampling as a stochastic process
Let us get back to the case of QS, and consider an agent solving this problem using a gambler ruin strategy [77]. We assume that, in the case of a two-alternative forced choice task, this agent has equal initial expectations that either a1 or a2 will lead to the highest reward, represented by a start point z0 = ζ/2, where ζ will be described shortly. The gambler ruin process works as follows: this agent samples a value and a value and assess which one is higher. If beats , she computes the number of wins of a1 until now as z1 = z0 + 1, and displaces her belief the other way (z0−1) if a2 beats a1. Then, she starts again and moves in the direction indicated by sign(r(a1) − r(a2)) until she reaches one of the two arbitrary thresholds situated at 0 or ζ that symbolize the two actions available. It is easy to see that the number of wins and losses generated by this procedure gives a Monte Carlo sample of p(r(a1) > r(a2)|x<j). We show in Appendix F.1 that this process tends to deteministically select the best option as the threshold grows.
So far, we have studied the gambler ruin problem as a discrete process. If the interval between the realization of two samples tends to 0, this accumulation of evidence can be approximated by a continuous stochastic process [77]. When the rewards are normally distributed, as in the present case, this results in a Wiener process, or Drift-Diffusion model (DDM, [78]), since the difference between two normally distributed random variables follows a normal distribution. This stochastic accumulation model has a displacement rate (drift) that is given by
see [79].
Crucially, it enjoys the same convergence property of selecting almost surely the best option for high thresholds (see Appendix F.2).
Sequential Q-sampling as a Full-DDM model
This simple case of a fixed-parameters DDM, however, is not the one we have to deal with, as the agent does not know the true value of {μ(a), σ2(a)}a∈a, but she can only approximate it based on her posterior estimate. Assuming that the approximate posterior over the latent mean and variance of the reward distribution is a distribution, and keeping the original statement , we have
(19) |
and where, for the sake of sparsity of the notation, the indices are used to indicate the corresponding action-related variable.
To see how such evidence accumulation process evolves, one can discretize Eq 19: this would be equivalent to sample at each time t a displacement
(20) |
where Δx stands for xt − xt−1. The drift in Eq 20 is sampled as the difference between two sampled means and the squared noise is sampled as the sum of the two sampled variances . At each time step (or at each trial, quite similarly as we will show), the tuple of parameters is drawn from the current posterior distribution.
Hereafter, we will refer to this process as the Normal-Inverse-Gamma Diffusion Process, or NIGDM.
NIGDM as an exploration rule
Importantly, and similarly to QS, this process has the desired property of selecting actions with a probability proportional to their probability of being the best option. This favours exploratory behaviour since, assuming equivalent expected rewards, actions that are associated with large reward uncertainty will tend to be selected more often. We show that the NIGDM behaves like QS in Appendix F.3. There, it is shown (Proposition 3) that, as the threshold grows, the NIGDM choice pattern resembles more and more the QS algorithm. For lower values of the threshold, this algorithm is less accurate than QS (see further discussion of the property of NIGDM for low thresholds in Appendix E).
Cognitive cost optimization under the AC-HAFVF
Algorithm 2 summarizes the AC-HAFVF model. This model ties together a learning algorithm—that adapts how fast it forgets its past knowledge on the basis of its assessment of the stability of the environment—with a decision algorithm that makes full use of the posterior uncertainty of the reward distribution to balance exploration and exploitation.
Importantly, these algorithms make time and resource costs explicit: for instance, time constraints can make decisions less accurate, because they will require a lower decision threshold. Fig 3 illustrates this interpretation of the AC-HAFVF by showing how accuracy and speed of the model vary as a function of the variance of the reward estimates: the cognitive cost of making a decision using the AC-HAFVF is high when the agent is uncertain about the reward mean (low κj) but has a low expectation of the variance (low βj). In these situations, the choices were also more random, allowing the agent to explore the environment. Decisions are easier to make when the difference in mean rewards is clearer or when the rewards were more noisy.
Besides the decision stage, the computational cost of the inference step can also be determined. To this end, one must first consider that the HAFVF updates the variational posterior using a natural gradient-based [80] approach with a Fisher preconditioning matrix [58, 81, 82]: the learner computes a gradient of the ELBO wrt the variational parameters, then moves in the direction of this gradient with a step length proportional to the posterior variance of the loss function. Because of this, the divergence between the true posterior probability p(z|x≤j) and the prior probability p(z) (i.e. the mixture of the default distribution and previous posterior) conditions directly the expected number of updates required for the approximate posterior to converge to a minimum DKL[q||p] (see for instance [83]). Also, in a more frequentist perspective and at the between-trial time scale, convergence rate of the posterior probability towards the true (if any) model configuration is faster when the KL divergence between this posterior and the prior is small [84]: in other words, in a stable environment, a higher confidence in past experience will require less observations for the same rate of convergence, because it will tighten the distance between the prior and posterior at each time step.
These three aspects of computational cost (for decision, for within-trial inference and for across-trial inference) can justify the choice (or emergence) of lower flexible behaviours as they ultimately maximize the reward rate [74, 85]: indeed, such strategies will lead in a stable environment to faster decisions, to faster inference at each time step and to a more stable and accurate posterior.
Algorithm 2: AC-HAFVF. For simplicity, the NIGDM process has been discretized.
input: prior belief {θ0, ϕ0, β0}
1 for j = 1 to J do
2 Actor: NIGDM;
input: Start point z0, threshold ζ, non-decision time τ
3 k ← 0;
4 sample for i = {1, 2};
5 while 0 < zk < ζ do
6 k += 1;
7 sample ;
8 move zk += δz;
9 end
10 select ;
11 Get reward rj = r(sj, aj), Observe transition sj+1 = j(sj, aj);
12
13 Critic: HAFVF;
14 ELBO ← −∞;
15 while |δL| ≤ 10−3 do
16 update using CVMP;
17 update {ϕj, βj} using NCVMP;
18 ;
19 ;
20 end
21 end
Fitting the AC-HAFVF
So far, we have provided all the necessary tools to simulate behavioural data using the AC-HAFVF. It is now necessary to show how to fit model parameters to an acquired dataset. We will first describe how this can be done in a Maximum Likelihood framework, before generalizing this method to Bayesian inference using variational methods.
The problem of fitting the AC-HAFVF to a dataset can be seen as a State-Space model fitting problem. We consider the following family of models:
(21) |
and tj,n stands for the reaction time associated with the state-action pair (s, a) of the subject n at the trial j. Unlike many State-Space models, we have made the assumption in Eq 21 that the transition model Ωj,n = f(Ωj−1,n, Ω0,n, xj−1,n) is entirely deterministic given the subject prior Ω0,n and the observations x<j, which is in accordance with the model of decision making presented in the The actor: Decision making under the HAFVF section. Note that the Bayesian procedure we will adopt hereafter is formally identical to considering that the drift and noise are drawn according to the rules defined in the The actor: Decision making under the HAFVF section, making this model equivalent to:
Quite importantly, we have made the assumption in Eq 21 that the threshold, the non-decision time and the start-point were fixed for each subject throughout the experiment. This is a strong assumption, that might be relaxed in practice. To simplify the analysis, and because it is not a mandatory feature of the model exposed above, we do not consider this possibility here and leave it for further developments.
We can now treat the problem of fitting the AC-HAFVF to behavioural data as two separate sub-problems: first, we will need to derive a differentiable function that, given an initial prior Ω0,n and a set of observations xn produces a sequence , and second (Maximum a Posteriori estimate of the AC-HAFVF section) a function that computes the probability of the observed behaviour given the current variational parameters.
Maximum a Posteriori estimate of the AC-HAFVF
The update equations described in the Update equation section enable us to generate a differentiable sequence of approximate posterior parameters Ωj,n given some prior Ω0,n and a sequence of choices-rewards x. We can therefore reduce Eq 21 to a loss function of the form
whose gradient wrt Ω0,n can be efficiently computed using the chain rule:
Recall that is the jacobian (i.e. matrix of partial derivative) of Ωj,n wrt each of the elements of Ω0,n that are optimized, and is the gradient of the loss function (i.e. the NIGDM) wrt the output of f(⋅).
As the variational updates that lead to the evaluation of Ωj,n are differentiable, the use of VB makes it possible to use automatic Differentiation to compute the Jacobian of Ωj,n wrt Ω0,n.
The next step, is to derive the loss function log p(yj,n|Ωj,n, ζn, z0n, τn). In this log-probability density function, the local, trial-wise parameters
(22) |
have been marginalized out. This makes its evaluation hard to implement with conventional techniques. Variational methods can be used to retrieve an approximate Maximum A Posteriori (MAP) in these cases [86]. The method is detailed in Appendix G. Briefly, VB is used to compute a lower bound (ℓj,n ≤ log p(yj,n|Ωj,n, ζn, z0n, τn)) to the marginal posterior probability described above for each trial. Instead of optimizing each variational parameters independently, we optimize the parameters ρ of an inference network [87] that maps the current HAFVF approximate posterior parameters Ωj,n and the data yj,n to each trial-specific approximate posterior. This amortizes greatly the cost of the optimization (hence the name Amortized Variational Inference), as the nonlinear mapping (e.g. multilayered perceptron) h(yj,n;ρ) can provide the approximate posterior parameters of any datapoint, even if it has not been observed yet. We chose qρ(χj,n|yj,n) to be a multivariate Gaussian distribution, which leads to the following form of variational posterior:
(23) |
and is the lower Cholesky factor of . Another consideration is that, in order to use the multivariate normal approximate posterior, the unbounded variances sample must be transformed to an unbounded space. We used the inverse softplus transform λ ≜ log(exp(⋅) − 1), as this function has a bounded gradient, in contrasts with the exponential mapping, which prevents numerical overflow. We found that this simple trick could regularize greatly the optimization process. However, this transformation of the normally distributed λ−1(⋅) variables requires us to correct the ELBO by the log-determinant of the Jacobian of the transform [54], which for the sofplus transform of x is simply .
The same transformation can be used for the parameters of θ0 that are required to be greater than 0 (i.e. all parameters except μ0), which obviously do not require any log-Jacobian correction.
In Eq 22, we have made explicit the fact that we used the three latent variables: the drift rate and the two action-specific noise parameters. This is due to the fact that, unfortunately, the distribution of the sum of two Inverse-Gamma distributed random variables does not have a closed form formula, making the use of single random variable challenging, whereas it can be done easily for ξj,n = μj,n(a1) − μj,n(a2), which is normally distributed (see Eq 19).
The final step to implement a MAP estimation algorithm is to set a prior for the parameters of the model. We used a simple L2 regularization scheme, which consists trivially in a normal prior over all parameters, mapped onto an unbounded space if needed.
Algorithm 3 shows how the full optimization proceeds.
Algorithm 3: MAP estimate of AC-HAFVF parameters.
input: Data x = {rj,n, yj,n for j = 1 to J, n = 1 to N}
1 initialize Ωn = {θ0, ϕ0, β0}n for n = 1 to N and IN§ weights ρ;
2 repeat
3 set L ← 0;
4 for n = 1 to N do
5 Set ;
6 for j = 1 to J do
7 Learning Step: HAFVF;
8 Get Ωj,n = f(Ωj−1,n, Ω0,n, xj−1,n)
9 and Jacobian wrt Ω0,n: using FAD*;
10 Decision Step: HAFVF;
11 Get ELBO ℓj,n of using AVI†;
12 and corresponding gradient using RAD‡;
13 Increment L += ℓj,n;
14 Compute gradient wrt Ω0,n using the chain rule:
15 ;
16 Increment gradient of DDM and IN parameters
17 ;
18 end
19 L2-norm regularization:
20 ;
21 ;
22 end
23 Perform gradient step for a small η;
24 until Some convergence criterion is met;
§ IN = Inference Network, * FAD = Forward Automatic Differenciation, † AVI = Amortized Variational Inference, ‡ RAD = Reverse Automatic Differentiation (i.e. backpropagation).
Results
We now present four simulated examples of the (AC-)HAFVF in various contexts. The first example compares the performance of the HAFVF to the HGF [23, 24] in a simple contingency change scenario. The second example provides various case scenarios in a changing environment, illustrating the trade-off between flexibility and the precision of the predictions (Learning and flexibility assessment section), including cases where agents fail to adapt to contingency changes following prolonged training in a stable environment, as commonly observed in behavioural experiments [5]. The third example shows that the HAFVF can be efficiently fitted to a RL dataset using the method described in Fitting the AC-HAFVF section. The fourth and final example shows how this model behaves in multi-stage environments, and compares various implementations.
Adaptation to contingency changes and comparison with the HGF
In order to compare the performance of our model to the HGF, we generated a simple dataset consisting of a noisy square-wave signal of two periods of 200 trials, alternating between two normally distributed random variables ( and ). We fitted both a Gaussian HGF and the HAFVF to this simple dataset by finding the MAP parameter configuration for both models. The default configuration of the HGF was used, whereas in our case we put a normal hyperprior of on the parameters (with inverse softplus transform for parameters needing positive domains).
This fit constituted the first part of our experiment, which is displayed on the left part of Fig 4. We compared the quadratic approximations of the Maximum Log-model evidences [88] for both models, which for the HAFVF reads
where H is the hessian of the log-joint at the mode and M is the number of parameters of the model. We found a value of -186.42 for HAVFV and -204.73 for the HGF, making the HAVFV a better model of the data, with a Bayes Factor [89] greater than 8 * 107.
The second part of the experiment consisted in adding to this 400-trial signal a 1200-trial signal of input situated at y = 5. We evaluated for both models the quality of the fit obtained when using the parameter configurations resulting from the fit of the first part of the experiment (first 400 trials, or training dataset) (Fig 4, right part), on the remaining dataset (following 1200 trials, i.e. testing dataset). An optimal agent in such a situation should first account for the surprise associated with the sudden contingency change, and then progressively reduce its expected variance estimate to reflect the steadiness of the environment. We considered the capacity of both models to account for new data for a given parameter configuration as a measure of their flexibility. This test was motivated by the observation that a change detection algorithm has to be able to detect changes at test time that might be qualitatively different from changes at training time. A financial crisis is for instance an event that is in essence singular and unseen in the past (otherwise it would have been prevented). The algorithm should nevertheless be able to detect it efficiently.
The HGF was unable to exhibit the expected behaviour: it hardly adapted its estimated variance to the contingency change and did not adjust it significantly afterwards. This contrasted with the HAFVF, in which we observed initially an increase in the variance estimate at the point of contingency change (reflecting a high surprise), followed by progressively decreasing variance estimate, reflecting the adaptation of the model to the newly stable environment.
Together, these results are informative of the comparative performance of the two algorithms. The Maximum Log-model Evidence was larger for the HAFVF than for the HGF by several orders of magnitude, showing that our approach modelled better the data at hand than the HGF. Moreover, the lack of generalization of the HGF to a simple, new signal not used to fit the parameters, shows that this model tended to overfit the data, as can be seen from the estimated variance at the time of the contingency change.
Importantly, this capability of the HAFVF to account for unseen volatility changes did not need to be instructed through the selection of the model hyperparameters: it is a built in feature of the model.
Learning and flexibility assessment
In the following datasets, we simulated the learning process of four hypothetical subjects differing in their prior distribution parameters ϕ0, β0, whereas we kept θ0 fixed for all of them (Table 2). The choice of the subject parameters was made to generate limit and opposite cases of each expected behaviour. With these simulations, we aimed at showing how the prior belief on the two levels of forgetting conditioned the adaptation of the subject in case of contingency change (CC, Experiment 1) or isolated expected event (Experiment 2).
Table 2. This table summarizes the parameters of the beta prior of the two forgetting factors w and b used in the Learning and flexibility assessment section, as well as the initial prior over the mean and variance.
Subjects | μ0 | κ0 | α0 | β0 | |||||
---|---|---|---|---|---|---|---|---|---|
(1) | LL | 0 | 0.1 | 1 | 1 | 4.5 | 0.5 | 4.5 | 0.5 |
(2) | SL | 0.5 | 4.5 | 4.5 | 0.5 | ||||
(3) | LS | 4.5 | 0.5 | 0.5 | 4.5 | ||||
(4) | SS | 0.5 | 4.5 | 0.5 | 4.5 |
In both experiments, these agents were confronted with a stream of univariate random variables from which they had to learn the trial-wise posterior distribution of the mean and standard deviation.
In Experiment 1, we simulated the learning of these agents in a steady environment followed by an abrupt CC, occurring either after a long (900 trials) or a short (100 trials) training. The signal r = {r1, r2, …, rn} was generated according to a Gaussian noise with mean μ = 3 before the CC and μ = −3 after the CC, and a constant standard deviation σ = 1. Fig 5 summarizes the results of this first simulation. During the training phase, the subjects with a long memory on the first level learned the observation value faster than others. Conversely, the SS subject took a long time to learn the current distribution. More interesting is the behaviour of the four subjects after the CC. In order to see which strategy reflected best the data at hand, we computed the average of the ELBOs for each model. The winning agent was the Long-Short memory, irrespective of training duration, because it was better able to adapt its memory to the contingency.
The two levels had a different impact on the flexibility of the subjects: the first level indicated how much a subject should trust his past experience when confronted with a new event, and the second level measured the stability of the first level. On the one hand, subjects with a low prior on first-level memory were too cautious about the stability of the environment (i.e. expected volatile environments) and failed to learn adequately the contingency at hand. On the other hand, after a long training, subjects with a high prior on second-level memory tended to over-trust environment stability, compared to subjects with a low prior on second level memory, impairing their adaptation after the CC.
The expected forgetting factors also shed light on the underlying learning process occurring in the four subjects: grew or was steady until the CC for the four subjects, even for the SL subject which showed a rapid growth of during the first trials, but failed to reduce it at the CC. In contrast, the LS subject did not exhibit this weakness, but rapidly reduced its expectation over the stability of the environment after the CC thanks to her pessimistic prior belief over b.
In Experiment 2, we simulated the effect of an isolated, unexpected event (rj = −3) after long and short training with the same distribution as before. For both datasets, we focused our analysis on the value of the expected forgetting factors and , as well as the effective memory of the agents, represented by the parameter κμ. As noted earlier, the value of sets an upper bound (the efficient memory) on κμ, which represented the actual number of trials kept in memory up to the current trial.
Fig 6 illustrates the results of this experiment. Here, the flexible agents (with a low memory on either the first or second memory level, or both) were disadvantaged wrt the low flexibility agent (mostly LL). Indeed, following the occurrence of a highly unexpected observation, one can observe that the LS learner memory dropped after either long or short training. The LL learner, instead, was able to cushion the effect of this outlier, especially after a long training, making it the best learner of the four to learn these datasets (Table 3).
Table 3. Average ELBOs for Experiment 1 and 2.
Experiment 1 | Experiment 2 | |||||||
---|---|---|---|---|---|---|---|---|
Dataset | LL | SL | LS | SS | LL | SL | LS | SS |
1 | -1.719 | -1.657 | -1.62 | -1.863 | -1.459 | -1.652 | -1.501 | -1.863 |
2 | -1.526 | -1.649 | -1.519 | -1.866 | -1.453 | -1.648 | -1.495 | -1.866 |
Fitting the HAFVF to a behavioural task
The model we propose has a large number of parameters, and overfitting could be an issue. To show that inference about the latent variables of the model depicted in the Fitting the AC-HAFVF section is possible, we simulated a dataset of 64 subjects performing a simple one-stage behavioural task, that consisted in trying to choose at each trial the action leading to the maximum reward. In any given trial , the two possible actions (e.g. left or right button press) were associated to different, normally distributed, reward probabilities with a varying mean and a fixed standard deviation for the first action (a1) the reward had a mean of 0 for the first 500 trials, then switched abruptly to + 2 for 100 trials and then to −2 for the rest of the experiment. The second action value was identically distributed but in the opposite order and with the opposite sign (Fig 7). This pattern was chosen in order to test the flexibility of each agent after an abrupt CC: after the first CC, an agent discarding completely exploration in favour of exploitation would miss the CC. The second and third CC tested how fast did the simulated agents adapt to the observed CC.
Individual prior parameters and thresholds ζn were generated as follows: we first looked for the L2-regularized MAP estimates of these parameters that led to the maximum total reward:
using a Stochastic Gradient Variational Bayes (SGVB) [90] optimization scheme. With a negligible loss of generality, the prior mean μ0 was considered to be equal to 0 for all subjects for both the data generation and the fitting procedures. With a negligible loss of generality, the prior mean μ0 was considered to be equal to 0 for all subjects for both the data generation and the fitting procedures.
We then simulated individual priors centered around this value with a covariance matrix arbitrarily sampled as
where is an inverse-Wishart distribution with n degrees of freedom. This choice of prior lead to a large variability in the values of the AC-HAFVF parameters, except for the NIGDM threshold whose variance was set to a sufficiently low value (hence the 0.1 value in the prior scale matrix) to keep the learning performance high.
This method ensured that each and every parameter set was centered around an unknown optimal policy. This approach was motivated by the need to prevent strong constrains on the data generation pattern while keeping behaviour close to optimal, as might be expected from healthy population. The other DDM parameters, νn and τn, were generated according to a Gaussian distribution centered on 0 and 0.3, respectively. Simulated subjects with a performance lower than 70% were rejected and re-sampled to avoid irrelevant parameter patterns.
Learning was simulated according to the Continuous Learning strategy (see Counterfactual learning), because it was supposed to link more comprehensively the tendency to explore the environment with the choice of prior parameters Ω0. Choices and RT where then generated according to the decision process described in the The actor: Decision making under the HAFVF section using the algorithm described by [91]. Fig 7 shows two examples of the simulated behavioural data.
The behavioural results showed a clear tendency of subjects with large expected variance in action selection to act faster and less precisely than others. This follows directly from the structure of the NIGDM: larger variance of the drift-rate leads to faster but less precise policies. More interesting is the negative correlation between the expected stability and the reward-rate and average reaction time: this shows that the AC-HAFVF was able to encode a form of subject-wise computational complexity of the task. Indeed, large stability expectation leads subjects to trust more their past experience, thereby decreasing the expected reward variance after a long training, but it also leads to a lower capacity to adapt to CCs. For subjects with low expectation of stability, the second level memory was able to instruct the first-level to trust past experience when needed, as the positive correlation between accuracy and upper level memory shows.
Fitting results
The fit was achieved with the Adam [92] Stochastic Gradient Descent optimizer with parameters s = 0.005, β1 = 0.9, β2 = 0.99, where s decreased with a rate where i is the iteration number of the SGD optimizer. We used the following annealing procedure to avoid local minima: at each iteration, the whole set of parameters was sampled from a diagonal Gaussian distribution with covariance matrix 1/i. This simple manipulation greatly improved the convergence of the algorithm.
The MAP fit of the model is displayed in Fig 8. In general, the posterior estimates of the prior parameters were well correlated with their true value, except for the prior shape parameter α0. This lack of correlation did not, however, harm much the fit of the variance (see below), showing that the model fit was able to accurately recover the expected prior variability of each subject, which depended on α0. The NIGDM parameters were highly correlated with their original value.
In order to evaluate the identifiability of our model, we performed the quadratic approximation to the posterior covariance of the fitted HAFVF parameters. The average result, displayed in Fig 9, shows that covariance between model parameters was low, except at the top level. This indicates that each of parameter had a distinguishable effect on the loss. The higher variance and covariance of the parameters αβ and ββ relates to the fact that the influence of these prior parameters vanishes as more and more data is observed, in accordance with the Bernstein-Von-Mises theorem. In practice, this means that αβ and ββ should not be given behavioural interpretation but should rather be viewed as regularizers of the model.
We also looked at how the true prior expected value of w and b and variance correlated with their estimate from the posterior distribution. All of these correlated well with their generative correspondent, with the notable exception of the expected value of the second-level memory (Fig 10). This confirms again the role of regulizer of b over w.
TD learning with the HAFVF
To study the TD learning described in the Temporal difference learning section, we built two similar Markov Decision Processes (MDP) which are described in Fig 11. In brief, they consist of 5 different states where the agent has to choose the best action in order to reach a reward (r = 5) delivered once a specific state has been reached (hereafter, we will use “rewarded state” and “rewarded state-action” interchangeably). The task consisted for the agent to learn the current best policy during a 1000-trial experiment where the contingency was fixed to the first MDP during the first 500 trials, and then switched abruptly to the second MDP during the second half of the experiment.
The same prior was used for all subjects: the mean and variance prior was set to μ0 = 0, κ0 = 0.5, α0 = 3, β0 = 0.5. The forgetting factors shared the same flat prior . The priors on the discounting factor were set to a high value in order to discourage myopic strategies. The policy prior (see Appendix D) was set to a high value (π0 = 5.) in order to limit the impact of initial choices on the computation of the state-action value.
Results are displayed in Fig 12. In both experiments, agents had a similar behaviour during the first phase of the experiment: they both learned well the first contingency by assigning an accurate value to each state-action pair in order to reach the rewarded state more often. As expected, after the contingency change, the agents in experiment (i) took a longer time to adapt than they took to learn the initial contingency, which can be seen from the steeper slope of the reward rate during the first half of the experiment wrt the second. This feature can only be observed if the agent weights its belief by some measure of certainty, which is not modelled in classical, non-Bayesian RL.
In experiment (ii), the CC changes less the environment structure than the CC in experiment (i). The agents were able to take this difference into account: the value of the effective memories dropped less, and so did the reward rate.
An important feature observed in these two experiments is that the expected value of γ adapted efficiently to the contingency: although we used a prior skewed towards high values of γ, its value tended to be initially low as the agents had no knowledge of the various action values. Also, this value increased afterwards, reflecting a gain in predictive accuracy. The drop of had the effect of pushing towards its prior, which in this case increased the posterior expectation of γ. At the same time, the uncertainty about γ increased, thereby enhancing the flexibility of this parameter.
Discussion
In this paper, we propose a new Bayesian Reinforcement Learning (RL) algorithm aimed at accounting for the adaptive flexibility of learning observed in animal and human subjects. This algorithm adapts continuously its learning rate to inferred environmental variability, and this adaptive learning rate is optimal under some assumptions about statistical properties of the environment. These assumptions take the form of prior distributions on the parameters of the latent and mixing weight variables. We illustrate different types of behaviour of the model when facing unexpected contingency changes by taking extreme case scenarios. These scenarios implemented four types of assumptions on the tendency of the environment to vary over time (first-level memory) and on the propensity of this environmental variability to change itself over time (second-level memory). This approach allowed us to reproduce the emergence of inflexible behaviour following prolonged experience of a stable environment, similar to empirical observations in animals. Indeed, it has long been known that extensive training leads to automatization of behaviour, called habits [5, 8, 9, 93–96], or procedural “system 1” actor ([95, 97]), which is characterized by a lack of flexibility (i.e. failure to adapt to contingency changes) and by reduction of computational costs, illustrated by the capacity to perform these behaviours concomitantly to other tasks [98–100]. These automatic types of behaviour are opposed to Goal-Directed behaviours ([99–101]) and share the common feature of being inflexible, either in terms of planning (for Model-Free RL for instance) or in terms of adaptation in general.
Regarding the actor part, we implemented a general, Bayesian decision-making algorithm that reflects in many ways the Full-DDM proposed by [3], as it samples the reward distribution associated to each action and selects at each time step the best option. These elementary decisions are integrated until a decision boundary is reached. Therefore, the actor maps the cognitive predictions of the critic onto specific behavioural outputs such as choice and reaction time (RT). Importantly, this kind of Bayesian evidence accumulation process for decision making is biologically plausible [102, 103], well suited for decision making in RL [104–107] and makes predictions that are in accordance with physiological models of learning and decision making [108]. Other noteworthy attempts have been made to integrate sampling-based decision-making and RL [109, 110] using the DDM. However, the present work is the first, to our knowledge, to frame the DDM as an optimal, Bayesian decision strategy to maximize long-term utility on the basis of value distributions inferred from a RL algorithm. We show that this RL-DDM association, and especially in the framework of the Full DDM [3], finds a grounded algorithmic justification in a Bayesian perspective, as the resulting policy mimics the one of an agent trying to infer the best decision given its posterior belief about the reward distribution.
Interestingly, under some slight modifications (i.e. assuming that the sampled rewards are not simulated but retrieved from the subject memory), it is similar to the model recently proposed by Bornstein et al. [111, 112]. More specifically, while our decision making scheme used a heuristic based on the asymptotic property of MCMC, the scheme of decision making proposed by Bornstein and colleagues might recall other approximation techniques such as Approximate Bayesian Computation (ABC [113]). Following this approach, data samples are generated according to some defined rule, and only the samples that match the actual previous observations are kept in memory to approximate the posterior distribution or, in our case, to evaluate the option with the greatest reward. This simple trick in the decision making process keeps the stochastic nature of the accumulation process, while directly linking the level of evidence to items retrieved from memory.
The HAFVF also recalls the learning model proposed by Behrens and colleagues [32] who studied the variations of human learning rate in volatile environments and showed that activity in Anterior Cingulate Cortex reflected their model estimate of environmental volatility. The AC-HAFVF exhibits several differences with respect to this model: first, and crucially, it uses a Stabilized Forgetting framework to modify the belief that the agent has in the parameter values at each level, unlike Behrens et al. who used a purely forward model, similar in this sense to the HGF. Second, our model used Mean-field VB to make inference about the parameter values, allowing it to approximate the posterior distribution at low cost. The drawback of this approach, however, is that the AC-HAFVF presented here does not allow us to compute posterior covariance of their parameters at each trial, in contrast to Behrens and colleagues. Given these differences, it would interesting to compare how these various models of adaptation to volatility fit actual behavioural data and how well their parameters follow recorded neurobiological signals.
An important feature of the HAFVF is that it can account for unstructured changes of contingency. In other words, it allows the agent to learn anew the state of the environment even if the transition that leads to this state has never been experienced before. This is an important feature that contrasts with Kalman filters and Hidden Markov Models [1]. Both approaches have their pros and cons: learning state transition probabilities makes sense in environment that enjoy specific regularity conditions, but if the environment is chaotic, they can lead to poor adaptation performance. The approach we adopted here makes sense in situations in which one expects that the environment may change in an unstructured way, e.g. in which an environment that has remained stable for a long period of time may (suddenly or progressively) change in a random and hence unpredictable manner. An intermediate approach could however be developed [35].
One major advantage of our model is that its parameters are easily interpretable as reflecting hidden behavioural features such as trial-wise effective memory, prior and posterior expected stability, etc. We detail how model parameters can be fitted to data in order to recover these behavioural features at the trial, subject and population levels. This approach could be used, for instance, to cluster subjects in high-stability seeking and low-stability seeking sub-populations, and correlate these behaviours to health conditions, neurophysiological measures or training condition (stress, treatment etc.) We show that, for a simulated dataset, the fitted posterior distribution of the parameters correlate well with their original value. Moreover, each of the layers of the model (learning, first level memory and second level memory) have interesting behavioural correlates in terms of accuracy and RT. These results show that the model is identifiable and that there is a low redundancy in the various layers of the model. Also, different priors over the expected distribution of rewards and environment stability led also to different outcomes in terms of reward rate and RT: subjects whom assumed large environmental stability were likely to act faster, but also to be less flexible and to gain less on average, than subjects whom assumed high likelihood of contingency changes. The second-level memory had different effect, in the sense that large memory tended to be associated with flexible or inflexible behavior, depending on whether past experience corresponded to volatile or stable environment, respectively.
This finding also resonates with the habitual learning literature, in which decreased flexibility is also typically associated with short average RT, reflecting lower cognitive cost (see also [74]). However, in contrast to our approach, Keramati and colleagues suggest that adaptations to changes in volatility would rely on switching between two alternative decision strategies: an information-seeking goal-directed controller and a greedy habitual system. Habits would consist in bypassing computationally expensive inference steps in order to maximize reward rate when information gathering is too costly. The AC-HAFVF does not require to achieve model selection prior to making a decision: on the contrary, computational cost of the decision process is optimized automatically during learning and inference. For instance, VPI-guided evidence accumulation (see Appendix E) is achieved at a cost virtually identical of inference using a Q-value sampling strategy. Also, the switch from information-gathering to pure value-based selection strategy is natural when the posterior variance of the action values decreases, and the VPI vanishes as the average return becomes certain. Furthermore, using Q-value sampling only, we can see that the behaviour turns from being stochastic and explorative to being deterministic through training, again confirming that the learning and decision scheme we propose can account for the emergence of automatic behaviours. In turn, this could only happen if the environment is considered as stable by the agent. Further developments of the model could show how the threshold (e.g. [114]) and the start point (e.g. [115]) could be adapted to optimize the exploration policy and the cost of decisions.
Finally, we extended our work to MDP and multi-stages tasks. More than adding a mere reparameterization of TD learning, we propose a framework in which the time discount factor is considered by the agent as a latent variable: this opens the possibility of studying how animals and humans adapt their long-term / short-term return balance in different conditions of environmental variability. We show in the results that deep CC provoke a reset of the parameters, and lead subjects to erase their acquired knowledge of the posterior value of γ. We also show that this ability to adapt γ to the uncertainty of the environment can also be determinant at the beginning of the task, where nothing is known about the long term return of each action, and where individuals might benefit of a low expectation of γ that contrasts with the high expectation of subjects that have a deeper knowledge of the environment.
The present model is based on forgetting, which is an important feature in RL that should be differentiated from learning. In signal processing and related fields where online learning is required, learning can be seen as the capacity to build on previous knowledge (or assess a posterior probability distribution in a Bayesian context) to perform inference at the present step. Forgetting, on the contrary, is the capacity to erase the learning to come back at a naive, initial state of learning. The algorithm we propose learns a posterior belief of the data distribution and infers how likely this belief is to be valid in the next time step, on the basis of past environmental stability. This allows the algorithm to decide when and how much to forget its past belief in order to adapt to new contingencies. This feature sets our algorithm apart from previous proposals such as HGF, in which the naive prior looses its importance as learning goes on, and where the learner has no possibility of coming back to his initial knowledge. This lack of capacity to forget implies that the agent can be easily fooled by its past experience, whereas our model is more resistant in such cases, as its point of reference is fixed (which is the common feature of SF algorithms, see above). We have shown in the results that our model outperforms the HGF both in its fitting capability and its capacity to learn new observations with a given prior configuration. The AC-HAFVF should help to flexibly model learning in these contexts, and find correlates between physiological measures, such as dopaminergic signals [116, 117], and precise model predictions in term of memory and flexibility.
The SF scheme we have used, where the previous posterior is compared with a naive prior to optimize the forgetting factor, is widely diffused in the signal processing community [2, 35–37, 40, 42, 118, 119] and finds grounded mathematical justifications for error minimization in recursive Bayesian estimation [120]. However, it is the first time, to our knowledge, that this family of algorithms is applied to the study of RL in animals. We show that the two algorithms (RL and SF) share deep common features: for instance, the HAFVF and other similar algorithms ([57]) can be used with a naive prior θ0 set to 0, in which case the update equations reduce somehow to a classical Q-learning algorithm ([1] and Appendix B). Another interesting bound between the two fields emerges when the measure of the environment volatility is built hierarchically: an interesting consequence of the forgetting algorithm we propose is that, when observations are not made, the agent erases progressively its memory of past events. This leads to counterfactual learning schedule that favors exploration over exploitation at a rate dictated by the learned stability of the environment (see Appendix C for a development). Crucially, this updating scheme, and the consecutive exploration policy, flows directly from the hierarchical implementation of the SF scheme.
This work provides a tool to investigate learning rate adaptation in behaviour. Previous work has shown, for instance that the process of learning rate adaptation can be decomposed into various components that relate to different brain areas or networks [32, 121]. Nassar and colleagues [122] have also looked at the impact of age on learning rate adaptation, and found that older subjects were more likely to have a narrow expectation of the variance of the data they were observing, impairing thereby their ability to detect true CC. The AC-HAFVF shares many similarities with the algorithm proposed by Nassar and McGuire: it is designed to detect how likely an observation is to be caused by an abrupt CC, and adapts its learning rate accordingly. Also, this detection of CC depends in both models not only on the first and second moment of the observations, but also on their prior average and variability. We think, however, that our model is more flexible and biologically plausible than the model of Nassar and McGuire for three reasons. First, it is fully Bayesian: when fitting the model, we do not fit an expected variance of the outcome observed, but a prior distribution on this variance. This important difference is likely to predict better the observed data, as the subjects performing an experiment have probably some prior uncertainty about the variability of the outcome they will witness. Second, we considered the steadiness of the environment as another Bayesian estimate, meaning that the subjects will have some confidence (and posterior distribution) in the fact that the environment has truly changed or not. Third, we believe that our model is more general (i.e. less ad hoc) than the model of Nassar, as the general form of Eq 13 encompasses many models that can be formulated using distributions issued from the exponential family. This includes behavioural models designed to evolve in multi-stage RL tasks, such as TD learning or Model-Based RL [123].
It is important to emphasize that the model we propose does not intend to be universal. In simple situations, fitting a classical Q-learning algorithm could lead to similar or even better predictions than those provided by our model. We think, however, that our model complexity makes it useful in situations where abrupt changes occur during the experiment, or where long sequences (several hundreds of trials) of data are acquired. The necessity to account for this adaptability could be determined by comparing the accuracy of the HAFVF model (i.e. the model evidence) to the one obtained from a simpler Bayesian Q-Learning algorithm without forgetting.
The richness and generality of the AC-HAFVF opens countless possibilities of future developments. The habitual/goal-directed duality has been widely framed in terms of Model-Free/Model-Based control separation (e.g. [4, 11, 96, 124–128]). Although we do not model here a Model-Based learning algorithm, we think our work will ultimately help to discriminate various forms of inflexibility, and complete the whole picture of our understanding of human RL: in the usual Model-Free/Model-Based control duality, overconfidence in a Model-Free controller means that the subject will need to go through the same sequences of state-action transitions over and over to downweight actions situated early in a sequence. This contrasts with Model-Based control, which can immediately adapt its policy when an action situated far from the current state is devaluated [129]. The current implementation of the AC-HAFVF can model a lack of flexibility due to an overconfidence in the volatility of the environment, whereas adding a Model-Based component to the model might help to discriminate a lack of flexibility due to the overuse of a Model-Free strategy that characterizes the Model-Free/Model-Based paradigm. This balance could be learned and, in turn, be subject to forgetting. In short, implementation of Model-Based RL in an AC-HAFVF context might enrich greatly our understanding of how the balance between Model-Based and Model-Free RL works. This is certainly a development we intend to implement in the near future.
In conclusion, we provide a new Model-Free RL algorithm aimed at modelling behavioural adaptation to continuous and abrupt changes in the environment in a fully Bayesian way. We show that this model is flexible enough to reflect very different behavioural predictions in case of isolated unexpected events and prolonged change of contingencies. We also provide a biologically plausible decision making model that can be integrated elegantly in our learning algorithm, and completes elegantly the toolbox to simulate and fit datasets.
Supporting information
Acknowledgments
We thank the Consortium des Équipements de Calcul Intensif (CECI) in Belgium for providing us with the computational resources that made possible the initial explorations that led to this manuscript.
Data Availability
The full code that was used to simulate, fit and plot data is available at: https://figshare.com/s/21c266bd84b83f01c06b.
Funding Statement
This work was performed at the Institute of Neuroscience (IoNS) of the Université catholique de Louvain (Brussels, Belgium); it was supported by grants from the ARC (Actions de Recherche Concertées, Communauté Francaise de Belgique), from the Fondation Médicale Reine Elisabeth (FMRE), from the Fonds de la Recherche Scientifique (F.R.S.-FNRS) and from IdEx Bordeaux. A.Z. was a Senior Research Associate supported by INNOVIRIS and is currently 1st grade researcher for the french National Center for Scientific Research (CNRS). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Moens V. The Hierarchical Adaptive Forgetting Variational Filter. Proceedings of the 35th international conference on Machine learning - ICML’18. 2018;.
- 2. KULHAVÝ R, ZARROP MB. On a general concept of forgetting. International Journal of Control. 1993;58(4):905–924. 10.1080/00207179308923034 [DOI] [Google Scholar]
- 3. Ratcliff R, Rouder JN. Modeling Response Times for Two-Choice Decisions. Psychological Science. 1998;9(5):347–356. 10.1111/1467-9280.00067 [DOI] [Google Scholar]
- 4. Daw ND, Niv Y, Dayan P. Uncertainty-based competition between prefrontal and dorsolateral striatal systems for behavioral control. Nature Neuroscience. 2005;8(12):1704–1711. 10.1038/nn1560 [DOI] [PubMed] [Google Scholar]
- 5. Dickinson A. Actions and Habits: The Development of Behavioural Autonomy. Philosophical Transactions of the Royal Society B: Biological Sciences. 1985;308(1135):67–78. 10.1098/rstb.1985.0010 [DOI] [Google Scholar]
- 6. Dickinson A, Balleine BW, Watt A, Gonzalez F, Boakes RA. Motivational control after extended instrumental training. Animal Learning & Behavior. 1995;23(2):197–206. 10.3758/BF03199935 [DOI] [Google Scholar]
- 7. Yin HH, Knowlton BJBJBJ. The role of the basal ganglia in habit formation. Nature reviews Neuroscience. 2006;7(6):464–476. 10.1038/nrn1919 [DOI] [PubMed] [Google Scholar]
- 8. Hull CL. Principles of Behavior: An Introduction to Behavior Theory. In: The Journal of Abnormal and Social Psychology; 1943. [Google Scholar]
- 9. Seger Ca, Spiering BJ. A critical review of habit learning and the Basal Ganglia. Frontiers in systems neuroscience. 2011;5(August):66 10.3389/fnsys.2011.00066 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Dezfouli A, Balleine BW. Habits, action sequences and reinforcement learning. European Journal of Neuroscience. 2012;35(7):1036–1051. 10.1111/j.1460-9568.2012.08050.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Gillan CM, Otto AR, Phelps Ea, Daw ND. Model-based learning protects against forming habits. Cognitive, Affective, & Behavioral Neuroscience. 2015;15(3):523–536. 10.3758/s13415-015-0347-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Morris LS, Kundu P, Dowell N, Mechelmans DJ, Favre P, Irvine MA, et al. Fronto-striatal organization: Defining functional and microstructural substrates of behavioural flexibility. Cortex. 2016;74:118–133. 10.1016/j.cortex.2015.11.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Economides M, Kurth-Nelson Z, Lübbert A, Guitart-Masip M, Dolan RJ. Model-Based Reasoning in Humans Becomes Automatic with Training. PLoS Computational Biology. 2015;11(9):1–19. 10.1371/journal.pcbi.1004463 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Hélie S, Waldschmidt JG, Ashby FG. Automaticity in rule-based and information-integration categorization. Attention, perception & psychophysics. 2010;72(4):1013–1031. 10.3758/APP.72.4.1013 [DOI] [PubMed] [Google Scholar]
- 15. MacLeod CM. Half a century of research on the Stroop effect: An integrative review. Psychological Bulletin. 1991;109(2):163–203. 10.1037/0033-2909.109.2.163 [DOI] [PubMed] [Google Scholar]
- 16. Clark A. Whatever next? Predictive brains, situated agents, and the future of cognitive science. The Behavioral and brain sciences. 2013;36(3):181–204. 10.1017/S0140525X12000477 [DOI] [PubMed] [Google Scholar]
- 17. Friston K, Kiebel S. Predictive coding under the free-energy principle. Philosophical transactions of the Royal Society of London Series B, Biological sciences. 2009;364:1211–1221. 10.1098/rstb.2008.0300 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Hesselmann G, Sadaghiani S, Friston KJ, Kleinschmidt A. Predictive coding or evidence accumulation? False inference and neuronal fluctuations. PLoS ONE. 2010;5(3):1–5. 10.1371/journal.pone.0009926 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Friston KJ, Stephan KE, Montague R, Dolan RJ. Computational psychiatry: the brain as a phantastic organ. The Lancet Psychiatry. 2014;1(2):148–158. 10.1016/S2215-0366(14)70275-5 [DOI] [PubMed] [Google Scholar]
- 20. Mayrhauser L, Bergmann J, Crone J, Kronbichler M. Neural repetition suppression: evidence for perceptual expectation in object-selective regions. Frontiers in Human Neuroscience. 2014;8(April):1–8. 10.3389/fnhum.2014.00225 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Limongi R, Silva AM, Góngora-Costa B. Temporal prediction errors modulate task-switching performance. Frontiers in Psychology. 2015;6(August):1–10. 10.3389/fpsyg.2015.01185 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Kneissler J, Drugowitsch J, Friston K, Butz MV. Simultaneous learning and filtering without delusions: a Bayes-optimal combination of Predictive Inference and Adaptive Filtering. Frontiers in Computational Neuroscience. 2015;9(April):1–12. 10.3389/fncom.2015.00047 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Mathys C. A Bayesian foundation for individual learning under uncertainty. Frontiers in Human Neuroscience. 2011;5(May):1–20. 10.3389/fnhum.2011.00039 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Mathys CD, Lomakina EI, Daunizeau J, Iglesias S, Brodersen KH, Friston KJ, et al. Uncertainty in perception and the Hierarchical Gaussian Filter. Frontiers in Human Neuroscience. 2014;8(November):1–24. 10.3389/fnhum.2014.00825 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Iglesias S, Mathys C, Brodersen KH, Kasper L, Piccirelli M, den Ouden HEM, et al. Hierarchical Prediction Errors in Midbrain and Basal Forebrain during Sensory Learning. Neuron. 2013;80(2):519–530. 10.1016/j.neuron.2013.09.009 [DOI] [PubMed] [Google Scholar]
- 26. Vossel S, Bauer M, Mathys C, Adams RA, Dolan RJ, Stephan KE, et al. Cholinergic stimulation enhances Bayesian belief updating in the deployment of spatial attention. The Journal of neuroscience: the official journal of the Society for Neuroscience. 2014;34(47):15735–15742. 10.1523/JNEUROSCI.0091-14.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Hauser TU, Iannaccone R, Ball J, Mathys C, Brandeis D, Walitza S, et al. Role of the Medial Prefrontal Cortex in Impaired Decision Making in Juvenile Attention-Deficit/Hyperactivity Disorder. JAMA Psychiatry. 2014;71(10):1165 10.1001/jamapsychiatry.2014.1093 [DOI] [PubMed] [Google Scholar]
- 28. Diaconescu AO, Mathys C, Weber LAE, Daunizeau J, Kasper L, Lomakina EI, et al. Inferring on the Intentions of Others by Hierarchical Bayesian Learning. PLoS Computational Biology. 2014;10(9). 10.1371/journal.pcbi.1003810 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Schwartenbeck P, FitzGerald THB, Mathys C, Dolan R, Kronbichler M, Friston K. Evidence for surprise minimization over value maximization in choice behavior. Scientific Reports. 2015;5(1):16575 10.1038/srep16575 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Brazil IA, Mathys CD, Popma A, Hoppenbrouwers SS, Cohn MD. Representational uncertainty in the brain during threat conditioning and the link with psychopathic traits. Biological Psychiatry: Cognitive Neuroscience and Neuroimaging. 2017;(14):1–7. [DOI] [PubMed] [Google Scholar]
- 31. Yu AJ, Dayan P. Uncertainty, neuromodulation, and attention. Neuron. 2005;46(4):681–692. 10.1016/j.neuron.2005.04.026 [DOI] [PubMed] [Google Scholar]
- 32. Behrens TEJ, Woolrich MW, Walton ME, Rushworth MFS. Learning the value of information in an uncertain world. Nature neuroscience. 2007;10(9):1214–1221. 10.1038/nn1954 [DOI] [PubMed] [Google Scholar]
- 33. Doucet A, Johansen AM. A Tutorial on Particle filtering and smoothing: Fiteen years later. The Oxford handbook of nonlinear filtering. 2011;(December 2008):656–705. [Google Scholar]
- 34. Doucet A, De Freitas N, Gordon N. Sequential Monte Carlo Methods in Practice. Springer; New York: 2001; p. 178–195. 10.1198/tech.2003.s23 [DOI] [Google Scholar]
- 35.Azizi S, Quinn A. A data-driven forgetting factor for stabilized forgetting in approximate Bayesian filtering. In: 2015 26th Irish Signals and Systems Conference (ISSC). vol. 11855. IEEE; 2015. p. 1–6. Available from: http://ieeexplore.ieee.org/document/7163747/.
- 36. Smidl V, Quinn A. Variational Bayesian Filtering. IEEE Transactions on Signal Processing. 2008;56(10):5020–5030. 10.1109/TSP.2008.928969 [DOI] [Google Scholar]
- 37.Smidl V, Gustafsson F. Bayesian estimation of forgetting factor in adaptive filtering and change detection. In: 2012 IEEE Statistical Signal Processing Workshop (SSP). 1. IEEE; 2012. p. 197–200. Available from: http://ieeexplore.ieee.org/document/6319658/.
- 38. Özkan E, Šmídl V, Saha S, Lundquist C, Gustafsson F. Marginalized adaptive particle filtering for nonlinear models with unknown time-varying noise parameters. Automatica. 2013;49(6):1566–1575. 10.1016/j.automatica.2013.02.046 [DOI] [Google Scholar]
- 39. Laar TVD, Cox M, Diepen AV, Vries BD. Variational Stabilized Linear Forgetting in State-Space Models. 2017;(Section II):848–852. [Google Scholar]
- 40. Smidl V, Quinn A. Mixture-based extension of the AR model and its recursive Bayesian identification. IEEE Transactions on Signal Processing. 2005;53(9):3530–3542. 10.1109/TSP.2005.853103 [DOI] [Google Scholar]
- 41. Masegosa A, Nielsen TD, Langseth H, Ramos-Lopez D, Salmeron A, Madsen AL. Bayesian Models of Data Streams with Hierarchical Power Priors. International Conference on Machine Learning (ICM). 2017;70:2334–2343. [Google Scholar]
- 42. Dedecius K, Hofman R. Autoregressive model with partial forgetting within Rao-Blackwellized particle filter. Communications in Statistics: Simulation and Computation. 2012;41(5):582–589. 10.1080/03610918.2011.598992 [DOI] [Google Scholar]
- 43. Sutton RS, Barto AG. Introduction to Reinforcement Learning. Learning. 1998;4:1–5. [Google Scholar]
- 44.Dearden R, Friedman N, Russell S. Bayesian Q-Learning. In: American Association of Artificial Intelligence (AAAI)-98; 1998. p. 761–768.
- 45.Dearden R, Dearden R, Friedman N, Friedman N, Andre D, Andre D. Model based Bayesian exploration. Proceedings of the fifteenth Conference on Uncertainty in Artificial Intelligence. 1999;(Howard 1966):150–159.
- 46. Bishop CM. Pattern Recognition and Machine Learning; 2006. [Google Scholar]
- 47. Jaakkola TS, Jordan MI. A variational approach to Bayesian logistic regression models and their extensions. Aistats. 1996;(AUGUST 2001). [Google Scholar]
- 48. Jaakkola TS, Jordan MI. Bayesian parameter estimation via variational methods. Statistics And Computing. 2000;10(1):25–37. 10.1023/A:1008932416310 [DOI] [Google Scholar]
- 49.Blei DM, Kucukelbir A, McAuliffe JD. Variational Inference: A Review for Statisticians. arXiv. 2016; p. 1–33.
- 50. Paisley J, Blei D, Jordan M. Variational Bayesian Inference with Stochastic Search. Icml. 2012;(2000):1367–1374. [Google Scholar]
- 51.Salimans T, Kingma DP, Welling M. Markov Chain Monte Carlo and Variational Inference: Bridging the Gap. International Conference on Machine Learning. 2015;.
- 52. Kingma DP, Rezende DJ, Mohamed S, Welling M. Semi-Supervised Learning with Deep Generative Models. 2014; p. 1–9. [Google Scholar]
- 53.Ranganath R, Tran D, Blei DM. Hierarchical Variational Models. arXiv. 2014; p. 1–9.
- 54.Rezende DJ, Mohamed S. Variational Inference with Normalizing Flows. Proceedings of the 32nd International Conference on Machine Learning. 2015;37:1530–1538.
- 55. Blei DM. Variational Inference. CsPrincetonEdu. 2002; p. 1–12. [Google Scholar]
- 56.V Smidl AQ. Bayesian estimation of non-stationary AR model parameters via an unknown forgetting factor. In: 3rd IEEE Signal Processing Education Workshop. 2004 IEEE 11th Digital Signal Processing Workshop, 2004. 6. IEEE; 2004. p. 221–225. Available from: http://staff.utia.cas.cz/smidl/files/publ/taos04.pdfhttp://ieeexplore.ieee.org/document/1437946/.
- 57. Smidl V. The Variational Bayes Approach in Signal Processing; 2004. Available from: http://staff.utia.cz/smidl/Public/Thesis-final.pdf. [Google Scholar]
- 58. Knowles D, Minka TP. Non-conjugate variational message passing for multinomial and binary regression. Nips. 2011; p. 1–9. [Google Scholar]
- 59. Bottou L, Peters J, Ch P, Quiñonero-Candela J, Charles DX, Chickering DM, et al. Counterfactual Reasoning and Learning Systems: The Example of Computational Advertising. Journal of Machine Learning Research. 2013;14:3207–3260. [Google Scholar]
- 60.Foerster J, Farquhar G, Afouras T, Nardelli N, Whiteson S. Counterfactual Multi-Agent Policy Gradients. Arxiv. 2017; p. 1–12.
- 61.Lawrence C, Sokolov A, Riezler S. Counterfactual Learning from Bandit Feedback under Deterministic Logging: A Case Study in Statistical Machine Translation. 2017;.
- 62. Mischel W, Ebbesen EB, Raskoff Zeiss A. Cognitive and attentional mechanisms in delay of gratification. Journal of Personality and Social Psychology. 1972;21(2):204–218. 10.1037/h0032198 [DOI] [PubMed] [Google Scholar]
- 63. Weatherly JN. On several factors that control rates of discounting. Behavioural Processes. 2014;104:84–90. 10.1016/j.beproc.2014.01.020 [DOI] [PubMed] [Google Scholar]
- 64. Story GW, Vlaev I, Seymour B, Darzi A, Dolan RJ. Does temporal discounting explain unhealthy behavior? A systematic review and reinforcement learning perspective. Frontiers in behavioral neuroscience. 2014;8(March):76 10.3389/fnbeh.2014.00076 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65. McClure SM, Laibson DI, Loewenstein G, Cohen JD. Separate neural systems value immediate and delayed monetary rewards. Science (New York, NY). 2004;306(5695):503–507. 10.1126/science.1100907 [DOI] [PubMed] [Google Scholar]
- 66. Schultz W. Updating dopamine reward signals. Current opinion in neurobiology. 2013;23(2):229–238. 10.1016/j.conb.2012.11.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67. Bermudez MA, Schultz W. Timing in reward and decision processes. Philosophical transactions of the Royal Society of London Series B, Biological sciences. 2014;369(1637):20120468 10.1098/rstb.2012.0468 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68. Takahashi T. Loss of self-control in intertemporal choice may be attributable to logarithmic time-perception. Medical Hypotheses. 2005;65(4):691–693. 10.1016/j.mehy.2005.04.040 [DOI] [PubMed] [Google Scholar]
- 69. Vincent BT. Hierarchical Bayesian estimation and hypothesis testing for delay discounting tasks. Behavior Research Methods. 2015. 10.3758/s13428-015-0672-2 [DOI] [PubMed] [Google Scholar]
- 70. Kurth-Nelson Z, Bickel W, Redish AD. A theoretical account of cognitive effects in delay discounting. European Journal of Neuroscience. 2012;35(7):1052–1064. 10.1111/j.1460-9568.2012.08058.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Wyatt J. Exploration and Inference in Learning From Reinforcement; 1998. Available from: https://www.era.lib.ed.ac.uk/handle/1842/532.
- 72. Thompson WR. On the Likelihood that One Unknown Probability Exceeds Another in View of the Evidence of Two Samples. Biometrika. 1933. 10.1093/biomet/25.3-4.285 [DOI] [Google Scholar]
- 73. Kaufmann E, Korda N, Munos R. Thompson Sampling: An Asymptotically Optimal Finite Time Analysis. International Conference on Algorithmic Learning Theory. 2012;(1):199–213. 10.1007/978-3-642-34106-9_18 [DOI] [Google Scholar]
- 74. Keramati M, Dezfouli A, Piray P. Speed/accuracy trade-off between the habitual and the goal-directed processes. PLoS computational biology. 2011;7(5):e1002055 10.1371/journal.pcbi.1002055 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Viejo G, Khamassi M, Brovelli A, Girard B. Modelling choice and reaction time during instrumental learning through the coordination of adaptive working memory and reinforcement learning. Fourth Symposium on Biology of Decision—Making (SBDM 2014). 2014;9(August). [DOI] [PMC free article] [PubMed]
- 76. Mcallister R, Dziugaite K. Bayesian Reinforcement Learning. 2013;35(March):1–21. [Google Scholar]
- 77. Feller W. An Introduction to Probability Theory and Its Applications. Wiley; 1968;2:509. [Google Scholar]
- 78. Ratcliff R. A theory of memory retrieval. Psychological Review. 1978;85(2):59–108. 10.1037/0033-295X.85.2.59 [DOI] [Google Scholar]
- 79. Smith PL. Stochastic Dynamic Models of Response Time and Accuracy: A Foundational Primer. Journal of Mathematical Psychology. 2000;44(3):408–463. 10.1006/jmps.1999.1260 [DOI] [PubMed] [Google Scholar]
- 80. Amari Si. Natural Gradient Works Efficiently in Learning. Neural Computation. 1998;10(2):251–276. 10.1162/089976698300017746 [DOI] [Google Scholar]
- 81. Sato MA. Online Model Selection Based on the Variational Bayes. Neural Comput. 2001;13(7):1649–1681. 10.1162/089976601750265045 [DOI] [Google Scholar]
- 82. Hoffman M, Blei DM, Wang C, Paisley J. Stochastic Variational Inference. 2012;. [Google Scholar]
- 83.Martens J. New insights and perspectives on the natural gradient method. 2014;.
- 84. Ghosal S, Ghosh JK, van der Vaart AW. Convergence rates of posterior distributions. The Annals of Statistics. 2000;28(2):500–531. 10.1214/aos/1016218228 [DOI] [Google Scholar]
- 85. Zenon A, Solopchuk O, Pezzulo G. An information-theoretic perspective on the costs of cognition. bioRxiv. 2018; p. 208280. 10.1101/208280 [DOI] [PubMed] [Google Scholar]
- 86. Moens V, Zenon A. Recurrent Auto-Encoding Drift Diffusion Model. bioRxiv. 2018. 10.1101/220517 [DOI] [Google Scholar]
- 87. Mnih A, Gregor K. Neural Variational Inference and Learning in Belief Networks. ArXiv statML. 2014;32(October):1–20. [Google Scholar]
- 88. Friston K, Mattout J, Trujillo-Barreto N, Ashburner J, Penny W. Variational free energy and the Laplace approximation. NeuroImage. 2007;34(1):220–234. 10.1016/j.neuroimage.2006.08.035 [DOI] [PubMed] [Google Scholar]
- 89. Daw ND. Trial-by-trial data analysis using computational models. Attention & Performance XXIII. 2011; p. 1–26. [Google Scholar]
- 90.Kingma DP, Welling M. Auto-Encoding Variational Bayes. 2013;.
- 91. Ratcliff R, Tuerlinckx F. Estimating parameters of the diffusion model: Approaches to dealing with contaminant reaction times and parameter variability. Psychonomic Bulletin & Review. 2002;9(3):438–481. 10.3758/BF03196302 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 92.Kingma DP, Ba JL. Adam: a Method for Stochastic Optimization. International Conference on Learning Representations 2015. 2015; p. 1–15.
- 93. Dickinson A, Nicholas DJ. Irrelevant incentive learning during instrumental conditioning: The role of the drive-reinforcer and response-reinforcer relationships. The Quarterly Journal of Experimental Psychology Section B. 1983. [Google Scholar]
- 94. Wood W, Ruenger D. Psychology of Habit. Annual Review of Psychology. 2015;(September):1–26. [DOI] [PubMed] [Google Scholar]
- 95. Dayan P. Goal-directed control and its antipodes. Neural Networks. 2009;22(3):213–219. 10.1016/j.neunet.2009.03.004 [DOI] [PubMed] [Google Scholar]
- 96. Dolan RJ, Dayan P. Goals and habits in the brain. Neuron. 2013;80(2):312–325. 10.1016/j.neuron.2013.09.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97. Kahneman D, Frederick S. Representativeness Revisited: Attribute Substitution in Intuitive Judgment In: Gilovich T, Griffin D, Kahneman D, editors. Heuristics and Biases. Cambridge University Press; 2001. p. 49–81. Available from: http://ebooks.cambridge.org/ref/id/CBO9780511808098A012. [Google Scholar]
- 98. Schneider W, Shiffrin RM. Controlled and Automatic Human Information Processing: I. Detection, Search, and Attention. Psychological Review. 1977;84(1):1–66. 10.1037/0033-295X.84.1.1 [DOI] [Google Scholar]
- 99. Moors A, De Houwer J, Houwer JD. Automaticity: A Theoretical and Conceptual Analysis. Psychological Bulletin. 2006;132(2):297–326. 10.1037/0033-2909.132.2.297 [DOI] [PubMed] [Google Scholar]
- 100. Ashby FG, Crossley MJ. Automaticity and multiple memory systems. Wiley Interdisciplinary Reviews: Cognitive Science. 2012;3(3):363–376. 10.1002/wcs.1172 [DOI] [PubMed] [Google Scholar]
- 101. Waldschmidt JG, Ashby FG. Cortical and striatal contributions to automaticity in information-integration categorization. NeuroImage. 2011;56(3):1791–1802. 10.1016/j.neuroimage.2011.02.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102. Hanes DP, Schall JD. Neural control of voluntary movement initiation. Science. 1996;274(5286):427–430. 10.1126/science.274.5286.427 [DOI] [PubMed] [Google Scholar]
- 103. Roitman JD, Shadlen MN. Response of neurons in the lateral intraparietal area during a combined visual discrimination reaction time task. The Journal of neuroscience: the official journal of the Society for Neuroscience. 2002;22(21):9475–9489. 10.1523/JNEUROSCI.22-21-09475.2002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104. Soltani A, Wang XJ. Synaptic computation underlying probabilistic inference. Nature Neuroscience. 2010;13(1):112–119. 10.1038/nn.2450 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 105. Gluth S, Rieskamp J, Buchel C. Deciding When to Decide: Time-Variant Sequential Sampling Models Explain the Emergence of Value-Based Decisions in the Human Brain. Journal of Neuroscience. 2012;32(31):10686–10698. 10.1523/JNEUROSCI.0727-12.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 106. Rombouts JO, Bohte SM, Roelfsema PR. Neurally Plausible Reinforcement Learning of Working Memory Tasks. Nips. 2012; p. 1–9. [Google Scholar]
- 107. Kurzawa N, Summerfield C, Bogacz R. Neural Circuits Trained with Standard Reinforcement Learning Can Accumulate Probabilistic Information during Decision Making. Neural Computation. 2017;29(2):368–393. 10.1162/NECO_a_00917 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 108. Smith PL, Ratcliff R. Diffusion and Random Walk Processes In: Elsevier Ltd, editor. International Encyclopedia of the Social & Behavioral Sciences. vol. 6 Elsevier; 2015. p. 395–401. Available from: http://www.sciencedirect.com/science/article/pii/B0080430767006203http://linkinghub.elsevier.com/retrieve/pii/B9780080970868430370. [Google Scholar]
- 109. Frank MJ, Gagne C, Nyhus E, Masters S, Wiecki TV, Cavanagh JF, et al. fMRI and EEG Predictors of Dynamic Decision Parameters during Human Reinforcement Learning. Journal of Neuroscience. 2015;35(2):485–494. 10.1523/JNEUROSCI.2036-14.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 110. Pedersen ML, Frank MJ, Biele G. The drift diffusion model as the choice rule in reinforcement learning. Psychonomic Bulletin & Review. 2017;24(4):1234–1251. 10.3758/s13423-016-1199-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- 111. Bornstein AM, Khaw MW, Shohamy D, Daw ND. Reminders of past choices bias decisions for reward in humans. Nature Communications. 2017;8(May 2015):15958 10.1038/ncomms15958 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 112. Bornstein AM, Norman KA. Reinstated episodic context guides sampling-based decisions for reward. Nature Neuroscience. 2017;20(7):997–1003. 10.1038/nn.4573 [DOI] [PubMed] [Google Scholar]
- 113. Lintusaari J, Gutmann MU, Dutta R, Kaski S, Corander J. Fundamentals and recent developments in approximate Bayesian computation. Systematic Biology. 2017;66(1):e66–e82. 10.1093/sysbio/syw077 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 114. Cavanagh JF, Wiecki TV, Cohen MX, Figueroa CM, Samanta J, Sherman SJ, et al. Subthalamic nucleus stimulation reverses mediofrontal influence over decision threshold. Nature Neuroscience. 2011;14(11):1462–1467. 10.1038/nn.2925 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 115. Mulder MJ, Wagenmakers EJ, Ratcliff R, Boekel W, Forstmann BU. Bias in the brain: a diffusion model analysis of prior probability and potential payoff. The Journal of neuroscience: the official journal of the Society for Neuroscience. 2012;32(7):2335–2343. 10.1523/JNEUROSCI.4156-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 116. Morita K, Kato A. Striatal dopamine ramping may indicate flexible reinforcement learning with forgetting in the cortico-basal ganglia circuits. Frontiers in Neural Circuits. 2014;8(April):1–15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 117. Kato A, Morita K. Forgetting in Reinforcement Learning Links Sustained Dopamine Signals to Motivation. PLoS Computational Biology. 2016;12(10):1–41. 10.1371/journal.pcbi.1005145 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 118. Kulhavy R, Karny M. Tracking of slowly varying parameters by directional forgetting. Preprints 9ih IFAC Congress. 1984;10:178–183. [Google Scholar]
- 119. Kulhavý R, Kraus FJ. On Duality of Exponential and Linear Forgetting. IFAC Proceedings Volumes. 1996;29(1):5340–5345. 10.1016/S1474-6670(17)58530-4 [DOI] [Google Scholar]
- 120. Kárný M. Approximate Bayesian recursive estimation. Information Sciences. 2014;285(1):100–111. [Google Scholar]
- 121. McGuire JT, Nassar MR, Gold JI, Kable JW. Functionally Dissociable Influences on Learning Rate in a Dynamic Environment. Neuron. 2014;84(4):870–881. 10.1016/j.neuron.2014.10.013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 122. Nassar MR, Bruckner R, Gold JI, Li SC, Heekeren HR, Eppinger B. Age differences in learning emerge from an insufficient representation of uncertainty in older adults. Nature Communications. 2016;7(May 2015):1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 123. Doll BB, Simon DA, Daw ND. The ubiquity of model-based reinforcement learning. Current opinion in neurobiology. 2012;22(6):1075–1081. 10.1016/j.conb.2012.08.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 124. Daw ND, Gershman SJ, Seymour B, Dayan P, Dolan RJ. Model-based influences on humans’ choices and striatal prediction errors. Neuron. 2011;69(6):1204–1215. 10.1016/j.neuron.2011.02.027 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 125. Wunderlich K, Smittenaar P, Dolan RJ. Dopamine enhances model-based over model-free choice behavior. Neuron. 2012;75(3):418–424. 10.1016/j.neuron.2012.03.042 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 126. Otto AR, Raio CM, Chiang A, Phelps EA, Daw ND. Working-memory capacity protects model-based learning from stress. Proceedings of the National Academy of Sciences. 2013;110(52):20941–20946. 10.1073/pnas.1312011110 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 127. Schad DJ, Jünger E, Sebold M, Garbusow M, Bernhardt N, Javadi AH, et al. Processing speed enhances model-based over model-free reinforcement learning in the presence of high working memory functioning. Frontiers in Psychology. 2014;5(December):1–10. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 128. Kool W, Gershman SJ, Cushman FA. Cost-Benefit Arbitration Between Multiple Reinforcement-Learning Systems. Psychological Science. 2017; p. 095679761770828. 10.1177/0956797617708288 [DOI] [PubMed] [Google Scholar]
- 129. Gläscher J, Daw N, Dayan P, O’Doherty JP, Doherty JPO, Gläscher J, et al. Article States versus Rewards: Dissociable Neural Prediction Error Signals Underlying Model-Based and Model-Free Reinforcement Learning. Neuron. 2010;66(4):585–595. 10.1016/j.neuron.2010.04.016 [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The full code that was used to simulate, fit and plot data is available at: https://figshare.com/s/21c266bd84b83f01c06b.