Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2023 Jan;144:None. doi: 10.1016/j.neubiorev.2022.104959

Computational perspectives on human fear and anxiety

Yumeya Yamamori a,, Oliver J Robinson a,b
PMCID: PMC10564627  PMID: 36375584

Abstract

Fear and anxiety are adaptive emotions that serve important defensive functions, yet in excess, they can be debilitating and lead to poor mental health. Computational modelling of behaviour provides a mechanistic framework for understanding the cognitive and neurobiological bases of fear and anxiety, and has seen increasing interest in the field. In this brief review, we discuss recent developments in the computational modelling of human fear and anxiety. Firstly, we describe various reinforcement learning strategies that humans employ when learning to predict or avoid threat, and how these relate to symptoms of fear and anxiety. Secondly, we discuss initial efforts to explore, through a computational lens, approach-avoidance conflict paradigms that are popular in animal research to measure fear- and anxiety-relevant behaviours. Finally, we discuss negative biases in decision-making in the face of uncertainty in anxiety.

Keywords: Anxiety, Fear, Computational modelling, Generative models, Reinforcement learning, Approach-avoidance conflict, Uncertainty, Decision-making

Highlights

  • There is increasing interest in computational approaches to human fear and anxiety.

  • We focus on ‘generative models’ of fear-/anxiety-related behaviours.

  • Key themes of aversive learning, approach-avoidance conflict, and uncertainty.

  • We discuss candidate computational mechanisms of pathological fear/anxiety.

1. Introduction

Fear and anxiety are adaptive states which elicit defensive behaviours that help organisms avoid harm and ultimately survive (LeDoux and Pine, 2016). They are commonly distinguished in the literature on the basis of the immediacy, proximity and/or certainty of the threat, with fear and fear-related responses evoked by proximal threats and anxiety/anxiety responses evoked by distal and uncertain threats (Grillon, 2008, Perusini and Fanselow, 2015, Mobbs et al., 2020) (although this may be an oversimplification at the neurobiological level (Fox and Shackman, 2019, Perusini and Fanselow, 2015, Daniel-Watanabe and Fletcher, 2021). Critically, while fear, anxiety and the responses they elicit (such as avoidance) can be adaptive, their excessive and/or inappropriate expression can have severe negative impacts on daily functioning. This, in conjunction with an individual’s set of beliefs and attitudes (Beck and Clark, 1988), may lead to fear- and anxiety-related disorders such as phobias or generalised anxiety disorder, respectively (Kessler et al., 2005) which collectively constitute leading causes of disability (Vos et al., 2017).

The computational approach, which we define here as inference of the unobserved causes of behaviour through generative models (Stephan and Mathys, 2014) (see Box 1 for the Glossary of bolded terms), has seen a surge in interest in the study of fear and anxiety (Bach and Dayan, 2017). This approach aims to derive the computations underlying defensive behaviours, as well as how these computations may be implemented at the neural level. This focus on computation aims to move the field beyond simply describing symptoms and towards understanding the cognitive and neurobiological mechanisms underlying fear and anxiety behaviours. Moreover, computational modelling of behaviour may offer superior measurement properties compared to more traditional summary-statistic-based methods (Price et al., 2019, Tipples, 2015). In this mini-review, we discuss recent themes emerging from this field, namely the roles of a) learning; b) approach-avoidance conflict; and c) uncertainty in human fear and anxiety.

Box 1. Glossary.

Computational concepts.

Active inference (Friston et al., 2017). Active inference is a Bayesian framework which assumes perception and action as problems of inference (Friston et al., 2013). Here, perception is defined as inferring the latent states of the world that cause observable outcomes. Action is defined as inferring policies (sequences of actions) that must be adopted to obtain certain outcomes. These problems are separated by assuming that actions are consequent on the predictions formed by perception. Like model-based reinforcement learning, the generative model includes information about future states and policies, which can be used to discern the optimal actions to obtain favourable outcomes. Optimising behaviour is achieved by maximising the observable evidence for this generative model (via the minimisation of variational free energy).

The drift-diffusion model (Ratcliff, 1978). The drift-diffusion model is a generative model which integrates choice and response time data. In the simplest form, it describes the accumulation of evidence (perceptual or value-based) for two competing responses (for example, whether a stimulus is large or small, or whether one stimulus is more valuable than the other). Evidence is accumulated until a threshold is reached, indicating more evidence for one response over the other. The form of evidence accumulated dictates choice, while the rate of evidence accumulation affects response times (faster evidence accumulation leads to shorter response times, and vice versa). The basic model has four parameters: a drift rate (controlling evidence accumulation); a threshold boundary (controlling how much evidence is required for a decision); initial evidence (for a priori biases in evidence); and a non-decision time (for decision-irrelevant factors preceding decisions).

Generative models. Mathematical models which describe how certain data (for example behaviour, response times, neural responses, or a combination of these data) are ‘generated'. With respect to generative modelling of behaviour, this consists of building models which represent a researcher’s hypotheses about the unobserved causes (e.g. computations that the brain performed) of the observed behavioural data. These models typically include free parameters which can reflect individual differences in cognitive processes (for example, a learning rate). The term ‘generative’ also implies that one can generate trial-by-trial artificial data from the model, with which to compare to the original data (often referred to as ‘posterior predictive checking’ (Gelman and Hill, 2007)) to determine how well the model captures the data. Examples include reinforcement learning models, logistic choice models, and drift-diffusion models.

Prospect theory (Kahneman and Tversky, 1979). Prospect theory is a generative model of risky economical decision-making. The model describes how to translate some monetary value (say the potential to earn £100) and its associated probability of occurrence (say 50% likelihood of earning the £100) into individual-specific subjective values. Biases in decision-making such as loss aversion can be modelled by transformations (e.g. scaling) of the value, which can also account for individual differences in sensitivity to value. Similar transformations can be performed for probabilities to model risk aversion.

Reinforcement learning. Reinforcement learning provides a framework for modelling reward-driven (or indeed punishment-driven) behaviour. In the context of modelling human or animal behaviour, a generative model takes the form of an ‘agent’ that performs certain actions within a certain environment, where the agent attempts to accrue as much reward as possible whilst minimising losses/punishment. Different agents can be specified to rely on different behavioural strategies (or ‘policies’). In the field of artificial intelligence, these strategies are commonly categorised into model-free and model-based (see below).

Reinforcement learning; model-free reinforcement learning. Model-free learning entails learning, from experience and trial-and-error, what actions are favourable (i.e. produce reward/avoid punishment, or otherwise lead to other favourable states) given the current state of the environment. This is a computationally efficient form of behavioural control since a model-free agent only needs to rely on state-action values to perform favourable actions. Put another way, a model-free agent does not require an explicit causal understanding of which actions lead to certain outcomes/other states (unlike model-based agents, see below).

Reinforcement learning; model-based reinforcement learning. Model-based strategies rely on an understanding of the causal structure of the environment, or more specifically an understanding of the most likely outcome given the current state and candidate action. Therefore, this causal structure can be used to plan future actions and consequent outcomes. Compared to model-free learning (see above), model-based planning is potentially computationally costly (scaling with the complexity of the environment) but confers greater behavioural flexibility, especially in dynamic environments.

State values. Information about an agent’s predictions about what positive or negative outcomes might occur in a certain ‘state’ of the environment is encoded in the state value. For example, this could be the agent’s prediction of an electric shock occurring when the agent observes a certain stimulus. In single-step reinforcement learning, this represents the predicted likelihood or magnitude of reward or threat associated with the state, whereas in multi-step reinforcement learning, it also includes future rewards/threats that might occur given the occurrence of the state. Typically, positively valenced outcomes (i.e. things one wants to obtain/approach such as food) lead to positive state values, whilst negatively valenced outcomes (i.e. things to be avoided such as pain) lead to negative values.

State-action values. In a similar fashion to state values, state-action values encode an agent’s predictions about what outcome might occur, given the state of the environment and the agent’s action. For example, this could be the agent’s prediction of an electric shock occurring if the agent presses a button (say rather than not pressing the button) in response to a certain stimulus.

Forms of anxiety - across this mini-review, we have used a number of terms that refer to different forms of anxiety. For clarity, we provide a brief description of the general terms used.

Pathologically anxious individuals. We use this phrase to denote individuals who report clinically relevant symptoms of anxiety, for example as assessed by clinical interviews.

Somatic vs cognitive anxiety. Previous research has proposed a distinction between these two dimensions of anxiety (Koksal and Power, 1990, Ree et al., 2008). Somatic anxiety involves physical symptoms of anxiety, such as hyperventilation, sweating and muscle tension. On the other hand, cognitive anxiety involves symptoms relating to thoughts, such as worry and difficulty concentrating.

State anxiety. The transient experience of anxiety, which may be induced by experimental procedures (e.g. anxiogenic tasks/manipulations) or experienced naturally.

Trait anxiety. Anxiety symptoms and related thoughts which are experienced over a relatively long period by an individual, which can sometimes be considered as part of their stable characteristics or personality.

2. Aversive learning

Learning to predict threat from the environment and act accordingly to avoid it is of fundamental importance for the survival of organisms. Building on the seminal work of Rescorla-Wagner (Rescorla and Wagner, 1971) and Pearce-Hall (Pearce and Hall, 1980), reinforcement learning (Sutton and Barto, 2018) (see Box 1) represents one of the most popular frameworks for quantitatively modelling learning about threats and punishments. It can be broadly divided into Pavlovian fear conditioning and instrumental learning (for a comprehensive classification of threat learning/behaviour, we direct the reader to this recent review (LeDoux and Pine, 2016)). In this section, we describe various reinforcement learning strategies that humans employ when learning to predict or avoid threat, and how these relate to symptoms of fear and anxiety.

2.1. Fear conditioning

Pavlovian fear conditioning was one of the earliest experimental models of fear and anxiety (Watson and Rayner, 1920), and is thought to describe the acquisition of fear in psychopathology. Briefly, Pavlovian conditioning involves the development of a learned reflexive response to a neutral cue, when paired with a biologically significant (e.g. threatening) stimulus/event. The Rescorla-Wagner learning model provides an elegant explanation of fear conditioning, by ascribing values (state values; see Box 1), Vs, to certain states of the environment, s, such as the presence of a cue predicting danger, at time, t (Fig. 1). Agents learn from discrepancies between their expected outcome and the actual outcome, ot (e.g. an electric shock which can be represented as a value of −1 [or sometimes 1 if there are no rewards present in the task, see Box 1]), which results in a prediction error, PEt=otVtst. This prediction error is used to update the state value via Vt+1(st)=Vt(st)+α·PEt, where learning is scaled by the learning rate, α0,1. Negative Vs values signal that the agent expects an aversive outcome to occur given s, modelling fear. The neurobiological validity of the Rescorla-Wagner learning model is supported by neural correlates of prediction errors during fear conditioning tasks in the human striatum (Delgado et al., 2008, Robinson et al., 2013, Seymour et al., 2007), with similar models (McNally et al., 2011) and neural correlates (McNally and Westbrook, 2006, Stanley et al., 2021) observed in the non-human animal literature. Further, autonomic measures (e.g. skin conductance) track model-predicted changes in Vs (Li et al., 2011, Zhang et al., 2016) supporting the notion of state-values as signalling fear.

Fig. 1.

Fig. 1

Pavlovian fear conditioning and instrumental avoidance learning. A) In generative models of Pavlovian fear conditioning, agents update an expectation of the value, Vs, of a certain state of the environment, s. In a simple fear conditioning experiment, this could be the expectation of a shock given the presentation of a certain cue. B) In instrumental learning, agents instead learn the expected value, Qs,a, of an action, a, given state, s. This could be where a participant learns to press a button after seeing a certain cue to avoid shock. C) Conceptual data in a common reversal learning paradigm (which involves elements of both value acquisition and extinction), where participants learn to associate a cue (Pavlovian conditioning) or a button-press after observing a cue (instrumental learning) with shock or no shock. The lines represent how the expected values, Vs and Qs,a, change after each observed outcome (red and grey dots represent shock and no shock outcomes on each trial, respectively). Initially in the task, the cue/cue and button-press leads to shock with 80% probability, but in the second half of the task, they only lead to shock on 20% of trials. For demonstration, we force the agent to button-press on each trial, but in practice, participants can choose between multiple actions. The dynamics of Vs and Qs,a can be modelled through similar mechanisms such as the Rescorla-Wagner learning rule. We demonstrate two models: a simple Rescorla-Wager model which involves a fixed learning rate (in solid lines), and a ‘hybrid Rescorla-Wagner Pearce-Hall’ model which allows for a dynamic learning rate (In dashed lines). In both models, the agents learn that the cue/cue and button-press is initially associated with the shock (as the expected values become negative). When the contingencies change midway through the task, the agents learn that they are no longer associated (i.e. the expected values become more positive). D) Trial-level learning rates in each model. The fixed learning rate stays constant. The hybrid model relies on a high learning rate at the start of the task and midway though (when the contingencies change), which decreases after these moments. The high learning rate captures the ‘unpredictability’ of recent outcomes, as the agent has not yet learned the current contingency. The learning rate decreases as the outcomes become more predictable. The hybrid model is typically a better explanation of human behavioural data (Li et al., 2011, Homan et al., 2019, Tzovara et al., 2018, Piray et al., 2019). For this simulation, we used an initial learning rate of 0.3 for both models, and a second-order learning rate (see η in the main text) of 0.3 in the hybrid model.

The classical Rescorla-Wagner model assumes a static learning rate, but models that allow for dynamic learning rates commonly provide more parsimonious accounts of fear conditioning (Li et al., 2011, Homan et al., 2019, Tzovara et al., 2018, Piray et al., 2019) (Fig. 1 C-D). Inspired by the Pearce-Hall learning model (Pearce and Hall, 1980), these models (which are referred to as ‘hybrid’ models as they combine Rescorla-Wagner and Pearce-Hall mechanisms) incorporate the notion of ‘predictability’ into the learning process, by scaling the rate of learning in proportion to the magnitude of recent prediction errors: αt+1st=η·PEt+(1η)·αtst, where the rate of scaling is parameterised by η0,1. This means that greater error magnitudes (i.e. low predictability) call for rapid learning, whereas learning is slower for low error magnitudes (i.e. high predictability). Neural correlates of model-predicted changes in learning rate in the amygdala (Li et al., 2011, Zhang et al., 2016) and dorsal anterior cingulate cortex (Piray et al., 2019) provide support for the potential neural implementation of this computational mechanism. These effects parallel findings from the animal literature, where for example amygdala inactivation leads to behaviour consistent with learning rate impairments (Roesch et al., 2010). With respect to pathological fear, PTSD symptoms in combat-exposed veterans were positively associated with learning rate variability (Homan et al., 2019) (i.e. greater values of η), meaning that those with more severe symptoms were more sensitive to changes in the predictability of outcomes. The opposite effect was reported, however, for trait social anxiety (Piray et al., 2019), perhaps indicating unique learning mechanisms across these disorders (see below for a further discussion of anxiety, learning rate and uncertainty). Finally, recent work has also shown that visual attention modulates the rate of learning, by selectively strengthening Vs for attended cues (Wise et al., 2019).

Fear conditioning can also describe how fear is overcome – a process referred to as fear extinction. During extinction, a cue previously associated with threat is repeatedly presented in its absence, which leads to decreases in cue-induced fear responses (Bouton, 1993). This is utilised clinically in exposure therapy (Craske et al., 2006, Hofmann, 2008), which aims to reduce pathological fear via extinction. Unfortunately, fear sometimes spontaneously re-emerges even after extinction learning, which is difficult to reconcile with Rescorla-Wagner accounts of learning and presents a hurdle for therapy. The current theoretical understanding of this effect is that extinction leads to a new ‘safety’ association which competes with the original association, potentially causing recovery of fear (Bouton et al., 2021). However, recent studies suggest that this might be subject to individual differences in learning about latent causes of the environment. Indeed, a greater tendency to form new safety associations predicted later recovery of fear in humans, whereas those who were more likely to modify original fear associations were less likely to re-experience fear (Gershman and Hartley, 2015). Convergent evidence from a rodent study also indicates that gradual extinction learning, compared to abrupt changes in outcome contingencies, promoted modification of the previous association over the formation of a new association, and reduced the likelihood of later fear (Gershman et al., 2013). Perhaps counterintuitively, re-experiencing symptoms in PTSD (which can be considered a clinical presentation of recovery of fear) were associated with a reduced tendency to form new safety associations (Norbury et al., 2021). This implies that although individual differences in latent cause learning are relevant to pathological fear, it is not a simple unidirectional relationship and other factors are likely involved (such as biases for which associations are evoked if multiple associations are stored).

2.2. Instrumental/avoidance learning

Instrumental learning tasks, in which an agent’s actions determine outcomes, can be used to model avoidance behaviours. These tasks can encompass both active avoidance, in other words situations where certain actions can avert aversive outcomes, and passive avoidance, where inaction avoids aversive outcomes (or conversely where actions may lead to aversive outcomes). In both cases, learning state-action values (see Box 1), Qs,a, under certain states, s, and following certain actions, a, can be modelled through model-free reinforcement learning (see Box 1), which relies on an error-dependent learning mechanism similar to that in fear conditioning: Qt+1st,at=Qtst,at+α·(otQtst,at) (Fig. 1; note close variants of model-free RL to that presented here which tie actions to cumulative expected reward can solve 'multistep' problems). This provides an account of how humans learn to act to avoid outcomes such as pain (Jepma et al., 2022, Eldar et al., 2016), with corresponding neural correlates of instrumental prediction errors in the striatum (Eldar et al., 2016) and periaqueductal gray (Roy et al., 2014). Further, when these models are extended to account for asymmetries in learning from safety and threat (the omission or occurrence of an aversive outcome, respectively), by implementing separate learning rate parameters for safety and danger, αsafety and αthreat, subtle differences can be found across different forms of anxiety on aversive learning performance. In one study (Wise and Dolan, 2020), a double dissociation emerged across trait cognitive vs somatic anxiety (see Box 1), where cognitive anxiety was associated with a bias for learning from threat, whilst the opposite was true for somatic anxiety and also trait compulsivity.

State values can provide a computationally efficient mechanism of instrumental responding, by biasing approach-relevant responses under positive Vs (i.e. when states predict positive outcomes) and avoidance-relevant responses under negative Vs in a Pavlovian manner (an effect referred to as Pavlovian-instrumental transfer (Dickinson and Balleine, 1994)). This is implemented by adding an action-weight to action-outcome associations, such that the overall value for the action is given by Qtst,at+π·Vtst. Here, π parameterises an agent’s tendency to behave according to Pavlovian mechanisms (Fig. 2). This can constitute an efficient mechanism of evading threats as a state value, which comprises two components (the state and its expected value), is sufficient to produce defensive behaviour. In contrast, instrumental learning requires a further ‘action’ component to build the state-action-value association. Humans readily learn state-action-value associations, in other words to emit or omit certain actions to avoid punishment (Guitart-Masip et al., 2012; Millner et al., 2018). At the same time, this learning is also biased by state values, which inhibit motor responses when learning to avoid future punishment in a manner consistent with disengagement/freezing (Guitart-Masip et al., 2012, Roelofs, 2017), but promote active escape behaviour when a threat is already present (Millner et al., 2018) (Fig. 2). These effects parallel classical learning accounts which posit that threat makes some actions more likely to be emitted (e.g. freezing) than others (Bolles, 1970, Hershberger, 1986). Finally, the disadvantage of Pavlovian mechanisms is that response biases can be challenging to overcome when Pavlovian and instrumental mechanisms conflict – for example, when one must make an active response in the face of potential threats (rather than freezing) leading to suboptimal behaviour in such situations (Guitart-Masip et al., 2012, Millner, 2018) (Fig. 2). This effect is exacerbated in pathologically anxious individuals (see Box 1), especially under acute anxiety induced by threat-of-shock (Mkrtchian et al., 2017), and this is thought to underlie the excessive avoidance behaviours observed in pathological anxiety.

Fig. 2.

Fig. 2

The reinforcement learning task of Millner et al. (2018). A) Task structure. Participants learnt to press a button (active response) or not (passive response) in order to avoid already-present or future threat. For the present threat condition, a continuous aversive sound was played until the correct action was made, which probabilistically terminated the sound. For the future threat condition, the trial began in silence and the correct action could probabilistically prevent the sound from being played (incorrect actions would instead result in the sound being played). Threat distance (present/future) was crossed with correct action (button-press/no button-press) to produce four conditions in the task. B) Illustration of data from Millner et al. (2018). Mean condition accuracy and standard errors extracted from the original paper. Participants were better at learning to make passive rather than active responses for future threat, but the opposite was true for already-present threat. C) A conceptual demonstration of the effect of Pavlovian response biases during instrumental avoidance learning. The curves reflect improved accuracy over trials as individuals learn the correct actions in each condition. If the correct (instrumental) action is congruent with Pavlovian responses, namely to emit active responses to avoid present threat and passive responses to avoid future threat, learning performance is augmented in healthy controls (orange line) and even further in pathological anxiety (red line). However, if the correct action conflicts with these Pavlovian response biases, learning performance is hindered and this effect is again greater in pathological anxiety. This can be captured by a Pavlovian action weight parameter, π - learning in individuals with pathological anxiety (and depression) is characterised by greater values of π (Mkrtchian et al., 2017).

The behaviours discussed so far are typically considered ‘single-step’ problems, where an agent is only required to learn the appropriate response to perform in a single state of the environment (e.g. how to respond to one particular cue). However, most real-world scenarios are more likely to be multi-step, where an agent must perform a sequence of actions across multiple states of the environment to obtain a particular outcome. In keeping with a distinction of fear and anxiety as relating to proximal/distal threats, single-step problems might provide better models of fear as the threat is potentially associated with the present state of the environment, whereas anxiety might be more associated with multi-step learning, where the potential threat is diffused across the multiple states and actions in the environment. How humans solve multi-step aversive learning is described through model-based reinforcement learning (see Box 1), which relies on a ‘model of the world’, in other words an understanding of the different states of the environment and the transitions across states, to infer optimal behaviour.

The dominant paradigm for measuring model-free and model-based learning strategies to date has been the two-step task (Daw et al., 2011) (Fig. 3), which involves a series of two actions to obtain a certain outcome. Model-free and model-based strategies imply separate predictions for the first action in the sequence, and thus individual choices here indicate a reliance on one strategy over the other. Humans tend to use both model-free and model-based learning strategies for threat avoidance (Sebold et al., 2019, Wang et al., 2018) (Fig. 3), as is the case with reward learning (Daw et al., 2011). The evidence relating model-based planning to anxiety is mixed: on one hand, some studies have reported that individuals favour model-free strategies over model-based in aversive environments (when viewing aversive images) (Sebold et al., 2019) and under social stress (Park et al., 2017). Yet, experimentally-induced (hypercapnic gas) and naturalistic anxiety (panic attacks/life stress) do not appear to impact model-based learning (Gillan et al., 2021) . Further research is needed to determine whether these discrepancies in findings are due to methodological differences or false positives. Instead, the evidence suggests that trait compulsivity, rather than trait anxiety (see Box 1), is associated with decreased reliance on model-based learning (Gillan et al., 2021, Gillan et al., 2016) – an effect which has been proposed to be driven by impairments in state-transition learning, in other words that compulsive individuals have difficulty learning the relationships between states of the environment (Sharp et al., 2021), which leads to reduced confidence in one’s ability to navigate these states.

Fig. 3.

Fig. 3

The two-step task of Sebold et al. (2019). A) Task structure. On each trial, participants made a series of binary choices. The first-step choice (grey cues) probabilistically leads to distinct second-stage states (orange or purple cues), with one choice leading to one second-stage state, for example the orange set, on 70% of trials (referred to as ‘common’ transitions; light grey arrows), and the other second-stage state, for example the purple set, on the remaining 30% of trials (‘rare’ transitions; black arrows). The opposite was true for the other first-stage choice. At the second stage, the second-step choice produced an outcome that depended on choice-specific outcome probabilities that varied over time. The outcome is depicted here as ‘Win’ for reward, but it is possible to also have punishments as the outcomes. B-C) Model-free and model-based reinforcement learning strategies entail different probabilities of repeating the first-stage choice of the previous trial, given the outcome on the previous trial. B) The model-free strategy does not take into account the transition structure of the task and tends to simply repeat first-stage choices if on the previous trial, that choice produced a reward or avoided a punishment. This manifests as a main effect of previous reward/punishment on first-stage choice repetition. C) The model-based strategy utilises the structure of the task and accounts for whether the previous trial involved a common or rare transition, leading to an interaction effect of previous reward/punishment and transition (common/rare) on first-stage choice repetition. D) Illustration of data from Sebold et al. (2019). Mean and error bars for repetition probabilities by condition. Dots represent individual participant data. When learning to avoid punishments in this task, participants typically show a mixture of model-free and model-based influences on choice (i.e. both a main effect of previous outcome and an interaction effect of transition type and previous outcome on choice repetition probability). The figures are reproduced from Sebold et al. (2019), which were published under CC BY 4.0.

Arguably, the two-step task may not be sufficiently complex to reveal impairments of planning in anxiety. Findings from a recent study (Sharp et al., 2022) implementing a multigoal pursuit task suggest that more naturalistic planning problems (balancing multiple goals) may be more sensitive to anxiety-related effects. In the task, healthy individuals learned the likelihoods of observing two tokens for each of two possible actions: one token was associated with monetary reward, and the other with monetary punishment. However, individuals were instructed on each trial that only one token would be relevant on each trial (i.e. on reward trials, obtaining the reward token could lead to a reward, while obtaining the punishment token had no consequence, and vice versa for punishment trials). Thus, the task required individuals to use only goal-relevant information in their choices. In the task, individuals struggled to disengage from punishment-relevant information on reward trials, and this effect was positively correlated with chronic worry, a central feature of pathological anxiety (American Psychiatric Association, 2013). This effect suggests an impairment in model-based planning, where individuals with more severe worry appear to plan for threat avoidance even in explicitly safe environments – a finding which parallels ideas from clinical theories of worry which posit that worry constitutes maladaptive planning to avoid imagined/distal threat (Mathews, 1990). Future work will be required to determine the generality of planning impairments to trait anxiety, or whether the effect is specific to forms of anxiety that involve higher cognitive functions (such as worry), compared to those that do not (such as somatic anxiety) (Sharp et al., 2015).

In brief, humans rely on multiple computational strategies to predict and avoid threat. There is strong evidence that fear is acquired and extinguished based on error-driven learning, and individual differences in beliefs about the latent state of the environment might predict failure to extinguish fear. Pavlovian and instrumental processes can also interact in driving behaviour, and this appears to be exacerbated by pathological anxiety. Anxiety, specifically, may be associated with differences in model-based planning, especially in worry. Similarly, different psychiatric symptoms (e.g. social anxiety vs PTSD, cognitive vs somatic anxiety) show subtly different learning characteristics, such as biases in learning selectively from safety or threat.

3. Approach-avoidance conflict

Real-world decisions are rarely reducible to evasion of a single threat; instead, the consequences of a single action often simultaneously involve positive and negative outcomes. This means that one needs to decide to pursue reward at risk of incurring a punishment, or avoid that punishment and forsake the potential reward. Such situations, which are referred to as involving approach-avoidance conflict, are exploited in animal models of anxiety such as the Vogel conflict test (Vogel et al., 1971), where the drive to approach reward (e.g. to obtain water) is pitted against the drive to avoid threat (e.g. to avoid an electrical shock). Excessive or consistent avoidance in situations of approach-avoidance conflict entails giving up reward in order to avoid potential threat, which reflects maladaptive avoidance in pathological fear/anxiety (Aupperle and Paulus, 2010, Loijen et al., 2020) where important things in life may be sacrificed in order to avoid stressors/threat. Here, we discuss some initial efforts to explore approach-avoidance conflict paradigms through a computational lens.

Recent studies have attempted to adapt rodent conflict tests for humans, typically in tasks where individuals decide on whether to accept or reject an offer of some aversive outcome (e.g. an electric shock) alongside some monetary reward (Aupperle et al., 2011, Ironside et al., 2020) (Fig. 4 A). Using an active inference model (Friston et al., 2013, Friston et al., 2017) (see Box 1), a Bayesian generative model of decision-making, one study compared the roles of ‘emotional conflict’,1 defined as the relative value of the aversive outcome compared to the reward, and ‘decision uncertainty’, defined as the difficulty in making the decision (Smith et al., 2021) (Fig. 4B). In keeping with the notion of the task as measuring anxiety-related behaviour, state anxiety during the task was correlated with greater emotional conflict, meaning that those who found the aversive outcomes more unpleasant were more anxious during the task. However, this effect did not extend to group differences between pathologically anxious and depressed individuals and non-symptomatic controls. Interestingly, the symptomatic group did show greater decision uncertainty, relative to controls, perhaps reflecting findings from clinical psychology that anxiety symptoms, specifically chronic worry, are associated with indecision (Snyder et al., 2014).

Fig. 4.

Fig. 4

The approach-avoidance conflict task of Smith et al. (2021). A) Task structure. Participants decided on their preference for viewing an aversive image and hearing an aversive sound (represented by the raincloud) in return for varying levels of monetary reward (represented by the proportion of a bar filled in red), OR viewing a pleasant image and hearing a pleasant sound (represented by the sun) but with no monetary reward (see the empty bar). Responses were made by moving an avatar left or right on a runway which corresponds to probabilities that participants will see either the left or right outcomes, reflecting their relative preference. B) A conceptual demonstration of the effect of the key parameters of the generative model. The aversive stimuli and reward outcome is referred to as the ‘Approach’ outcome, and the pleasant stimuli with no reward outcome is referred to as the ‘Avoid’ outcome. The subjective value of each position of the runway is represented in greyscale, with darker tones representing greater value. Subjective value depends on two parameters. Firstly, the emotional conflict parameter captures the expected aversiveness of the aversive stimuli relative to the monetary reward, with higher emotional conflict indicating greater aversiveness. Low emotional conflict (top-left, bottom-left) means that individuals value the ‘Approach’ outcome more than the ‘Avoid’ outcome because the expected aversiveness of the aversive stimuli is low – this reverses as emotional conflict increases (top-right, bottom-right). Secondly, the decision uncertainty parameter reflects how confident participants are in their choices. High decision uncertainty (bottom-left, bottom-right) indicates lower confidence in discerning the value of each runway position, leading to a broader distribution of value across the runway. Conversely, low decision uncertainty (top-left, top-right), or greater confidence, leads to a narrower distribution of value. Fig. A is reproduced from McDermott et al. (2021), which was published under CC BY 4.0.

Approach-avoidance conflict has also been modelled in tasks simulating foraging under predation, where individuals collect rewards at risk of being caught by a predator (Bach, 2021). Avoidance behaviour in these tasks scales with threat probability (Bach, 2015, Loh et al., 2017, Qi et al., 2018), with greater threat probability leading to greater avoidance. Even when individuals make approach responses, response latencies are greater under high vs. low threat probabilities, demonstrating that behavioural inhibition occurs even during approach behaviour, and this effect positively correlates with trait anxiety (Bach, 2015). Neuroimaging studies using these tasks have repeatedly implicated the ventral hippocampus (Ito and Lee, 2016, O'Neil et al., 2015, Bach et al., 2014), consistent with previous reports of its involvement in anxiety/fear and reward circuitry (Shin and Liberzon, 2010, Russo and Nestler, 2013), but its precise computational role has yet to be established.

Lastly, asymmetries in reward and aversive learning have been studied in instrumental learning paradigms where individuals learn to simultaneously pursue reward and avoid punishment, specifically where on each trial, an action could potentially lead to the delivery of both the reward and the punishment. Compared to healthy controls, the learning performance of pathologically anxious and depressed individuals was characterised by higher punishment learning rates (Aylward et al., 2019), meaning that they relied on fewer observations to estimate the aversive value of a certain action, but there was no group difference for the reward value of actions. The group difference in punishment learning rates was corroborated by a recent meta-analysis of 27 reinforcement learning studies across more than 3000 participants (Pike and Robinson, 2022). This study used a novel simulation-based approach that allowed for pathologically anxious/depressed individuals to be compared with controls across a range of different learning tasks, and by using simulations of their behaviour, the study found that the symptomatic group showed not only higher punishment learning rates, but also lower reward learning rates, compared to controls. This implies that anxious/depressed individuals show negative biases when learning about rewards and punishments, learning slower about the former and faster about the latter. Of note, however, this analysis pooled across anxiety and depression studies, leaving open the question of the effect’s specificity to anxiety.

In sum, despite the reliance of many animal models of fear and anxiety on approach-avoidance conflict (Campos et al., 2013), there is surprisingly limited work applying computational approaches to its understanding. Initial findings indicate roles of uncertainty in making decisions under conflict and asymmetries in learning about rewards and punishments in pathological anxiety (and depression), but more work will be required to better understand how individuals trade off rewards and punishments, perhaps by better use of translational paradigms (Kirlic et al., 2017).

4. Decision-making under uncertainty

Central to the definition of anxiety is the notion of uncertainty of threat. Pathologically anxious individuals often report negative beliefs about uncertainty and its implications, a concept referred to as intolerance of uncertainty (Dugas et al., 1997) and this is considered by some to be a core component of certain anxiety disorders (Dugas et al., 1998, Carleton et al., 2012). Indeed, multiple cognitive mechanisms may contribute to maladaptive responses to uncertainty in pathological anxiety (Grupe and Nitschke, 2013). In the next section, we discuss negative biases in decision-making in the face of uncertainty in anxiety.

4.1. Negative biases for uncertain outcomes

A key finding from early cognitive research was that anxious individuals demonstrate a negative interpretational bias (Beck and Clark, 1988, Beck and Clark, 1997, Hirsch et al., 2016), that is, a tendency to interpret ambiguous information in a negative light. A computational basis for this has been proposed in a perceptual task (Aylward et al., 2020) translated from an animal model of negative bias (Harding et al., 2004, Hales et al., 2016). Pathologically anxious and depressed individuals were more likely to interpret ambiguous stimuli as if they would lead to the worst of two possible outcomes (Fig. 5), relative to healthy controls. Drift-diffusion modelling (see Box 1) revealed that this bias was driven by group differences in drift rates; in other words, the symptomatic group was more likely to accumulate negative evidence about the ambiguous stimulus during deliberation. This effect parallels findings from lexical decision-making (White et al., 2010) and facial discrimination tasks (Glasgow et al., 2022), which together suggest that pathologically anxious individuals tend to build more negative representations of the environment, which may lead to maladaptive behaviour especially under uncertainty.

Fig. 5.

Fig. 5

The perceptual decision-making task of Aylward et al. (2020). A) Task structure. Participants were first trained to discriminate between two stimuli (here, a small and large circle) using two response keys. Pressing the correct response key for one of the stimuli deterministically leads to a high reward, whereas the correct response key for the other stimulus leads to a smaller reward (response-reward associations were counter-balanced across participants). In the test phase, an intermediate-sized stimulus was presented, for which both response keys lead to their respective rewards (high reward for the high reward response, low reward for the low reward response) with equal probability. B) Illustration of data from Aylward et al. (2020). Mean proportion of high-reward responses to the intermediate stimulus by group (pathologically anxious vs healthy controls) and standard errors extracted from the original paper. Pathologically anxious individuals were less likely to respond to the ambiguous stimulus with the high-reward response, and healthy controls showed the opposite effect. C) This effect was computationally captured in a drift-diffusion model, specifically through the drift rate parameter which controls the rate and direction by which evidence is accumulated (i.e. for the high-reward vs the low-reward response). In the study, pathologically anxious individuals (red line) showed more negative drift rates, compared to the controls (orange line).

What might drive negative biases under uncertainty? A plausible explanation is that anxiety is associated with a prior belief that negative events occur more frequently than positive events (Butler and Mathews, 1983). A signal detection theory analysis of the perceptual task described above (Aylward et al., 2020) is consistent with negatively biased prior beliefs about outcome probabilities (Locke and Robinson, 2021). Further, modelling individual risk sensitivity using prospect theory (Kahneman and Tversky, 1979) (see Box 1) showed that anxious individuals are more sensitive to risk in economical decision-making compared to healthy controls, leading to preferences for smaller yet guaranteed (i.e. more certain) rewards over gambles for greater rewards (Charpentier et al., 2017), which can be interpreted as over-weighting of the probability of disadvantageous outcomes. Relatedly, recently proposed computational models of anxiety (Zorowitz et al., 2020) and obsessive-compulsive disorder (Fradkin et al., 2020) show through simulations that maladaptive avoidance behaviour can stem from pessimistic beliefs about personal ability to avoid threat in the future (which is intrinsically uncertain), leading individuals to make inappropriate/excessive avoidance responses in the present.

4.2. Stochasticity vs volatility

Another line of research has focused on dissecting uncertainty into two components: stochasticity and volatility (Piray and Daw, 2021). Stochasticity arises when learning the value of a state or state-action pair if an agent observes variance in the outcomes (i.e. if the relationships are non-deterministic, as much the work in the previous section involved), leading to uncertainty in the precision of the value estimate. On the other hand, volatility involves changes in the state or state-action values themselves, which leads to uncertainty in the estimate of the value. Stochasticity and volatility have diverging consequences for efficient learning: high volatility, in other words when there is a lot of change in the values of states/state-action pairs, requires agents to update their values estimates more frequently, which can be achieved by increasing one’s learning rate, α. High stochasticity, or where there is high variance in the observed outcomes (assuming the value is stable) calls for slower learning and lower α, as each individual outcome is less informative about the true value. Humans can adaptively adjust their learning rates to variable stochasticity (Lee et al., 2020) and volatility (Behrens et al., 2007).

The previous section of this review showed that anxiety is associated with negative biases when outcomes are stochastic. There is also evidence associating anxiety with impaired learning in volatile environments. Specifically, when learning to avoid threat under conditions of low and high volatility (with stable stochasticity across conditions), individuals with low trait anxiety adjusted their learning rates across conditions according to the optimal strategy, but high trait anxiety was associated with less flexibility in learning rates, leading to sub-optimal performance (Browning et al., 2015). Similar effects have also been found during reward learning (Huang et al., 2017, Gagne et al., 2020) and social decision-making (Lamba et al., 2020), indicating a domain-general impairment in adapting to volatility. Obsessive-compulsive and fear symptoms have also been associated with impairments in state-transition learning under low and high volatility, specifically with sub-optimally fast learning in stable conditions and slow learning in volatile conditions (Sharp et al., 2021), indicating a transdiagnostic role of learning under volatility in threat-related psychopathology. Towards a neural mechanism of learning rate inflexibility, one study found that the dorsal anterior cingulate cortex in healthy controls tracks changes in learning rate over trials, but this was not the case for anxious individuals (Piray et al., 2019). Finally, a recent simulation-based approach suggests that learning rate inflexibility can be explained if an agent is biased to perceived stochasticity as low and constant but volatility is estimated inappropriately highly (Piray and Daw, 2021). This implies that anxious individuals are insensitive to changes in true volatility and have impaired learning in environments involving low volatility - indeed, both of these effects have empirical support (Piray et al., 2019, Browning et al., 2015).

These two streams of findings, specifically that anxiety is associated with negatively-biased expectations of potential outcomes, and with impairments in adapting to environmental volatility, have begun to identify the computational mechanisms that might underlie pathologically anxious individuals’ altered reactivity to uncertain outcomes. The causal role of the anxious state (i.e. the anticipation for distal and uncertain threat) in the misestimation of uncertainty is unclear, but speculatively, these ‘impairments’ might have some adaptive value if an agent believes there may be incoming threat. Firstly, overestimating threat likelihood is adaptive as it is safer to overestimate than underestimate threat. Secondly, a bias to attribute environmental uncertainty to volatility rather than stochasticity allows an agent to respond more readily to changes in the environment – for example, an action which successfully avoided threat in the past may not be so effective in the present, which would be important to adjust for. Future work using induced- fear/anxiety designs will be important to understand if individual differences in threat estimation and dealing with uncertainty drive anxious states, or if indeed they are goal-directed changes in perception in line with threat avoidance.

5. Summary

We have presented an overview of the key themes of the computational literature on human fear and anxiety. Specifically, we discussed multiple computational processes involved in learning to predict and avoid threat, namely error-driven learning, the interactions of Pavlovian and instrumental learning and model-based planning. Given the diversity and complexity of these processes, the evidence suggests that there are multiple ways that learning can go awry with respect to pathological fear and anxiety, and different symptoms are associated with divergent computational mechanisms. We also introduced the few studies to date to study approach-avoidance conflict behaviour from a computational perspective, which have begun to suggest roles of decision uncertainty and asymmetries in learning about rewards and punishment, but there is not yet a clear computational account of approach-avoidance conflict. We also argued that uncertainty is a key motif implicated in anxiety, where converging evidence from different paradigms suggest that anxious individuals overestimate the likelihood of disadvantageous outcomes and struggle to learn in volatile environments. However, there are a number of challenges and opportunities for the future. Whilst computational approaches have been useful in deepening our understanding of the specific neurocognitive processes underlying defensive behaviour and how disruptions in these processes can lead to psychopathology, little progress has been made in informing treatment. Looking forward, computational approaches could be extended to better understand basic mechanisms and treatments for fear-and anxiety-related disorders. Further, better cross-species paradigms of defensive behaviour, especially those amenable to computational analysis (Redish, 2022), will be important in integrating findings across the human and animal literature and potentially spurring the development of psychiatric interventions (Pike et al., 2021) (Box 2).

Box 2. Challenges and opportunities.

The factors driving pessimism under uncertainty. There are multiple potential reasons why anxious individuals might be pessimistic during decision-making under uncertainty, or more specifically, accumulate negative evidence from the environment. This effect may be driven by individual differences in the representations of outcomes, for example by overweighting negative outcomes over positive ones (i.e. loss aversion) especially in fearful/anxious states, although this does not appear to fully explain the data (Charpentier et al., 2017). Alternatively, pessimism may be learned from one’s environment if negative outcomes historically occurred more frequently than positive outcomes. This leads to the question of whether pessimism can be unlearned, for example by exposure to an environment in which positive outcomes occur more frequently. Signal detection theory and Bayes’ theorem have been proposed as suitable frameworks to address these questions (Locke and Robinson, 2021, Huys et al., 2021).

Translational approaches. Some of the studies reviewed above have implemented translational behavioural paradigms to bridge the gap between the animal and human literature (Smith et al., 2021, Aylward et al., 2020). Computational approaches may be especially well-suited for translational studies of behaviour (Redish, 2022). The computations necessary to solve certain tasks (e.g. to avoid threat via reinforcement learning), even if these tasks are outwardly very different, may serve as a better criterion for translational validity as opposed to other criteria such as face or predictive validity, when the aim is to better understand the cognitive and neurobiological mechanisms underlying behaviour. This approach will be especially important for approach-avoidance conflict tasks, which are some of the most commonly employed rodent anxiety models (Campos et al., 2013). Future work should develop fear/anxiety tasks that are explicitly designed to engage similar computational processes in both animals and humans, which have been referred to as ‘common currency’ tasks (Pike et al., 2021). These tasks can also act as preclinical tests that will help to spur drug discovery for fear/anxiety disorders, which is especially important given that psychiatric drug development has slowed over the last decade (Hyman, 2012, Kesselheim et al., 2015).

Computational mechanisms of anxiolytic/anxiogenic interventions. Computational approaches are also well-suited for elucidating the cognitive and physiological mechanisms of anxiolytic and anxiogenic interventions. For example, traditional analyses based on summary-statistics of behaviour may demonstrate that drug X increased approach responses in an approach-avoidance conflict task, but this effect could be driven by changes in reward sensitivity or punishment sensitivity (or indeed both). Understanding pharmacological and psychological interventions at the level of cognitive/computational mechanisms will help to make treatments more targeted and therefore effective. A large-sample meta-analysis comparing pathologically anxious and depressed individuals to controls has already identified one potential computational target for therapy in punishment learning rates, where the learning behaviour of the symptomatic group was characterised by faster learning from punishment compared to the controls, but not sensitivity to punishment (Pike and Robinson, 2022). This suggests that treatment approaches in these patients could specifically target how they learn and adapt to negative events in the world, for example through cognitive-behavioural therapy (CBT) and this may improve treatment efficacy. Such approaches will be an important avenue for future research to potentially drive innovation in psychiatric treatment.

Funding

This research was funded by a PhD studentship to YY from the Wellcome Trust (222268/Z/20/Z) and a Senior Non-Clinical Fellowship to OJR from the Medical Research Council (MR/R020817/1). For the purpose of Open Access, the authors have applied a CC BY public copyright license to any Author Accepted Manuscript version arising from this submission.

Footnotes

1

Although the term, emotional conflict, has been previously used to describe the interference effect of emotionally salient stimuli on cognitive processing (e.g. Bush et al., 2000; Etkins et al., 2006), the use of the term here reflects the degree to which individuals may be conflicted in approaching reward and avoiding affective (i.e. emotional) punishment. We use the exact name of the model parameter from the original study by Smith and colleagues.

References

  1. American Psychiatric Association . Vol. 5. American psychiatric association,; Washington, DC: 2013. (Diagnostic and statistical manual of mental disorders: DSM-5). [Google Scholar]
  2. Aupperle R.L., et al. A reverse translational approach to quantify approach-avoidance conflict in humans. Behav. Brain Res. 2011;225(2):455–463. doi: 10.1016/j.bbr.2011.08.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Aupperle R.L., Paulus M.P. Neural systems underlying approach and avoidance in anxiety disorders. Dialog-. Clin. Neurosci. 2010;12(4):517–531. doi: 10.31887/DCNS.2010.12.4/raupperle. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Aylward J., et al. Altered learning under uncertainty in unmedicated mood and anxiety disorders. Nat. Hum. Behav. 2019;3(10):1116–1123. doi: 10.1038/s41562-019-0628-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Aylward J., et al. Translating a rodent measure of negative bias into humans: the impact of induced anxiety and unmedicated mood and anxiety disorders. Psychol. Med. 2020;50(2):237–246. doi: 10.1017/S0033291718004117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Bach D.R., et al. Human Hippocampus Arbitrates Approach-Avoidance Conflict (vol 24, pg 541, 2014) Curr. Biol. 2014;24(12) doi: 10.1016/j.cub.2014.01.046. 1435-1435. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bach D.R. Anxiety-like behavioural inhibition is normative under environmental threat-reward correlations. Plos Comput. Biol. 2015;11:12. doi: 10.1371/journal.pcbi.1004646. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Bach D.R. Cross-species anxiety tests in psychiatry: pitfalls and promises. Mol. Psychiatry. 2021 doi: 10.1038/s41380-021-01299-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Bach D.R., Dayan P. OPINION Algorithms for survival: a comparative perspective on emotions. Nat. Rev. Neurosci. 2017;18(5):311–319. doi: 10.1038/nrn.2017.35. [DOI] [PubMed] [Google Scholar]
  10. Beck A.T., Clark D.A. Anxiety and depression: an information processing perspective. Anxiety Res. 1988;1(1):23–36. [Google Scholar]
  11. Beck A.T., Clark D.A. An information processing model of anxiety: automatic and strategic processes. Behav. Res. Ther. 1997;35(1):49–58. doi: 10.1016/s0005-7967(96)00069-1. [DOI] [PubMed] [Google Scholar]
  12. Behrens T.E.J., et al. Learning the value of information in an uncertain world. Nat. Neurosci. 2007;10(9):1214–1221. doi: 10.1038/nn1954. [DOI] [PubMed] [Google Scholar]
  13. Bolles R.C. Species-specific defense reactions and avoidance learning. Psychol. Rev. 1970;77(1):32–48. [Google Scholar]
  14. Bouton M.E. Context, time, and memory retrieval in the interference paradigms of pavlovian learning. Psychol. Bull. 1993;114(1):80–99. doi: 10.1037/0033-2909.114.1.80. [DOI] [PubMed] [Google Scholar]
  15. Bouton M.E., Maren S., McNally G.P. Behavioral and neurobiological mechanisms of pavlovian and instrumental extinction learning. Physiol. Rev. 2021;101(2):611–681. doi: 10.1152/physrev.00016.2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Browning M., et al. Anxious individuals have difficulty learning the causal statistics of aversive environments. Biol. Psychiatry. 2015;77(9):47s–48s. doi: 10.1038/nn.3961. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Butler G., Mathews A. Cognitive-processes in anxiety. Adv. Behav. Res. Ther. 1983;5(1):51–62. [Google Scholar]
  18. Campos A.C., et al. Animal models of anxiety disorders and stress. Rev. Bras. De. Psiquiatr. 2013;35:S101–S111. doi: 10.1590/1516-4446-2013-1139. [DOI] [PubMed] [Google Scholar]
  19. Carleton R.N., et al. Increasingly certain about uncertainty: intolerance of uncertainty across anxiety and depression. J. Anxiety Disord. 2012;26(3):468–479. doi: 10.1016/j.janxdis.2012.01.011. [DOI] [PubMed] [Google Scholar]
  20. Charpentier C.J., et al. Enhanced risk aversion, but not loss aversion, in unmedicated pathological anxiety. Biol. Psychiatry. 2017;81(12):1014–1022. doi: 10.1016/j.biopsych.2016.12.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Craske M.G., Hermans D.E., Vansteenwegen D.E. American Psychological Association,; 2006. Fear and learning: from basic processes to clinical implications. [Google Scholar]
  22. Daniel-Watanabe, L. and P.C. Fletcher, Are Fear and Anxiety Truly Distinct? Biological Psychiatry Global Open Science, 2021. [DOI] [PMC free article] [PubMed]
  23. Daw N.D., et al. Model-based influences on humans' choices and striatal prediction errors. Neuron. 2011;69(6):1204–1215. doi: 10.1016/j.neuron.2011.02.027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Delgado M.R., et al. The role of the striatum in aversive learning and aversive prediction errors. Philos. Trans. R. Soc. B-Biol. Sci. 2008;363(1511):3787–3800. doi: 10.1098/rstb.2008.0161. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Dickinson A., Balleine B. Motivational Control of Goal-Directed Action. Anim. Learn. Behav. 1994;22(1):1–18. [Google Scholar]
  26. Dugas M.J., et al. Generalized anxiety disorder: a preliminary test of a conceptual model. Behav. Res. Ther. 1998;36(2):215–226. doi: 10.1016/s0005-7967(97)00070-3. [DOI] [PubMed] [Google Scholar]
  27. Dugas M.J., Freeston M.H., Ladouceur R. Intolerance of uncertainty and problem orientation in worry. Cogn. Ther. Res. 1997;21(6):593–606. [Google Scholar]
  28. Eldar E., et al. Striatal structure and function predict individual biases in learning to avoid pain. Proc. Natl. Acad. Sci. USA. 2016;113(17):4812–4817. doi: 10.1073/pnas.1519829113. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Fox A.S., Shackman A.J. The central extended amygdala in fear and anxiety: Closing the gap between mechanistic and neuroimaging research. Neurosci. Lett. 2019;693:58–67. doi: 10.1016/j.neulet.2017.11.056. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Fradkin I., et al. Searching for an anchor in an unpredictable world: a computational model of obsessive compulsive disorder. Psychol. Rev. 2020;127(5):672–699. doi: 10.1037/rev0000188. [DOI] [PubMed] [Google Scholar]
  31. Friston K., et al. Active inference: a process theory. Neural Comput. 2017;29(1):1–49. doi: 10.1162/NECO_a_00912. [DOI] [PubMed] [Google Scholar]
  32. Friston K.J., et al. The anatomy of choice: active inference and agency. Frontiers in Human. Neuroscience. 2013:7. doi: 10.3389/fnhum.2013.00598. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Gagne C., et al. Impaired adaptation of learning to contingency volatility in internalizing psychopathology. Elife. 2020:9. doi: 10.7554/eLife.61387. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Gelman A., Hill J. Analytical methods for social research. Cambridge University Press,; Cambridge; New York: 2007. Data analysis using regression and multilevel/hierarchical models; p. xxii, 625. [Google Scholar]
  35. Gershman S.J., et al. Gradual extinction prevents the return of fear: implications for the discovery of state. Front. Behav. Neurosci. 2013:7. doi: 10.3389/fnbeh.2013.00164. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Gershman S.J., Hartley C.A. Individual differences in learning predict the return of fear. Learn. Behav. 2015;43(3):243–250. doi: 10.3758/s13420-015-0176-z. [DOI] [PubMed] [Google Scholar]
  37. Gillan C.M., et al. Characterizing a psychiatric symptom dimension related to deficits in goal-directed control. Elife. 2016:5. doi: 10.7554/eLife.11305. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Gillan C.M., et al. Experimentally induced and real-world anxiety have no demonstrable effect on goal-directed behaviour. Psychol. Med. 2021;51(9):1467–1478. doi: 10.1017/S0033291720000203. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Glasgow S., et al. Threat and uncertainty in the face of perceptual decision-making in anxiety. J. Psychopathol. Clin. Sci. 2022;131(3):265–277. doi: 10.1037/abn0000729. [DOI] [PubMed] [Google Scholar]
  40. Grillon C. Models and mechanisms of anxiety: evidence from startle studies. Psychopharmacol. (Berl. ) 2008;199(3):421–437. doi: 10.1007/s00213-007-1019-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Grupe D.W., Nitschke J.B. Uncertainty and anticipation in anxiety: an integrated neurobiological and psychological perspective. Nat. Rev. Neurosci. 2013;14(7):488–501. doi: 10.1038/nrn3524. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Guitart-Masip M., et al. Go and no-go learning in reward and punishment: interactions between affect and effect. Neuroimage. 2012;62(1):154–166. doi: 10.1016/j.neuroimage.2012.04.024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Hales C.A., Robinson E.S.J., Houghton C.J. Diffusion modelling reveals the decision making processes underlying negative judgement bias in rats. Plos One. 2016;11:3. doi: 10.1371/journal.pone.0152592. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Harding E.J., Paul E.S., Mendl M. Animal behavior - Cognitive bias and affective state. Nature. 2004;427(6972) doi: 10.1038/427312a. 312-312. [DOI] [PubMed] [Google Scholar]
  45. Hershberger W.A. An approach through the looking-glass. Anim. Learn. Behav. 1986;14(4):443–451. [Google Scholar]
  46. Hirsch C.R., et al. Resolving ambiguity in emotional disorders: the nature and role of interpretation biases. Annu. Rev. Clin. Psychol. 2016;Vol 12(12):281. doi: 10.1146/annurev-clinpsy-021815-093436. [DOI] [PubMed] [Google Scholar]
  47. Hofmann S.G. Cognitive processes during fear acquisition and extinction in animals and humans: Implications for exposure therapy of anxiety disorders. Clin. Psychol. Rev. 2008;28(2):199–210. doi: 10.1016/j.cpr.2007.04.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Homan P., et al. Neural computations of threat in the aftermath of combat trauma. Nat. Neurosci. 2019;22(3):470. doi: 10.1038/s41593-018-0315-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Huang H., Thompson W., Paulus M.P. Computational dysfunctions in anxiety: failure to differentiate signal from noise. Biol. Psychiatry. 2017;82(6):440–446. doi: 10.1016/j.biopsych.2017.07.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Huys Q.J.M., et al. Advances in the computational understanding of mental illness. Neuropsychopharmacology. 2021;46(1):3–19. doi: 10.1038/s41386-020-0746-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Hyman S.E. Revolution Stalled. Sci. Transl. Med. 2012;4:155. doi: 10.1126/scitranslmed.3003142. [DOI] [PubMed] [Google Scholar]
  52. Ironside M., et al. Approach-Avoidance Conflict in Major Depressive Disorder: Congruent Neural Findings in Humans and Nonhuman Primates. Biol. Psychiatry. 2020;87(5):399–408. doi: 10.1016/j.biopsych.2019.08.022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Ito R., Lee A.C.H. The role of the hippocampus in approach-avoidance conflict decision-making: Evidence from rodent and human studies. Behav. Brain Res. 2016;313:345–357. doi: 10.1016/j.bbr.2016.07.039. [DOI] [PubMed] [Google Scholar]
  54. Jepma M., et al. Different brain systems support learning from received and avoided pain during human pain-avoidance learning. Elife. 2022:11. doi: 10.7554/eLife.74149. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Kahneman D., Tversky A. Prospect theory: an analysis of decision under risk. Econometrica. 1979;47(2):263–291. [Google Scholar]
  56. Kesselheim A.S., Hwang T.J., Franklin J.M. Two decades of new drug development for central nervous system disorders. Nat. Rev. Drug Discov. 2015;14(12):815–816. doi: 10.1038/nrd4793. [DOI] [PubMed] [Google Scholar]
  57. Kessler R.C., et al. Prevalence, severity, and comorbidity of 12-month DSM-IV disorders in the national comorbidity survey replication. (vol 62, pg 617, 2005) Arch. Gen. Psychiatry. 2005;62(7) doi: 10.1001/archpsyc.62.6.617. 709-709. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Kirlic N., Young J., Aupperle R.L. Animal to human translational paradigms relevant for approach avoidance conflict decision making. Behav. Res. Ther. 2017;96:14–29. doi: 10.1016/j.brat.2017.04.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Koksal F., Power K.G. Four systems anxiety questionnaire (FSAQ): a self-report measure of somatic, cognitive, behavioral, and feeling components. J. Pers. Assess. 1990;54(3–4):534–545. doi: 10.1080/00223891.1990.9674018. [DOI] [PubMed] [Google Scholar]
  60. Lamba A., Frank M.J., FeldmanHall O. Anxiety impedes adaptive social learning under uncertainty. Psychol. Sci. 2020;31(5):592–603. doi: 10.1177/0956797620910993. [DOI] [PubMed] [Google Scholar]
  61. LeDoux J.E., Pine D.S. Using neuroscience to help understand fear and anxiety: a two-system framework. Am. J. Psychiatry. 2016;173(11):1083–1093. doi: 10.1176/appi.ajp.2016.16030353. [DOI] [PubMed] [Google Scholar]
  62. Lee S., Gold J.I., Kable J.W. The human as delta-rule learner. Decis. -Wash. 2020;7(1):55–66. [Google Scholar]
  63. Li J., et al. Differential roles of human striatum and amygdala in associative learning. Nat. Neurosci. 2011;14(10):1250–1252. doi: 10.1038/nn.2904. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Locke S.M., Robinson O.J. Affective bias through the lens of Signal Detection Theory. Comput. Psychiatry (Camb., Mass. ) 2021;5(1):4. doi: 10.5334/cpsy.58. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Loh E., et al. Parsing the role of the hippocampus in approach-avoidance conflict. Cereb. Cortex. 2017;27(1):201–215. doi: 10.1093/cercor/bhw378. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Loijen A., et al. Biased approach-avoidance tendencies in psychopathology: a systematic review of their assessment and modification. Clin. Psychol. Rev. 2020:77. doi: 10.1016/j.cpr.2020.101825. [DOI] [PubMed] [Google Scholar]
  67. Mathews A. Why worry? The cognitive function of anxiety. Behav. Res Ther. 1990;28(6):455–468. doi: 10.1016/0005-7967(90)90132-3. [DOI] [PubMed] [Google Scholar]
  68. McDermott T.J., et al. Test-retest reliability of approach-avoidance conflict decision-making during functional magnetic resonance imaging in healthy adults. Hum. Brain Mapp. 2021;42(8):2347–2361. doi: 10.1002/hbm.25371. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. McNally G.P., Westbrook R.F. Predicting danger: The nature, consequences, and neural mechanisms of predictive fear learning. Learn. Mem. 2006;13(3):245–253. doi: 10.1101/lm.196606. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. McNally G.P., Johansen J.P., Blair H.T. Placing prediction into the fear circuit. Trends Neurosci. 2011;34(6):283–292. doi: 10.1016/j.tins.2011.03.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Millner A.J., et al. Pavlovian control of escape and avoidance. J. Cogn. Neurosci. 2018;30(10):1379–1390. doi: 10.1162/jocn_a_01224. [DOI] [PubMed] [Google Scholar]
  72. Mkrtchian A., et al. Modeling avoidance in mood and anxiety disorders using reinforcement learning. Biol. Psychiatry. 2017;82(7):532–539. doi: 10.1016/j.biopsych.2017.01.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Mobbs D., et al. Space, time, and fear: survival computations along defensive circuits. Trends Cogn. Sci. 2020;24(3):228–241. doi: 10.1016/j.tics.2019.12.016. [DOI] [PubMed] [Google Scholar]
  74. Norbury A., et al. Latent cause inference during extinction learning in trauma-exposed individuals with and without PTSD. Psychol. Med. 2021 doi: 10.1017/S0033291721000647. [DOI] [PubMed] [Google Scholar]
  75. O'Neil E.B., et al. Examining the role of the human hippocampus in approach-avoidance decision making using a novel conflict paradigm and multivariate functional magnetic resonance imaging. J. Neurosci. 2015;35(45):15039–15049. doi: 10.1523/JNEUROSCI.1915-15.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Park H., Lee D., Chey J. Stress enhances model-free reinforcement learning only after negative outcome. Plos One. 2017;12:7. doi: 10.1371/journal.pone.0180588. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Pearce J.M., Hall G. A model for pavlovian learning - variations in the effectiveness of conditioned but not of unconditioned stimuli. Psychol. Rev. 1980;87(6):532–552. [PubMed] [Google Scholar]
  78. Perusini J.N., Fanselow M.S. Neurobehavioral perspectives on the distinction between fear and anxiety. Learn Mem. 2015;22(9):417–425. doi: 10.1101/lm.039180.115. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Pike, A.C. and O.J. Robinson, Reinforcement Learning in Patients With Mood and Anxiety Disorders vs Control Individuals A Systematic Review and Meta-analysis. Jama Psychiatry, 2022. [DOI] [PMC free article] [PubMed]
  80. Pike A.C., Lowther M., Robinson O.J. The Importance of Common Currency Tasks in Translational Psychiatry. Curr. Behav. Neurosci. Rep. 2021;8(1):1–10. doi: 10.1007/s40473-021-00225-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Piray P., et al. Emotionally Aversive Cues Suppress Neural Systems Underlying Optimal Learning in Socially Anxious Individuals. J. Neurosci. 2019;39(8):1445–1456. doi: 10.1523/JNEUROSCI.1394-18.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Piray P., Daw N.D. A model for learning based on the joint estimation of stochasticity and volatility. Nat. Commun. 2021;12:1. doi: 10.1038/s41467-021-26731-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Price R.B., Brown V., Siegle G.J. Computational modeling applied to the dot-probe task yields improved reliability and mechanistic insights. Biol. Psychiatry. 2019;85(7):606–612. doi: 10.1016/j.biopsych.2018.09.022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Qi S., et al. How cognitive and reactive fear circuits optimize escape decisions in humans. Proc. Natl. Acad. Sci. USA. 2018;115(12):3186–3191. doi: 10.1073/pnas.1712314115. [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Ratcliff R. A theory of memory retrieval. Psychol. Rev. 1978;85(2):59. [Google Scholar]
  86. Redish A.D., et al. Computational validity: using computation to translate behaviours across species. Philos. Trans. R. Soc. B-Biol. Sci. 2022;1844:377. doi: 10.1098/rstb.2020.0525. [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Ree M.J., et al. Distinguishing cognitive and somatic dimensions of state and trait anxiety: development and validation of the state-trait inventory for cognitive and somatic anxiety (STICSA) Behav. Cogn. Psychother. 2008;36(3):313–332. [Google Scholar]
  88. Rescorla, R.A. and A.R. Wagner, A theory of Pavlovian conditioning. Classical Conditioning II: Current Theory and Research, 1971.
  89. Robinson O.J., et al. Stress increases aversive prediction error signal in the ventral striatum. Proc. Natl. Acad. Sci. USA. 2013;110(10):4129–4133. doi: 10.1073/pnas.1213923110. [DOI] [PMC free article] [PubMed] [Google Scholar]
  90. Roelofs K. Freeze for action: neurobiological mechanisms in animal and human freezing. Philos. Trans. R. Soc. B-Biol. Sci. 2017;372:1718. doi: 10.1098/rstb.2016.0206. [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Roesch M.R., et al. Neural correlates of variations in event processing during learning in basolateral amygdala. J. Neurosci. 2010;30(7):2464–2471. doi: 10.1523/JNEUROSCI.5781-09.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Roy M., et al. Representation of aversive prediction errors in the human periaqueductal gray. Nat. Neurosci. 2014;17(11):1607–1612. doi: 10.1038/nn.3832. [DOI] [PMC free article] [PubMed] [Google Scholar]
  93. Russo S.J., Nestler E.J. The brain reward circuitry in mood disorders (vol 14, pg 609, 2013) Nat. Rev. Neurosci. 2013;14(10) doi: 10.1038/nrn3381. 736-736. [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Sebold M., et al. Reward and avoidance learning in the context of aversive environments and possible implications for depressive symptoms. Psychopharmacology. 2019;236(8):2437–2449. doi: 10.1007/s00213-019-05299-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  95. Seymour B., et al. Differential encoding of losses and gains in the human striatum. J. Neurosci. 2007;27(18):4826–4831. doi: 10.1523/JNEUROSCI.0400-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Sharp P.B., et al. Humans perseverate on punishment avoidance goals in multigoal reinforcement learning. Elife. 2022:11. doi: 10.7554/eLife.74402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Sharp P.B., Miller G.A., Heller W. Transdiagnostic dimensions of anxiety: Neural mechanisms, executive functions, and new directions. Int. J. Psychophysiol. 2015;98(2):365–377. doi: 10.1016/j.ijpsycho.2015.07.001. [DOI] [PubMed] [Google Scholar]
  98. Sharp P.B., Dolan R.J., Eldar E. Disrupted state transition learning as a computational marker of compulsivity. Psychol. Med. 2021 doi: 10.1017/S0033291721003846. [DOI] [PMC free article] [PubMed] [Google Scholar]
  99. Shin L.M., Liberzon I. The neurocircuitry of fear, stress, and anxiety disorders. Neuropsychopharmacology. 2010;35(1):169–191. doi: 10.1038/npp.2009.83. [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. Smith R., et al. Greater decision uncertainty characterizes a transdiagnostic patient sample during approach-avoidance conflict: a computational modelling approach. J. Psychiatry Neurosci. 2021;46(1):E74–E87. doi: 10.1503/jpn.200032. [DOI] [PMC free article] [PubMed] [Google Scholar]
  101. Snyder H.R., et al. Opposite effects of anxiety and depressive symptoms on executive function: the case of selecting among competing options. Cogn. Emot. 2014;28(5):893–902. doi: 10.1080/02699931.2013.859568. [DOI] [PMC free article] [PubMed] [Google Scholar]
  102. Stanley A.T., et al. Roles for the dorsal striatum in aversive behavior. Front. Cell. Neurosci. 2021:15. doi: 10.3389/fncel.2021.634493. [DOI] [PMC free article] [PubMed] [Google Scholar]
  103. Stephan K.E., Mathys C. Computational approaches to psychiatry. Curr. Opin. Neurobiol. 2014;25:85–92. doi: 10.1016/j.conb.2013.12.007. [DOI] [PubMed] [Google Scholar]
  104. Sutton R.S., Barto A.G. Reinforcement learning: an introduction, 2nd edition. Reinf. Learn.: Introd., 2nd Ed. 2018:1–526. [Google Scholar]
  105. Tipples J. Rapid temporal accumulation in spider fear: evidence from hierarchical drift diffusion modelling. Emotion. 2015;15(6):742–750. doi: 10.1037/emo0000079. [DOI] [PubMed] [Google Scholar]
  106. Tzovara A., Korn C.W., Bach D.R. Human pavlovian fear conditioning conforms to probabilistic learning. PLoS Comput. Biol. 2018;14(8) doi: 10.1371/journal.pcbi.1006243. [DOI] [PMC free article] [PubMed] [Google Scholar]
  107. Vogel J.R., Beer B., Clody D.E. A simple and reliable conflict procedure for testing anti-anxiety agents. Psychopharmacologia. 1971;21(1):1–7. doi: 10.1007/BF00403989. [DOI] [PubMed] [Google Scholar]
  108. Vos T., et al. Global, regional, and national incidence, prevalence, and years lived with disability for 328 diseases and injuries for 195 countries, 1990-2016: a systematic analysis for the Global Burden of Disease Study 2016. Lancet. 2017;390(10100):1211–1259. doi: 10.1016/S0140-6736(17)32154-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Wang O., et al. Model-based and model-free pain avoidance learning. Brain Neurosci. Adv. 2018;2 doi: 10.1177/2398212818772964. 2398212818772964. [DOI] [PMC free article] [PubMed] [Google Scholar]
  110. Watson J.B., Rayner R. Conditioned emotional reactions. J. Exp. Psychol. 1920;3:1–14. doi: 10.1037//0003-066x.55.3.313. [DOI] [PubMed] [Google Scholar]
  111. White C.N., et al. Anxiety enhances threat processing without competition among multiple inputs: a diffusion model analysis. Emotion. 2010;10(5):662–677. doi: 10.1037/a0019474. [DOI] [PubMed] [Google Scholar]
  112. Wise T., et al. A computational account of threat-related attentional bias. Plos Comput. Biol. 2019;15:10. doi: 10.1371/journal.pcbi.1007341. [DOI] [PMC free article] [PubMed] [Google Scholar]
  113. Wise T., Dolan R.J. Associations between aversive learning processes and transdiagnostic psychiatric symptoms in a general population sample. Nat. Commun. 2020;11:1. doi: 10.1038/s41467-020-17977-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Zhang S.Y., et al. Dissociable learning processes underlie human pain conditioning. Curr. Biol. 2016;26(1):52–58. doi: 10.1016/j.cub.2015.10.066. [DOI] [PMC free article] [PubMed] [Google Scholar]
  115. Zorowitz S., Momennejad I., Daw N.D. Anxiety, Avoid., Seq. Eval. Comput. Psychiatr. 2020:4. doi: 10.1162/cpsy_a_00026. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES