Skip to main content
Journal of Mathematical Neuroscience logoLink to Journal of Mathematical Neuroscience
. 2019 Jul 3;9:5. doi: 10.1186/s13408-019-0073-4

Drift–diffusion models for multiple-alternative forced-choice decision making

Alex Roxin 1,2,
PMCID: PMC6609930  PMID: 31270706

Abstract

The canonical computational model for the cognitive process underlying two-alternative forced-choice decision making is the so-called drift–diffusion model (DDM). In this model, a decision variable keeps track of the integrated difference in sensory evidence for two competing alternatives. Here I extend the notion of a drift–diffusion process to multiple alternatives. The competition between n alternatives takes place in a linear subspace of n1 dimensions; that is, there are n1 decision variables, which are coupled through correlated noise sources. I derive the multiple-alternative DDM starting from a system of coupled, linear firing rate equations. I also show that a Bayesian sequential probability ratio test for multiple alternatives is, in fact, equivalent to these same linear DDMs, but with time-varying thresholds. If the original neuronal system is nonlinear, one can once again derive a model describing a lower-dimensional diffusion process. The dynamics of the nonlinear DDM can be recast as the motion of a particle on a potential, the general form of which is given analytically for an arbitrary number of alternatives.

Keywords: Decision making, Networks, Winner-take-all

Introduction

Perceptual decision-making tasks require a subject to make a categorical decision based on noisy or ambiguous sensory evidence. A computationally advantageous strategy in doing so is to integrate the sensory evidence in time, thereby improving the signal-to-noise ratio. Indeed, when faced with two possible alternatives, accumulating the difference in evidence for the two alternatives until a fixed threshold is reached is an optimal strategy, in that it minimizes the mean reaction time for a desired level of performance. This is the computation carried out by the sequential probability ratio test devised by Wald [1], and its continuous-time variant, the drift–diffusion model (DDM) [2]. It would be hard to overstate the success of these models in fitting psychophysical data from both animals and human subjects in a wide array of tasks, e.g. [25], suggesting that brain circuits can implement a computation analogous to the DDM.

At the same time, neuroscientists have characterized the neuronal activity in cortical areas of monkey, which appear to reflect an integration process during DM tasks [6], although see [7]. The relevant computational building blocks, as revealed from decades of in-vivo electrophysiology, seem to be neurons, the activity of which selectively increases with increasing likelihood for a given upcoming choice. Attractor network models, built on this principle of competing, selective neuronal populations, generate realistic performance and reaction times; they also provide a neuronal description which captures some salient qualitative features of the in-vivo data [8, 9].

While focus in the neuroscience community has been almost exclusively on two-alternative DM (although see [10]), from a computational perspective there does not seem to be any qualitative difference between two or more alternatives. In fact, in a model, increasing the number of alternatives is as trivial as adding another neuronal population to the competition. On the other hand, how to add an alternative to the DDM framework does not seem, on the face of things, obvious. Several groups have sought to link the DDM and attractor networks for two-alternative DM. When the attractor network is assumed linear, one can easily derive an equation for a decision variable, representing the difference in the activities of the two competing populations, which precisely obeys a DDM [11]. For three-alternative DM previous work has shown that a 2D diffusion process can be defined by taking appropriate linear combinations of the three input streams [12]. The general n-alternative case for leaky accumulators has also been treated [13]. In the first section of the paper I will summarize and build upon this previous work to illustrate how one can obtain an equivalent DDM, starting from a set of linear firing rate equations which compete through global inhibitory feedback. The relevant decision variables are combinations of the activity of the neuronal populations, and which represent distinct modes of competition. Specifically, I will propose a set of “competition” basis functions which allow for a simple, systematic derivation of the DDMs for any n. I will also show how a Bayesian implementation of the the multiple sequential probability ratio test (MSPRT) [1416] is equivalent in the continuum limit to these same DDMs, but with a moving threshold.

Of course, linear models do not accurately describe the neuronal data from experiments on DM. However, previous work has shown that attractor network models for two-alternative DM operate in the vicinity of pitchfork bifurcation, which is what underlies the winner-take-all competition leading to the decision dynamics [17]. In this regime the neuronal dynamics is well described by a stochastic normal-form equation which right at the bifurcation is precisely equivalent to the DDM with an additional cubic nonlinearity. This nonlinear DDM fits behavioral data extremely well, including both correct and error reaction times. In the second part of the paper I will show how such normal-form equations can be derived for an arbitrary number of neuronal populations. These equations can be thought of as nonlinear DDMs and, in fact, are identical to the linear DDMs with the addition of quadratic nonlinearities (for n>2). Amazingly, the dynamics of such a nonlinear DDM can be recast as the diffusion of particle on a potential, which is obtained analytically, for arbitrary n.

Results

The canonical drift diffusion model (DDM) can be written

τX˙=μ+ξ(t), 1

where X is the decision variable, and μ is the drift or the degree of evidence in favor of one choice over the other: we can associate choice 1 with positive values of X and choice 2 with negative values. The Gaussian process ξ(t) represents noise and/or uncertainty in the integration process, with ξ(t)=0 and ξ(t)ξ(t)=σ2δ(tt). I have also explicitly included a characteristic time scale τ, which will appear naturally if one derives Eq. (1) from a neuronal model. The decision variable evolves until reaching one of two boundaries ±θ at which point a decision for the corresponding choice has been made.

It is clear that a single variable can easily be used to keep track of two competing processes by virtue of its having two possible signs. But what if there are three or more alternatives? In this case it is less clear. In fact, if we consider a drift–diffusion process such as the one in Eq. (1) as an approximation to an actual integration process carried out by neuronal populations, then there is a systematic approach to deriving the corresponding DDM. The value of such an approach is that one can directly tie the DDM to the neuronal dynamics, thereby linking behavior to neuronal activity.

I will first consider the derivation of a DDM starting from a system of linear firing rate equations. This analysis is similar to that found in Sect. 4.4 of [13], although the model of departure is different. In this case the derivation involves a rotation of the system so as to decouple the linear subspace for the competition between populations from the subspace which describes non-competitive dynamical modes. This rotation is equivalent to expressing the firing rates in terms of a set of orthogonal basis functions: one set for the competition, and another for the non-competitive modes. I subsequently consider a system of nonlinear firing rate equations. In this case one can once again derive a reduced set of equations to describe the decision-making dynamics. The reduced models have precisely the form of the corresponding DDM for a linear system, but now with additional nonlinear terms. These terms reflect the winner-take-all dynamics which emerge in nonlinear systems with multi-stability. Not only do the equations have a simple, closed-form solution for any number of alternatives, but they can be succinctly expressed in terms of a multivariate potential.

Derivation of a DDM for two-alternative DM

The DDM can be derived from a set of linear equations which model the competition between two populations of neurons, the activity of each of which encodes the accumulated evidence for the corresponding choice. The equations are

τr˙1=r1+sr1crI+I1+ξ1(t),τr˙2=r2+sr2crI+I2+ξ2(t),τIr˙I=rI+g2(r1+r2)+II+ξI(t), 2

where rI represents the activity of a population of inhibitory neurons. The parameter s represents the strength of excitatory self-coupling and c is the strength of the global inhibition. The characteristic time constants of excitation and inhibition are τ and τI respectively. A choice is made for 1 (2) whenever r1=rth (r2=rth).

It is easier to work with the equations if they are written in matrix form, which is

graphic file with name 13408_2019_73_Equ3_HTML.gif 3

where

graphic file with name 13408_2019_73_Equa_HTML.gif

In order to derive the DDM, I express the firing rates in terms of three orthogonal basis functions: one which represents a competition between the populations e1, a second for common changes in population rates ec and a third which captures changes in the activity of the inhibitory cells eI. Specifically I write

r(t)=e1X(t)+eCmC(t)+eImI(t), 4

where e1=(1,1,0) and eC=(1,1,0) and eI=(0,0,1). The decision variable will be X, while mC and mI stand for the common mode and inhibitory mode respectively.

The dynamics of each of these modes can be isolated in turn by projecting Eq. (2) onto the appropriate eigenvector. For example, the dynamics for the decision variable X are found by projecting onto e1, namely

graphic file with name 13408_2019_73_Equ5_HTML.gif 5

and similarly the dynamics for mC and mI and found by projecting onto eC and eI respectively. Doing so results in the set of equations

τX˙=(1s)X+I1I22+ξ1(t)ξ2(t)2,τm˙C=(1s)mCcmI+I1+I22+ξ1(t)+ξ2(t)2,τIm˙I=mI+gmC+II+ξI(t). 6

If the self-coupling sits at the critical value s=1, then these equations simplify to

τX˙=I1I22+ξ1(t)ξ2(t)2,τm˙C=cmI+I1+I22+ξ1(t)+ξ2(t)2,τIm˙I=mI+gmc+II+ξI(t), 7

from which it is clear that the equation for X describes a drift–diffusion process. It is formally identical to Eq. (1) with μ=I1I22 and ξ(t)=ξ1(t)ξ2(t)2. Importantly, X is uncoupled from the common and inhibitory modes, which themselves form a coupled subsystem. For s1 the decision variable still decouples from the other two equations, but the process now has a leak term (or ballistic for s>1) [18]. It has therefore been argued that obtaining a DDM from linear neuronal models requires fine tuning, a drawback which can be avoided in nonlinear models in which the linear dynamics is approximated via multi-stability; see e.g. [19, 20]. If one ignores the noise terms, the steady state of this linear system is (X,mC,mI)=(X0,Mc,MI) where

MC=I1+I22cgIIg,MI=I1+I22c. 8

One can study the stability of this solution by considering a perturbation of the form (X,mC,mI)=(X0,Mc,MI)+(δX,δmC,δmI)eλt. Plugging this into Eq. (7) one finds that there is a zero eigenvalue associated with the decision variable, i.e. λ1=0, whereas the eigenvalues corresponding to the subsystem comprising the common and inhibitory modes are given by

λ2,3=12τI(1±14τIτcg), 9

which always have a negative real part. Therefore, as long as τI is not too large, perturbations in the common mode or in the inhibition will quickly decay away. This allows one to ignore their dynamics and assume they take on their steady-state values. Finally, the bounds for the decision variable are found by noting that r1=X+MC and r2=X+MC. Therefore, given that the neuronal threshold for a decision is defined as rth, we find that θ=±(rthMC).

Derivation of a DDM for three-alternative DM

I will go over the derivation of a drift–diffusion process for three-choice DM in some detail for clarity, although conceptually it is very similar to the two-choice case. Then the derivation can be trivially extended to n-alternative DM for any n.

The linear rate equations are once again given by Eq. (3), with

graphic file with name 13408_2019_73_Equb_HTML.gif

One once again writes the firing rates in terms of orthogonal basis functions, of which there must now be four. The common and inhibitory modes are the same as before, whereas now there will be two distinct modes to describe the competition between the three populations, in contrast to just a single decision variable. Any orthogonal basis in the 2D space for competition is equally valid. However, in order to make the choice systematic, I assume that the first vector is just the one from the two-alternative case, namely (1,1,0,0), from which it follows that the second must be (up to an amplitude) (1,1,2,0). Then for n alternatives I will always take the first n2 basis vectors to be those from the n1 case. The last eigenvector must be orthogonal to these. Specifically, for n=3, r=e1X1(t)+e2X2(t)+eCmC(t)+eImI, where

e1=(1,1,0,0),e2=(1,1,2,0),eC=(1,1,1,0),eI=(0,0,0,1). 10

One projects Eq. (3) onto the four relevant eigenvectors, which leads to the following equations:

τX1˙=(1s)X1+I1I22+ξ1(t)ξ2(t)2,τX2˙=(1s)X2+I1+I22I36+ξ1(t)+ξ2(t)2ξ3(t)6,τmC˙=(1s)mCcmI+I1+I2+I33+ξ1(t)+ξ2(t)+ξ3(t)3,τImI˙=mI+gmC+II+ξI(t). 11

When s=1 then the first two equations in Eq. (11) describe a drift–diffusion process in a two-dimensional subspace, while the coupled dynamics of the common and inhibitory modes are once again strongly damped. The DDM for three-alternative DM can therefore be written

τX˙1=I1I22+ξ1(t)ξ2(t)2, 12
τX˙2=I1+I22I36+ξ1(t)+ξ2(t)2ξ3(t)6. 13

Note that the dynamics of the two decision variables X1 and X2 are coupled through the correlation in their noise sources. The decision boundaries are set by noting that

r1=X1+X2+MC,r2=X1+X2+MC,r3=2X2+MC. 14

Therefore, given that the neuronal threshold for a decision is defined as rth we can set three decision boundaries: 1 Population 1 wins if X2=X1+rthMC, 2 Population 2 wins if X2=X1+rthMC and 3 Population 3 wins if X2=(rthMC)/2. These three boundaries define a triangle in (X,Y)-space over which the drift–diffusion process take place.

Derivation of DDMs for n-alternative DM

The structure of the linear rate equations Eq. (3) can be trivially extended to any number of competing populations. In order to derive the corresponding DDM one need only properly define the basis functions for the firing rates, which was described above. The common and inhibitory modes always have the same structure. If the basis functions are ei and the corresponding decision variables Xi for i=[1,n1], then the firing rates are r=i=1n1eiXi(t) and it is easy to show that the dynamics for the kth decision variable is given by

τX˙k=ekIekek+ekξekek, 15

as long as s=1. The decision boundaries are defined by setting the firing rates equal to their threshold value for a decision, i.e.

i=1n1eiXi,b=rthu, 16

where u=(1,1,,1).

The basis set proposed here for n-alternative DM is to take for the kth eigenvector

ek=(1,1,,1,k,0,,0), 17

where the element −k appears in the (k+1)st spot, and which is a generalization of the eigenvector basis taken earlier for two- and three-alternative DM. With this choice, the equation for the kth decision variable can be written

τX˙k=ekIk+k2+ekξk+k2. 18

The firing rate for the ith neuronal population can then be expressed in terms of the decision variables as

ri=(i1)xi1+l=in1xl+MC, 19

which, given a fixed neuronal threshold rth, directly gives the bounds on the decision variables. Namely, the ith neuronal population wins when (i1)xi1+l=in1xl+MC>rth.

The n-alternative DDM reproduces the well-known Hick’s law [21], which postulates that the mean reaction time (RT) increases as the logarithm of the number of alternatives, for fixed accuracy; see Fig. 1.

Figure 1.

Figure 1

Hick’s Law for multi-alternative DDMs. If the accuracy is held fixed as the number of alternatives n is increased, then the mean reaction time increases as the logarithm of n. This is shown for both an accuracy of 0.8 (circles) and 0.6 (squares). The inset shows that the variance in the RT increases proportional to the mean RT. Parameter values are σ=1, τ=20 ms. The thresholds θPc=0.8=3 and θPc=0.6=0.9 for n=2, and are increased for n>2 to achieve the same accuracy. The initial condition is always taken to be Xi=0 for all i. The symbols are the average of 10,000 runs. The solid curves in the main panel are fits to the function RT=a+bln(c+n). The solid lines in the inset are best-fit linear regressions

DDM for multiple alternatives as the continuous-time limit of the MSPRT

In the previous section I have illustrated how to derive DDMs starting from a set of linear differential equations describing an underlying integration process of several competing data streams. An alternative computational framework would be to apply a statistical test directly to the data, without any pretensions with regards to a neuronal implementation. In fact, the development of an optimal sequential test for multiple alternatives followed closely on the heels of Wald’s work for two alternatives in the 1940s [22]. Subsequent work proposed a Bayesian framework for a multiple sequential probability ratio test (MSPRT) [1416]. Here I will show how such a Bayesian MSPRT is equivalent to a corresponding DDM in the continuum limit, albeit with moving thresholds.

I will follow the exposition from [16] in setting up the problem, although some notation differs for clarity. I assume that there are n possible alternatives, and that the instantaneous evidence for an alternative i is given by zi(t), which is drawn from a Gaussian distribution with mean Ii and variance σ2. The total observed input over all alternatives n and up to a time t is written I and the hypothesis that alternative i has the highest mean Hi. Then, using Bayes theorem, the probability that alternative i has the highest mean given the total observed input is

Pr(Hi|I)=Pr(I|Hi)Pr(Hi)Pr(I). 20

Furthermore one has Pr(I)=k=1nPr(I|Hk)Pr(Hk). Given equal priors on the different hypotheses, Eq. (20) can be simplified to

Pr(Hi|I)=Pr(I|Hi)k=1nPr(I|Hk). 21

Finally, the log-likelihood of the alternative i having the largest mean up to a time t is Li(t)=lnPr(Hi|I), which is given by

Li(t)=ln(Pr(I|Hi))ln(k=1nelnPr(I|Hk)). 22

The choice of Gaussian distributions for the input leads to a simple form for the log-likelihoods. Specifically, the time rate-of-change of the log-likelihood for alternative i is given by

L˙i(t)=zik=1nzkeykk=1neyk, 23

where yi(t)=0tdtzi(t). Note that zi(t)=Ii+ξi(t), where ξi is a Gaussian white noise process with zero mean and variance σ2.

I can now write the Li in terms of the orthogonal basis used in the previous section,

L=j=1n1ejXj(t)+eCMC(t). 24

Formally projecting onto each mode in turn leads precisely to the DDMs of Eq. (18) (with τ=1) for the decision variables, while the dynamics for the common mode is

M˙C=1nj=1nzjk=1nzkeykk=1neyk. 25

For Bayes-optimal behavior, a choice i should be chosen if the log-likelihood exceeds a given threshold, namely if

Li(t)=(i1)Xi1+l=in1Xl>θMC(t). 26

Note that the log-likelihood is always negative and hence does not represent a firing rate, as in the differential equations studied in the previous section. This does not pose a problem since we are simply implementing a statistical test. On the other hand, an important difference with the case studied previously is the fact that, for the neuronal models, the common mode was stable and converged to a fixed point. Therefore, the decision variable dynamics was equivalent to the original rate dynamics with a shift of threshold. Here, that is not the case. The common mode represents the normalizing effect of the log-marginal probability of the stimulus, which always changes in time. Specifically, if we assume that Il>Ii for all i, namely that the mean of the distribution is greatest for alternative l, then the expected dynamics of the common mode at long times are

M˙C=1nj=1nIjk=1nIkeIktk=1neIkt1nj=1nIjIl=1nj=1n|IjIl|. 27

Therefore the DDMs are only equivalent to applying the Bayes theorem if the threshold is allowed to vary in time. In this case they are, in fact, mathematically identical and thus give the same accuracy and reaction time distributions, as shown in Fig. 2.

Figure 2.

Figure 2

A Bayesian MSPRT is equivalent to DDMs with moving threshold. Accuracy and reaction time for three alternatives. A. Accuracy as a function of the mean of input 1 for the Bayesian test, Eq. (23) (solid line), for the DDM with moving threshold (dotted line) and the DDM with fixed threshold (dashed line). Note that the Bayesian test and the DDM with MT give identical results and are, in fact, mathematically identical. B. Reaction times for the Bayesian test (solid line), DDM with MT (dotted line) and DDM with fixed threshold (dashed line). Orange lines are for error reaction times. Error bars indicate the standard error of the mean. Other parameter values are: σ=0.03, Ii=0 for i=2,3, θ=1.05 with MT and θ=1.073 with fixed threshold. The thresholds were chosen to give identical reaction times for I1=0. The values shown are averages over 10,000 trials for each value of the input

On the other hand, an equivalent DDM with a fixed threshold has worse accuracy and shorter reaction times. The way I choose the parameters for a fair comparison is to set the fixed threshold such that the mean reaction time is identical to the Bayesian model for zero coherence. Another important difference is that the error RTs are longer than the correct RTs for the Bayesian model, see Fig. 2B, an effect which is commonly seen in experiment (and is also reproduced by the nonlinear DDMs studied in the next section of the paper) [17]. On the other hand correct and error RTs for the DDMs are always the same.

Derivation of a reduced model for two-choice DM for a nonlinear system

A more realistic firing rate model for a decision-making circuit allows for a nonlinear input-output relationship in neuronal activity. For two-alternative DM the equations are

τr˙1=r1+ϕ(sr1crI+I1)+ξ1(t),τr˙2=r2+ϕ(sr2crI+I1)+ξ2(t),τIr˙I=rI+ϕI(g2(r1+r2)+II)+ξI(t). 28

The nonlinear transfer function ϕ (ϕI) does not need to be specified in the derivation. The noise sources ξi are taken to be Gaussian white noise and hence must sit outside of the transfer function; they therefore directly model fluctuations in the population firing rate rather than input fluctuations. Input fluctuations can be modeled by allowing for a non-white noise process and including it directly as an additional term in the argument of the transfer function. Note that here I assume the nonlinearity is a smooth function. This is a reasonable assumption for a noisy system such as a neuron or neuronal circuit. Non-smooth systems, such as piecewise linear equations for DM, require a distinct analytical approach; see, e.g., [23].

The details of the derivation for two alternatives can be found in [17], but here I give a flavor for how one proceeds; the process will be similar when there are three or more choices, although the scaling of the perturbative expansion is different. One begins by ignoring the noise sources and linearizing Eq. (28) about a fixed-point value for which the competing populations have the same level of activity, and hence also I1=I2. Specifically one takes (r1,r2,rI)=(R,R,RI)+(δr1,δr2,δrI)eλt, where δr1. In vector form this can be written r=R+δreλt. Plugging this ansatz into Eq. (28) and keeping only terms linear in the perturbations leads to the following system of linear equations:

graphic file with name 13408_2019_73_Equ29_HTML.gif 29

where

graphic file with name 13408_2019_73_Equ30_HTML.gif 30

and the slope of the transfer function ϕ is calculated at the fixed point. Note that the matrix Eq. (30) has a very similar structure to the linear operator in Eq. (3). This system of equations only has a solution if the determinant of the matrix is equal to zero; this yields the characteristic equation for the eigenvalues λ. These eigenvalues are

λ1=(1sϕ)τ, 31
λ2,3=(τ+τI(1sϕ))2ττI±12ττI(ττI(1sϕ))24ττIcgϕϕI. 32

Note that λ1=0 if 1sϕ=0, while the real part of the other two eigenvalues is always negative. This indicates that there is an instability of the fixed point in which the activity of the neuronal populations is the same, and that the direction of this instability can be found by setting 1sϕ=0 and λ=0 in Eq. (30). This yields Inline graphic, where

graphic file with name 13408_2019_73_Equ33_HTML.gif 33

the solution of which can clearly be written δr=(1,1,0). This is the same competition mode as found earlier for the linear system.

A brief overview of normal-form derivation

At this point it is still unclear how one can leverage this linear analysis to derive a DDM. Specifically, and unlike in the linear case, one cannot simply rotate the system to uncouple the competition dynamics from the non-competitive modes. Also, note that the steady states in a nonlinear system depend on the external inputs, whereas that is not the case in a linear system. In particular, the DDM has a drift term μ which ought to be proportional to the difference in inputs I1I2, whereas to perform the linear stability we assumed I1=I2. Indeed, if one assumes that the inputs are different, then the fixed-point structure is completely different. The solution is to assume that the inputs are only slightly different, and formalize this by introducing the small parameter ϵ. Specifically, we write I1=I0+ϵ2ΔI+ϵ3I¯1, and I2=I0+ϵ2ΔI+ϵ3I¯2. In this expansion, I0 is the value of the external input which places the system right at the bifurcation in the zero-coherence case. In order to describe the dynamics away from the bifurcation we also allow the external inputs to vary. Specifically, ΔI represents the component of the change in input which is common to both populations (overall increased or decreased drive compared to the bifurcation point), while I¯i is a change to the drive to population i alone, and hence captures changes in the coherence of the stimulus. The particular scaling of these terms with ϵ is enforced by the solvability conditions which appear at each order. That is, the mathematics dictates what these are; if one chose a more general scaling one would find that only these terms would remain.

The firing rates are then also expanded in orders of ϵ and written

r=r0+ϵe1X(T)+O(ϵ2), 34

where r0 are the fixed-point values, e1 is the eigenvector corresponding to the zero eigenvalue and X is the decision variable which evolves on a slow-time scale, T=ϵ2t. The slow-time scale arises from the fact that there is an eigenvector with zero eigenvalue; when we change the parameter values slightly, proportional to ϵ, the growth rate of the dynamics along that eigenvector is no longer zero, but still very small, in fact proportional to ϵ2 in this case.

The method for deriving the normal-form equation, i.e. the evolution equation for X, involves expanding Eq. (28) in ϵ. At each order in ϵ there is a set of equations to be solved; at some orders, in this case first at order O(ϵ3), the equations cannot be solved and a solvability condition must be satisfied, which leads to the normal-form equation.

The normal-form equation for two choices

Following the methodology described in the preceding section leads to the evolution equation for the decision variable X,

τTX=η(I¯1I¯2)+μΔIX+γX3+ξ(t), 35

where for the case of Eq. (28), η=ϕ/2, ξ(t)=(ξ1(t)ξ2(t))/2 and the coefficients μ and γ are

μ=s2ϕcgϕI,γ=s3(ϕ)22cgϕϕI(scgϕI)+s3ϕ6, 36

see [17] for a detailed calculation. Equation (35) provides excellent fits to performance and reaction times for monkeys and human subjects; see Fig. 3 from [17].

It is important to note that the form of Eq. (35) only depends on there being a two-way competition, not on the exact form of the original system. As an example, consider another set of firing rate equations

τr˙1=r1+ϕ(sr1cr2+I1),τr˙2=r2+ϕ(sr2cr1+I2), 37

where rather than model the inhibition explicitly, an effective inhibitory interaction between the two populations is assumed. In this case the resulting normal-form equation is still Eq. (35). In fact, performing a linear stability analysis on Eq. (37) yields a null eigenvector e1=(1,1). This indicates that the instability causes one population to grow at the expense of the other, in a symmetric fashion, as before. This is the key point which leads to the normal-form equation. More specifically we see that for both systems r1=R+X while r2=RX, which means that if we flip the sign on the decision variable X and switch the labels on the neuronal populations, the dynamics is once again the same. This reflection symmetry ensures that all terms in X will have odd powers in Eq. (35) [17]. It is broken only when the inputs to the two populations are different, i.e. by the first term on the r.h.s. in Eq. (35).

As we shall see, the stochastic normal-form equation, Eq. (35), which from now on I will refer to as a nonlinear DDM, has a very different form from the nonlinear DDMs for n>2. The reason is, again, the reflection symmetry in the competition subspace for n=2, which is not present for n>2. Therefore, for n>2 the leading-order nonlinearity is quadratic, and, in fact, a much simpler function of the original neuronal parameters.

Three-alternative forced-choice decision making

The derivation of the normal form, and the corresponding DDM for three-choice DM differs from that for two-alternative DM in several technical details; these differences continue to hold for n-alternative DM for all n3. Therefore I will go through the derivation in some detail here and will then extend it straightforwardly to the other cases.

Again I will make use of a particular system of firing rate equations to illustrate the derivation. I take a simple extension of the firing rate equations for two-alternative DM Eq. (28). The equations are

τr˙1=r1+ϕ(sr1crI+I1)+ξ1(t),τr˙2=r2+ϕ(sr2crI+I2)+ξ2(t),τr˙3=r3+ϕ(sr3crI+I3)+ξ3(t),τIr˙I=rI+ϕI(g3(r1+r2+r3)+II)+ξI(t). 38

I first ignore the noise terms and consider the linear stability of perturbations of the state in which all three populations have the same level of activity (and so I1=I2=I3=I0), i.e. r=R+δreλt, where R=(R,R,R,RI). This once again leads to a set of linear equations Inline graphic. The fourth-order characteristic equation leads to precisely the same eigenvalues as in the two-choice case, Eqs. (31) and (32), with the notable difference that the first eigenvalue has multiplicity two. This means that if 1sϕ=0 then there will be two eigenvalues identically equal to zero and two stable eigenvalues. This is the first indication that the integration process underlying the DM process for three choices will be two-dimensional. The eigenvectors for the DM process are found by solving Inline graphic, where

graphic file with name 13408_2019_73_Equ39_HTML.gif 39

There are many possible solutions; a simple choice would be e1=(1,1,0,0) and e2=(1,1,2,0), and so e1Te2=0. Note that in this linear subspace any valid choice of eigenvector will have the property that the sum of all the elements will equal zero; this will be true whatever the dimensionality of the DM process and reflects the fact that all of the excitatory populations excite the inhibitory interneurons in equal measure.

To derive the normal form I once again assume that the external inputs to the three populations differ by a small amount, namely (I1,I2,I3)=(I0,I0,I0)+ϵ2(I¯1,I¯2,I¯3), and then expand the firing rates as r=R+ϵ(e1X1(T)+e2X2(T))+O(ϵ2), where the slow time is T=ϵt. Note that the inputs are only expanded to second order in ϵ, as opposed to third order as in the previous section. The reason is that the solvability condition leading to the normal-form equation for the 2-alternative case arises at third order. This is due to the fact that the bifurcation has a reflection symmetry, i.e. it is a pitchfork bifurcation and so only odd terms in the decision variable are allowed. The lowest-order nonlinear term is therefore the cubic one. On the other hand, for more than two alternatives there is no such reflection symmetry in the corresponding bifurcation to winner-take-all behavior. Therefore the lowest-order nonlinear term is quadratic, as in a saddle-node bifurcation.

I expand Eq. (38) in orders of ϵ. In this case a solvability condition first arises at order ϵ2, which also accounts for the different scaling of the slow time compared to two-choice DM. Note that there are two solvability conditions, corresponding to eliminating the projection of terms at that order onto both of the left-null eigenvectors of Inline graphic. As before, the left-null eigenvectors are identical to the right-null eigenvectors. Applying the solvability condition yields the normal-form equations

τX˙1=ϕ2(I¯1I¯2)+s2ϕX1X2+12(ξ1(t)ξ2(t)),τX˙2=ϕ6(I¯1+I¯22I¯3)+s2ϕ6(X123X22)+16(ξ1(t)+ξ2(t)2ξ3(t)). 40

The nonlinear DDM Eq. (40) provides an asymptotically correct description of the full dynamics in Eq. (38) in the vicinity of the bifurcation leading to the decision-making behavior. Figure 3(A) shows a comparison of the firing rate dynamics with the nonlinear DDM right at the bifurcation. The appropriate combinations of the two decision variables X1 and X2 clearly track the rate dynamics accurately, including the correct choice (here population 2) and reaction time. The nonlinear drift–diffusion process evolves in a triangular section of the plane; see Fig. 3(B).

Figure 3.

Figure 3

Decision-making dynamics for three alternatives. (A) A comparison of the dynamics from integration of the coupled firing rate equations Eq. (38) (solid lines) and from the DDM, Eq. (40) (dotted lines). (B) The dynamics from three separate simulations of the DDM, Eq. (40), indicating three different winning alternatives. Parameters are τ=20 ms, τI=10 ms, s=c=g=1, σ=0.01. The function ϕ(x)=0 for x<0, x2 for 0x1 and 2x3/4 for x>1. For this choice of transfer function and these parameter values, there is a bifurcation for I0=1/2, which is the value taken here. The steady-state firing rate R=1/4. The threshold θ=1/2

A note on the difference between the nonlinear DDM for 2A and 3A DM

The dynamics of the nonlinear DDM for 2A, Eq. (35), depends strongly on the sign of the cubic coefficient γ. Specifically, when γ<0 the bifurcation is supercritical, while for γ>0 it is subcritical, indicating the existence of a region of multi-stability for ΔI<0. In fact, in experiment, cells in parietal cortex which exhibit ramping activity during perceptual DM tasks, also readily show delay activity in anticipation of the sacade to their response field [24]. One possible mechanism for this would be precisely this type of multi-stability. When ΔI=0, i.e. when the system sits squarely at the bifurcation, Eq. (35) is identical to its linear counterpart with the sole exception of the cubic term. For γ<0 the state X=0 is stabilized. In fact, the dynamics of the decision variable can be viewed as the motion of a particle in a potential, which for γ<0 increases rapidly as X grows, pushing the particle back. On the other hand, for γ>0 the potential accelerates the motion of X, pushing it off to ±∞. This is very similar to the potential for the linear DDM with absorbing boundaries. Therefore, the nonlinear DDM for two-alternatives is qualitatively similar to the linear DDM when it is subcritical, and hence when the original neuronal system is multi-stable.

On the other hand, the nonlinear DDM for three alternatives, Eq. (40), has a much simpler, quadratic nonlinearity. The consequence of this is that there are no stable fixed points and the decision variables always evolve to ±∞. Furthermore, to leading order there is no dependence on the mean input, indicating that the dynamics is dominated by the behavior right at the bifurcation.1 The upshot is that Eq. (40) is as similar to the corresponding linear DDM with absorbing boundaries as possible for a nonlinear system without fine tuning. This remains true for all n>2.

This also means that neuronal systems with inhibition-mediated winner-take-all dynamics are generically multi-stable for n>2, although for n=2 they need not be. This is due to the reflection symmetry present only for n=2.

n-alternative forced-choice decision making

One can now extend the analysis for three-alternative DM to the more general n-choice case. Again I start with a set of firing rate equations

τr˙1=r1+ϕ(sr1crI+I1)+ξ1(t),τr˙2=r2+ϕ(sr2crI+I2)+ξ2(t),τr˙n=rn+ϕ(srncrI+In)+ξn(t),τIr˙I=rI+ϕI(gnj=1nrj+II)+ξI(t).

A linear stability analysis shows that the eigenvalues of perturbations of the state r=(R,R,,R,RI) are given by Eqs. (31) and (32), where the first eigenvalue has multiplicity n1. Therefore the decision-making dynamics evolves on a manifold of dimension n1. The linear subspace associated with this manifold is spanned by n1 eigenvectors which are mutually orthogonal and the elements of which sum to zero. For n alternatives, we take n1 eigenvectors of the form ek=(1,1,,1,k,0,,0), for the kth eigenvector (again, the −k sits in the (k+1)-st spot). Therefore one can write

r=R+ϵi=1n1eiXi(T)+O(ϵ2). 41

Following the same procedure as in the case of three-choice DM and applying the n1 solvability conditions at order O(ϵ2), one arrives at the following normal-form equation for the kth decision variable for n alternatives:

τX˙k=ϕekTIk+k2+s2ϕ2(k+k2)(j=1k1(j+j2)Xj2k(k21)Xk2+2(k+k2)Xkj=k+1n1Xj)+ekTξ(t)k+k2. 42

It is illuminating to write this formula out explicitly for some of the decision variables:

τX˙1=ϕ12(I¯1I¯2)+s2ϕX1j=2n1Xj,τX˙2=ϕ16(I¯1+I¯22I¯3)+s2ϕ6(X123X22+6X2j=3n1Xj),τX˙3=ϕ112(I¯1+I¯2+I¯33I¯4)+s2ϕ12(X12+3X2212X32+12X3j=4n1Xj),τX˙n1=ϕ1n(n1)(j=1n1I¯j(n1)I¯n)τX˙n1=+s2ϕn(n1)(j=1n2(j+j2)Xj2n(n1)2Xn2), 43

where I have left off the noise terms for simplicity.

Surprisingly, the n1 equations for the decision variables in n-alternative DM can all be derived from a single, multivariate function:

f(X1,,Xn)=aj=1n1ejIXjb2(j=1n2(j+j2)Xj2i=j+1n1Xij=1n1j(j21)3Xj3). 44

For the system of firing rate equations studied here the parameters a=ϕ and b=s2ϕ. Then the equation for the kth decision variable is simply

τX˙k=1k+k2fXk+ekξ(t)k+k2. 45

The dynamics of the function f is given by

τft=j=1n1fXjX˙j=j=1n11k+k2(fXj)2<0, 46

where I have ignored the effect of noise. Therefore the dynamics of the decision variables can be thought of as the motion of a particle on a potential landscape, given by f. Noise sources lead to a diffusion along this landscape.

Discussion

In this paper I have illustrated how to derive drift–diffusion models starting from models of neuronal competition for n-alternative decision-making tasks. In the case of linear systems, the derivation consists of nothing more than a rotation of the dynamics onto a subspace of competition modes. This idea is not new, e.g. [13], although I have made the derivation explicit here, and have chosen as a model of departure one in which inhibition is explicitly included as a dynamical variable. It turns out that a Bayesian implementation of a multiple sequential probability ratio test is also equivalent to a DDM in the continuum limit, albeit with time-varying thresholds.

For nonlinear systems, the corresponding DDM is a stochastic normal form, which is obtained here using the method of multiple-scales [25]. The nonlinear DDM was obtained earlier for the special case of two-alternative DM [17]. For four-alternative DM the nonlinear DDM was obtained with a different set of competitive basis functions [26] to describe performance and reaction time from experiments with human subjects. This led to a different set of coupled normal-form equations from those given by Eq. (42), although the resulting dynamics is, of course, the same. The advantage of the choice I have made in this paper for the basis functions, is that they are easily generalizable for any n, leading to a simple, closed-form expression for the nonlinear DDM for any arbitrary number of alternatives, Eq. (42).

An alternative approach to describing the behavior in DM tasks, is to develop a statistical or probabilistic description of evidence accumulation; see, e.g., [1, 13, 15, 22, 27]. Such an approach also often leads to a drift–diffusion process in some limit, as is the case for the Bayesian MSPRT studied here, and see also [28]. In fact, recent work has shown that an optimal policy for multiple-alternative decision making can be approximately implemented by an accumulation process with time-varying thresholds, similar to the Bayesian model studied in this manuscript [29]. From a neuroscience perspective, however, it is of interest to pin down how the dynamics of neuronal circuits might give rise to animal behavior which is well described by a drift–diffusion process. This necessitates the analysis of neuronal models at some level. What I have shown here is that the dynamics in a network of n neuronal populations which compete via a global pool of inhibitory interneurons, can in general be formally reduced to a nonlinear DDM of dimension n1. The nonlinear DDMs differ from the linear DDMs through the presence of quadratic (or cubic for n=2) nonlinearities which accelerate the winner-take-all competition. In practical terms this nonlinear acceleration serves the same role as the hard threshold in the linear DDMs. Therefore the two classes of DDMs have quite similar behavior.

The DDM is one of the most-used models for fitting data from two-alternative forced-choice decision-making experiments. In fact it provides fits to decision accuracy and reaction time in a wide array of tasks, e.g. [2, 3, 30]. Here I have illustrated how the DDM can be extended to n alternatives straightforwardly. It remains to be seen if such DDMs will fit accuracy and reaction times as well as their two-alternative cousin, although one may refer to promising results from [12] for three alternatives. Note also that the form of the nonlinear DDMs, Eqs. (44) and (45) does not depend on the details of the original neuronal equations; this is what is meant by a normal-form equation. The only assumptions needed for the validity of the normal-form equations are that there be global, nonlinear competition between n populations. Of course, if the normal form is derived from a given neuronal model, then the parameters a and b of the nonlinear potential Eq. (44) will depend on the original neuronal parameters.

As stated earlier, the nonlinear DDMs can have dynamics quite similar to the standard, linear DDM with hard thresholds. Nonetheless, there are important qualitative differences between the two classes of models. First of all, both correct and error reaction-time distributions are identical in the linear DDMs, given unbiased initial conditions, whereas the nonlinear DDMs generically show longer error reaction times [17], also a feature of the Bayesian MSPRT. Because error reaction times in experiment indeed tend to be longer than correct ones, the linear DDM cannot be directly fit to data. Rather, variability in the drift rate across trials can be assumed in order to account for differences in error and correct reaction times; see, e.g., [30]. Secondly, nonlinear DDMs exhibit intrinsic dynamics which reflect the winner-take-all nature of neuronal models with strong recurrent connectivity. As a consequence, as the decision variables increase (or decrease) from their initial state, they undergo an acceleration which does not explicitly depend on the value of the external input. This means that the response of the system to fluctuations in the input is not the same late in a trial as it is early on. Specifically, later fluctuations will have lesser impact. Precisely this effect has been seen in the response of neurons in parietal area LIP in monkeys in two-alternative forced-choice decision-making experiments; see Fig. 10B in [31]. Given that network models of neuronal activity driving decision-making behavior lead to nonlinear DDMs, fitting such models to experimental data in principle allows one to link behavioral measures to the underlying neuronal parameters. In fact, it may be that the linear DDM has been so successful in fitting behavioral data over the years precisely because it is a close approximation to the true nonlinear DDM which arises in neuronal circuits with winner-take-all dynamics.

Acknowledgments

Acknowledgements

I acknowledge helpful conversations with Klaus Wimmer and valuable suggestions for improvements from the reviewers.

Availability of data and materials

Data sharing not applicable to this article as no datasets were generated or analyzed during the current study.

Abbreviations

DDM

Drift Diffusion Model

DM

Decision Making

Appendix

1.1 Derivation of the normal form for three-alternative DM

I assume the inputs to the three competing neuronal populations in Eq. (38) are slightly different, namely Ii=I0+ϵ2I¯i for i={1,2,3}, which defines the small parameter ϵ1. The firing rates are expanded as r=R+ϵr1+ϵ2r2+O(ϵ3) where R are the values of the rates at the fixed point. Finally, we define a slow time T=ϵt. Plugging the expansions for the inputs and rates into Eq. (38) and expanding leads to the following system of equations:

graphic file with name 13408_2019_73_Equ47_HTML.gif 47

where to place the system at the bifurcation to winner-take-all behavior I choose 1sϕ=0. Then we have the linear operator

graphic file with name 13408_2019_73_Equ48_HTML.gif 48

and

L1=(τT0000τT0000τT0000τIT), 49
N2=ϕ(I¯1I¯2I¯30)+12(ϕ(sr11crI1)2ϕ(sr21crI1)2ϕ(sr31crI1)2ϕI′′g29(r11+r21+r31)2). 50

Now one solves the equations order by order. At O(ϵ)

graphic file with name 13408_2019_73_Equ51_HTML.gif 51

The solution can be written r1=e1X1(T)+e2X2(T), where the right-null eigenvectors of Inline graphic are e1=(1,1,0,0) and e2=(1,1,2,0). The left-null eigenvectors, which satisfy Inline graphic are identical to the right-null eigenvectors in this case. Up to O(ϵ2):

graphic file with name 13408_2019_73_Equ52_HTML.gif 52

Given the solution r1, the forcing term has the form

N2=ϕ(I¯1I¯2I¯30)+s2ϕ2((X1+X2)2(X1+X2)24X220). 53

This forcing term contains a projection in the left-null eigenspace of Inline graphic. This projection cannot give rise to a solution in r2 and so must be eliminated. The solvability conditions are therefore

graphic file with name 13408_2019_73_Equ54_HTML.gif 54

which yield Eq. (40). Note that the resulting normal form has detuning or bias terms, proportional to differences in inputs; these would be called ‘drifts’ in the context of a DDM, but no term linearly proportional to the normal-form amplitudes, or ‘decision variables’, as in the normal form for two-choice DM. Such a term describes the growth rate of the critical mode away from the bifurcation point: attracting to one side and repelling to the other. To derive these terms we will need to continue the calculation to next order. First, we must solve for r2. We note that, while r2 cannot have components in the directions e1 or e2, the components in the directions eC=(1,1,1,0) and eI=(0,0,0,1) (the common and inhibitory modes which we discussed in the section on DDMs) can be solved for.

Therefore we take r2=eCR2C+eIR2I and project

graphic file with name 13408_2019_73_Equ55_HTML.gif 55

which yields

R2C=1gϕIR2I,R2I=13c(I¯1+I¯2+I¯3)+s2ϕ3cϕ(X12X22). 56

Up to O(ϵ3):

graphic file with name 13408_2019_73_Equ57_HTML.gif 57

where

N3=(sϕ(X1+X2)((scgϕI)R2C+I¯1)sϕ(X1+X2)((scgϕI)R2C+I¯2)2sϕX2((scgϕI)R2C+I¯3)0)+s3ϕ6((X1+X2)3(X1+X2)38X230). 58

Applying the solvability conditions gives the equations

0=e1N3,0=e2N3. 59

If we add the resulting terms to the normal-form equations derived at O(ϵ2) we have

τTX1=ϕ2(I¯1I¯2)+sϕ(I¯1+I¯22+(scgϕI)3cgϕI(I¯1+I¯2+I¯3))X1+sϕ2(I¯1I¯2)X2+sϕX1X2+(scgϕI)3cgϕϕIs3(ϕ)2X1(X12X22)+s3ϕ6X1(X12+3X22),τTX2=ϕ6(I¯1+I¯22I¯3)+sϕ(I¯1+I¯2+4I¯36+(scgϕI)3cgϕI(I¯1+I¯2+I¯3))X2+sϕ6(I¯1+I¯2)X1+s2ϕ6(X123X22)+(scgϕI)3cgϕϕIs3(ϕ)2X2(X12X22)+s3ϕ6X2(X12+3X22). 60
1.1.1 Normal form for three-choice DM starting with a different neuronal model

If we consider a different neuronal model for three-choice DM we can still arrive at the same normal-form equations. The only important factor is the presence of a linear subspace associated with a zero eigenvalue of multiplicity two and which is spanned by eigenvectors representing competition modes. As an example, consider the model

τr˙1=r1+ϕ(sr1c2(r2+r3)+I1),τr˙2=r2+ϕ(sr2c2(r1+r3)+I2),τr˙3=r3+ϕ(sr3c2(r1+r2)+I3), 61

in which the inhibition is not explicitly modeled as before. A linear stability analysis reveals that there is a zero eigenvalue with multiplicity two when cϕ2=1sϕ and 1sϕ>0. Therefore, the instability mechanism here is the cross-inhibition, while before it was the self-coupling. Nonetheless, it is only the structure of the corresponding eigenvectors which matters. In this case the two eigenvectors with zero eigenvalue are e1=(1,1,0) and e2=(1,1,2). Following the methodology described above leads to the identical normal-form equations Eq. (40) with s replaced by s+c/2.

1.2 Normal-form equations for n-choice DM

A neuronal model for n-choice DM (n>3) is

τr˙1=r1+ϕ(sr1crI+I1)+ξ1(t),τr˙2=r2+ϕ(sr2crI+I2)+ξ2(t),τr˙n=rn+ϕ(srncrI+In)+ξn(t),τIr˙I=rI+ϕI(gnj=1nrj+II)+ξI(t). 62

I again consider small differences in the inputs, Ii=I0+ϵ2ΔIi for i=[1,n], expand the rates as r=R+ϵr1+O(ϵ2), where R are the rates at the fixed point, and define the slow time T=ϵt. If we take 1sϕ=0 then the matrix of the linearized system has a zero eigenvalue with multiplicity n1. Carrying out an expansion up to O(ϵ2) as described above and applying the n1 solvability conditions leads to the normal-form equations Eq. (42).

Authors’ contributions

All authors read and approved the final manuscript.

Funding

Grant number MTM2015-71509-C2-1-R from the Spanish Ministry of Economics and Competitiveness. Grant 2014 SGR 1265 4662 for the Emergent Group “Network Dynamics” from the Generalitat de Catalunya. This work was partially supported by the CERCA program of the Generalitat de Catalunya.

Ethics approval and consent to participate

Not applicable.

Competing interests

The author declares to have no competing interests.

Consent for publication

Not applicable

Footnotes

1

Changing the mean input leads to higher-order corrections which can be calculated straightforwardly; see the Appendix.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Wald A. Sequential tests of statistical hypotheses. Ann Math Stat. 1945;16:117–186. doi: 10.1214/aoms/1177731118. [DOI] [Google Scholar]
  • 2.Ratcliff R. A theory memory retrieval. Psychol Rev. 1978;85:59–108. doi: 10.1037/0033-295X.85.2.59. [DOI] [Google Scholar]
  • 3.Ratcliff R, McKoon G. The diffusion decision model: theory and data for two-choice decision tasks. Neural Comput. 2008;20:873–922. doi: 10.1162/neco.2008.12-06-420. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Shadlen MN, Kiani R. Decision making as a window on cognition. Neuron. 2013;80:791–806. doi: 10.1016/j.neuron.2013.10.047. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Kira S, Yang T, Shadlen MN. A neural implementation of Wald’s sequential probability test. Neuron. 2015;84:861–873. doi: 10.1016/j.neuron.2015.01.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Roitman JD, Shadlen MN. Response of neurons in the lateral intraparietal area during a combined visual discrimination reaction time task. J Neurosci. 2002;22(21):9475–9489. doi: 10.1523/JNEUROSCI.22-21-09475.2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Latimer KW, Yates JL, Meister MLR, Huk AC, Pillow JW. Single-trial spike trains in parietal cortex reveal discrete steps during decision making. Science. 2015;349:184–187. doi: 10.1126/science.aaa4056. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Wang X-J. Pacemaker neurons for the theta rhythm and their synchronization in the septohippocampal reciprocal loop. J Neurophysiol. 2002;87:889–900. doi: 10.1152/jn.00135.2001. [DOI] [PubMed] [Google Scholar]
  • 9. Wong KF, Huk AC, Shadlen MN, Wang X-J. Neural circuit dynamics underlying accumulation of time-varying evidence during perceptual decision-making. Front Comput Neurosci. 2007:neuro.10.006.2007. [DOI] [PMC free article] [PubMed]
  • 10.Churchland AK, Kiani R, Shadlen MN. Decision-making with multiple alternatives. Nat Neurosci. 2008;11:693–702. doi: 10.1038/nn.2123. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Bogacz R, Brown E, Moehlis J, Holmes P, Cohen JD. The physics of optimal decision making: a formal analysis of models of performance in two-alternative forced-choice tasks. Psychol Rev. 2006;113(4):700–765. doi: 10.1037/0033-295X.113.4.700. [DOI] [PubMed] [Google Scholar]
  • 12.Niwa M, Ditterich J. Perceptual decisions between multiple directions of visual motion. J Neurosci. 2008;28:4435–4445. doi: 10.1523/JNEUROSCI.5564-07.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.McMillen T, Holmes P. The dynamics of choice among multiple alternatives. J Math Psychol. 2006;50:30–57. doi: 10.1016/j.jmp.2005.10.003. [DOI] [Google Scholar]
  • 14.Baum CW, Veeravalli VV. A sequential procedure for multihypothesis testing. IEEE Trans Inf Theory. 1994;40:1996–2007. doi: 10.1109/18.340472. [DOI] [Google Scholar]
  • 15.Dragalin VP, Tertakovsky AG, Veeravalli VV. Multihypothesis sequential probability ratio tests—part I: asymptotic optimality. IEEE Trans Inf Theory. 1999;45:2448–2461. doi: 10.1109/18.796383. [DOI] [Google Scholar]
  • 16.Bogacz R, Gurney K. The basal ganglia and cortex implement optimal decision making between alternative options. Neural Comput. 2007;19:442–477. doi: 10.1162/neco.2007.19.2.442. [DOI] [PubMed] [Google Scholar]
  • 17.Roxin A, Ledberg A. Neurobiological models of two-choice decision making can be reduced to a one-dimensional nonlinear diffusion equation. PLoS Comput Biol. 2008;4:1000046. doi: 10.1371/journal.pcbi.1000046. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Usher M, McClelland JL. The time course of perceptual choice: the leaky, competing accumulator model. Psychol Rev. 2001;108:550–592. doi: 10.1037/0033-295X.108.3.550. [DOI] [PubMed] [Google Scholar]
  • 19.Koulakov AA, Raghavachari S, Kepecs A, Lisman JE. Model for a robust neural integrator. Nat Neurosci. 2002;5:775–782. doi: 10.1038/nn893. [DOI] [PubMed] [Google Scholar]
  • 20.Goldman MS, Levine JH, Major G, Tank DW, Seung HS. Robust persistent neural activity in a model integrator with multiple hysteretic dendrites per neuron. Cereb Cortex. 2003;13:1185–1195. doi: 10.1093/cercor/bhg095. [DOI] [PubMed] [Google Scholar]
  • 21.Hick WE. On the rate of gain of information. Q J Exp Psychol. 1952;4:11–26. doi: 10.1080/17470215208416600. [DOI] [Google Scholar]
  • 22.Armitage P. Sequential analysis with more than two alternative hypotheses, and its relation to discriminant function analysis. J R Stat Soc, Ser B. 1950;12:137–144. [Google Scholar]
  • 23.Bogacz R, Usher M, Zhang J, McClelland JL. Extending a biologically inspired model of choice: multi-alternatives, nonlinearity and value-based multidimensional choice. Philos Trans R Soc Lond B, Biol Sci. 2007;362:1655–1670. doi: 10.1098/rstb.2007.2059. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Shadlen MN, Newsome WT. Neural basis of a perceptual decision in the parietal cortex (area lip) of the rhesus monkey. J Neurophysiol. 2001;86:1916–1936. doi: 10.1152/jn.2001.86.4.1916. [DOI] [PubMed] [Google Scholar]
  • 25.Wiggins S. Introduction to applied nonlinear dynamical systems and chaos. Berlin: Springer; 2003. [Google Scholar]
  • 26.Ueltzhoeffer K, Armbruster-Genç DJN, Flebach CJ. Stochastic dynamics underlying cognitive stability and flexibility. PLoS Comput Biol. 2015;11:1004331. doi: 10.1371/journal.pcbi.1004331. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Nguyen KP, Josić K, Kilpatrick ZP. Optimizing sequential decision in the drift–diffusion model. J Math Psychol. 2019;88:32–47. doi: 10.1016/j.jmp.2018.11.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Radillo AE, Veliz-Cuba A, Josić K, Kilpatrick ZP. Evidence accumulation and change rate inference in dynamic environments. Neural Comput. 2017;29:1561–1610. doi: 10.1162/NECO_a_00957. [DOI] [PubMed] [Google Scholar]
  • 29. Tajima S, Drugowitsch J, Patel N, Pouget A. Optimal policy for multi-alternative decisions. 2019;biorxiv:595843. [DOI] [PubMed]
  • 30.Ratcliff R, van Zandt T, McKoon G. Connectionist and diffusion models of reaction time. Psychol Rev. 1999;106:261–300. doi: 10.1037/0033-295X.106.2.261. [DOI] [PubMed] [Google Scholar]
  • 31.Huk AC, Shadlen MN. Neural activity in macaque parietal cortex reflects temporal integration of visual motion signals during perceptual decision making. J Neurosci. 2005;25(45):10420–10436. doi: 10.1523/JNEUROSCI.4684-04.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Data sharing not applicable to this article as no datasets were generated or analyzed during the current study.


Articles from Journal of Mathematical Neuroscience are provided here courtesy of Springer-Verlag

RESOURCES