Skip to main content
Frontiers in Neuroscience logoLink to Frontiers in Neuroscience
. 2012 Oct 26;6:146. doi: 10.3389/fnins.2012.00146

How Costs Influence Decision Values for Mixed Outcomes

Deborah Talmi 1,*, Alex Pine 2
PMCID: PMC3481112  PMID: 23112758

Abstract

The things that we hold dearest often require a sacrifice, as epitomized in the maxim “no pain, no gain.” But how is the subjective value of outcomes established when they consist of mixtures of costs and benefits? We describe theoretical models for the integration of costs and benefits into a single value, drawing on both the economic and the empirical literatures, with the goal of rendering them accessible to the neuroscience community. We propose two key assays that go beyond goodness of fit for deciding between the dominant additive model and four varieties of interactive models. First, how they model decisions between costs when reward is not on offer; and second, whether they predict changes in reward sensitivity when costs are added to outcomes, and in what direction. We provide a selective review of relevant neurobiological work from a computational perspective, focusing on those studies that illuminate the underlying valuation mechanisms. Cognitive neuroscience has great potential to decide which of the theoretical models is actually employed by our brains, but empirical work has yet to fully embrace this challenge. We hope that future research improves our understanding of how our brain decides whether mixed outcomes are worthwhile.

Keywords: cost-benefit analysis, decision-making, decision-making and neuroeconomics, economic models, reward, punishment, aversive decision-making


When faced with many possible courses of action humans and animals must evaluate their expected future costs and benefits in order to decide optimally. Every action is associated with a cost because every action requires, at minimum, some energy expenditure for execution. The things that we hold dearest often require a sacrifice, as epitomized in the maxim “no pain, no gain.” We struggle to be included in our peer group, study hard to increase our career prospects, work to provide for our families, pay to go on vacation, subject ourselves to painful health tests to maintain our physical well-being, and spend considerable energy on caring for our loved ones. Understanding how costs are integrated with benefits to ultimately reach a decision is therefore of paramount importance.

Value-based theories of decision-making suggest that people are thought to evaluate courses of action on the basis of their predictions about the future happiness a choice will engender (Von Neumann and Morgenstern, 1947; Vlaev et al., 2011). Because people are notoriously bad at predicting their future emotions (Hsee and Hastie, 2006) their decision utility is often different from their experienced utility at the time the consequences of their action come to fruition (Kahneman et al., 1997). Here we focus on decision utility, the time when agents decide between different prospects. Our question is how the subjective value of outcomes that are mixtures of costs and benefits is established.

We define costs and benefits as outcome attributes that decrease or increase, respectively, the decision value of that outcome at the time of decision-making. The costs and benefits most often studied in cognitive neuroscience include primary reinforcers such as food, drink, physical effort, and pain; secondary reinforcers such as monetary gains and losses; and mental events such as cognitive effort (Kool et al., 2010) and emotional suffering, such as the pain of regret (Bell, 1982). Although in some situations effort may be rewarding (Kivetz, 2003; Kim and Labroo, 2011), it is normally considered a cost (Hull, 1943).

While the definition of benefits is straightforward, our definition of costs may be controversial because it excludes some aversive outcome attributes. For instance, a decision may be risky because it entails a chance that a reward is not obtained, or it may prolong the time until reward is available for consumption. Yet we do not consider risk and delay to be true costs because they do not produce a negative subjective utility on their own, in the absence of other rewards or costs. Both risk and delay derive their meaning from the nature of the outcome and modulate its utility; their emotional valence depends on whether the outcome is rewarding or costly (see Loewenstein, 1987, for discussion on the valence of delay). As we will see, all available models of value integration make a similar distinction in that they treat risk and delay differently to “true” costs.

There are several theoretical valuation models for integrating costs and benefits. Cognitive neuroscience has great potential to decide which one is actually employed by the brain. Yet the burgeoning behavioral and neurobiological work on decisions between mixed outcomes often employs a single model of valuation, and model comparison work is rare. Our aim in this paper is to encourage empirical scientists to design behavioral and neurobiological experiments that can uncover the functional form of value integration in the brain.

In the first section we review the dominant model of value, which assumes an additive integration of costs and benefits. Alternative cost-benefit integration models draw substantially on our understanding of how risk and delay influence valuation. This influence is therefore reviewed briefly in the second section. The third section describes alternative models of cost-benefit integration, all interactive in nature, with the aim of rendering them accessible for the neuroscience community. These three sections concentrate on theoretical models but draw on some pertinent behavioral work. In the final two sections we provide a selective review of relevant neurobiological work from a computational perspective, focusing on those studies that illuminate the underlying valuation mechanisms.

The Additive Model of Value Integration

A dominant model for cost-benefit decision-making is expected utility theory (Von Neumann and Morgenstern, 1947). Here the subjective value (V) of a choice is computed as the sum of the probability (p) weighted utility (U) of each of its possible outcomes:

V=k=1npk×Umk (1)

In this equation m signifies the magnitude of rewards and costs associated with the outcomes the choice entails. The utility function in this theory is typically plotted for positive ms where it is concave, with diminishing sensitivity for larger ms. Costs are represented by negative Us. This is an additive model of valuation because the disutilities of costs are summed with the utilities of beneficial outcomes. A positive V favors a decision to act, and if there is more than one option under consideration the action with the greatest expected utility is chosen.

The additive model of valuation is dominant, but there are a number of interesting alternatives (see Table 1; Figure 1). Crucially, there are situations in which the additive model may not be valid. Multi-attribute utility theory (Keeney and Raiffa, 1993), for example, allows additive integration only under the assumption of “additive independence” (Thurston, 2006). Consider a lottery where an agent may obtain one of two outcomes with a 50% probability. Both outcomes are a mixture of two attributes, x and y, each with two values – for example, a large sandwich for $4 or a smaller sandwich for $2. Under additive independence an agent who is indifferent between the two outcomes of Lottery A, [x2,y1] and [x1,y2] would also be indifferent between the two outcomes of Lottery B where the same attributes and values are recombined [x1,y1] and [x2,y2]. Clearly, an agent who is indifferent between the possible outcomes of lottery A – the high-reward/high-cost outcome and the low-reward/low-cost outcome – is unlikely to be indifferent between the two outcomes of lottery B where the high-reward/low-cost outcome (a large sandwich for $2) clearly dominates the low-reward/high-cost outcome (a small sandwich for $4). In this scenario reward and cost are not additively independent, suggesting that they should not always be combined according to the additive model.

Table 1.

Models of value integration.

Additive models Interactive models
Expected utility theory: V = pr × U(mr) − pc × U(mc) Trade-off model: V = p × U(mr)/U(mc)
Prospect theory: V = P(pr) × U(mr) − P(pc) × U(mc) Hyperbolic discounting: V = U(mr)/[1 + k × U(mc)]
Discounted utility theory: V = P(pr) × D(dr) × U(mr)− P(pc) × D(dc) × U(mc) Right-shift of the utility curve: V = P(p) × U(mr − mc)
Bilinear model: V = x1 × U(mr) − x2 × U(mc)− x3 × U(mc) × U(mr)

The subjective value of a single mixed outcome with probability p, one rewarding attribute mr and one costly attribute mc (coded positively so that greater mc represents greater cost), delivered after a time delay d. U, P, and D are functions that transform externally given quantities m, p, and d into internal representations. x1, x2, and x3 are constants.

Figure 1.

Figure 1

Additive and. interactive models of value integration. Value was computed according to the bilinear model with x1 = 1 and x2 = 1. Additive models are depicted on the left, with x3 = 0. Interactive models are depicted on the right, with x3 = 0.06. Top panels employed a linear utility function for rewards and costs (those that were employed by Talmi et al., 2009); bottom panels employed a power function with exponent 0.8. Additive models with non-linear utility (bottom left) represent the predictions of expected utility theory with p = 1. Figure adapted from Park et al. (2011).

Before we describe alternative models of cost-benefit integration we discuss in a little more detail how risk and delay are thought to modulate the value of an outcome with either a rewarding or a costly attribute.

Modulation of Costs and Benefits by Risk and Delay

Consider a patient who must evaluate a treatment option. The improvement in health this treatment brings and the painful procedure it involves must both be weighed against the chance that it is not efficacious and will only yield benefits after a long recovery period. In this section we discuss how risk and delay influence the subjective value of a reward or a cost.

Modeling risk

Expected utility theory, a prescriptive model for decision-making under risk, fits empirical data less well than the descriptive Prospect theory (Kahneman and Tversky, 1979; Tversky and Kahneman, 1992). In Prospect theory utilities are again computed as a product of two functions, one transforming gains and losses (the utility function) and the other transforming given probabilities (the probability weighing function). The utility function is concave in the gain domain and convex in the loss domain with a steeper slope in the latter, so that the disutility of losses is greater than the utility of gains, allowing prospect theory to account for loss aversion. Prospect theory also proposes a non-fixed reference point for changes in utility, rather than a fixed point representing final wealth states; this feature is important for our discussion in Section “Modeling costs as right-shifts of the utility function”. These transformations allow the theory to account for a number of expected utility violations, such as the reflection and framing effects and biases in the perception of small and large outcome probabilities (Kahneman and Tversky, 1979, 2000; Tversky and Kahneman, 1981). Both expected utility and Prospect theory entail a multiplicative integration of utility with its probability, such that utility is weighted (or discounted) in accordance with its decreasing likelihood of occurrence.

The form of probability discounting proposed by expected utility theory does not account adequately for an array of anomalies in decision-making under uncertainty, but prospect theory can account for most of those. The mathematical form of the probability weighting and utility functions in prospect theory are not formally specified beyond their qualitative predictions, but more precise formulations derived from a body of animal and human probability discounting experiments indicate that probabilistic rewards are discounted in a hyperbolic or quasi-hyperbolic manner as their likelihood diminishes (Green and Myerson, 2004; Green et al., 2004, 2011). These experiments typically employ psychophysical, “adjusting amount” procedures (Rachlin et al., 1991). In a standard procedure, participants are required to choose between a smaller-certain reward and a larger-probabilistic reward. In each trial the amount of the smaller reward is adjusted until the participant is indifferent between the two options. Under the assumption that indifference entails equality of subjective value the subjective value of the risky option can be quantified in terms of the certain (risk-free) option. The probability of the larger reward is then altered such that the probability discount function can be estimated from a number of indifference points across the probability spectrum. Results of these procedures consistently show that hyperbolic functions provide a superior fit to these indifference points, in contrast with the predictions of expected utility theory (Figure 2). To take this into account p in Eq. 1 can be replaced by P(p) where

Figure 2.

Figure 2

Hyperbolic discounting of risk. In this basic hyperbolic model individuals steeply discount reward value with initial decreases in probability of occurrence and only gradually as they get more unlikely. h is a risk aversion parameter. In the case of h = 1 discounting conforms to EU theory, i.e., an equal decrease in value for every percent decline in probability. h < 1 equates to risk aversion and when greater than one to risk seeking. Value here is represented as a proportion of its initial (certain) value or alternatively as the discount factor.

P(p)=11+h×Θ(1.1)

With

Θ=1-pp(1.2)

Θ is termed the “odds ratio” and is computed as the probability of non-occurrence divided by the probability of occurrence. An odds ratio of 1 therefore corresponds to outcomes that occur 50% of the time. h is a discount rate parameter which determines the rate of probability discounting. If h = 1 the individual is risk neutral and values the reward in accordance with EU theory, so P(p) = p. When h < 1 the individual is described as risk averse [V < (p × M)] and when h > 1 as risk seeking (Figure 2). P(p) can be considered as a discount factor between zero and one by which the reward is discounted in accordance with its “odds against.”

The main feature of a hyperbolic discount function is that the reward loses a gradually smaller proportion of its value per increasing unit in odds against – so it loses a larger proportion of its value when the probability changes from 90 to 80% than when it changes from 60 to 50%. This can explain why a person who chooses a smaller but more certain reward over a larger but more risky option can switch their preferences when the probability of both options is reduced by a constant – similar to the Allais or certainty paradox (Allais, 1953; Kahneman and Tversky, 1979). The smaller (below 1) is h, the greater is the steepness of the initial devaluation relative to the later devaluation. Note that this formulation is consistent with the predictions of prospect theory, for example, the overweighting of events with small probabilities and the underweighting events with large probabilities (Figure 2). A number of similar but more complex functions have been proposed that attempt to capture other features of subjective valuation by adding extra parameters (Green and Myerson, 2004; Myerson et al., 2011).

Modeling delay

While it is difficult to extend EU and prospect theory to account for situations where people choose between rewards or punishments that are available at different points in time (termed intertemporal choice), discounted utility theory models this situation explicitly. The subjective value of a single, certain outcome m that is delayed by d can be expressed as

V=D(d)×U(m) (2)

With U(m) representing the instantaneous utility of m and D(d) representing a discount factor ranging from zero to one by which U(m) is discounted in accordance with its objective delay.

In fact, the probability discounting approach outlined above derives from an older and richer literature on temporal discounting (Loewenstein and Elster, 1992; Frederick et al., 2002; Green and Myerson, 2004), and relates to a debate as to which is the primary discounting mechanism – probability (because delay entails uncertainty) or delay (because the resolution of uncertainty takes time, Prelec and Loewenstein, 1991; Rachlin et al., 1991). In discounted utility theory the utility of a reward is discounted exponentially as a function of its delay, namely with a constant percentage decrease in value per unit time (Samuelson, 1937; Koopmans, 1960).

D(d)=e(-k*d)(2.1)

k is a free parameter which represents the individual’s discount rate. Thus k quantifies an individual’s tendency to discount future costs and benefits. An individual with a high k value devalues future costs and benefits more steeply than a lower k individual, i.e. with a greater percentage decrease in value per unit time. k is thought to relate to impulsivity in the same manner as h relates to an individual’s risk profile (Ainslie, 1975, 2001) because individuals with a large k are more likely to choose the smaller-sooner over larger-later option.

Although people do discount exponentially in some situations (Schweighofer et al., 2006), there is a wealth of empirical evidence against exponential discounting, primarily in the robust finding that the discount rate is not constant but decreases with time. In a simple demonstration (Thaler, 1981) asked subjects to specify the amount of money they would require in 1 month, 1 year or 10 years to make them indifferent between that option and receiving $15 now. Their median responses ($20, $50, $100) implied an average annual discount rate of 19% over a 10 year horizon, 120% over a 1 year horizon and 345% over a 1 month horizon. Similar observations have been made for in non-monetary domains such as health and credit markets (Redelmeier and Heller, 1993; Chapman and Elstein, 1995; Chapman, 1996, 2001; Pender, 1996). A noted manifestation of this feature is that humans and animals are prone to preference reversals when a constant delay is added to both options of an intertemporal choice (Prelec and Loewenstein, 1991; Loewenstein and Elster, 1992). For example, people who prefer $10 today over $11 tomorrow often also prefer $11 in 31 days to $10 in 30 days (Green et al., 1994). As we have seen, the same reversals also characterize choices between certain and probabilistic outcomes.

When mathematical functions are fit to intertemporal choice data (for example indifference points between smaller-sooner and larger-later options) a multitude of studies have demonstrated that hyperbolic or quasi-hyperbolic discount functions provide a superior fit compared to exponential functions in both humans and animals, for delayed monetary, health-related, and other forms of reward, and punishment (reviewed in Rachlin et al., 1991; Ho et al., 1999; Frederick et al., 2002; Green and Myerson, 2004, but see Kable and Glimcher, 2007, for a different model). The standard and most widely used functional form for hyperbolic discounting in the behavioral literature was proposed by Mazur (1987) and based on earlier work by Ainslie and Herrnstein (1981), Ainslie (1975), Herrnstein (1981). According to this work,

D(d)=11+k×d(2.2)

so that

V=m1+k×d(2.3)

If we taker U(m) to be a better representation of the instantaneous value of M (Pine et al., 2009) then

V=U(m)1+k×d(2.4)

As with probability discounting, other functional forms which capture decreasing rates of discounting and the non-linearity of the relationship between objective and subjective delay have also been proposed (Phelps and Pollak, 1968; Loewenstein and Prelec, 1991; Frederick et al., 2002; Green and Myerson, 2004; Myerson et al., 2011).

Expected utility, prospect theory and discounted utility theory entail an attenuation of reward sensitivity with risk or delay. This means that when reward is risky or delayed, the utility gained by increasing it by a constant reduces, so that U(m + 1) − U(m) is greater than U(m + 1) × D(d) − U(m) × D(d) or U(m + 1) × P(p) − U(m) × P(p). Figure 3B illustrates this point. By combining these forms of discounting, the subjective value of either costs or benefits can also therefore be represented as a product of utility with two discount factors, one based on probability and the other on delay (Prelec and Loewenstein, 1991; Rachlin and Raineri, 1992; Ho et al., 1999):

Figure 3.

Figure 3

Hyperbolic discounting of delay. This function describes theoretical data from an experiment in which two groups of animals are given a choice between two arms of a maze, one of which contains a larger-later reward, four food pellets that necessitate a wait of 15 s, and one which contained a smaller-sooner reward, two food pellets that were available after 3 s. (A) Value as a function of delay for a single reward magnitude, two food pellets, computed according to Eq. 2.3. The discount function is depicted for two values of k with higher k indicating steeper, (more impulsive) discounting. (B) Value as a function of reward magnitude for the two different and delays, 3 and 15 s, computed using the same equation. The circles show the two options presented to the two groups of animals. In this example the value of the larger-later reward is greater for the less impulsive group, and the value of the smaller-sooner reward is greater for the more impulsive group. (C) Value as a function of both delay and reward magnitude.

V=k=1nPpk×Ddk×Umk(2.5)

Alternative models for the effects of risk and delay

A key challenge for the models we presented for the integration of reward with probability and delay (Eq. 2.5) concerns the effect of reward magnitude on valuation. The “magnitude effect” refers to a prevalent finding in intertemporal choice that small magnitudes are discounted more steeply than large ones. A person who is indifferent between $60 today and $120 in 1 year is thus more likely to choose $1200 in 1 year to $600 today. The magnitude effect has been documented in numerous studies involving both real and hypothetical rewards (reviewed in Frederick et al., 2002; Green and Myerson, 2004) For instance, Thaler (1981) asked his participants to decide between a given immediate monetary reward and a delayed monetary reward they would receive in a year’s time. Participants were required to declare how much money they would want in a year for them to be indifferent between the immediate and the delayed options. He found that the immediate amounts of $4000, $350, and $60 were discounted by 29, 34, and 39%, respectively. Although the magnitude effect has also been documented in non-monetary reward domains such as medical treatments, drugs, job choices, vacations, and restaurant tips (Raineri and Rachlin, 1993; Chapman and Elstein, 1995; Chapman, 1996; Chapman and Winquist, 1998; Baker et al., 2003; Schoenfelder and Hantula, 2003) it has not been observed in species other than humans, for example in rats and pigeons and primates using food rewards (Richards et al., 1997; Grace, 1999; Green and Myerson, 2004; Freeman et al., 2012, but see Grace et al., 2012). In humans the magnitude effect has not been reliably observed in the loss domain (Estle et al., 2006) and in some (but not all) studies seems to level off when the magnitudes involved are fairly large (Shelley, 1993; Green et al., 1997).

Although less explored, there is evidence that reward magnitude has the opposite effect on probability discounting compared to temporal discounting. The “peanuts effect” describes the finding that larger magnitude rewards are discounted more steeply than smaller rewards, implying that people tend to be less risk averse when they are “playing for peanuts” (Weber and Chapman, 2005; Chapman and Weber, 2006). For example, an individual may prefer a 10% probability of obtaining $100 over a certain $10, but may also prefer a certain $100 to a 10% probability of winning $1000. The reward magnitude effect humans display when they evaluate risky and delay outcomes poses a challenge for the multiplicative P × D × U approach in Eq. 2.5, since it suggests that D and P themselves depend on U. The double dissociation between the effect of reward magnitude on delay and probability exacerbate the challenge because it suggests that magnitude effects cannot simply be explained by a property of the discount function U. We briefly review two approaches to this challenge and in Section “Using Neurobiological Evidence to Decide Between Models of Risk and Delay” we discuss how neurobiological data can help decide between them.

Green and Myerson (2004), Myerson et al. (2011) posit that magnitude scales the temporal discount rate parameter k in Eq. 2.3 such that k decreases with increasing magnitude. By contrast, in probability discounting magnitude scales an exponent of the denominator of the discount function in Eq. 1.1. Because this renders the two discount functions D and P partially a function of m, the subjective value V can no longer be thought of as the product of a multiplication of separate utility and discount functions U, D, and P.

Prelec and Loewenstein (1991) offer a general scheme with which to view multi-attribute choice. Their model, an extension of prospect theory, similarly relies on the decomposition of valuation into separate discount and utility functions, and explains magnitude effects in delay discounting in terms of the utility function. They suggest that agents represent or “frame” each attribute of a multi-attribute outcome as a pair of values. The first value is the absolute magnitude of the attribute and the second is what they call the “polarity,” namely, whether it is beneficial or detrimental. For example, $50 in 2 weeks is encoded as (50, 2) with the polarities (+, −). The importance of attributes relative to each other can be altered by certain linear transformations. One such linear transformation is adding a constant to the magnitude of all values of an attribute. The consequence of this transformation is “decreasing absolute sensitivity,” a decrease in the importance of that attribute relative to others. A second transformation involves multiplying all values of an attribute by a constant. The consequence of this transformation is “increasing proportional sensitivity,” an increase in the importance of that attribute. The magnitude effect in intertemporal choice follows from increasing proportional sensitivity because multiplying the monetary attribute increases its importance relative to the delay attribute, leading to the appearance that larger magnitudes are discounted less. These features of multi-attribute framing can explain many of the anomalies common to decision-making under uncertainty and intertemporal choice. Yet because increased proportional sensitivity will always increases the importance of the monetary attribute this effect cannot explain the opposite effects of reward magnitude on delay and probability discounting.

To account for the peanuts effect Prelec and Loewenstein (1991) invoke “interaction effects.” These are emotional processes that can influence cost-benefit decisions by changing the importance of the attributes of an outcome through valuation processes unrelated to utility and discount functions (Berns et al., 2007). Disappointment, one of many interaction effects, accounts for the magnitude effect in probability discounting. The notion here is that anticipation of disappointment – should a losing outcome occur – increases the greater the potential gain, and that increasing disappointment decreases the importance of the money relative to the probability attribute in a manner that accounts for preference reversals (Prelec and Loewenstein, 1991; Weber and Chapman, 2005; Chapman and Weber, 2006). Disappointment does not enter into delay discounting since there are no probabilistic outcomes. Interaction effects are useful, however, when we consider other interesting phenomena in intertemporal choice. The interaction effects of anticipation and dread are invoked to explain why in some cases people prefer to speed up punishments to “get them over with,” and savor rewards by delaying them, phenomena which are incompatible with standard discounted utility theory (Loewenstein, 1987; Berns et al., 2006).

In summary, the discounted utility theory notion of a decision value-based on the multiplication of separate utility and discount functions has been challenged in light of opposing magnitude effects. In one view discount functions accept magnitude as an argument, with no requirement for a separate utility function. Although two separate mechanisms are required to account for opposing magnitude effects this is perhaps a more parsimonious account, but it does not explain a host of other influences on valuation that are captured by interaction effects. In another view additional mechanisms are invoked with magnitude solely acting on the utility function, and delay and risk are treated within separate weighting functions.

Interactive Models of Value Integration

While the functional form of decision-making that involves risk and delay costs is well described, and a rich empirical literature delineates the neurobiology of effort-based decision-making (Salamone et al., 2007; Floresco et al., 2008a; Kurniawan et al., 2011), less research has been devoted to uncovering the functional form of valuation when effort and other costs are mixed with reward. In this section we review non-additive models for decision values when outcomes include costs. Because the empirical evidence for these models is more limited than that of the additive model we describe them within the context of experiments that corroborate them. Our aim is to expose these models for scrutiny by the neuroscience community and encourage further model comparison work to decide between them. Table 1 lists all the valuation models discussed in this paper.

A strong alternative to the additive model is the trade-off model (Simonson, 1989), where decision values are expressed as the ratio of costs and benefits. The subjective value of a single mixed outcome with probability p, one rewarding attribute mr and one costly attribute mC, could be expressed as:

V=p×UmrUmC (3)

Let us take two examples to illustrate how this model has been employed. Soman (2004) used the trade-off model to account for the results of an experiment where participants made hypothetical decisions between differently priced products that required more and less effort, such as an expensive desk that was already assembled and a cheaper desk that required customer assembly. Soman did not utilize model comparison but the trade-off model fitted his data well and accounted for the influence of delay on choosing between these mixed outcomes. Another example comes from work on foraging, where both the additive and the trade-off models are prevalent (Stephens and Krebs, 1986). Bautista et al. (2001) modeled cost-benefit decisions of starlings deciding how to forage for food: to walk (low-reward/low effort) or fly (high-reward/high effort). In that set-up rewards consisted of the energy gain from the food, and costs consisted of the energy loss associated with the chosen travel method. The authors compared the additive model and the trade-off models, which they termed net rate and efficiency, respectively, and found that the additive model accounted best for the starlings’ decision.

One key difference between the additive and trade-off models is how they treat situations that do not offer any reward but impose a cost. The additive model allows subjective values to become negative in these situations, while the trade-off model does not. The importance of this feature depends on the situation. While non-rewarding decisions – between two costs, or a cost and the status quo – are rare in economic studies, they are common outside the laboratory. Berns et al. (2008) used an example of a patient deciding between risky treatment options, where the best outcome is maintaining current levels of health, and contended that it is difficult to explore such decisions in typical financial decision-making experiments because people are unlikely to take part in a study where they might end up losing out financially. The difficulty with financial costs is partially ameliorated in experiments that require participants to exert effort or suffer experimentally induced pain, although here too there is an implicit, unmodelled reward that draws participants to take part in the experiment in the first place. Clearly, though, according to the trade-off model the patient’s decision or its laboratory equivalents, do not have a negative decision utility. The next models of valuation that we review differ in their approach to decisions between costs. The first does not model negative utilities in the absence of reward, the second predicts negative utilities, and the third allows for either zero or negative utilities. Therefore, in decisions between “bad” and “worse” an empirical demonstration of negative decision utilities will constrain model selection.

Another key difference between the additive and the trade-off models is whether reward sensitivity changes with costs. This is an important and under-appreciated difference between models. The additive model predicts that costs do not alter reward sensitivity, while the trade-off model predicts that they do. The models we review below differ in this respect too. The first predicts decreased reward sensitivity with cost, the second predicts increased sensitivity, and the third allows changes in either direction. Measuring changes in reward sensitivity is therefore another assay for model selection. Taken together, three aspects of the decision – whether it adheres to the additive independence assumption, the presence or absence of negative decision utilities, and whether reward sensitivity changes with cost – distinguishes integration across costs and benefit from valuation of all-rewarding multi-attribute outcomes.

Hyperbolic discounting of rewards associated with costs

Brañas-Garza et al. (2012) report a result that illuminates the social importance of understanding decision values of mixed outcomes in the field of health. When people consider a painful medical procedure their decision values should integrate over the pain costs of the procedure as well as the value of consequent future health benefits. How steeply one discounts the future will therefore impinge on the integrated value of the procedure. They found that the more impatient participants, those that discounted the future more steeply in an intertemporal choice task, reported more frequently that they experience a negative feeling as soon as they decide to undergo the procedure. This feeling may derive from the disutility of the decision value, and bias these participants against some health-promoting behaviors such as necessary painful medical procedures.

The success of the hyperbolic discount functions in accounting for the effect of delay and risk costs on choice makes it rather tempting to consider whether other costs devalue reward hyperbolically. The subjective value of a single, certain, mixed outcome with one rewarding attribute mr and one costly attribute mc, could be expressed as:

V=Umr1+k×UmC (4)

Prevost et al. (2010) used hyperbolic discounting (Eq. 2.4) to model how participants decided between cost-benefit mixtures. Participants first viewed fuzzy erotic images and then decided between two options: either viewing a clear image of the same content for a low-cost, or viewing it for a longer duration for a higher cost. The low-cost involved a short wait or exerting minimal physical effort, while the high-cost involved waiting longer or exerting more effort. The hyperbolic discount function described choices equally well for both delay and effort cost, and fared better in doing so than the exponential discount function.

We have seen that hyperbolic models are not ideal for situations that require a cost without providing a reward, because they do not allow negative decision utility when only costs are on offer. However, all trials in the paradigm used by Prevost et al. (2010) included at least a small reward, possibly contributing to the fit of this model for their data.

Clearly, if effort also modulates U(mR) hyperbolically, reward sensitivity will decrease under effort. But a hyperbolic interaction of effort with reward may not be detected in studies that only consider a linear form of interaction. Kurniawan et al. (2010), for example, obtained a different result from Prevost et al. (2010) in a task that similarly required participants to choose between a low-reward, low effort option and a large reward, large-effort option. Both effort and reward had the expected effect on decisions, and a similar effect on ratings of choice likability, but the interaction between reward and effort was not significant for either measurement. The null interaction effect appears to go against interactive models, but as Kurniawan and colleagues used the general linear model in their analysis it is possible that they could not detect a non-linear interaction between effort and reward.

Prelec and Loewenstein (1991) speculated that their rules for transformation of attribute weighting should apply to all multi-attribute choices. Consequently, if effort and pain discount reward hyperbolically it would be natural to predict that effort and pain discounting will resemble risk and delay discounting, and generate preference reversals when a constant is added to both options. In decisions between outcomes that mix reward and pain, for example, adding a constant amount of pain to both options should shift preference in favor of the high-reward high-pain option. This is an example of how consideration of underlying models yields useful hypotheses for experimentation.

Modeling costs as right-shifts of the utility function

In prospect theory, choice options are coded as gains and losses relative to the “status quo,” a neutral point of reference that is assigned a value of zero on the reward magnitude axis and is where the utility function crosses that axis (Kahneman and Tversky, 1979). Utility is zero when people do not expend any effort, and therefore do not expect to be rewarded. Kivetz (2003) argued that effort requirements create an expectation for reward, which should be modeled as a right shift of the reference point. For example, when Thea is asked to mow the lawn she forms an expectation for reward. Agreeing to do this chore has a negative decision utility and is experienced as a loss relative to her revised reference point. If she is promised $5 for completing this chore and considers this sum fair this reward merely brings her decision value back to zero.

Prospect theory can be extended to account for effort costs by shifting utility function to the right under effort (Kivetz, 2003). People who expend effort U(mc0) expect a fair reward, U(mr0) in return. According to this formulation people should be indifferent between no reward/no effort and U(mro)/U(mc0). Kivetz (2003) was interested in frequency programs such as frequent-flyer miles, a marketing tool that requires customers to invest effort for future rewards. He showed that replacing U(m) in Eq. 1 by U(mrmc) provided an adequate account for customers’ choices and for the influence of risk on their decisions. The subjective value of a single mixed outcome with probability p, one rewarding attribute mr and one costly attribute mc, could be expressed as:

V=P(p)×Umr-mC (5)

Because of the concavity of U, right-shifting it under effort means that U now increases more steeply with mR. Consequently, this model implies that effort increases reward sensitivity. For example, Thea may be just a little more delighted with a gift of $10 than $5, but after she is asked to mow the lawn her increased happiness with $10 relative to $5 is greater.

Kivetz (2003) argued that his model can be extended to all costs that people perceive as having an inherent disutility, and mentions delay and pain costs. Beyond the conceptual problem of considering the passage of time as inherently negative we would argue that the empirical evidence base for decreased reward sensitivity with delay (Figure 3) means that Kivetz’ model, which predicts increased sensitivity under cost, is unlikely to account well for intertemporal choice.

A bilinear model for cost-benefit analysis

Phillips et al. (2007) based their proposed valuation model on their review of animal research concerning the role of dopamine in cost-benefit analysis. Their model was intended to be applicable for delay, effort, risk, and other aversive outcomes. They did not provide the functional form of the value function, perhaps because there is limited evidence for two of the central components of their model, namely, exactly how dopamine levels and the indifference functions vary with reward magnitude. Noting these reservations, and making some assumptions of our own, we derived their value function (Appendix). The function we derived in this way is somewhat unwieldy. However, for small rewards, within the linear portion of the utility function, their model can be expressed more simply as:

V=x1×Umr+x2×UmC+x3×Umr×UmC (6)

With positive constants x1 and x2, and a constant x3 that can be either positive or negative. Thus, although the value function proposed by Phillips et al. (2007) may appear to model value additively (Botvinick et al., 2009), a closer look shows that it includes an interaction between reward and cost, albeit of a different form than that in Eqs 35. Figure 1 depicts this model and compares it to the additive model both for a linear and for a non-linear utility function.

Two aspect of the bilinear model are important for our discussion. First, in contrast with the other models discussed here this model allows reward sensitivity to either increase or decrease when outcomes include costs. The direction of change depends crucially on the functional forms of the reward utility function and the indifference function (see Appendix). Second, in contrast to the models in Eqs 3 and 4 this model allows utility to become negative when the choice options do not offer any reward.

Although Phillips et al. (2007) offered a very specific functional form to describe the interaction of rewards, costs, and value, they did not describe direct empirical evidence for that particular form. Two studies that examined decisions involving pain costs observed that the bilinear interaction model fitted their data well. In the first study (Talmi et al., 2009) participants chose between two options, one that maximized and one that minimized the chances for the delivery of a mixed outcome. That outcome included a monetary gain or a loss as well as an electrical stimulation of the skin that could be either painful or mild. Participants experienced the pain a few seconds after they made their choice, at which time they were also informed that the promised amount of money was added to their account. When Talmi and colleagues compared the additive model with the bilinear interaction model they found that the addition of an interactive term significantly improved the model fit, with the interaction parameter x3 suggesting that physical pain attenuated the sensitivity of participants to monetary reward. This conclusion was corroborated by another study (Park et al., 2011) where participants were asked to accept or reject mixed outcomes that involved varying monetary reward and one of five pain magnitudes. The authors replicated Talmi et al.’s (2009) finding that the bilinear model accounted for, behavioral choice better than the additive model. Notably, participants in both studies likely experienced disutility in some of the experimental trials, because Talmi et al. paired pain with either gains, losses, or zero rewards, and Park et al. used a very low amount of 1 cent in some of the trials, paired with both low and high levels of pain. This aspect of their paradigm may explain the importance of the parameter x2 in their data.

While Talmi et al. argued that because the monetary rewards were very small, under $0.60, the utility of reward and the disutility of pain should be modeled linearly (Rabin and Thaler, 2001), Park et al. (2011) tested this hypothesis formally. Although they also employed small monetary rewards, up to €0.99, they found that modeling reward utility using a power function explained their data better than when a linear function was used. When this change was implemented, behavioral choice data no longer favored the bilinear over the additive model. Their fMRI results, however, fit the bilinear model better than the additive model regardless of whether reward utility was modeled with a linear or a power function. Figure 1 (adapted from their Figure 2) compares four models: two interactive models, computed either with a linear or a non-linear utility function, and their additive model counterparts, which are identical but omit the interaction term. In fact, a detailed analysis of the bilinear model suggests that the importance of the interaction term depends on the relationship between the utility functions for rewards and costs, and is likely not to be important when they are identical. In order to decide, for a particular situation, whether an interaction term is present it is sufficient to determine the form of three functions: the utility function of rewards without costs; the utility function of costs without rewards; and the indifference function – the relationship between rewards and costs. In future work we plan to investigate the conditions for an interaction in more detail.

In summary, while the additive model is dominant in the economic literature, there are several alternative models that feature an interaction between costs and benefits such that costs alter sensitivity to reward. According to four of these models, described in this section, costs modulate the subjective utility of mixed outcomes. Each of these models proposes a different functional form to describe this interaction. Further empirical work can help determine which model best captures the decision values of mixed outcomes in animals and humans. In Section “Using Neurobiological Evidence to Decide between Additive and Interactive Valuation Models” we explore how neurobiological data can assist in this endeavor, but first we explore how such data can shed light on the modulation of costs and benefits by risk and delay.

Using Neurobiological Evidence to Decide between Models of Risk and Delay

A large amount of empirical work in animals has been dedicated to uncovering the neural structures that mediate decision-making when mixed outcomes involve both costs and benefits. There is strong evidence in human neuroimaging for an abstract representation of subjective utility in the ventromedial prefrontal cortex (vmPFC) across many different kinds of commodities (Padoa-Schioppa, 2011; Levy and Glimcher, 2012), but controversy on where this abstract representation is expressed in animals (Padoa-Schioppa, 2011; Roesch and Bryden, 2011). The regions involved in the effects of delay, risk, and effort on decision-making have been described, with more limited investigations of other costs (Phillips et al., 2007; Salamone et al., 2007; Floresco et al., 2008a; Roesch and Bryden, 2011). Much of this literature, however, does not speak directly to the issues we focus on here, the functional form of integrative valuation of mixed outcomes. In this section we examine how empirical neurobiological data could help decide between the models for the modulation of costs or benefits, separately, by risk and delay. In the final section we discuss data relevant for models of cost-benefit integration.

Evidence for separate representations of D and U

Animal and human neuroimaging studies have identified a relatively large set of regions that are involved in intertemporal decision-making (for reviews in animal studies see Cardinal et al., 2004; Winstanley et al., 2006; Floresco et al., 2008a,b; and in humans Tanaka et al., 2004, 2007; McClure et al., 2004, 2007; Kable and Glimcher, 2007; Gregorios-Pippas et al., 2009; Luhmann et al., 2008; Ballard and Knutson, 2009; Wittmann et al., 2007; Prevost et al., 2010; Hariri et al., 2006; Pine et al., 2009, 2010. They include ventromedial and medial prefrontal cortex, dorsal, and ventral striatum (VS), posterior cingulate cortex, and insula, as well as dorsolateral PFC, amygdala, and lateral OFC. There is no agreement on the particular contribution of each region to intertemporal choice and value construction. Recent literature has started to address regional specificity by correlating behaviorally derived model parameters with BOLD responses or single cell electrophysiological recordings. We will demonstrate how this approach has led to an increasingly sophisticated view of functional specificity and model implementation in the brain in animal electrophysiological recording and human neuroimaging studies.

McClure et al. (2004) performed the first neuroimaging study of intertemporal choice to provide a neurobiological account of temporal discounting and preference reversals. The disproportionate valuation of rewards available in the immediate future, and other evidence, led them to postulate the differential activation of distinguishable neural systems – specifically, that impatience is driven by the limbic system which responds to immediate rewards and is less sensitive to the value of future rewards, whereas patience is mediated by the lateral PFC which is able to evaluate trade-offs between more abstract rewards, including those in the more distant future. This proposed struggle between an affective and a deliberative decision-making system was based theoretically on a quasi-hyperbolic time discounting function which splices together two different discounting functions, one exponential and another which distinguishes sharply between present and future rewards, modeled by a parameter termed beta. Beta represents the special value placed on immediate rewards relative to those received at any other time. The hypothesis then was that activity in lateral PFC areas should correspond with the rational, deliberative processes, and limbic activity should represent the beta parameter. To test this hypothesis, they scanned the brains of subjects as they made a series of different hypothetical intertemporal choices. Critically, they split the trials into two types – those where both rewards were delayed in the future, and those where the small reward could be received immediately following the experiment.

When they compared these two conditions in their analysis they found that whereas lateral PFC (dorsal and ventral) and intraparietal regions were similarly active across all trial types, limbic structures including the VS (NAc), mPFC, posterior cingulate, and medial OFC (regions they defined as beta regions) were preferentially activated in response to choices where there was an option for immediate reward. Furthermore, when they analyzed all the choices where there was an immediate component, they could predict the choice outcome – a greater activation of limbic areas led to choice of the immediate small reward, whereas choice of the delayed reward followed a greater activation of the lateral PFC areas relative to the limbic ones. However, the hypothesis that the limbic system mediates impulsivity by its preference for immediate rewards is difficult to reconcile with animal work indicating that the integrity of the NAc and OFC is crucial for self-control and the ability to choose delayed rewards (Cardinal et al., 2004). In McClure et al.’s (2004) account, NAc or OFC lesions should, in theory, promote delayed choice as long as the DLPFC is left intact.

Kable and Glimcher (2007) have argued that McClure et al.’s (2004) study provided insufficient evidence for a dual valuation system since they did not demonstrate activity corresponding to different discount rates in the limbic and lateral PFC regions, and critically, that the discount rate in the beta regions was greater than the observed behavioral discount rate of the subjects. Without such evidence the results of their analysis could simply be explained by proposing that limbic regions track the subjective value of rewards at all delays and this preferential activity simply reflects the fact that sooner rewards are more valuable than later rewards. Thus, to explicitly determine the neural correlates of subjective value in intertemporal choice Kable and Glimcher employed a model-based approach in their analyses. They scanned participants while they were deciding between a constant smaller-sooner option and a variable larger-later option. The crux of their analysis was regressing the BOLD response against the hyperbolically discounted values of the larger-later option, derived from the choices each subject made, by estimating individuals’ discount rate parameter (k) according to Mazur’s (1987) hyperbolic discounting function (Eq. 2.3). These regressors identified a network of three regions which correlated with the subjective discounted value of the delayed reward – the VS, medial prefrontal cortex, and posterior cingulate cortex. This network did not exclusively value immediate rewards, as hypothesized by McClure et al. (2004) but tracked the subjective value of rewards at all delays, leading Kable and Glimcher to conclude that there is a single valuation system for delayed rewards.

Kable and Glimcher (2007) along with subsequent studies (Peters and Buchel, 2010; Prevost et al., 2010) successfully established a neural correlate of the subjective value of delayed rewards under the assumption of a single valuation process, but did not attempt to tease apart the putative subcomponents of this process. Therefore, their data cannot distinguish between a single valuation system in Green and Myerson’s (2004) model and the multiplicative model in Eq. 2.5. To examine the architecture of valuation in more detail Pine et al. (2009) scanned participants while they were deciding between serially presented options that differed in both monetary amount and delay. Here the BOLD responses during the presentation of each option were modeled with the three key subcomponents U, D, and V, derived from subjects’ choices according to Eq. 2.5. To ensure that no brain activity could be misattributed to a particular regressor by virtue of correlation with another, these regressors were orthogonalized. Pine et al. found that U correlated with activity in ventral tegmental area (VTA), striatum, and anterior cingulate cortex (ACC); D with VTA, striatum, insula, posterior cingulate cortex, inferior frontal gyrus, and vmPFC; and V with dorsal striatum and subgenual ACC/vmPFC. Interestingly, there was one anatomical region in the dorsal striatum where all three independent correlations overlapped, that is this area correlated independently with the discount factor, utility, and subjective value.

These results demonstrated that the brain evaluates delayed rewards in an integrative fashion. They suggest that the determinants of value are estimated separately, both with a system which relates instantaneous, undiscounted subjective value to the magnitude dimension, and with a system which calculates a discount factor to evaluate the subjective value of rewards based on their delay. Finally, a further set of regions encodes the multiplicatively integrated value of these subcomponents, which is then used to guide decisions. It was demonstrated that the dorsal striatum is the site where information from the individual value systems is integrated, because only this region represented all three subcomponents. Pine et al. (2009) thus replicated Kable and Glimcher’s (2007) findings of subjective value coding in the striatum and medial PFC, but extended them to demonstrate an expanded network of regions involved in discrete aspects of the multiplicative valuation process. To illustrate the difference consider the possibility that the activity in the posterior cingulate cortex reported by Kable and Glimcher expresses the discount factor rather than overall subjective value (i.e. implicating this region solely in discounting) – a possibility supported by Pine et al.’s findings that this region only correlated with D, and not V. These results are thus consistent with the separation of D and U (Eq. 2), and support the Prelec and Loewenstein (1991) framework over the notion of a single valuation process.

Working on a similar premise, over a number of experiments, Roesch et al. (2007), Roesch and Bryden (2011) recorded from single units in rats as they made intertemporal choices. They manipulated magnitude and delay over different blocks in order to understand how these two determinants of value are represented and integrated in various brain regions. They found that activity in the majority of the OFC neurons they recorded declined as the delay to the reward increased, implicating this region in the temporal discounting of rewards. This activity also correlated with a decreased tendency for rats to choose the larger-later option trials. Interestingly, Roesch et al. (2007) argue that the OFC does not represent V and that the OFC is not a site of “common value currency” because neurons that encoded delay did not also encode reward magnitude (a requirement if they encoded discounted value). These results conflict with other electrophysiological findings, which observed delay-discounted values of reward in single neurons in pigeon OFC analog (Kalenscher et al., 2005) and in primate OFC (Roesch and Olson, 2005). Additionally, they seem to contradict the finding that OFC lesions in rats influence both temporal discount rates and sensitivity to reward (D and U, Kheramin et al., 2002, 2003, 2004). Taken together, it is not yet clear whether the animal medial PFC represents U, D, and V separately or in an integrated fashion.

Of all the single neurons recorded in rodents by Roesch and Bryden (2011) only dopamine neurons in the midbrain, especially in VTA, appeared to integrate magnitude and delay in that they encoded to both variables and their responses to the two variables were highly correlated. As a population, neurons in VS also encoded both delay and magnitude (see also forced choice voltametry data in Day et al., 2011), and some neurons did responded to both variables in a correlated manner, but overall the correlation between the response of single cells to these two variables was low. Kobayashi and Schultz (2008) demonstrated more specifically that the activity of midbrain dopamine neurons in primates tracks the discounted value of rewards in accordance with a hyperbolic discount function. Neural firing in response to Pavlovian conditioned stimuli that predicted rewards of differing delays decreased with longer delays, at a rate similar to the discount rate measured when the same animal performed a separate choice task, and followed a pattern akin to a hyperbolic decline. These neurons were also responsive to the magnitude of the predicted reward. The site of integration Pine et al. (2009) localized in the striatum may therefore reflect the output of midbrain dopamine neurons in the rodent electrophysiological recordings (Logothetis et al., 2001). Alternatively, Pine et al.’s findings could be related to coding of temporally discounted rewards in the primate dorsal striatum reported by Cai et al. (2011).

In summary, animal and human work converge on a network comprising VTA, striatum, and medial PFC which is involved in computing and representing subjective value, but the exact role of each of these regions in constructing value remains debated. Evidence from animal single unit recordings corroborates the hierarchical model of separate encoding of D and U with an integration of the two to inform subjective value, but there is a controversy as to how downstream one has to record to locate the site of this integration. In human fMRI studies, by contrast, correlates of subjective value derived from hyperbolic discount functions are fairly consistent, but only one paper so far has investigated each component of the multiplicative model individually to delineate regions implementing separate valuation functions for delay and magnitude and their integration.

Evidence for separate representations of P and U

As with temporal discounting, a burgeoning human neuroimaging literature has shed a great deal of light on the neurobiological mechanisms of valuation and decision-making under uncertainty. Early studies examined BOLD responses to anticipation versus outcome of probabilistic rewards of varying magnitude (Breiter et al., 2001; Knutson et al., 2001a,b). Subsequent studies have identified the neural regions performing computations relating to the subjective value of risky prospects based on their expected values – correlates of which have been observed in the striatum and OFC (Preuschoff et al., 2006; Rolls et al., 2007; Tobler et al., 2007). The utility and probability functions which are multiplied to calculate subjective value are typically assumed to be linear and their neural correlates have also been examined individually (Knutson et al., 2001a examined magnitude and Abler et al., 2006 examined probability). In addition, more sophisticated Prospect theory-like utility and probability functions have been associated with BOLD signal in the striatum and vmPFC when participants decide between gambles (Tom et al., 2007; Hsu et al., 2009). Another school of thought proposes that in addition to expected values/utilities, the “riskiness” of probabilistic rewards is encoded by the brain and has some role in valuation. Such properties can be modeled by statistical moments such as their mathematical variance and skewness. Correlates of the former have been found in the lateral OFC, insula, and striatum (Dreher et al., 2006; Preuschoff et al., 2006, 2008; Rolls et al., 2007; Tobler et al., 2007; for detailed reviews see Rangel et al., 2008; Schultz et al., 2008).

Surprisingly however, to our knowledge only one study has attempted to tease apart the three components of the model in Eq. 1 within the same task. Tobler et al. (2007) presented their participants with stimuli predictive of outcomes which varied in their probability and magnitude. The value of each cue was computed as the product of the given, objective probability of the associated outcomes and their utility, which was modeled simply as a linear function of magnitude. They found that both the magnitude and probability of the predicted outcome correlated positively with separate regions in the striatum (dorsal and ventral respectively; Figure 4). In contrast, the medial PFC was only responsive to the probability of each reward. When correlating the expected value of each option predicted by the cue, Tobler et al. (2007) observed a third and separate region in the striatum. Critically, this region also overlapped with the individual probability and magnitude sensitive regions (Figure 4), strongly suggesting the striatum a site for the integration of probability and utility. To make the case more convincing Tobler et al. (2007) also showed that the striatal BOLD response to a particular expected value was the same whether the value was a product of a large magnitude and low probability (e.g., 200 points with a 10% probability) or vice versa (e.g., 20 points with a 100% probability). In a prior study (Fiorillo et al., 2003) the same group showed that VTA neurons exhibited similar characteristics, correlating positively with probability, magnitude, and their multiplicative integration. Tobler et al.’s results are analogous to those of Pine et al. in the intertemporal sphere, in that both studies revealed that each component of the integrative model in Eq. 2 was represented separately in the brain. The findings of Tobler, Fiorillo, Pine, and their colleagues therefore support the multiplicative integration of separable value components rather than a single valuation function, as in the model advocated by Green and Myerson (2004).

Figure 4.

Figure 4

Separate and partially overlapping striatal regions encoding unique valuation components of probabilistic rewards. Activity in dorsal striatum positively correlated with U (in this case magnitude), and in ventral striatum with probability. Their product, that is expected utility, positively correlated with a third, and overlapping striatal region. Figure adapted from Tobler et al. (2007).

Berns and Bell (2012) recently utilized a task where information about the probability and magnitude of rewards were presented sequentially. Although they did not look for correlates of expected value they did assess probability and magnitude independently and showed that whereas magnitude correlated with the ventral striatum BOLD responses, probability correlated with dorsal striatum activity. In this instance however the two regions did not overlap. This led the authors to conclude that although magnitude and probability are processed by distinct neural systems, their integration is not achieved by any of the models discussed above. We note, however, that none of the studies discussed in this section fully embraced the model in Eq. 1 in that they all used objective magnitudes and probabilities instead of their transformation by the utility and the probability weighting functions. This is more of a concern for null results which could be more easily explained by parameter misspecification. Therefore, before accepting Berns and Bell’s conclusion it would be necessary to check whether an overlap could be found when BOLD signal is regressed against participant-specific, behaviorally derived utilities, and subjective probabilities.

Evidence for separate representations of D and P

Both Tobler et al. (2007) and Pine et al. (2009) found that magnitude correlated predominantly with the striatum, whereas the modulators of magnitude – delay and risk – were expressed in the striatum and vmPFC among other regions. In both studies the striatum was considered the critical site of value integration. This begs the question of whether probability and delay share neural mechanisms, and even whether all modulators of reward are integrated in the striatum.

In support of the dissociation between D and P, Simon et al. (2009) observed poor correlation between the two across individuals. They allowed rats to choose between a safe small reward and a larger reward that was associated with varied probability of an electric shock. As expected, increasing magnitude and probability of shock biased rats to prefer the safe reward, and individual animals showed a stable pattern of responding across training sessions. The same rats also took part in probability discounting and delay discounting tasks that entailed choice between rewards without risk of shock. The sensitivity of individual animals to the risk of shock – namely, their tendency to opt for the safe reward in the main task – was correlated with sensitivity to risk in the probability discounting task but not with sensitivity to delay in the delay discounting task.

Relatively few fMRI studies have studied probabilistic and intertemporal choices in the same task (Luhmann et al., 2008; Ballard and Knutson, 2009; Peters and Buchel, 2009). Of these, only Peters and Buchel employed a model-based approach. They scanned participants while they were deciding between an immediate-small and larger-later rewards, and separately, between a certain-small and larger-probabilistic rewards. Decisions were modeled using hyperbolic discount functions to infer participant-specific parameters for k and h, which were subsequently used to calculate the subjective value of the dynamic (larger-later, larger-risky) options. Utility was modeled as a linear function of reward magnitude. To demonstrate that subjective value here was equivalent across choice types participants also performed a separate behavioral experiment where they decided between delayed and risky rewards. Indeed, Peters and Buchel found that their participants were indifferent between delayed and probabilistic rewards which had the same subjective value as calculated from the separate tasks, indicating the rewards had comparable intrinsic values. Analyses of the BOLD response revealed both overlapping and diverging activations. Whereas overlapping regions in the striatum and OFC were correlated with subjective value in both cases, other regions correlated with the subjective value of either risk- or delay-discounted reward. The authors concluded that the striatum and OFC are domain-general valuation regions which integrate results from domain specific subjective valuation systems into a common “neural currency” of value – that is a metric of value which can be used to compare the utilities of various multi-attribute options.

Though this is certainly a feasible interpretation, it is somewhat unintuitive to assume an integration of different subjective values (Vs) rather than an integration of different sub-components of a common subjective value (such as U, D, and P). Indeed, the analysis performed by Peters and Buchel (2009) does not eliminate the possibility that some of the regions identified as correlating with V could in fact have a more specific role in the representation of one of the subcomponents of subjective value1. Had they have compared brain activity which could only be explained by utility, the discount factors D and P (rather than inverse delay and probability), and subjective value we would have a clearer picture of the relationship between each domain specific discount system, if they are common or separable, and where and they are integrated with utility to calculate subjective value.

A critical experiment to elucidate integration across reward and its modulators, delay and probability, would be to examine the BOLD correlates when participants decide between options with all three attributes, i.e., delayed probabilistic rewards. Modeling such choices will enable a neurobiological evaluation of the multiplicative V = U × D × P approach outlined in Eq. 2.5.

Using Neurobiological Evidence to Decide between Additive and Interactive Valuation Models

The claim that animals and humans represent a mixed outcome with a single value, and that this value is intimately tied to subsequent choice, is dominant in neuroeconomics and fundamental for the models we discussed here (although it is not without its critics: Vlaev et al., 2011). Neurobiological data can provide converging evidence for this claim by showing that single cells, populations, or brain regions represent reward and other costs together. In a second step the form of neural integration could be determined. As discussed in Section “Using Neurobiological Evidence to Decide between Models of Risk and Delay” the influence of risk and delay on reward has been described in detail, and great strides have been made in our understanding of the neurobiological underpinnings of this process. We know relatively little, however, about how rewards are integrated with other costs.

We begin with a brief review of the regions thought to be involved in this process when outcomes involve a mix of rewarding food and physical effort. The dopamine system and the nucleus accumbens (NAc) are central to animals’ motivation to overcome effort in order to obtain a larger reward (Phillips et al., 2007; Salamone et al., 2007; Floresco et al., 2008a,b). Dopamine antagonists and lesions of the NAc decrease the probability that high-reward/high effort options are chosen, while dopamine agonists make this choice more likely (Cousins and Salamone, 1994; Bardgett et al., 2009; Ghods-Sharifi and Floresco, 2010; Mai et al., 2012). Although exerting effort often takes time and therefore delays reward delivery, there is evidence that the dopamine system and the NAc are important for overcoming effort even when the delay to reward is controlled (Floresco et al., 2008b; Ghods-Sharifi and Floresco, 2010). For example, Day et al. (2011) used voltametry to show that in forced choice trials NAc dopamine release expressed the discounted reward value of future outcomes associated with either effort or delay. When food reward was associated with low effort or delivered immediately, dopamine release was higher than when the same amount of food was associated with high effort or delivered after a longer delay. One difficulty in interpreting these results was that exerting effort inevitably resulted in a time delay between the cue and reward delivery. However, because dopamine release was significantly lower in the high effort relative to the long delay trials the authors could conclude that the attenuation of dopamine release in high effort trials was not solely due to the time delay associated with the effort cost. The role of dopamine is not limited to physical effort but also extends to cognitive effort (Cocker et al., 2012).

The NAc is part of a network of inter-connected regions that play a role in effort-based decision-making which includes the ACC (Walton et al., 2002, 2003; Schweimer and Hauber, 2005; Rudebeck et al., 2006) and basolateral amygdala (Floresco and Ghods-Sharifi, 2007; Ghods-Sharifi et al., 2009). Animals can overcome the effects of ACC lesions with additional training (Rudebeck et al., 2006) or when the ratio between the easy-smaller and the harder-larger rewards increases (Walton et al., 2002), and ACC lesions do not always alter effort-based choices (Schweimer and Hauber, 2005), suggesting that it plays less of a key role than the NAc (Floresco et al., 2008a).

In line with this animal work, Talmi et al. (2009) and Park et al. (2011) found that a medial PFC region extending from the subgenual/perigenual ACC to vmPFC/OFC expressed the bilinear interaction between reward and pain cost at the time of decision (Eq. 6). Talmi et al. (2009) showed that activation in the subgenual ACC that was parametrically modulated by monetary reward was attenuated when the rewarding outcome also involved pain (Figure 5). Park et al. (2011) replicated these results, demonstrating that pain-discounted values in this region fitted the bilinear model (Eq. 6) better than the additive model (Eq. 1), and even more so when the utility function in Eq. 6 was modeled as a power function. Talmi et al. also observed the same pattern in VS, in the region of the NAc. Park et al. did not find such activation in the VS but reported increased connectivity between the subgenual ACC and the amygdala when outcomes involved high compared to low pain. Notably, the VS and amygdala regions reported by these authors were only 13 mm apart. In summary, these two datasets suggest that the vmPFC and possibly the VS and amygdala express the modulation of reward by pain costs. The convergence on these regions is not surprising given their ubiquitous role in representing subjective value across a variety of paradigms (Levy and Glimcher, 2012). Rather, these studies are important because of their computational approach, which allowed them to demonstrate that neural signals in these regions conformed better to an interactive than an additive valuation of mixed outcomes.

Figure 5.

Figure 5

Neurobiological evidence for the bilinear model. BOLD signal in the ventral striatum (top) and subgenual cingulate gyrus (bottom) covaried positively with reward in the low-cost conditions (blue), in which participants decided between mixtures of money and mild electric stimulation. This correlation was attenuated in the high-cost condition (red), in which participants decided between mixtures of money and painful electric stimulation. Figure adapted from Talmi et al. (2009).

Hare et al. (2008) pointed out that decision values often correlate with the utility of reward U(m) and with reward prediction errors, and optimized their task to decorrelate these three factors. They observed that the ACC/vmPFC region close to the regions where decision value was expressed in the studies of Talmi et al. (2009) and Park et al. (2011) expressed U(m), not V, casting some doubt on the interpretation of the above findings. This concern is addressed, however, by the pattern of results in Talmi et al.’s study. In that study pain costs significantly interacted with reward value in both ventral ACC and VS, suggesting that this signal does not merely express reward utility.

Amemori and Graybiel (2012) report results that seem at first glance to challenge the bilinear model (Eq. 6). They recorded from pregenual ACC (close to the region that expressed the bilinear interaction in Talmi et al., 2009) when monkeys decided between accepting a minimum amount of food (“avoidance” decisions) and a reward that was paired with an aversive air puff (“approach” decisions). Both food amount and the strength of the air puff were manipulated parametrically. The additive model (Eq. 1) fitted behavioral choice best, better than the interactive model (Eq. 6) and more complex second and third order models.

An additional result from the same study, however, suggests that an additive model may not tell the whole story. Neuronal activity in pregenual ACC when monkeys anticipated mixed outcomes correlated with subjective value, computed according to the winning additive model, with one population of neurons coding V positively and the other negatively. Those positive-coding and negative-coding neural populations were, for the most part, evenly intermixed within the recording area, but a subzone in the ventral bank of the cingulate sulcus had a higher concentration of negatively coding neurons. Microstimulation of this subzone increased avoidance behavior, biasing monkeys to forego the mixed outcome (large reward and air puff) in favor of smaller rewards. The authors plotted cost as a function of reward, noting the decision boundary – the mixture of costs and benefits that resulted in indifference. Trials where the stimulation was “on” had shallower indifference function slopes than trials where the stimulation was “off,” indicating reduced sensitivity to reward under stimulation. The anxiolitic drug diazepam abolished this effect of the stimulation.

To interpret these more complex data it may be helpful to consider the individual differences in Talmi et al.’s (2009) dataset. Only about half of their participants exhibited shallower reward sensitivity under pain; the other participants appeared to decide according to an additive model. The former participants, those who exhibited the interaction, likely experienced the pain cost as more threatening, because their SCR responses to pain were higher and they activated the anterior insula – a region associated with the emotional response to pain (Craig, 2003) – more strongly than those whose reward sensitivity did not change. We speculate that microstimulation in Amemori and Graybiel (2012) may have similarly rendered the aversive air puff more threatening for the monkeys, resulting in attenuated reward sensitivity.

A different map of activation altogether was obtained by Prevost et al. (2010). They used an elegant design where delay and effort costs were manipulated as closely as possible using identical trial structures. Subjective value for mixtures of reward and delay or effort was computed according to the hyperbolic model (Eq. 4). Attesting to the effectiveness of their paradigm, their delay data localized subjective value to the VS and vmPFC, replicating commonalities across many previous studies (Levy and Glimcher, 2012), and closely resembling those of another dataset which also employed the same hyperbolic model to model delay-discounted reward (Peters and Buchel, 2010). This makes the dissociation they observed between the representation of subjective value in the delay and the effort condition particularly striking. Prevost et al. reported that effort-discounted value negatively correlated with signal in the dorsal ACC and anterior insula. Signals in these regions increased when outcomes were more effortful and subjectively less valuable.

Croxson et al. (2009) have similarly observed an interaction between effort and reward in the dorsal ACC. Their task employed a forced choice paradigm, allowing a clear distinction between valuation per se and decision-making. They were interested in the location of BOLD responses associated with the subjective value of cues that signaled mixtures of reward and effort. The subjective value was modeled according to a variant of the trade-off model in Eq. 3, and correlated with activity in the dorsal ACC, striatum, and midbrain; yet only the dorsal ACC expressed the interaction of effort and reward, while dopaminergic midbrain and VS expressed both reward and effort but not their interaction. The difficulty in relating Prevost et al.’s and Croxson et al.’s datasets to each other is that although both studies observed an interaction between reward and effort in the same region of the ACC, the correlation between that signal and subjective value was negative in the former and positive in the latter study. By contrast, the correlation between that signal and the level of effort required in each trial was positive in the former and negative in the latter study.

While effort is often associated with activation in the ACC, animal data provides less evidence for effort representation in the insula (Floresco et al., 2008a). Yet a relationship between insula activation and value, in the same direction as that reported by Prevost et al., was also observed in two recent studies. Brooks et al. (2010) required participants to decide between a standard delivery of 10 painful shocks and a gamble, in which either more than 10 or less than 10 shocks could be delivered in equal probabilities. As in Prevost et al. (2010) Brooks et al. also reported negative correlations between subjective gamble values and activation in the dorsal ACC and insula. In keeping with the dominant pattern in the literature, however, they also observed a positive correlation between subjective value and activity in VS. In the second study, a PET study with [18F] fallypride and d-amphetamine challenge, individual differences in dopamine function in the bilateral insula were correlated with their tendency to choose to spend more time and exert more effort in order to win larger rewards (Treadway et al., 2012). Participants who were willing to spend more time and effort for larger rewards – those who presumably evaluated this choice to have a higher subjective value than other participants – exhibited reduced dopamine function in the insula. At the same time, in line with the prevalent pattern in the literature, dopamine function in the striatum and vmPFC correlated positively with this individual difference.

The negative correlation of the insula signal with reward in Treadway et al. (2012), Brooks et al. (2010), and Prevost et al. (2010) may be related to the salience of the more effortful trials. Participants in Prevost et al.’s study may have perceived effort but not delay costs to be salient; the more effortful and time-consuming options were also likely more salient to those of Treadway et al.’s participants who chose them only infrequently, and high probability of receiving more painful shocks could also have been more salient. The “salience network” (Seeley et al., 2007), intriguingly, is identified with conjoint activation in the very same regions, dorsal ACC and bilateral insula, observed by Prevost, Brooks, and their colleagues. This reverse-inference does not, however, explain why effort-discounted value in Prevost et al.’s study did not activate the VS and vmPFC, as delay-discounted value did.

Clearly, even if the hyperbolic model does account both for the effect of effort on decision value and for the effect of delay on these values, it does not necessitate that the two share a neurobiological mechanism. A well-known set of studies demonstrated that Marmosets were willing to wait longer than Tamarins for a larger food reward, but preferred a food reward that was closer in distance to food reward that was further from them in spatial distance (Stevens et al., 2005). Because all animals grew up in captivity with limited exposure to predators, the most likely interpretation for the discounting, in Marmosets, of spatially distant food rewards was the energetic cost (effort) involved in obtaining that reward, rather than risk of predation. The double dissociation may suggest separable mechanisms for effort and delay discounting, or it could indicate differences in valuing these two costs upstream to the decision-making process, as per the discussion of D and U in Section “Evidence for Separate Representations of D and U.”

The direction of the correlation between outcome mixtures and VS activity is also not without controversy. On the one hand, Kurniawan et al. (2010) also reported positive value coding in VS for mixtures of reward and effort. Although behaviorally, effort and reward did not interact significantly, the fMRI data suggested a neurobiological interaction. NAc activity was positively correlated with reward magnitude, a correlation that was only significant when participants chose to expend effort for large rewards, but not when they chose to expend effort for smaller rewards or when they chose the low-reward, low effort option. On the other hand, Botvinick et al. (2009) observed stronger NAc activation when a cue signaled a more effortful task. One possibility is that because participants were not offered any reward in that study, the direction of value coding in the VS may have reversed; but in Brooks et al. (2010) the choices were also between “bad” and “worse,” and VS activity correlated positively with reward. Interestingly, Botvinick and colleagues interpreted their data according to Kivetz’ model (Eq. 5), suggesting that NAc activation signals the obligatory shift of the reference point of the utility function to the right in effortful blocks.

We have reviewed neurobiological evidence that accords with interactive models of valuation, however the additive model dominates the human imaging literature. To take just one example from a particularly elegant study, Hare et al. (2008) used an additive model (Eq. 1) to compute the decision value, which in their study was the difference between the true value of a food item, established according to the elicited “willingness to pay” for that item in a Becker–DeGroot–Marschak (BDM) auction procedure, and the price at which the food item was offered to the participant. Hare et al. optimized their task to decorrelate this decision value from U(m) and a reward prediction error signal, and observed a positive correlation between central OFC activity and this decision value (see also Plassmann et al., 2007). Similarly to studies that used only one model to fit their data, or compared the fit only between two models, this converging evidence for the additive model of valuation cannot rule out the possibility that an interaction term would have improved the fit.

In summary, computationally inspired studies of decision valuation, in participants deliberating between mixed outcomes, have produced converging evidence for additive as well as interactive models when correlating the value computed according to these models with neural activity. However, since most neuroimaging studies compare two models at most, it is possible that more convergence could be achieved by greater employment of model comparison.

Conclusion

The goal of much neurobiological work on valuation is to understand internal reward representations, namely, how the brain represents costs and benefits that are present in the environment (Dayan, 2012). Here we asked how the subjective value of outcomes is established when they consist of mixtures of costs and benefits. This is a surprisingly under-researched topic despite a large empirical and computational body of work on decision-making.

The way people value costs and benefits, individually, has been studied extensively. Here we reviewed current thinking and empirical data concerning the subjective value of monetary gains and losses, and the influence of risk and time delay on this value. We discussed data that support and challenge available models, and the potential for neurobiological work to illuminate some open questions. By comparison, the functional form of cost-benefit analysis – the decision between mixtures of rewards and costs – is relatively unknown. We described two general classes of models – additive and interactive – for the process of integrating rewards and costs into a single decision value. The economic literature typically assumes that costs and benefits are integrated additively, but there is also support for a variety of interactive models. Yet only a handful of studies directly compare additive and interactive models, or between interactive model variants. Modeling-informed empirical work is clearly necessary in order to enhance understanding of the neurobiological mechanism that allows animals and humans to integrate costs and benefits. Empirical work on this intriguing question should proceed with caution; not assuming that integrated representations of the subjective value of anticipated outcomes are natural kinds, but to demonstrate their existence empirically (Vlaev et al., 2011). We hope that by clarifying some of the main candidates for valuation, and the way neurobiological data can support or challenge them, we will encourage further empirical work on the mechanism that allows animals and humans to decide optimally in a complex environment.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We would like to thank J. Dennett-Thorpe for help with modeling, C. Webb and A. Howes for useful advice, and P. Dayan for comments on a draft of this manuscript. The work was partly funded by an ESRC First Grant (Deborah Talmi).

Appendix

Phillips et al. (2007) suggested that when a participant is faced with a choice of taking or foregoing an action which would yield both rewards MR and costs MC, the value of the action can be expressed as the value V(MR, MC) relative to the value of the same action if cost was not involved, V(MR, MC = 0). When there are no costs the value is maximal because V(MR, MC) = V(MR, MC = 0).

They observed that there is a specific cost Ci which translates to V(MR, MC) = 0, namely, lead the participant to be indifferent about the choice. When MC = Ci participants are not motivated either to act or avoid acting; consequently, they act 50% of the time. Costs higher than Ci mean that V(MR,MC) < 0 and bias the participant against the action; costs lower than Ci mean that V(MR,MC) > 0 and favor taking the action. The authors note that there is limited evidence as to the exact functional forms of the indifference function Ci and offered a tentative, plausible form in their Figure 1, where cost is a linear function of dopamine:

MC=a1×DA (A1)

With the constant a1 > 0. Given evidence that dopamine level (DA) is a function of currently available reward and the maximum levels of dopamine DAmax observed in the task context, and that the utility of reward is a decelerating function of reward magnitude, DA was described according to the following function:

DA=DAmax×MRMR+a2 (A2)

With the constant a2 > 0. Therefore,

Ci=a3×MRMR + a2 (A3)

With the constant a3 > 0.

The authors further proposed a specific form for a cost-benefit function Z (Figure 2), which depicts the ratio of outcome with cost to the same outcome without cost.

Z=V(MR,MC)V(MR,MC=0). (A4)

Because when there is no response cost the two outcomes are equivalent (the ratio is maximal), and because when the ratio is 0.5 decision-makers are indifferent between the two outcomes, we determine that Z passes between (Ci, 0) and (0,1). Therefore

Z=1-MCCi (A5)

The authors did not provide an explicit model for V(MR, MC = 0), the utility of rewarding outcomes that are not accompanied by costs. On the basis of their choice of Eq. 2 to represent DA and their reasoning for that choice we selected the same form to also model reward utility:

VMR,MC=0=Rmax×MRMR+a4 (A6)

Combining the above equations provides us with a value function for mixed outcomes:

VMR,MC=Rmax×MRMR+a4-Rmaxa3×MC×MR+a2(MR+a4) (A7)

To understand its implications we considered the case where Mr < < a4, namely when the utility of Mr, V(MR, MC = 0), resembles a linear function. In this case, some algebra will show that

V(MR,MC)=Rmaxa4×MR-a1A3×MC-MR×MCa31-a2a4-R2a4×1-MCa3 (A8)

Now, as MR < < a4, we could probably drop the last term as it is small (it is always a positive term) and “tidy up,” introducing constants x1, x2, and x3’:

V(MR,MC)=X1×MR-X2×MC-X3×MR×MC (A9)

With x3 a constant that is either positive or negative:

X3=Rmaxa4×1-a2a4a3 (A10)

And x1 and x2 positive constants.

Footnotes

1Peters and Buchel included magnitude (m), probability (p) and the inverse of delay (1/d) as regressors in their model. These regressors were orthogonalised with respect to the regressor that coded value (V). Crucially, because m, d and p would, to some extent, be correlated with V, and because the analysis model ensured that V was assigned all variance that correlated with it, it is difficult to ascertain which of the regions that seemingly coded V actually coded m, d or P. This perspective may explain why the authors obtained a different map of activation when they reversed the order of orthogonalization.

References

  1. Abler B., Walter H., Erk S., Kammerer H., Spitzer M. (2006). Prediction error as a linear function of reward probability is coded in human nucleus accumbens. Neuroimage 31, 790–795 10.1016/j.neuroimage.2006.01.001 [DOI] [PubMed] [Google Scholar]
  2. Ainslie G. (1975). Specious reward – behavioral theory of impulsiveness and impulse control. Psychol. Bull. 82, 463–496 10.1037/h0076860 [DOI] [PubMed] [Google Scholar]
  3. Ainslie G. (2001). Breakdown of Will. Cambridge: Cambridge University Press [Google Scholar]
  4. Ainslie G., Herrnstein R. J. (1981). Preference reversal and delayed reinforcement. Anim. Learn. Behav. 9, 476–482 10.3758/BF03209777 [DOI] [Google Scholar]
  5. Allais M. (1953). Le comportement de l’homme rationnel devant le risque: critique des postulats et axioms de l’École Américaine. Econometrica 21, 503–546 10.2307/1907921 [DOI] [Google Scholar]
  6. Amemori K., Graybiel A. M. (2012). Localized microstimulation of primate pregenual cingulate cortex induces negative decision-making. Nat. Neurosci. 15, 776–785 10.1038/nn.3088 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Baker F., Johnson M. W., Bickel W. K. (2003). Delay discounting in current and never-before cigarette smokers: similarities and differences across commodity, sign, and magnitude. J. Abnorm. Psychol. 112, 382–392 10.1037/0021-843X.112.3.382 [DOI] [PubMed] [Google Scholar]
  8. Ballard K., Knutson B. (2009). Dissociable neural representations of future reward magnitude and delay during temporal discounting. Neuroimage 45, 143–150 10.1016/j.neuroimage.2008.10.052 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Bardgett M. E., Depenbrock M., Downs N., Points M., Green L. (2009). Dopamine modulates effort-based decision making in rats. Behav. Neurosci. 123, 242–251 10.1037/a0014625 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Bautista L. M., Tinbergen J., Kacelnik A. (2001). To walk or to fly? How birds choose among foraging modes. Proc. Natl. Acad. Sci. U.S.A. 98, 1089–1094 10.1073/pnas.98.3.1089 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bell D. E. (1982). Regret in decision-making under uncertainty. Oper. Res. 30, 961–981 10.1287/opre.30.5.961 [DOI] [Google Scholar]
  12. Berns G. S., Bell E. (2012). Striatal topography of probability and magnitude information for decisions under uncertainty. Neuroimage 59, 3166–3172 10.1016/j.neuroimage.2011.11.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Berns G. S., Capra C. M., Moore S., Noussair C. (2008). Three studies on the neuroeconomics of decision-making when payoffs are real and negative. Adv. Health Econ. Health Serv. Res. 20, 1–29 10.1016/S0731-2199(08)20001-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Berns G. S., Chappelow J., Cekic M., Zink C. F., Pagnoni G., Martin-Skurski M. E. (2006). Neurobiological substrates of dread. Science 312, 754–758 10.1126/science.1123721 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Berns G. S., Laibson D., Loewenstein G. (2007). Intertemporal choice – toward an integrative framework. Trends Cogn. Sci. (Regul. Ed.) 11, 482–488 10.1016/j.tics.2007.08.011 [DOI] [PubMed] [Google Scholar]
  16. Botvinick M. M., Huffstetler S., McGuire J. T. (2009). Effort discounting in human nucleus accumbens. Cogn. Affect. Behav. Neurosci. 9, 16–27 10.3758/CABN.9.1.16 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Brañas-Garza P., Espinosa M. P., Repollés M. (2012). Time discounting and pain anticipation: experimental evidence. Revista Internacional de Sociología 70, 73–81 10.3989/ris.2011.04.1A [DOI] [Google Scholar]
  18. Breiter H. C., Aharon I., Kahneman D., Dale A., Shizgal P. (2001). Functional imaging of neural responses to expectancy and experience of monetary gains and losses. Neuron 30, 619–639 10.1016/S0896-6273(01)00303-8 [DOI] [PubMed] [Google Scholar]
  19. Brooks A. M., Pammi V. S., Noussair C., Capra C. M., Engelmann J. B., Berns G. S. (2010). From bad to worse: striatal coding of the relative value of painful decisions. Front. Neurosci. 4, 176. 10.3389/fnins.2010.00176 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Cai X. Y., Kim S., Lee D. (2011). Heterogeneous coding of temporally discounted values in the dorsal and ventral striatum during intertemporal choice. Neuron 69, 170–182 10.1016/j.neuron.2010.11.041 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Cardinal R. N., Winstanley C. A., Robbins T. W., Everitt B. J. (2004). Limbic corticostriatal systems and delayed reinforcement. Ann. N. Y. Acad. Sci. 1021, 33–50 10.1196/annals.1308.004 [DOI] [PubMed] [Google Scholar]
  22. Chapman G. B. (1996). Temporal discounting and utility for health and money. J. Exp. Psychol. Learn. Mem. Cogn. 22, 771–791 10.1037/0278-7393.22.3.771 [DOI] [PubMed] [Google Scholar]
  23. Chapman G. B. (2001). Time preferences for the very long term. Acta Psychol. (Amst.) 108, 95–116 10.1016/S0001-6918(01)00030-0 [DOI] [PubMed] [Google Scholar]
  24. Chapman G. B., Elstein A. S. (1995). Valuing the future – temporal discounting of health and money. Med. Decis. Making 15, 373–386 10.1177/0272989X9501500408 [DOI] [PubMed] [Google Scholar]
  25. Chapman G. B., Weber B. J. (2006). Decision biases in intertemporal choice and choice under uncertainty: testing a common account. Mem. Cognit. 34, 589–602 10.3758/BF03193582 [DOI] [PubMed] [Google Scholar]
  26. Chapman G. B., Winquist J. R. (1998). The magnitude effect: temporal discount rates and restaurant tips. Psychon. Bull. Rev. 5, 119–123 10.3758/BF03209466 [DOI] [Google Scholar]
  27. Cocker P. J., Hosking J. G., Benoit J., Winstanley C. A. (2012). Sensitivity to cognitive effort mediates psychostimulant effects on a novel rodent cost/benefit decision-making task. Neuropsychopharmacology 37, 1825–1837 10.1038/npp.2012.30 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Cousins M. S., Salamone J. D. (1994). Nucleus-accumbens dopamine depletions in rats affect relative response allocation in a novel cost/benefit procedure. Pharmacol. Biochem. Behav. 49, 85–91 10.1016/0091-3057(94)90460-X [DOI] [PubMed] [Google Scholar]
  29. Craig A. D. (2003). Interoception: the sense of the physiological condition of the body. Curr. Opin. Neurobiol. 13, 500–505 10.1016/S0959-4388(03)00090-4 [DOI] [PubMed] [Google Scholar]
  30. Croxson P. L., Walton M. E., O’Reilly J. X., Behrens T. E. J., Rushworth M. F. S. (2009). Effort-based cost-benefit valuation and the human brain. J. Neurosci. 29, 4531–4541 10.1523/JNEUROSCI.4515-08.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Day J. J., Jones J. L., Carelli R. M. (2011). Nucleus accumbens neurons encode predicted and ongoing reward costs in rats. Eur. J. Neurosci. 33, 308–321 10.1111/j.1460-9568.2010.07531.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Dayan P. (2012). How to set the switches on this thing. Curr. Opin. Neurobiol. [Epub ahead of print]. 10.1016/j.conb.2012.05.011 [DOI] [PubMed] [Google Scholar]
  33. Dreher J. C., Kohn P., Berman K. F. (2006). Neural coding of distinct statistical properties of reward information in humans. Cereb. Cortex 16, 561–573 10.1093/cercor/bhj004 [DOI] [PubMed] [Google Scholar]
  34. Estle S. J., Green L., Myerson J., Holt D. D. (2006). Differential effects of amount on temporal and probability discounting of gains and losses. Mem. Cognit. 34, 914–928 10.3758/BF03193437 [DOI] [PubMed] [Google Scholar]
  35. Fiorillo C. D., Tobler P. N., Schultz W. (2003). Discrete coding of reward probability and uncertainty by dopamine neurons. Science 299, 1898–1902 10.1126/science.1077349 [DOI] [PubMed] [Google Scholar]
  36. Floresco S. B., Ghods-Sharifi S. (2007). Amygdala-prefrontal cortical circuitry regulates effort-based decision making. Cereb. Cortex 17, 251–260 10.1093/cercor/bhj143 [DOI] [PubMed] [Google Scholar]
  37. Floresco S. B., Onge J. R. S., Ghods-Sharifi S., Winstanley C. A. (2008a). Cortico-limbic-striatal circuits subserving different forms of cost-benefit decision making. Cogn. Affect. Behav. Neurosci. 8, 375–389 10.3758/CABN.8.4.375 [DOI] [PubMed] [Google Scholar]
  38. Floresco S. B., Tse M. T. L., Ghods-Sharifi S. (2008b). Dopaminergic and glutamatergic regulation of effort- and delay-based decision making. Neuropsychopharmacology 33, 1966–1979 10.1038/sj.npp.1301565 [DOI] [PubMed] [Google Scholar]
  39. Frederick S., Loewenstein G., O’Donoghue T. (2002). Time discounting and time preference: a critical review. J. Econ. Lit. 40, 351–401 10.1257/002205102320161311 [DOI] [Google Scholar]
  40. Freeman K. B., Nonnemacher J. E., Green L., Myerson J., Woolverton W. L. (2012). Delay discounting in rhesus monkeys: equivalent discounting of more and less preferred sucrose concentrations. Learn. Behav. 40, 54–60 10.3758/s13420-011-0045-3 [DOI] [PubMed] [Google Scholar]
  41. Ghods-Sharifi S., Floresco S. B. (2010). Differential effects on effort discounting induced by inactivations of the nucleus accumbens core or shell. Behav. Neurosci. 124, 179–191 10.1037/a0018932 [DOI] [PubMed] [Google Scholar]
  42. Ghods-Sharifi S., Onge J. R. S., Floresco S. B. (2009). Fundamental contribution by the basolateral amygdala to different forms of decision making. J. Neurosci. 29, 5251–5259 10.1523/JNEUROSCI.0315-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Grace R. C. (1999). The matching law and amount-dependent exponential discounting as accounts of self-control choice. J. Exp. Anal. Behav. 71, 27–44 10.1901/jeab.1999.71-27 [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Green L., Fristoe N., Myerson J. (1994). Temporal discounting and preference reversals in choice between delayed outcomes. Psychon. Bull. Rev. 1, 383–389 10.3758/BF03213979 [DOI] [PubMed] [Google Scholar]
  45. Grace R. C., Sargisson R. J., White K. G. (2012). Evidence for a magnitude effect in temporal discounting with pigeons. J. Exp. Psychol. Anim. Behav. Process. 38, 102–108 10.1037/a0026345 [DOI] [PubMed] [Google Scholar]
  46. Green L., Myerson J. (2004). A discounting framework for choice with delayed and probabilistic rewards. Psychol. Bull. 130, 769–792 10.1037/0033-2909.130.5.769 [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Green L., Myerson J., Critchfield T. S. (2011). Introduction to the special issue: translational research on discounting. Psychol. Rec. 61, 523–525 [Google Scholar]
  48. Green L., Myerson J., Holt D. D., Slevin J. R., Estle S. J. (2004). Discounting of delayed food rewards in pigeons and rats: is there a magnitude effect? J. Exp. Anal. Behav. 81, 39–50 10.1901/jeab.2004.81-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Green L., Myerson J., McFadden E. (1997). Rate of temporal discounting decreases with amount of reward. Mem. Cognit. 25, 715–723 10.3758/BF03211314 [DOI] [PubMed] [Google Scholar]
  50. Gregorios-Pippas L., Tobler P. N., Schultz W. (2009). Short-term temporal discounting of reward value in human ventral striatum. J. Neurophysiol. 101, 1507–1523 10.1152/jn.90730.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Hare T. A., O’Doherty J., Camerer C. F., Schultz W., Rangel A. (2008). Dissociating the role of the orbitofrontal cortex and the striatum in the computation of goal values and prediction errors. J. Neurosci. 28, 5623–5630 10.1523/JNEUROSCI.1309-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Hariri A. R., Brown S. M., Williamson D. E., Flory J. D., De Wit H., Manuck S. B. (2006). Preference for immediate over delayed rewards is associated with magnitude of ventral striatal activity. J. Neurosci. 26, 13213–13217 10.1523/JNEUROSCI.3446-06.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Herrnstein R. J. (1981). “Self-control as response strength,” in Quantification of Steady-State Operant Behaviour, eds Bradshaw C. M., Szabadi E., Lowe C. F. (North Holland: Elsevier; ), 3–20 [Google Scholar]
  54. Ho M. Y., Mobini S., Chiang T. J., Bradshaw C. M., Szabadi E. (1999). Theory and method in the quantitative analysis of “impulsive choice” behaviour: implications for psychopharmacology. Psychopharmacology (Berl.) 146, 362–372 10.1007/PL00005482 [DOI] [PubMed] [Google Scholar]
  55. Hsee C. K., Hastie R. (2006). Decision and experience: why don’t we choose what makes us happy? Trends Cogn. Sci. (Regul. Ed.) 10, 31–37 10.1016/j.tics.2005.11.007 [DOI] [PubMed] [Google Scholar]
  56. Hsu M., Krajbich I., Zhao C., Camerer C. F. (2009). Neural response to reward anticipation under risk is nonlinear in probabilities. J. Neurosci. 29, 2231–2237 10.1523/JNEUROSCI.5296-08.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Hull C. (1943). Principles of Behavior. New York: Appleton-Century-Crofts [Google Scholar]
  58. Kable J. W., Glimcher P. W. (2007). The neural correlates of subjective value during intertemporal choice. Nat. Neurosci. 10, 1625–1633 10.1038/nn2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Kahneman D., Tversky A. (eds). (2000). Choices, Values and Frames. New York: Cambridge University Press [Google Scholar]
  60. Kahneman D., Tversky A. (1979). Prospect theory – analysis of decision under risk. Econometrica 47, 263–291 10.2307/1914185 [DOI] [Google Scholar]
  61. Kahneman D., Wakker P. P., Sarin R. (1997). Back to Bentham? – Explorations of experienced utility. Q. J. Econ. 112, 375–405 10.1162/003355397555235 [DOI] [Google Scholar]
  62. Kalenscher T., Windmann S., Diekamp B., Rose J., Gunturkun O., Colombo M. (2005). Single units in the pigeon brain integrate reward amount and time-to-reward in an impulsive choice task. Curr. Biol. 15, 594–602 10.1016/j.cub.2005.02.052 [DOI] [PubMed] [Google Scholar]
  63. Keeney R. L., Raiffa H. (1993). Decisions with Multiple Objectives: Preferences and Value Tradeoffs. Cambridge: Cambridge University Press [Google Scholar]
  64. Kheramin S., Body S., Ho M., Velazquez-Martinez D. N., Bradshaw C. M., Szabadi E., Deakin J. F., Anderson I. M. (2003). Role of the orbital prefrontal cortex in choice between delayed and uncertain reinforcers: a quantitative analysis. Behav. Processes 64, 239–250 10.1016/S0376-6357(03)00142-6 [DOI] [PubMed] [Google Scholar]
  65. Kheramin S., Body S., Ho M. Y., Velazquez-Martinez D. N., Bradshaw C. M., Szabadi E., et al. (2004). Effects of orbital prefrontal cortex dopamine depletion on inter-temporal choice: a quantitative analysis. Psychopharmacology (Berl.) 175, 206–214 10.1007/s00213-004-1813-y [DOI] [PubMed] [Google Scholar]
  66. Kheramin S., Body S., Mobini S., Ho M. Y., Velazquez-Martinez D. N., Bradshaw C. M., et al. (2002). Effects of quinolinic acid-induced lesions of the orbital prefrontal cortex on inter-temporal choice: a quantitative analysis. Psychopharmacology (Berl.) 165, 9–17 10.1007/s00213-002-1228-6 [DOI] [PubMed] [Google Scholar]
  67. Kim S., Labroo A. A. (2011). From inherent value to incentive value: when and why pointless effort enhances consumer preference. J. Consum. Res. 38, 712–742 10.1086/658148 [DOI] [Google Scholar]
  68. Kivetz R. (2003). The effects of effort and intrinsic motivation on risky choice. Marketing Sci. 22, 477–502 10.1287/mksc.22.4.477.24911 [DOI] [Google Scholar]
  69. Knutson B., Adams C. M., Fong G. W., Hommer D. (2001a). Anticipation of increasing monetary reward selectively recruits nucleus accumbens. J. Neurosci. 21, RC159. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Knutson B., Fong G. W., Adams C. M., Varner J. L., Hommer D. (2001b). Dissociation of reward anticipation and outcome with event-related fMRI. Neuroreport 12, 3683–3687 10.1097/00001756-200112040-00016 [DOI] [PubMed] [Google Scholar]
  71. Kobayashi S., Schultz W. (2008). Influence of reward delays on responses of dopamine neurons. J. Neurosci. 28, 7837–7846 10.1523/JNEUROSCI.1656-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Kool W., McGuire J. T., Rosen Z. B., Botvinick M. M. (2010). Decision making and the avoidance of cognitive demand. J. Exp. Psychol. Gen. 139, 665–682 10.1037/a0020198 [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Koopmans T. C. (1960). Stationary ordinal utility and impatience. Econometrica 28, 287–309 10.2307/1907722 [DOI] [Google Scholar]
  74. Kurniawan I. T., Guitart-Masip M., Dolan R. J. (2011). Dopamine and effort-based decision making. Front. Neurosci. 5:81. 10.3389/fnins.2011.00081 [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Kurniawan I. T., Seymour B., Talmi D., Yoshida W., Chater N., Dolan R. J. (2010). Choosing to make an effort: the role of striatum in signaling physical effort of a chosen action. J. Neurophysiol. 104, 313–321 10.1152/jn.00027.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Levy D. J., Glimcher P. W. (2012). The root of all value: a neural common currency for choice. Curr. Opin. Neurobiol. [Epub ahead of print]. 10.1016/j.conb.2012.06.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Loewenstein G. (1987). Anticipation and the valuation of delayed consumption. Econ. J. 97, 666–684 10.2307/2232929 [DOI] [Google Scholar]
  78. Loewenstein G., Elster J. (eds). (1992). Choice Over Time. New York: Russell Sage Foundation [Google Scholar]
  79. Loewenstein G., Prelec D. (1991). Negative time preference. Am. Econ. Rev. 81, 347–352 [Google Scholar]
  80. Logothetis N. K., Pauls J., Augath M., Trinath T., Oeltermann A. (2001). Neurophysiological investigation of the basis of the fMRI signal. Nature 412, 150–157 10.1038/35084005 [DOI] [PubMed] [Google Scholar]
  81. Luhmann C. C., Chun M. M., Yi D. J., Lee D., Wang X. J. (2008). Neural dissociation of delay and uncertainty in intertemporal choice. J. Neurosci. 28, 14459–14466 10.1523/JNEUROSCI.5058-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Mai B., Sommer S., Hauber W. (2012). Motivational states influence effort-based decision making in rats: the role of dopamine in the nucleus accumbens. Cogn. Affect. Behav. Neurosci. 12, 74–84 10.3758/s13415-011-0068-4 [DOI] [PubMed] [Google Scholar]
  83. Mazur J. E. (1987). “An adjusting procedure for studying delayed reinforcement,” in Quantitative Analyses of Behavior. The Effect of Delay and of Intervening Events on Reinforcement Value, eds Commons M. L., Mazur J. E., Nevin J. A., Rachlin H. (Hillsdale, NJ: Lawrence Earlbaum; ), 55–73 [Google Scholar]
  84. McClure S. M., Ericson K. M., Laibson D. I., Loewenstein G., Cohen J. D. (2007). Time discounting for primary rewards. J. Neurosci. 27, 5796–5804 10.1523/JNEUROSCI.4246-06.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. McClure S. M., Laibson D. I., Loewenstein G., Cohen J. D. (2004). Separate neural systems value immediate and delayed monetary rewards. Science 306, 503–507 10.1126/science.1100907 [DOI] [PubMed] [Google Scholar]
  86. Myerson J., Green L., Morris J. (2011). Modeling the effect of reward amount on probability discounting. J. Exp. Anal. Behav. 95, 175–187 10.1901/jeab.2011.95-175 [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Padoa-Schioppa C. (2011). Neurobiology of economic choice: a good-based model. Annu. Rev. Neurosci. 34, 333–359 10.1146/annurev-neuro-061010-113648 [DOI] [PMC free article] [PubMed] [Google Scholar]
  88. Park S. Q., Kahnt T., Rieskamp J., Heekeren H. R. (2011). Neurobiology of value integration: when value impacts valuation. J. Neurosci. 31, 9307–9314 10.1523/JNEUROSCI.4973-10.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  89. Pender J. L. (1996). Discount rates and credit markets: theory and evidence from rural India. J. Dev. Econ. 50, 257–296 10.1016/S0304-3878(96)00400-2 [DOI] [Google Scholar]
  90. Peters J., Buchel C. (2009). Overlapping and distinct neural systems code for subjective value during intertemporal and risky decision making. J. Neurosci. 29, 15727–15734 10.1523/JNEUROSCI.3684-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Peters J., Buchel C. (2010). Neural representations of subjective reward value. Behav. Brain Res. 213, 135–141 10.1016/j.bbr.2010.04.031 [DOI] [PubMed] [Google Scholar]
  92. Phelps E. S., Pollak R. A. (1968). 2nd-Best national saving and game-equilibrium growth. Rev. Econ. Stud. 35, 185–199 10.2307/2296547 [DOI] [Google Scholar]
  93. Phillips P. E. M., Walton M. E., Jhou T. C. (2007). Calculating utility: preclinical evidence for cost-benefit analysis by mesolimbic dopamine. Psychopharmacology (Berl.) 191, 483–495 10.1007/s00213-006-0626-6 [DOI] [PubMed] [Google Scholar]
  94. Pine A., Seymour B., Roiser J. P., Bossaerts P., Friston K. J., Curran H. V., et al. (2009). Encoding of marginal utility across time in the human brain. J. Neurosci. 29, 9575–9581 10.1523/JNEUROSCI.1126-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  95. Pine A., Shiner T., Seymour B., Dolan R. J. (2010). Dopamine, time, and impulsivity in humans. J. Neurosci. 30, 8888–8896 10.1523/JNEUROSCI.6028-09.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Plassmann H., O’Doherty J., Rangel A. (2007). Orbitofrontal cortex encodes willingness to pay in everyday economic transactions. J. Neurosci. 27, 9984–9988 10.1523/JNEUROSCI.2131-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Prelec D., Loewenstein G. (1991). Decision-making over time and under uncertainty – a common approach. Manage. Sci. 37, 770–786 10.1287/mnsc.37.7.770 [DOI] [Google Scholar]
  98. Preuschoff K., Bossaerts P., Quartz S. R. (2006). Neural differentiation of expected reward and risk in human subcortical structures. Neuron 51, 381–390 10.1016/j.neuron.2006.06.024 [DOI] [PubMed] [Google Scholar]
  99. Preuschoff K., Quartz S. R., Bossaerts P. (2008). Human insula activation reflects prediction errors as well as risk. J. Neurosci. 28, 2745–2752 10.1523/JNEUROSCI.4286-07.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. Prevost C., Pessiglione M., Metereau E., Clery-Melin M. L., Dreher J. C. (2010). Separate valuation subsystems for delay and effort decision costs. J. Neurosci. 30, 14080–14090 10.1523/JNEUROSCI.2752-10.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  101. Rabin M., Thaler R. H. (2001). Anomalies – risk aversion. J. Econ. Perspect. 15, 219–232 10.1257/jep.15.4.219 [DOI] [Google Scholar]
  102. Rachlin H., Raineri A. (1992). “Irrationality, impulsiveness and selfishness as discount reversal effects,” in Choice Over Time, eds Loewenstein G., Elster J. (New York: Russell Sage; ), 93–118 [Google Scholar]
  103. Rachlin H., Raineri A., Cross D. (1991). Subjective-probability and delay. J. Exp. Anal. Behav. 55, 233–244 10.1901/jeab.1991.55-233 [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Raineri A., Rachlin H. (1993). The effect of temporal constraints on the value of money and other commodities. J. Behav. Decis. Mak. 6, 77–94 10.1002/bdm.3960060202 [DOI] [Google Scholar]
  105. Rangel A., Camerer C., Montague P. R. (2008). A framework for studying the neurobiology of value-based decision making. Nat. Rev. Neurosci. 9, 545–556 10.1038/nrn2357 [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Redelmeier D. A., Heller D. N. (1993). Time preference in medical decision-making and cost-effectiveness analysis. Med. Decis. Mak. 13, 212–217 10.1177/0272989X9301300306 [DOI] [PubMed] [Google Scholar]
  107. Richards J. B., Mitchell S. H., Dewit H., Seiden L. S. (1997). Determination of discount functions in rats with an adjusting-amount procedure. J. Exp. Anal. Behav. 67, 353–366 10.1901/jeab.1997.67-353 [DOI] [PMC free article] [PubMed] [Google Scholar]
  108. Roesch M. R., Bryden D. W. (2011). Impact of size and delay on neural activity in the rat limbic corticostriatal system. Front. Neurosci. 5:130. 10.3389/fnins.2011.00130 [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Roesch M. R., Calu D. J., Burke K. A., Schoenbaum G. (2007). Should I stay or should I go? Transformation of time-discounted rewards in orbitofrontal cortex and associated brain circuits. Ann. N. Y. Acad. Sci. 1104, 21–34 10.1196/annals.1390.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  110. Roesch M. R., Olson C. R. (2005). Neuronal activity in primate orbitofrontal cortex reflects the value of time. J. Neurophysiol. 94, 2457–2471 10.1152/jn.00373.2005 [DOI] [PubMed] [Google Scholar]
  111. Rolls E. T., McCabe C., Redoute J. (2007). Expected value, reward outcome, and temporal difference error representations in a probabilistic decision task. Cereb. Cortex 18, 652–663 10.1093/cercor/bhm097 [DOI] [PubMed] [Google Scholar]
  112. Rudebeck P. H., Walton M. E., Smyth A. N., Bannerman D. M., Rushworth M. F. (2006). Separate neural pathways process different decision costs. Nat. Neurosci. 9, 1161–1168 10.1038/nn1756 [DOI] [PubMed] [Google Scholar]
  113. Salamone J. D., Correa M., Farrar A., Mingote S. M. (2007). Effort-related functions of nucleus accumbens dopamine and associated forebrain circuits. Psychopharmacology (Berl.) 191, 461–482 10.1007/s00213-006-0668-9 [DOI] [PubMed] [Google Scholar]
  114. Samuelson P. A. (1937). A note on measurement of utility. Rev. Econ. Stud. 4, 155–161 10.2307/2967612 [DOI] [Google Scholar]
  115. Schoenfelder T. E., Hantula D. A. (2003). A job with a future? Delay discounting, magnitude effects, and domain independence of utility for career decisions. J. Vocat. Behav. 62, 43–55 10.1016/S0001-8791(02)00032-5 [DOI] [Google Scholar]
  116. Schultz W., Preuschoff K., Camerer C., Hsu M., Fiorillo C. D., Tobler P. N., et al. (2008). Explicit neural signals reflecting reward uncertainty. Philos. Trans. R. Soc. Lond. B Biol. Sci. 363, 3801–3811 10.1098/rstb.2008.0152 [DOI] [PMC free article] [PubMed] [Google Scholar]
  117. Schweighofer N., Shishida K., Han C. E., Okamoto Y., Tanaka S. C., Yamawaki S., Doya K. (2006). Humans can adopt optimal discounting strategy under real-time constraints. PLoS Comput. Biol. 2, 1349–1356 10.1371/journal.pcbi.0020152 [DOI] [PMC free article] [PubMed] [Google Scholar]
  118. Schweimer J., Hauber W. (2005). Involvement of the rat anterior cingulate cortex in control of instrumental responses guided by reward expectancy. Learn. Mem. 12, 334–342 10.1101/lm.90605 [DOI] [PMC free article] [PubMed] [Google Scholar]
  119. Seeley W. W., Menon V., Schatzberg A. F., Keller J., Glover G. H., Kenna H., Reiss A. L., Greicius M. D. (2007). Dissociable intrinsic connectivity networks for salience processing and executive control. J. Neurosci. 27, 2349–2356 10.1523/JNEUROSCI.5587-06.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  120. Shelley M. K. (1993). Outcome signs, question frames and discount rates. Manage. Sci. 39, 806–815 10.1287/mnsc.39.7.806 [DOI] [Google Scholar]
  121. Simon N. W., Gilbert R. J., Mayse J. D., Bizon J. L., Setlow B. (2009). Balancing risk and reward: a rat model of risky decision making. Neuropsychopharmacology 34, 2208–2217 10.1038/npp.2009.48 [DOI] [PMC free article] [PubMed] [Google Scholar]
  122. Simonson I. (1989). Choice based on reasons – the case of attraction and compromise effects. J. Consum. Res. 16, 158–174 10.1086/209205 [DOI] [Google Scholar]
  123. Soman D. (2004). The effect of time delay on multi-attribute choice. J. Econ. Psychol. 25, 153–175 10.1016/j.joep.2003.09.002 [DOI] [Google Scholar]
  124. Stephens D. W., Krebs J. R. (1986). Foraging Theory. Princeton, NJ: Princeton University Press [Google Scholar]
  125. Stevens J. R., Rosati A. G., Ross K. R., Hauser M. D. (2005). Will travel for food: spatial discounting in two new world monkeys. Curr. Biol. 15, 1855–1860 10.1016/j.cub.2005.10.055 [DOI] [PMC free article] [PubMed] [Google Scholar]
  126. Talmi D., Dayan P., Kiebel S. J., Frith C. D., Dolan R. J. (2009). How humans integrate the prospects of pain and reward during choice. J. Neurosci. 29, 14617–14626 10.1523/JNEUROSCI.2026-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  127. Tanaka S. C., Doya K., Okada G., Ueda K., Okamoto Y., Yamawaki S. (2004). Prediction of immediate and future rewards differentially recruits cortico-basal ganglia loops. Nat. Neurosci. 7, 887–893 10.1038/nn1279 [DOI] [PubMed] [Google Scholar]
  128. Tanaka S. C., Schweighofer N., Asahi S., Shishida K., Okamoto Y., Yamawaki S., et al. (2007). Serotonin differentially regulates short- and long-term prediction of rewards in the ventral and dorsal striatum. PLoS ONE 2, e1333. 10.1371/journal.pone.0001333 [DOI] [PMC free article] [PubMed] [Google Scholar]
  129. Thaler R. (1981). Some empirical-evidence on dynamic inconsistency. Econ. Lett. 8, 201–207 10.1016/0165-1765(81)90067-7 [DOI] [Google Scholar]
  130. Thurston D. L. (2006). “Multi-attribute utility analysis of conflicting preferences,” in Decision Making in Engineering Design eds Lewis K. E., Chen W., Schmidt L. C. (New York: ASME Press; ), 125–133 [Google Scholar]
  131. Tobler P. N., O’Doherty J. P., Dolan R. J., Schultz W. (2007). Reward value coding distinct from risk attitude-related uncertainty coding in human reward systems. J. Neurophysiol. 97, 1621–1632 10.1152/jn.00745.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  132. Tom S. M., Fox C. R., Trepel C., Poldrack R. A. (2007). The neural basis of loss aversion in decision-making under risk. Science 315, 515–518 10.1126/science.1134239 [DOI] [PubMed] [Google Scholar]
  133. Treadway M. T., Buckholtz J. W., Cowan R. L., Woodward N. D., Li R., Ansari M. S., Baldwin R. M., Schwartzman A. N., Kessler R. M., Zald D. H. (2012). Dopaminergic mechanisms of individual differences in human effort-based decision-making. J. Neurosci. 32, 6170–6176 10.1523/JNEUROSCI.6459-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  134. Tversky A., Kahneman D. (1981). The framing of decisions and the psychology of choice. Science 211, 453–458 10.1126/science.7455683 [DOI] [PubMed] [Google Scholar]
  135. Tversky A., Kahneman D. (1992). Advances in prospect-theory – cumulative representation of uncertainty. J. Risk Uncertain 5, 297–323 10.1007/BF00122574 [DOI] [Google Scholar]
  136. Vlaev I., Chater N., Stewart N., Brown G. D. (2011). Does the brain calculate value? Trends. Cogn. Sci. 15, 546–554 10.1016/j.tics.2011.09.008 [DOI] [PubMed] [Google Scholar]
  137. Von Neumann J., Morgenstern O. (1947). Theory of Games and Economic Behavior. Princeton: Princeton University Press [Google Scholar]
  138. Walton M. E., Bannerman D. M., Alterescu K., Rushworth M. F. S. (2003). Functional specialization within medial frontal cortex of the anterior cingulate for evaluating effort-related decisions. J. Neurosci. 23, 6475–6479 [DOI] [PMC free article] [PubMed] [Google Scholar]
  139. Walton M. E., Bannerman D. M., Rushworth M. F. S. (2002). The role of rat medial frontal cortex in effort-based decision making. J. Neurosci. 22, 10996–11003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  140. Weber B. J., Chapman G. B. (2005). The combined effects of risk and time on choice: does uncertainty eliminate the immediacy effect? Does delay eliminate the certainty effect? Organ. Behav. Hum. Decis. Process. 96, 104–118 10.1016/j.obhdp.2005.01.001 [DOI] [Google Scholar]
  141. Winstanley C. A., Eagle D. M., Robbins T. W. (2006). Behavioral models of impulsivity in relation to ADHD: translation between clinical and preclinical studies. Clin. Psychol. Rev. 26, 379–395 10.1016/j.cpr.2006.01.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  142. Wittmann M., Leland D. S., Paulus M. P. (2007). Time and decision making: differential contribution of the posterior insular cortex and the striatum during a delay discounting task. Exp. Brain Res. 179, 643–653 10.1007/s00221-006-0822-y [DOI] [PubMed] [Google Scholar]

Articles from Frontiers in Neuroscience are provided here courtesy of Frontiers Media SA

RESOURCES