Skip to main content
Proceedings of the National Academy of Sciences of the United States of America logoLink to Proceedings of the National Academy of Sciences of the United States of America
. 2010 Nov 15;107(49):20911–20916. doi: 10.1073/pnas.1013305107

Neural basis of the undermining effect of monetary reward on intrinsic motivation

Kou Murayama a,1, Madoka Matsumoto b,c,d, Keise Izuma b,c, Kenji Matsumoto b,d,1
PMCID: PMC3000299  PMID: 21078974

Abstract

Contrary to the widespread belief that people are positively motivated by reward incentives, some studies have shown that performance-based extrinsic reward can actually undermine a person's intrinsic motivation to engage in a task. This “undermining effect” has timely practical implications, given the burgeoning of performance-based incentive systems in contemporary society. It also presents a theoretical challenge for economic and reinforcement learning theories, which tend to assume that monetary incentives monotonically increase motivation. Despite the practical and theoretical importance of this provocative phenomenon, however, little is known about its neural basis. Herein we induced the behavioral undermining effect using a newly developed task, and we tracked its neural correlates using functional MRI. Our results show that performance-based monetary reward indeed undermines intrinsic motivation, as assessed by the number of voluntary engagements in the task. We found that activity in the anterior striatum and the prefrontal areas decreased along with this behavioral undermining effect. These findings suggest that the corticobasal ganglia valuation system underlies the undermining effect through the integration of extrinsic reward value and intrinsic task value.

Keywords: crowding-out effect, dopamine, midbrain, neuroeconomics


Performance-based incentive systems have long been part of the currency of schools and workplaces. This predominance of incentive systems may reflect a widespread cultural belief that performance-based reward is a reliable and effective way to enhance motivation in students and workers. However, classic psychological experiments have repeatedly revealed that performance-based reward can also undermine people's intrinsic motivation (16), that is, motivation to voluntarily engage in a task for the inherent pleasure and satisfaction derived from the task itself (35). In a typical experiment of this “undermining effect” [also called the “motivation crowding-out effect” (79) or “overjustification effect” (2)], participants are randomly divided into a performance-based reward group and a control group, and both groups work on an interesting task. Participants in the performance-based reward group obtain (or expect) reward contingent on their performance, whereas participants in the control group do not. After the session, participants are left to engage in any activity, including more of the target task if they wish, for a brief period when they believe they are no longer being observed (i.e., “free-choice period”). A number of studies (46) found that the performance-based reward group spends significantly less time than the control group engaging in the target activity during the free-choice period, providing evidence that the performance-based reward undermines voluntary engagement in the task (i.e., intrinsic motivation for the task).

The undermining effect challenges normative economic theories, which assume that raising monetary incentives monotonically increases motivation and, more importantly, that increasing and then removing monetary incentives does not disturb underlying intrinsic motivation (79). It also challenges traditional operant learning theory and reinforcement learning theory, which currently constitutes the fundamental theoretical framework for human decision making (1012). These theories basically predict that performance-based rewards increase the likelihood that the behavior will be voluntarily performed again. It should be noted that, as thoroughly discussed in the literature (4, 5, 13), the undermining effect cannot be explained by the operant conditioning concept of extinction as a result of the withdrawal of reward (i.e., because the reward is no longer promised in the free-choice period, the reinforced response is extinguished and this produces the undermining effect) in several respects. Most importantly, the extinction account predicts that, when rewards are no longer in effect, behavior should revert to its original rate (the baseline) and never decrease below that level (14, 15). Studies on the undermining effect, however, showed less voluntary engagement in the target task in the performance-based reward group than in the control group, which serves as the strict baseline in randomized control design (5, 16). Given that normative economic theories and standard reinforcement learning theory have difficulty explaining the undermining effect, a better understanding of this effect has the potential to enrich and give new insight to these broad research fields (17). However, the neural basis of this provocative and important phenomenon remains unknown.

A source of intrinsic motivation is the intrinsic value of achieving success on a given task (3, 5). As such, the undermining effect may involve the interaction of two different types of subjective values when one succeeds at a task: the extrinsic value of obtaining a reward and the intrinsic value of achieving success. Many neuroscience studies have revealed that a dopaminergic reward network plays a pivotal role in representing and updating various types of subjective valuation (1012, 1824). In particular, recent studies have suggested that activation in response to feedback in the anterior part of the striatum (caudate head) is modulated by one's subjective belief in determining the outcome (23, 25), which is considered a key psychological factor in the undermining effect (35). Previous studies have also suggested that the midbrain, which has a strong anatomical connection with the anterior striatum (19), is responsive to both monetary reward feedback and cognitive feedback (feedback without monetary reward)—feedbacks that are related to the undermining effect (26, 27). Therefore, we expected that the undermining effect may manifest as brain activity changes in the reward network, especially in the anterior striatum and midbrain, in response to task feedback.

We expected the lateral prefrontal cortex (LPFC) to be another key structure mediating the undermining effect. When confronting an upcoming task, people tend to be more engaged in mental preparation for tasks with higher value (28, 29). As the LPFC is the center for the preparatory cognitive control to achieve goals (3034), and this function has been shown to be modulated by task value (28, 34), the undermining effect may be accompanied by a decrease in LPFC activity upon presentation of the task cue. Here, by using functional MRI (fMRI), we report evidence that these areas are involved in the undermining effect of monetary reward on intrinsic motivation.

The undermining effect is applicable only to interesting tasks that have an intrinsic value of achieving success (36). We developed a stopwatch (SW) task in which participants were presented with an SW that starts automatically, and the goal was to press a button with the right thumb so that the button press fell within 50 ms of the 5-s time point (Fig. 1A). A point was added to their score when they succeeded. A series of pilot studies showed that the SW task is moderately challenging and inherently interesting to Japanese university students (details provided in Materials and Methods). The control task was a watch-stop (WS) task, in which participants passively viewed a SW and were asked to simply press a button when it automatically stopped (Fig. 1A). Success and failure were not defined in this task; therefore the WS task was less interesting than the SW task. Both tasks were pseudorandomly intermixed and preceded by a cue that indicated which task to perform.

Fig. 1.

Fig. 1.

Experimental protocol and behavioral results. (A) Illustration of SW and WS tasks. (B) Depiction of the experimental procedure. (C) Means and SEs of the number of times participants voluntarily played the SW and WS tasks during the first and the second free-choice periods. Performance-based reward undermined the intrinsic motivation for the SW task for both free-choice periods (Mann–Whitney U = 52.5 and 54.5; P < 0.05).

Twenty-eight participants were randomly assigned to a control group or a reward group. Participants were scanned in two separate sessions (Fig. 1B). Before the first session, participants in the reward group were told that they would obtain 200 Japanese yen (approximately $2.20) for each successful trial of the SW task, and indeed they received the performance-based reward after the session. Participants in the control group were told nothing about the performance-based reward and received money just for task participation after the first session. For each control group participant, the amount of monetary reward for the task was matched to that received by another participant of the same sex in the reward group; thus, it was unrelated to the control participant's own task performance. This allowed us to examine the effect of performance-based reward apart from the amount of monetary reward offered. After being released from the scanner and receiving the monetary reward, participants were left alone in a quiet room for 3 min, where they could freely spend time playing the SW or WS task on a computer, read several booklets, or anything else (i.e., free-choice period). The number of times participants played the SW task during this free-choice period was used as the index of intrinsic motivation toward the task (16).

To track the brain activity associated with the undermining effect, we asked participants in both groups to perform the SW and WS tasks again after the free-choice period and without performance-based reward inside the scanner (second session; Fig. 1B). Both groups of participants were explicitly told in advance that no performance-based rewards would be provided. After being released from the scanner, the second free-choice period followed to confirm that the undermining effect persisted through the second scanning session.

Results

Behavioral Results.

We conducted a 2 × 2 mixed ANOVA on the number of times the voluntary SW task was played, with period (first or second free-choice period; within-subject) and group (control or reward group; between-subject) as factors. As predicted, the main effect of group was significant, (F1,26 = 6.59, P = 0.016). This result indicates the presence of the undermining effect: participants in the reward group played the SW task during the free-choice period significantly fewer times than did those in the control group (Fig. 1C). A significant group difference was observed in both the first and the second free-choice periods (P < 0.05; SI Results). To the contrary, neither the main effect of period nor the session-by-group interaction was significant (F < 1, P > 0.32). This suggests that no overall increase or decrease in the voluntary SW task play was observed and that the pattern of change in the voluntary SW task play did not differ across the groups. In fact, we observed no significant increase or decrease of the voluntary SW task play from the first to the second free-choice period in either group (P > 0.19).

We also conducted the same 2 × 2 mixed ANOVA on the number of times participants played the WS task during the free-choice period. Neither the main effects nor the interaction was significant (F < 1, P > 0.34; Fig. 1C). The numbers were quite small, suggesting that the WS task was not interesting to the participants.

fMRI Results.

In the fMRI analysis, we were interested in finding significant session-by-group interactions, which means that changes in activation across sessions showed different patterns between the two groups. Thus, we applied a 2 × 2 mixed ANOVA with session (first or second session) and group (control or reward group) as factors. The significant interactions reported here were based on the regions that survived both a whole-brain analysis (P < 0.001, uncorrected) and small-volume correction analysis (P < 0.05; details in Materials and Methods).

We first focused on a feedback period to examine the neural responses to the success feedback versus the failure feedback. A one-sample t test in the first session showed that the bilateral anterior striatum (caudate head) and midbrain were significantly activated, regardless of the group (P values < 0.05, small-volume corrected). This result indicates that the success feedback in the experimental task we developed involves reward network activation, regardless of whether the feedback was accompanied with monetary reward. This is consistent with previous work (21, 23, 25, 26) and supports the validity of our experimental task for examining brain activation in response to task feedback.

In the 2 × 2 ANOVA, as expected, the bilateral striatum activation showed a significant interaction between session (first or second session) and group (control or reward group) that is a striking parallel with the behavioral undermining effect (P < 0.05, small-volume-corrected; Fig. 2 and Fig. S1). During the first session, significant anterior striatum activation was observed in both groups: one-sample t13 values of 6.61 (control) and 8.43 (reward); P < 0.01 for both. However, the activation was significantly greater in the reward group than in the control group: two-sample t26 = 3.30 (P < 0.01). Previous studies have implied that the striatum functions as a hub of the human valuation process, by converting and integrating different types of reward values onto a common scale (11). Our result can be interpreted by this view such that the significant positive activation in the control group reflects the intrinsic value of achieving success (23, 25) and this activation was elevated by the additional performance-based monetary reward in the reward group. Importantly, whereas this activity during the second session was sustained in the control group (one-sample t13 = 7.33, P < 0.01; no between-session change was observed, P = 0.41), there was a dramatic decrease in activation of the bilateral anterior striatum in the reward group, and the activation was no longer significant (one-sample t13 = 0.41, P = 0.69; decrease in the activity from the first to the second session was significant, paired t13 = 7.35, P < 0.01). As a result, the between-group difference in the anterior striatal activation was reversed from the first session to the second session and became significantly smaller in the reward group compared with the control group during the second session (two-sample t26 = 3.75, P < 0.01). Also as predicted, the midbrain showed a similar pattern of interaction (P < 0.05, small-volume-corrected; Fig. 3), consistent with the strong anatomical connection between the midbrain and anterior striatum (19, 26).

Fig. 2.

Fig. 2.

Bilateral striatum responses elicited by success trials relative to failure trials plotted for each session/group. Left: Activations superimposed on transaxial sections (P < 0.001, one-sample t test for display). Right: Mean contrast values and SEs of the bilateral striatum (averaged) activation are plotted. During the first session, significant bilateral striatum activation was observed in both groups, although the activation was significantly greater in the reward group than in the control group (two-sample t26 = 3.30, P < 0.01). In contrast, during the second session, whereas the control group sustained significant activity, the activation of the bilateral striatum in the reward group decreased significantly below that of the control group (two-sample t26 = 3.75, P < 0.01) and the activation was no longer significant. This striatal response pattern is in parallel with the behavioral undermining effect. Asterisks represent the statistical significance of one-sample/two-sample t tests (**P < 0.01).

Fig. 3.

Fig. 3.

Midbrain activation (peak at −9, −7, −11) detected in the session-by-group interaction during the feedback period (success trials minus failure trials; P < 0.05, small-volume-corrected; the image is shown at P < 0.001, uncorrected). Neural responses are displayed in sagittal and transaxial formats. The midbrain was activated when performance-based monetary reward was expected (during the first session; two-sample t26 = 1.80, P < 0.10), but the activation decreased significantly below the control group in the second session (two-sample t26 = 2.63, P < 0.05). Asterisks represent the statistical significance of one-sample/two-sample t tests (+P < 0.10, *P < 0.05, **P < 0.01).

We next focused on a task cue period to investigate the brain activity associated with preparatory cognitive control in the SW task relative to the WS task. A one-sample t test in the first session revealed that the right LPFC was significantly activated regardless of the group (P < 0.05, small-volume corrected). This result indicates that the participants were cognitively engaged in the SW task relative to the WS task when the task cue was presented. This is consistent both with the observation that participants were more willing to engage in the SW task in the free-choice periods and with previous findings that the LPFC is activated in response to a task cue with high value (28, 34). In other words, this finding supports the validity of our experimental task for examining LPFC activation in response to task cue.

In the 2 × 2 ANOVA, as expected, the right LPFC showed a significant session-by-group interaction that is also a parallel with the behavioral undermining effect (P < 0.05, small-volume-corrected; Fig. 4). During the first session, the LPFC in the reward group showed significantly larger activation than that in the control group (two-sample t26 = 2.62, P < 0.05), suggesting that participants in the reward group prepared for the SW task more actively than those in the control group when they saw a task cue. However, during the second session, although significant activity in the control condition was sustained in the second session (one-sample t13 = 2.53, P < 0.05), the activation became significantly smaller in the reward group than in the control group (two-sample t26 = 2.27, P < 0.05), and the activation was no longer significant (P = 0.43). This result may indicate that the participants in the reward group were not motivated to prepare for the SW task during the second session in comparison with the control group participants. The bilateral striatum also showed a significant interaction for the task cue period, but unlike the activation pattern in the feedback period, no significant between-group difference in activation was detected during the second session (Fig. S2).

Fig. 4.

Fig. 4.

Right LPFC activation (peak at 39, 41, 40) detected in the session-by-group interaction during the task cue period (P < 0.05, small-volume-corrected; image is shown at P < 0.001, uncorrected for display). Neural responses are displayed in transaxial and coronal formats. The bar plot represents mean contrast values and SEs for each session/group. During the first session, the LPFC in the reward group showed significantly larger activation than that in the control group (two-sample t26 = 2.62, P < 0.05). However, the activation became significantly smaller in the reward group than in the control group during the second session (two-sample t26 = 2.27, P < 0.05).

Table S1 (for the feedback period) and Table S2 (for the task cue period) list all regions displaying a significant session-by-group interaction in a whole-brain analysis (P < 0.001, uncorrected, k > 5). The tables also describe the results of simple main effect analyses and one-sample t tests that quantify the pattern of interaction as we conducted for the striatum, midbrain, and LPFC (SI Results includes additional analyses focusing on a possible sex effect).

Brain–Behavior Relation.

We conjecture that the observed decreases in activity of the anterior striatum, midbrain, and LPFC are collectively related to the undermining effect. In fact, the magnitudes of the decreases in activation in these regions were highly correlated (mean r = 0.65). Accordingly, we calculated a “neural undermining index”—a composite score representing the between-session decreases in activity for these regions, and regressed it on the voluntary SW task play during the free-choice period. Specifically, we computed the magnitude of a decrease in activation by subtracting the contrast value in the second session from the contrast value in the first session for each region of interest (i.e., the striatum and the midbrain for feedback period and the LPFC for cue period) and submitted these values to principal component analysis. Principal component analysis is a statistical technique to compute optimal and reliable composite scores of a set of variables that are less susceptible to random noise by taking into account their variance and covariance (35). The first principal component explained a substantial portion (74%) of the total variance, and we used this component score as the neural undermining index.

As expected, regression analysis revealed a significant negative relationship between the amount of voluntary SW play and the neural undermining index in the reward group (standardized β = −0.49, P = 0.037, one-tailed), indicating that those who did not voluntarily try the SW task during the free choice period showed a larger decrease in activation of the corticobasal ganglia network (Fig. 5). The regression coefficient remained significant even when the confidence interval was based on a (bias-corrected) bootstrapping method to correct for the potential statistical biases resulting from nonnormality and outliers (36). The magnitude of relationship is large according to the Cohen established effect size criterion (37). In contrast, the relationship in the control group was not significant (standardized β = 0.09, P = 0.75). An additional regression analysis including group, the number of voluntary SW play trials, and their interaction as independent variables showed significant interaction (P = 0.045), indicating that the aforementioned regression coefficients in the reward and control groups were significantly different.

Fig. 5.

Fig. 5.

Relationship between behavioral choice during the first free-choice period and the neural undermining index. Significant negative relationship was observed in the reward group (β = −0.49, P = 0.037, one-tailed), indicating that those who did not voluntarily try the SW task during the free-choice period showed a larger decrease in activation of the corticobasal ganglia network. The relationship in the control group was not significant (β = 0.09, P = 0.75). Blue triangles represent participants in the control group; red squares represent participants in the reward group.

Discussion

Our study provides evidence that the corticobasal ganglia valuation system plays a central role in the undermining effect. Specifically, our neuroimaging results suggest that, when performance-based reward is no longer promised, (i) people do not feel subjective value in succeeding in the task, as indicated by the dramatic decreases in the activation of the striatum and midbrain in response to the success feedback; and (ii) they are not motivated to show cognitive engagement in facing the task, as indicated by the decrease in the LPFC activation in response to the task cue. A number of theories have been proposed to explain the undermining effect from value-based and cognitive perspectives (5, 6). Our findings clearly indicate that value-driven and cognitive processes are involved in the undermining effect, and they are linked. Notably, activation in the anterior part of the striatum, which has been implicated in subjective belief in determining outcomes (23, 25), corresponds particularly well to the pattern observed in the behavioral undermining effect. This lends support for the recent psychological theory that the undermining effect is closely linked to a decreased sense of self-determination (35).

The precise neural and computational mechanism that accounts for the striatal signal decrease in the reward group merits future inquiry. One explanation is that the striatum, in which incommensurable subjective values are aligned on a unidimensional common scale, integrates the intrinsic value of task success and monetary reward value through relative comparison and rescaling processes (38). Given the relatively stronger salience of monetary reward, the rescaled value of task success could become smaller than the original magnitude. In other words, the strong incentive value of monetary reward pushed down the intrinsic value of task success. As a result, when the monetary reward was no longer promised, the intrinsic task value was underestimated, resulting in decreased motivation relative to the control group (i.e., less frequent play of the SW task in the free-choice period). This interpretation underscores the importance for future empirical and theoretical work addressing the human value integration process (38, 39).

Neuroscience research has made considerable progress by incorporating concepts of motivation (40), yet most research to date has been confined to extrinsic rewards such as food or money. In comparison with our knowledge of extrinsic rewards, little light has been shed on intrinsic sources of motivation, and much less on the integration of the two. However, given the burgeoning of performance-based incentive systems in contemporary society, the interaction of these motivations is gaining practical importance in guiding human behavior. We believe an expanded understanding of this integration process is a key piece in aligning the undermining effect with leading economic and learning theories, and in reaching a deeper understanding of human behavior in general.

Materials and Methods

Participants.

Twenty-eight right-handed healthy participants [mean age, 20.6 ± 1.1 y (SD); 10 male and 18 female] recruited from a pool of Tamagawa University (Tokyo, Japan) students took part in the experiment. Participants were randomly assigned to a control group (n = 14) or a reward group (n = 14). All participants gave informed consent for the study and the protocol was approved by the Ethics Committee of Tamagawa University.

Experimental Tasks.

The undermining effect can be observed only when a task is interesting and has intrinsic value of achieving success; with boring tasks, there is little or no intrinsic motivation to undermine (5). Accordingly, an SW task was developed (Fig. 1A) to meet this criterion. A series of pilot studies were conducted to determine the time window for success so that participants can succeed on approximately half the trials on average. Previous literature indicated that people obtain the greatest sense of achievement for the tasks of intermediate difficulty (41, 42). In addition, this rate of success allows a sufficient number of success or failure trials to be obtained for proper fMRI statistical analysis. The participant's total score was displayed in the upper right corner of the display area, and when the participant succeeded in stopping the SW display between 4.95 s and 5.05 s, a point was added to their score (1,500 ms after the button press) and the updated score panel flashed for 1,500 ms. Another pilot study using an independent university student sample (n = 37) revealed that this task is sufficiently interesting without any extrinsic incentives (mean enjoyment rating, 4.14; SD, 0.82, on a five-point Likert scale). We also developed a WS task as the control task (Fig. 1A). Because success and failure were not defined in this task, no point was added on their response.

The experiment was composed of two separate scanning sessions (approximately 18 min each) and each session consists of 30 SW and 30 WS trials, which were pseudorandomly intermixed with the interstimulus interval alternated between 1,000 and 5,000 ms. Both tasks were preceded by a cue that indicates which of two tasks is to be performed in the next trial. The cue was presented for 1,500 ms and the SW starts 3,000 ms after the cue onset. The timing of the stop for a WS trial is approximately matched (alternating based on random noise) to the time achieved in a prespecified SW trial.

Experimental Procedure.

The experimental sessions were conducted individually. Upon arriving at the experiment room, participants were greeted by two experimenters. The experiment consisted of two separate sessions, each of which was conducted by one of the experimenters. We decided to use a different experimenter for each session to prevent the participants from being aware of the relationship between these two sessions. In the postexperimental interviews, no participant reported noticing the fact that the two experiments were conducted under a common purpose. Regardless of the experimental conditions, the payment for the task participation in the second session was provided to the participants before the experiment (2,000 Japanese yen, approximately equal to $22). This procedure allowed us to avoid any possible monetary incentive effects in the second session, the critical session to capture the brain activity of the undermining effect.

Before the first scanning session, participants in the reward group were informed that they would receive 200 Japanese yen (approximately equal to $2.20) for each point they obtained during the session. In contrast, no mention was made about the performance-based monetary reward in the control group. The instruction was provided through a computer program to prevent possible experimenter bias.

On completing the first scanning session, participants were released from the scanner and led to a small waiting room, where participants in the reward group were provided with the performance-based monetary reward. Participants in the control group received payment for task participation. For each participant in the control group, the amount of monetary reward for the task was matched to that of a participant in the reward group (i.e., both groups received the same amount of reward).

After the participants confirmed the amount of money they received, the experimenter asked the participants to wait for a couple of minutes because the other experimenter needed a few more minutes to prepare for the next experiment. There was a computer in the room and participants could freely choose to play the SW or WS task as many times as they wanted with that computer. There were also a few booklets on a desk and participants could freely read them. During this free-choice period, participants believed they were no longer observed by the experimenters, but the computer program confidentially recorded the number of SW and WS task trials they voluntarily played. The number of trials they played on the SW task was used as the index of intrinsic motivation. This is a standard way of assessing intrinsic motivation and has been used in many previous experiments on the undermining effect (15). After exactly 3 min, the door of the waiting room was opened and the participants were led to the scanning room again.

The second session was ostensibly conducted by the second experimenter. Before the task, participants were instructed that they would do the SW and WS tasks used in the previous experiment, but it was emphasized that the purpose of the experiment was completely independent. Both groups of participants were also explicitly told that no performance-based rewards would be provided in this experiment. As such, participants in both groups did not expect any performance-based rewards in the second session. After the scanning session, participants were released from the scanner and exposed to a 3-min free-choice period again.

fMRI Data Acquisition.

The functional imaging was conducted using a 3-T Trio A Tim MRI scanner (Siemens) to acquire gradient echo T2*-weighted echo-planar images (EPI) with blood oxygenation level-dependent contrast. Forty-two contiguous interleaved transversal slices of EPI images were acquired in each volume, with a slice thickness of 3 mm and no gap (repetition time, 2,500 ms; echo time, 25 ms; flip angle, 90°; field of view, 192 mm2; matrix, 64 × 64). Slice orientation was tilted −30° from the AC–PC line. We discarded the first three images before data processing and used statistical analysis to compensate for the T1 saturation effects.

fMRI Data Analysis.

Image analysis was performed by using Statistical Parametric Mapping software (version 8; http://www.fil.ion.ucl.ac.uk). Images were corrected for slice acquisition time within each volume, motion-corrected with realignment to the first volume, spatially normalized to the standard Montreal Neurological Institute EPI template, and spatially smoothed using a Gaussian kernel with a full width at half maximum of 8 mm.

For each participant, the blood oxygen-level dependent responses across the scanning run (including both sessions) were modeled with a general linear model. The model included the following regressors of interest: presentation of success feedback in the SW task, presentation of failure feedback in the SW task, presentation of SW task cue, and presentation of WS task cue. The motion parameters, error trials, and session effects were also included as regressors of no interest. The regressors (except for the motion parameters and the session effects) were calculated using a boxcar function convolved with a hemodynamic-response function. The estimates were corrected for temporal autocorrelation by using a first-order autoregressive model. To investigate the feedback effects and cue effects, our primary focus of interest, two contrast values were calculated: (i) contrast between success feedback and failure feedback effects (i.e., success minus failure), and (ii) contrast between SW task cue and WS task cue effects (i.e., SW minus WS).

We conducted a second-level, whole-brain 2 × 2 mixed ANOVA with session (first or second session; within-subject) and group (control or reward group; between-subject) as factors, once on the success/failure contrasts and once on the SW/WS contrasts. A number of regions showed a significant session-by-group interaction (P < 0.001, uncorrected, k > 5 voxels), including anterior striatum, midbrain (for the feedback period), and LPFC (for the task cue period), our primary region of interest (Tables S1 and S2). To confirm the reliability of the significant interaction effects obtained in the regions for which we had an a priori hypothesis (the anterior striatum and midbrain for the feedback period and the LPFC for the task cue period), we also performed a small-volume correction analysis with a corrected significance threshold of P < 0.05 within a 12-mm sphere centered on the coordinates identified in the previous empirical studies or meta-analyses (25, 26, 33, 43). All regions of interest survived this analysis.

To quantify the pattern of interaction, we further conducted a series of post-hoc analyses. Specifically, we extracted the contrast values from a 3-mm sphere centered on the peak voxel of each region using rfxplot (44), and subjected these extracted values to a series of simple main-effect analyses (i.e., test for the between-group difference within each session) and one-sample t tests (i.e., test for the significance of the absolute contrast values for each session/group).

Supplementary Material

Supporting Information

Acknowledgments

We thank J. Tanji, C. Camerer, J. O'Doherty, K. Samejima, H. Kim, A. Przybylski, and R. Pekrun for helpful comments; K. D'Ardenne and M. J. Tyszka for providing technical information; T. Haji, R. Iseki, S. Bray, and J. Gläscher for technical advice; J. Helen for English proofing; and Y. Otake for research assistance. This study was supported by Grant-in-Aid for Scientific Research C#21530773 (to K. Matsumoto), Grand-in-Aid for Scientific Research on Innovative Areas 22120515 (to K. Matsumoto); a Tamagawa University Global Center of Excellence grant from the Ministry of Education, Culture, Sports, Science and Technology, Japan; and an Alexander von Humboldt Foundation fellowship (to K. Murayama).

Footnotes

The authors declare no conflict of interest.

See Commentary on page 20849.

This article is a PNAS Direct Submission.

This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1013305107/-/DCSupplemental.

References

  • 1.Deci EL. Effects of externally mediated rewards on intrinsic motivation. J Pers Soc Psychol. 1971;18:105–115. [Google Scholar]
  • 2.Lepper MR, Greene D, Nisbett RE. Undermining children's intrinsic interest with extrinsic rewards: A test of the “overjustification” hypothesis. J Pers Soc Psychol. 1973;28:129–137. [Google Scholar]
  • 3.Ryan RM, Mims V, Koestner R. Relation of reward contingency and interpersonal context to intrinsic motivation: A review and test using cognitive evaluation theory. J Pers Soc Psychol. 1983;45:736–750. [Google Scholar]
  • 4.Deci EL, Koestner R, Ryan RM. A meta-analytic review of experiments examining the effects of extrinsic rewards on intrinsic motivation. Psychol Bull. 1999;125:627–668, discussion 692–700. doi: 10.1037/0033-2909.125.6.627. [DOI] [PubMed] [Google Scholar]
  • 5.Deci EL, Ryan RM. Intrinsic Motivation and Self-Determination in Human Behavior. New York: Plenum; 1985. [Google Scholar]
  • 6.Morgan M. Reward-induced decrements and increments in intrinsic motivation. Rev Educ Res. 1984;54:5–30. [Google Scholar]
  • 7.Camerer CF, Hogarth RM. The effects of financial incentives in experiments: A review and capital-labor-production framework. J Risk Uncertain. 1999;19:7–42. [Google Scholar]
  • 8.Frey BS, Jegen R. Motivation crowding theory. J Econ Surv. 2001;15:589–611. [Google Scholar]
  • 9.Kreps D. Intrinsic motivation and extrinsic incentives. Am Econ Rev. 1997;87:359–364. [Google Scholar]
  • 10.Rushworth MFS, Mars RB, Summerfield C. General mechanisms for making decisions? Curr Opin Neurobiol. 2009;19:75–83. doi: 10.1016/j.conb.2009.02.005. [DOI] [PubMed] [Google Scholar]
  • 11.Montague PR, Berns GS. Neural economics and the biological substrates of valuation. Neuron. 2002;36:265–284. doi: 10.1016/s0896-6273(02)00974-1. [DOI] [PubMed] [Google Scholar]
  • 12.Dayan P, Niv Y. Reinforcement learning: the good, the bad and the ugly. Curr Opin Neurobiol. 2008;18:185–196. doi: 10.1016/j.conb.2008.08.003. [DOI] [PubMed] [Google Scholar]
  • 13.Grolnick WS. The Psychology of Parental Control: How Well Mean Parenting Backfires. Hillsdale, NJ: Lawrence Erlbaum Associates; 2003. [Google Scholar]
  • 14.Skinner BF. The Behavior of Organisms: An Experimental Analysis. New York: Appleton-Century-Crofts; 1938. [Google Scholar]
  • 15.Hull C. Principles of Behavior. New York: Appleton-Century-Crofts; 1943. [Google Scholar]
  • 16.Kirk RE. Experimental Design: Procedures for the Behavioral Sciences. 3rd Ed. Pacific Grove, CA,: Brooks/Cole; 1995. [Google Scholar]
  • 17.Camerer CF, Loewenstein G, Rabin M. Advances in Behavioral Economics. Princeton, NJ: Princeton Univ Press; 2004. [Google Scholar]
  • 18.Seymour B, et al. Temporal difference models describe higher-order learning in humans. Nature. 2004;429:664–667. doi: 10.1038/nature02581. [DOI] [PubMed] [Google Scholar]
  • 19.Haber SN, Knutson B. The reward circuit: Linking primate anatomy and human imaging. Neuropsychopharmacology. 2010;35:4–26. doi: 10.1038/npp.2009.129. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Kringelbach ML, Berridge KC. Pleasures of the Brain. London: Oxford Univ Press; 2010. [Google Scholar]
  • 21.Izuma K, Saito DN, Sadato N. Processing of social and monetary rewards in the human striatum. Neuron. 2008;58:284–294. doi: 10.1016/j.neuron.2008.03.020. [DOI] [PubMed] [Google Scholar]
  • 22.Tobler PN, Fletcher PC, Bullmore ET, Schultz W. Learning-related human brain activations reflecting individual finances. Neuron. 2007;54:167–175. doi: 10.1016/j.neuron.2007.03.004. [DOI] [PubMed] [Google Scholar]
  • 23.Tricomi EM, Delgado MR, Fiez JA. Modulation of caudate activity by action contingency. Neuron. 2004;41:281–292. doi: 10.1016/s0896-6273(03)00848-1. [DOI] [PubMed] [Google Scholar]
  • 24.Schultz W. Behavioral theories and the neurophysiology of reward. Annu Rev Psychol. 2006;57:87–115. doi: 10.1146/annurev.psych.56.091103.070229. [DOI] [PubMed] [Google Scholar]
  • 25.Tricomi E, Delgado MR, McCandliss BD, McClelland JL, Fiez JA. Performance feedback drives caudate activation in a phonological learning task. J Cogn Neurosci. 2006;18:1029–1043. doi: 10.1162/jocn.2006.18.6.1029. [DOI] [PubMed] [Google Scholar]
  • 26.D'Ardenne K, McClure SM, Nystrom LE, Cohen JD. BOLD responses reflecting dopaminergic signals in the human ventral tegmental area. Science. 2008;319:1264–1267. doi: 10.1126/science.1150605. [DOI] [PubMed] [Google Scholar]
  • 27.Aron AR, et al. Human midbrain sensitivity to cognitive feedback and uncertainty during classification learning. J Neurophysiol. 2004;92:1144–1152. doi: 10.1152/jn.01209.2003. [DOI] [PubMed] [Google Scholar]
  • 28.Jimura K, Locke HS, Braver TS. Prefrontal cortex mediation of cognitive enhancement in rewarding motivational contexts. Proc Natl Acad Sci USA. 2010;107:8871–8876. doi: 10.1073/pnas.1002007107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Castel AD, Balota DA, McCabe DP. Memory efficiency and the strategic control of attention at encoding: impairments of value-directed remembering in Alzheimer's disease. Neuropsychology. 2009;23:297–306. doi: 10.1037/a0014888. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Matsumoto K, Suzuki W, Tanaka K. Neuronal correlates of goal-based motor selection in the prefrontal cortex. Science. 2003;301:229–232. doi: 10.1126/science.1084204. [DOI] [PubMed] [Google Scholar]
  • 31.Bunge SA. How we use rules to select actions: A review of evidence from cognitive neuroscience. Cogn Affect Behav Neurosci. 2004;4:564–579. doi: 10.3758/cabn.4.4.564. [DOI] [PubMed] [Google Scholar]
  • 32.Miller EK, Cohen JD. An integrative theory of prefrontal cortex function. Annu Rev Neurosci. 2001;24:167–202. doi: 10.1146/annurev.neuro.24.1.167. [DOI] [PubMed] [Google Scholar]
  • 33.Wager TD, Smith EE. Neuroimaging studies of working memory: A meta-analysis. Cogn Affect Behav Neurosci. 2003;3:255–274. doi: 10.3758/cabn.3.4.255. [DOI] [PubMed] [Google Scholar]
  • 34.Leon MI, Shadlen MN. Effect of expected reward magnitude on the response of neurons in the dorsolateral prefrontal cortex of the macaque. Neuron. 1999;24:415–425. doi: 10.1016/s0896-6273(00)80854-5. [DOI] [PubMed] [Google Scholar]
  • 35.Dunteman GH. Principal Component Analysis. Thousand Oaks, CA: Sage; 1989. [Google Scholar]
  • 36.Efron B, Tibshirani R. An Introduction to the Bootstrap. Boca Raton, FL: Chapman & Hall/CRC; 1993. [Google Scholar]
  • 37.Cohen J. Statistical Power Analysis for the Behavioral Sciences. 2nd Ed. Hillsdale, NJ,: Lawrence Erlbaum Associates; 1988. [Google Scholar]
  • 38.Seymour B, McClure SM. Anchors, scales and the relative coding of value in the brain. Curr Opin Neurobiol. 2008;18:173–178. doi: 10.1016/j.conb.2008.07.010. [DOI] [PubMed] [Google Scholar]
  • 39.FitzGerald TH, Seymour B, Dolan RJ. The role of human orbitofrontal cortex in value comparison for incommensurable objects. J Neurosci. 2009;29:8388–8395. doi: 10.1523/JNEUROSCI.0717-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Berridge KC. Motivation concepts in behavioral neuroscience. Physiol Behav. 2004;81:179–209. doi: 10.1016/j.physbeh.2004.02.004. [DOI] [PubMed] [Google Scholar]
  • 41.Atkinson JW. Motivational determinants of risk-taking behavior. Psychol Rev. 1957;64:359–372. doi: 10.1037/h0043445. [DOI] [PubMed] [Google Scholar]
  • 42.Csikszentmihalyi M. Flow: The Psychology of Optimal Experience. New York: Harper and Row; 1990. [Google Scholar]
  • 43.Tricomi E, Fiez JA. Feedback signals in the caudate reflect goal achievement on a declarative memory task. Neuroimage. 2008;41:1154–1167. doi: 10.1016/j.neuroimage.2008.02.066. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Gläscher J. Visualization of group inference data in functional neuroimaging. Neuroinformatics. 2009;7:73–82. doi: 10.1007/s12021-008-9042-x. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supporting Information

Articles from Proceedings of the National Academy of Sciences of the United States of America are provided here courtesy of National Academy of Sciences

RESOURCES