Common and distinct networks underlying reward valence and processing stages: A meta-analysis of functional neuroimaging studies (2011))

Published in final edited form as:

PMCID: PMC3395003

NIHMSID: NIHMS261816

The publisher’s final edited version of this article is available at Neurosci Biobehav Rev

See other articles in PMC that cite the published article.

Go to:

Abstract

To better understand the reward circuitry in human brain, we conducted activation likelihood estimation (ALE) and parametric voxel-based meta-analyses (PVM) on 142 neuroimaging studies that examined brain activation in reward-related tasks in healthy adults. We observed several core brain areas that participated in reward-related decision making, including the nucleus accumbens (NAcc), caudate, putamen, thalamus, orbitofrontal cortex (OFC), bilateral anterior insula, anterior (ACC) and posterior (PCC) cingulate cortex, as well as cognitive control regions in the inferior parietal lobule and prefrontal cortex (PFC). The NAcc was commonly activated by both positive and negative rewards across various stages of reward processing (e.g., anticipation, outcome, and evaluation). In addition, the medial OFC and PCC preferentially responded to positive rewards, whereas the ACC, bilateral anterior insula, and lateral PFC selectively responded to negative rewards. Reward anticipation activated the ACC, bilateral anterior insula, and brain stem, whereas reward outcome more significantly activated the NAcc, medial OFC, and amygdala. Neurobiological theories of reward-related decision making should therefore distributed and interrelated representations of reward valuation and valence assessment into account.

Keywords: meta-analysis, reward, nucleus accumbens, orbitofrontal cortex, anterior cingulate cortex, anterior insula

1. Introduction

People face countless reward-related decision making opportunities everyday. Our physical, mental, and socio-economical well-being critically depends on the consequences of the choices we make. It is thus crucial to understand what underlies normal functioning of reward-related decision making. Studying the normal functioning of reward-related decision making also helps us to better understand the various behavioral and mental disorders which arise when such function is disrupted, such as depression (Drevets, 2001), substance abuse (Bechara, 2005; Garavan and Stout, 2005; Volkow et al., 2003), and eating disorders (Kringelbach et al., 2003; Volkow and Wise, 2005).

Functional neuroimaging research on reward has become a rapidly growing field. We have observed a huge surge of neuroimaging research in this domain, with dozens of relevant articles showing up in the PubMed database every month. On the one hand, this is exciting because the mounting results are paramount to formalizing behavioral and neural mechanisms of reward-related decision making (Fellows, 2004; Trepel et al., 2005). On the other hand, the heterogeneity of the results in conjunction with the occasional opposing patterns make it difficult to obtain a clear picture of the reward circuitry in human brain. The mixture of results is partly due to diverse experimental paradigms developed by various research groups that aimed to address different aspects of reward-related decision making, such as the distinction between reward anticipation and outcome (Breiter et al., 2001; Knutson et al., 2001b; McClure et al., 2003; Rogers et al., 2004), valuation of positive and negative rewards (Liu et al., 2007; Nieuwenhuis et al., 2005; O’Doherty et al., 2003a; O’Doherty et al., 2001; Ullsperger and von Cramon, 2003), and assessment of risk (Bach et al., 2009; d’Acremont and Bossaerts, 2008; Hsu et al., 2009; Huettel, 2006).

Therefore, it is crucial to pool existing studies together and examine the core reward networks in human brain, from both data-driven and theory-driven approaches to test the commonality and distinction of different aspects of reward-related decision making. To achieve this goal, we employed and compared two coordinate-based meta-analysis (CBMA) methods (Salimi-Khorshidi et al., 2009), activation likelihood estimation (ALE) (Laird et al., 2005; Turkeltaub et al., 2002) and parametric voxel-based meta-analysis (PVM) (Costafreda et al., 2009), so as to reveal the concordance across a large number of neuroimaging studies on reward-related decision making. We anticipated that the ventral striatum and orbitofrontal cortex (OFC), two major dopaminergic projection areas that have been associated with reward processing, would be consistently activated.

In addition, from a theory-driven perspective, we aimed to elucidate whether there exist distinctions in the brain networks that are responsible for processing positive and negative reward information, and that are preferentially involved in different stages of reward processing such as reward anticipation, outcome monitoring, and decision evaluation. Decision making involves encoding and representation of the alternative options and comparing the values or utilities associated with these options. Across these processes, decision making is usually affiliated with positive or negative valence from either the outcomes or emotional responses toward the choices made. Positive reward valence refers to the positive subjective states we experience (e.g., happiness or satisfaction) when the outcome is positive (e.g., winning a lottery) or better than we anticipate (e.g., losing less value than projected). Negative reward valence refers to the negative feelings we go through (e.g., frustration or regret) when the outcome is negative (e.g., losing a gamble) or worse than what we expect (e.g., stock value increasing lower than projected). Although previous studies have attempted to distinguish reward networks that are sensitive to processing positive or negative information (Kringelbach, 2005; Liu et al., 2007), as well as those that are involved in reward anticipation or outcome (Knutson et al., 2003; Ramnani et al., 2004), empirical results have been mixed. We aimed to extract consistent patterns by pooling over a large number of studies examining these distinctions.

2. Methods

2.1 Literature search and organization

2.1.1 Study identification

Two independent researchers conducted a thorough search of the literature for fMRI studies examining reward-based decision making in humans. The terms used to search the online citation indexing service PUBMED (through June 2009) were “fMRI”, “reward”, and “decision” (by the first researcher), “reward decision making task”, “fMRI”, and “human” (by the second researcher). These initial search results were merged to yield a total of 182 articles. Another 90 articles were identified from a reference database of a third researcher accumulated through June 2009 using “reward” and “MRI” as filtering criteria. We also searched the BrainMap database using Sleuth, with “reward task” and “fMRI” as search terms, and found 59 articles. All of these articles were pooled into a database and redundant entries were eliminated. We then applied several exclusion criteria to further eliminate articles that are not directly relevant to the current study. These criteria are: 1) non-first hand empirical studies (e.g., review articles); 2) studies that did not report results in standard stereotactic coordinate space (either Talairach or Montreal Neurological Institute, MNI); 3) studies using tasks unrelated to reward or value-based decision making; 4) studies of structural brain analyses (e.g., voxel-based morphometry or diffusion tensor imaging); 5) studies purely based on region of interest (ROI) analysis (e.g., using anatomical masks or coordinates from other studies); 6) studies of special populations whose brain functions may be deviated from those of normal healthy adults (e.g., children, aging adults, or substance dependent individuals), although coordinates reported in these studies for the healthy adult group alone were included. Variability among methods with which subjects were instructed to report decisions during the tasks (i.e., verbal, nonverbal button-press) was accepted. This resulted in 142 articles in the final database (listed in the Appendix).

During the data extraction stage, studies were then grouped by different spatial normalization schemes according to coordinate transformations implemented in the GingerALE toolbox (http://brainmap.org, Research Imaging Center of the University of Texas Health Science Center, San Antonio, Texas): using FSL to report MNI coordinates, using SPM to report MNI coordinates, using other programs to report MNI coordinates, using Brett methods to convert MNI coordinates into Talairach space, using a Talairach native template. Lists of coordinates that were in Talairach space were converted into the MNI space according to their original normalization schemes. For the Brett-Talairach list, we converted the coordinates back into the MNI space using reverse transformation by Brett (i.e., tal2mni)(Brett et al., 2002). For the native Talairach list, we used BrainMap’s Talairach-MNI transformation (i.e., tal2icbm_other). A master list of all studies was created by combining all coordinates in MNI space in preparation for the ALE meta-analyses in GingerALE.

2.1.2 Experiment categorization

To test hypotheses with regards to the common and distinct reward pathways that are recruited by different aspects of reward-related decision making, we categorized coordinates according to two types of classification: reward valence and decision stages. We adopted the term of “experiments” used by the BrainMap database to refer to individual regressors or contrasts typically reported in fMRI studies. For reward valence, we organized the experiments into positive and negative rewards. For decision stages, we separated the experiments into reward anticipation, outcome, and evaluation. Coordinates in the master list that fit into these categories were put into sub-lists; those that were difficult to interpret or not clearly defined were omitted. Below we list some examples that were put into each of these categories.

The following contrasts were classified as processing of positive rewards: those in which subjects won money or points (Elliott et al., 2000)(reward during run of success); avoided losing money or points (Kim et al., 2006)(direct comparison between avoidance of an averse outcome and reward receipt); won the larger of two sums of money or points (Knutson et al., 2001a)(large vs. small reward anticipation); lost the smaller of two sums of money or points (Ernst et al., 2005)(no-win $0.50 > no-win $4); received encouraging words or graphics on the screen(Zalla et al., 2000) (increase for “win”); received sweet taste in their mouths (O’Doherty et al., 2002)(glucose > neutral taste); positively evaluated the choice (Liu et al., 2007)(right > wrong), or received any other type of positive rewards as a result of successful completion of the task.

Experiments classified for negative rewards included those in which subjects lost money or points (Elliott et al., 2000)(penalty during run of failure); did not win money or points (Ernst et al., 2005)(dissatisfaction of no-win); won the smaller of two sums of money or points (Knutson et al., 2001a)($1 vs. $50 reward); lost the larger of two sums of money or points (Knutson et al., 2001a)(large vs. small punishment anticipation); negatively evaluated the choice (Liu et al., 2007)(wrong > right); or received any other negative rewards such as the administration of a bitter taste in their mouths (O’Doherty et al., 2002)(salt > neutral taste) or discouraging words or images (Zalla et al., 2000)(increase for “lose” and decrease for “win”).

Reward anticipation was defined as the time period when the subject was pondering potential options before making a decision. For example, placing a bet and expecting to win money on that bet would be classified as anticipation (Cohen and Ranganath, 2005)(high-risk vs. low-risk decision). Reward outcome/delivery was classified as the period when the subject received feedback on the chosen option, such as a screen with the words “win x$” or “lose x$” (Bjork et al., 2004)(gain vs. non-gain outcome). When the feedback influenced the subject’s decision and behavior in a subsequent trial or was used as a learning signal, the contrast was classified as reward evaluation. For example, a risky decision that is rewarded in the initial trial may lead a subject to take another, perhaps bigger, risk in the next trial (Cohen and Ranganath, 2005)(low-risk rewards followed by high-risk vs. low-risk decisions). Loss aversion, the tendency for people to strongly prefer avoiding losses to acquiring gains, is another example of evaluation (Tom et al., 2007)(relation between lambda and neural loss aversion).

2.2 Activation likelihood estimation (ALE)

The algorithm of ALE is based on (Eickhoff et al., 2009). ALE models the activation foci as 3D Gaussian distributions centered at the reported coordinates, and then calculates the overlap of these distributions across different experiments (ALE treats each contrast in a study as a separate experiment). The spatial uncertainty associated with activation foci is estimated with respect to the number of subjects in each study (i.e., a larger sample produces more reliable activation patterns and localization; therefore the coordinates are convolved with a tighter Gaussian kernel). The convergence of activation patterns across experiments is calculated by taking the union of the above modeled activation maps. A null distribution that represents ALE scores generated by random spatial overlap across studies is estimated through permutation procedure. Finally the ALE map computed from the real activation coordinates is tested against the ALE scores from the null distribution, producing a statistical map representing the p values of the ALE scores. The nonparametric p values are then transformed into z scores and thresholded at a cluster-level corrected p<0.05.

Six different ALE analyses were conducted using GingerALE 2.0 (Eickhoff et al., 2009), one for the main analysis of all studies, and one for each of the five sub-lists characterizing brain activation by positive or negative rewards as well as anticipation, outcome, and evaluation. Two subtraction ALE analyses were conducted using GingerALE 1.2 (Turkeltaub et al., 2002), one for the contrast between positive and negative rewards, and the other for the contrast between anticipation and outcome.

2.2.1 Main analysis of all studies

All 142 studies were included in the main analysis, which consisted of 5214 foci from 655 experiments (contrasts). We used the algorithm implemented in GingerALE 2.0, which models the ALE based on the spatial uncertainty of each focus using an estimation of the inter-subject and inter-experiment variability. The estimation was constrained by a gray matter mask and estimated the above-chance clustering with the experiments as a random-effects factor, rather than using a fixed-effects analysis on foci (Eickhoff et al., 2009). The resulting ALE map was thresholded using the false discover rate (FDR) method with p<0.05 and a minimum cluster size of 60 voxels of 2×2×2 mm (for a total of 480 mm3) to protect against false positives of multiple comparisons.

2.2.2 Individual analyses of sub-lists

Five other ALE analyses were also conducted based on the sub-lists that categorize different experiments into positive and negative rewards, as well as reward anticipation, reward delivery (outcome), and choice evaluation. For the positive reward analysis, 2167 foci from 283 experiments were included. The negative reward analysis consisted of 935 foci from 140 experiments. The numbers of foci included in the analyses for anticipation, outcome, and choice evaluation were 1553 foci (185 experiments), 1977 (253), and 520 (97), respectively. We applied the same analysis and threshold approaches as we did for the main analysis above.

2.2.3 Subtraction analyses

We were also interested in contrasting the brain areas that were selectively or preferentially activated by positive versus negative rewards, and by reward anticipation versus reward delivery. GingerALE 1.2 was used to conduct these two analyses. ALE maps were smoothed with a kernel with a FWHM of 10 mm. A permutation test of randomly distributed foci with 10000 simulations was run to determine statistical significance of the ALE maps. To correct for multiple comparisons, the resulting ALE maps were thresholded using the FDR method with p<0.05 and a minimum cluster size of 60 voxels.

2.3 Parametric voxel-based meta-analysis (PVM)

We also analyzed the same coordinate lists using another meta-analysis approach, PVM. In contrast to the ALE analysis, which treats different contrasts within a study as distinct experiments, PVM analysis pools peaks from all different contrasts within a study and creates a single coordinate map for the specific study (Costafreda et al., 2009). Therefore, the random-effects factor in the PVM analysis is the studies, in comparison to individual experiments/contrasts in the ALE analysis. This further reduces estimation bias caused by studies with multiple contrasts that reporting similar activation patterns. Similar to the ALE approach, we conducted six different PVM analyses using the algorithms implemented in R statistical software (http://www.R-project.org) from a previous study (Costafreda et al., 2009), one for the main analysis of all studies, and one for each of the five sub-lists characterizing brain activation by different aspects of reward processing. Two additional PVM analyses were conducted using the same code base to compare between positive and negative rewards as well as between reward anticipation and outcome.

2.3.1 Main analysis of all studies

MNI coordinates (5214) from the same 142 studies used in the ALE analysis were transformed into a text table, with each study identified by a unique study identification label. Computations on the peak map were constrained within a mask in MNI space. The peak map was first smoothed with a uniform kernel (ρ = 10 mm) to generate the summary map, which represents the number of studies reporting overlapping activation peaks within a neighborhood of 10 mm radius. Next, random-effects PVM analysis was run to estimate statistical significance associated with each voxel in the summary map. The number of studies in the summary map was converted into the proportion of studies that reported concordant activation. We used the same threshold as used in ALE analysis to identify significant clusters for the proportion map (using the FDR method with p<0.05 and a minimum cluster size of 60 voxels).

2.3.2 Individual analyses of sub-lists

Five other PVM analyses were conducted on the sub-lists for positive and negative rewards, as well as reward anticipation, outcome, and evaluation. The positive reward analysis included 2167 foci from 111 studies whereas the negative reward analysis included 935 foci from 67 studies. The numbers of studies included in the analyses for anticipation, outcome, and choice evaluation were 1553 foci (65 studies), 1977 (86), and 520 (39), respectively. We applied the same analysis and threshold approaches as we did for the main analysis above.

2.3.3 Comparison analyses

We also conducted two PVM analyses to compare the activation patterns between positive and negative rewards as well as between reward anticipation and outcome. Two peak maps (e.g., one for positive and the other for negative) were first smoothed with a uniform kernel (ρ = 10 mm) to generate the summary maps, each representing the number of studies with overlapping activation peak within a neighborhood of 10 mm radius. These two summary maps were entered into a Fisher test to estimate the odds ratio and statistical significance p value for each contributing voxel within the MNI space mask. Since the Fisher test is not specifically developed for fMRI data analysis and empirically less sensitive than the other methods, we applied a relatively lenient threshold for the direct comparison PVM analysis, using uncorrected p<0.01 and a minimum cluster size of 60 voxels (Xiong et al., 1995), to correct for multiple comparison Type I error.

3. Results

3.1 ALE results

The all-inclusive analysis of 142 studies showed significant activation of a large cluster that encompassed the bilateral nucleus accumbens (NAcc), pallidum, anterior insula, lateral/medial OFC, anterior cingulate cortex (ACC), supplementary motor area (SMA), lateral prefrontal cortex (PFC), right amygdala, left hippocampus, thalamus, and brain stem (Figure 1A). Other smaller clusters included the right middle frontal gyrus and left middle/inferior frontal gyrus, bilateral inferior/superior parietal lobule, and posterior cingulate cortex (PCC) (Table 1).

Figure 1 

Concordance of brain activation from the ALE analyses. A. Core network activated by all contrasts/experiments. B. Overlay of brain areas separately involved in positive versus negative reward processing. C. Overlay of brain areas individually activated
Table 1 

Brain areas commonly activated by all studies from the ALE analysis (FDR p<0.05 and a minimum cluster size of 60 voxels).

Positive rewards activated a subset of the above mentioned networks, including the bilateral pallidum, anterior insula, thalamus, brain stem, medial OFC, ACC, SMA, PCC, and other frontal and parietal areas (Figure 1B and Table 2, also see Supplementary Materials – Figure S1A). Negative rewards showed activation in the bilateral NAcc, caudate, pallidum, anterior insula, amygdale, thalamus, brain stem, rostral ACC, dorsomedial PFC, lateral OFC, and right middle and inferior frontal gyrus (Figure 1B and Table 2, also see Supplementary Materials – Figure S1B). Contrasting activation by positive versus negative rewards, we found that positive rewards significantly activated the following regions to a great degree: bilateral NAcc, anterior insula, medial OFC, hippocampus, left putamen, and thalamus (Figure 1D and Table 4). None showed more activation by negative than positive rewards.

Table 2 

Brain areas activated by positive or negative rewards from the ALE analysis (FDR p<0.05 and a minimum cluster size of 60 voxels).
Table 4 

Brain areas differentially activated by positive and negative rewards from the ALE subtraction analysis (FDR p<0.05 and a minimum cluster size of 60 voxels).

Different reward processing stages shared similar brain activation patterns in the above-mentioned core networks, including the bilateral NAcc, anterior insula, thalamus, medial OFC, ACC, and dorsomedial PFC (Figure 1C and Table 3, also see Supplementary Materials – Figures S1C–E). Reward anticipation, as compared to reward outcome, revealed greater activation in the bilateral anterior insula, ACC, SMA, left inferior parietal lobule and middle frontal gyrus (Figure 1E and Table 5). Outcome preferential activation included bilateral NAcc, caudate, thalamus, and medial/lateral OFC (Table 5).

Table 3 

Brain areas activated by anticipation, outcome, and evaluation from the ALE analysis (FDR p<0.05 and a minimum cluster size of 60 voxels).
Table 5 

Brain areas differentially activated by anticipation and outcome from the ALE subtraction analysis (FDR p<0.05 and a minimum cluster size of 60 voxels).

3.2 PVM results

The main analysis of 142 studies showed significant activation in bilateral NAcc, anterior insula, lateral/medial OFC, ACC, PCC, inferior parietal lobule, and middle frontal Gyrus (Figure 2A and Table 6).

Figure 2 

Concordance of brain activation from the PVM analyses. A. Core network activated by all contrasts/experiments. B. Overlay of brain areas separately involved in positive versus negative reward processing. C. Overlay of brain areas individually activated
Table 6 

Brain areas commonly activated by all studies from the PVM analysis (FDR p<0.05 and a minimum cluster size of 60 voxels).

Positive rewards activated the bilateral NAcc, pallidum, putamen, thalamus, medial OFC, pregenual cingulate cortex, SMA, and PCC (Figure 2B and Table 7, also see Supplementary Materials – Figure S2A). Activation by negative rewards was found in the bilateral NAcc and anterior insula, pallidum, ACC, SMA, and middle/inferior frontal gyrus (Figure 2B and Table 7, also see Supplementary Materials – Figure S2B). Direct contrast between positive and negative rewards revealed preferential activation by positive rewards in the NAcc, pallidum, medial OFC, and PCC, and greater activation by negative rewards in ACC and middle/inferior frontal gyrus (Figure 2D and Table 9).

Table 7 

Brain areas activated by positive or negative rewards from the PVM analysis (FDR p<0.05 and a minimum cluster size of 60 voxels).
Table 9 

Brain areas differentially activated by positive and negative rewards from the PVM Fisher odds ratio analysis (voxel p<0.01 and a minimum cluster size of 60 voxels).

Different reward processing stages similarly activated the NAcc and ACC whereas they differentially recruited other brain areas such as medial OFC, anterior insula, and amygdala (Figure 2C and Table 8, also see Supplementary Materials – Figure S2C–E). Reward anticipation, as compared to reward outcome, revealed significant activation in the bilateral anterior insula, thalamus, precentral gyrus, and inferior parietal lobule (Figure 2E and Table 10). No brain area showed greater activation by reward outcome in comparison to anticipation.

Table 8 

Brain areas activated by anticipation, outcome, and evaluation from the PVM analysis (FDR p<0.05 and a minimum cluster size of 60 voxels).
Table 10 

Brain areas differentially activated by anticipation and outcome from the PVM Fisher odds ratio analysis (voxel p<0.01 and a minimum cluster size of 60 voxels).

3.3 Comparison of ALE and PVM results

The current study also showed that although ALE and PVM methods treated the coordinate-based data differently and adopted distinct estimation algorithms, the results for a single list of coordinates from these two meta-analysis approaches were very similar and comparable (Figures 1A–C and 2A–C, Table 11, also see Figures S1 and S2 in the Supplementary Materials). The improved ALE algorithm implemented in GingerALE 2.0, by design, treats experiments (or contrasts) as the random-effects factor, which significantly reduces the bias caused by experiments reporting more loci versus those with fewer loci. Different studies, however, include different number of experiments/contrasts. Therefore, the results of GingerALE 2.0 may still be affected by the bias that weighs more toward studies reporting more contrasts, potentially overestimating cross-study concordance. However, by choice, users can combine coordinates from different contrasts together so that GingerALE 2.0 can treat each study as a single experiment. This is what PVM implements, pooling coordinates from all contrasts within a study into a single activation map, thus weighing all studies equally to estimate activation overlap across studies.

Table 11 

Summary of ALE and PVM results on key regions of interest.

In contrast, comparison of two lists of coordinates differed significantly between ALE and PVM approaches (Table 11), as a result of their differences in sensitivity to within-study and cross-study convergence. Since the improved ALE algorithm has not been implemented for the subtractive ALE analysis, we used an earlier version, GingerALE 1.2, which treats the coordinates as the random-effects factor and experiments as the fixed-effects variable. Therefore differences in both the numbers of coordinates and experiments in two lists may affect the subtraction results. The subtractive ALE analysis biased toward the list with more experiments against the other with fewer (Figure 1D/E). Positive reward studies (2167 foci from 283 experiments) clearly predominated over negative studies (935 foci from 140 experiments). The difference between reward anticipation (1553 foci from 185 experiments) and outcome (1977 foci from 253 experiments) was smaller, but could have also caused the bias toward the outcome phase. On the other hand, the use of the Fisher test to estimate the odds ratio and assign voxels in one of the two lists by PVM seemed to be less sensitive in detecting activation difference between the two lists (Figure 2D/E).

4. Discussion

We are constantly making decisions in our everyday life. Some decisions involve no apparent positive or negative values of the outcomes whereas others have significant impacts on the valence of the results and our emotional responses toward the choices we make. We may feel happy and satisfied when the outcome is positive or our expectation is fulfilled, or feel frustrated when the outcome is negative or lower than what we anticipated. Moreover, many decisions must be made without advance knowledge of their consequences. Therefore, we need to be able to make predictions about the future reward, and evaluate the reward value and potential risk of obtaining it or being penalized. This requires us to evaluate the choice we make based on the presence of prediction errors and to use these signals to guide our learning and future behaviors. Many neuroimaging studies have examined reward-related decision making. However, given the complex and heterogeneous psychological processes involved in value-based decision making, it is no trivial task to examine neural networks that subserve representation and processing of reward-related information. We have observed a rapid growth in the number of empirical studies in the field of neuroeconomics, yet thus far it has been hard to see how these studies have converged so as to clearly delineate the reward circuitry in the human brain. In the current meta-analysis study, we have showed concordance across a large number of studies and revealed the common and distinct patterns of brain activation by different aspects of reward processing. In a data-driven fashion, we pooled over all coordinates from different contrasts/experiments of 142 studies, and observed a core reward network, which consists of the NAcc, lateral/medial OFC, ACC, anterior insula, dorsomedial PFC, as well as the lateral frontoparietal areas. A recent meta-analysis study focusing on risk assessment in decision making reported a similar reward circuitry (Mohr et al., 2010). In addition, from a theory-driven perspective, we contrasted neural networks that were involved in positive and negative valence across anticipation and outcome stages of reward processing, and elucidated distinct neural substrates subserving valence-related assessment as well as their preferential involvement in anticipation and outcome.

4.1 Core reward areas: NAcc and OFC

The NAcc and OFC have long been conceived as the major players in reward processing because they are the main projection areas of two distinct dopaminergic pathways, the mesolimbic and mesocortical pathways, respectively. However, it remains unknown how dopamine neurons distinctively modulate activity in these limbic and cortical areas. Previous studies have tried to differentiate the roles of these two structures in terms of temporal stages, associating the NAcc with reward anticipation and relating the medial OFC to receipt of reward (Knutson et al., 2001b; Knutson et al., 2003; Ramnani et al., 2004). Results from other studies questioned such a distinction (Breiter et al., 2001; Delgado et al., 2005; Rogers et al., 2004). Many studies also implied that the NAcc was responsible for detecting prediction error, a crucial signal in incentive learning and reward association (McClure et al., 2003; O’Doherty et al., 2003b; Pagnoni et al., 2002). Studies also found that the NAcc showed a biphasic response, such that activity in the NAcc would decrease and drop below the baseline in response to negative prediction errors (Knutson et al., 2001b; McClure et al., 2003; O’Doherty et al., 2003b). Although the OFC usually displays similar patterns of activity as the NAcc, previous neuroimaging studies in humans have suggested that the OFC serves to convert a variety of stimuli into a common currency in terms of their reward values (Arana et al., 2003; Cox et al., 2005; Elliott et al., 2010; FitzGerald et al., 2009; Gottfried et al., 2003; Kringelbach et al., 2003; O’Doherty et al., 2001; Plassmann et al., 2007). These findings paralleled those obtained from single cell recording and lesion studies in animals (Schoenbaum and Roesch, 2005; Schoenbaum et al., 2009; Schoenbaum et al., 2003; Schultz et al., 2000; Tremblay and Schultz, 1999, 2000; Wallis, 2007).

Our overall analyses showed that the NAcc and OFC responded to general reward processing (Figure 1A and Figure 2A). Activation in the NAcc largely overlapped across different stages, whereas the medial OFC was more tuned to reward receipt (Figure 1C/E and Figure 2C). These findings highlighted that the NAcc may be responsible for tracking both positive and negative signals of reward and using them to modulate learning of reward association, whereas the OFC mostly monitors and evaluates reward outcomes. Further investigation is needed to better differentiate the roles of the NAcc and OFC in reward-related decision making (Frank and Claus, 2006; Hare et al., 2008).

4.2 Valence-related assessment

In addition to converting various reward options into common currency and representing their reward values, distinct brain regions in the reward circuitry may separately encode positive and negative valences of reward. Direct comparisons across reward valence revealed that both the NAcc and medial OFC were more active in response to positive versus negative rewards (Figure 1B/D and Figure 2B/D). In contrast, the anterior insular cortex was involved in the processing of negative reward information (Figure 1B and Figure 2B). These results confirmed the medial-lateral distinction for positive versus negative rewards (Kringelbach, 2005; Kringelbach and Rolls, 2004), and were consistent with what we observed in our previous study on a reward task (Liu et al., 2007). Sub-regions of the ACC uniquely responded to positive and negative rewards. Pregenual and rostral ACC, close to the medial OFC, were activated by positive rewards whereas the caudal ACC responded to negative rewards (Figure 1B and Figure 2B). ALE and PVM meta-analyses also revealed that the PCC was consistently activated by positive rewards (Figure 1B and Figure 2B).

Interestingly, separate networks encoding positive and negative valences are similar to the distinction between two anti-correlated networks, the default-mode network and task-related network (Fox et al., 2005; Raichle et al., 2001; Raichle and Snyder, 2007). Recent meta-analyses found that the default-mode network mainly involved the medial prefrontal regions (including the medial OFC) and medial posterior cortex (including the PCC and precuneus), and the task-related network includes the ACC, insula, and lateral frontoparietal regions (Laird et al., 2009; Toro et al., 2008). Activation in the medial OFC and PCC by positive rewards mirrored the default-mode network commonly observed during the resting state, whereas activation in the ACC, insula, lateral prefrontal cortex by negative rewards paralleled the task-related network. This intrinsic functional organization of the brain was found to influence reward and risky decision making and account for individual differences in risk-taking traits (Cox et al., 2010).

4.3 Anticipation versus outcome

The bilateral anterior insula, ACC/SMA, inferior parietal lobule, and brain stem showed more consistent activation in anticipation in comparison to the outcome phase (Figure 1C/E and Figure 2C/E). The anterior insula and ACC have previously been implicated in interoception, emotion and empathy (Craig, 2002, 2009; Gu et al., 2010; Phan et al., 2002), and risk and uncertainty assessment (Critchley et al., 2001; Kuhnen and Knutson, 2005; Paulus et al., 2003), lending its role in anticipation. The anterior insula was consistently involved in risk processing, especially in anticipation of loss, as revealed by a recent meta-analysis (Mohr et al., 2010). Similar to the role of the OFC, the parietal lobule has been associated with valuation of different options (Sugrue et al., 2005), numerical representation (Cohen Kadosh et al., 2005; Hubbard et al., 2005), and information integration (Gold and Shadlen, 2007; Yang and Shadlen, 2007). Therefore, it is crucial for the parietal lobule to be involved in the anticipation stage of reward processing so as to plan and prepare for an informed action (Andersen and Cui, 2009; Lau et al., 2004a; Lau et al., 2004b).

On the other hand, the ventral striatum, medial OFC, and amygdala showed preferential activation during reward outcome in comparison to the anticipation stage (Figure 1C/E and Figure 2C). These patterns were consistent with what we and other investigators found previously (Breiter et al., 2001; Delgado et al., 2005; Liu et al., 2007; Rogers et al., 2004), standing against the functional dissociation between the ventral striatum and medial OFC in terms of their respective roles in reward anticipation and reward outcome (Knutson et al., 2001a; Knutson et al., 2001b; Knutson et al., 2003).

4.4 A schematic illustration of reward processing

Based on the findings of common and distinct networks involved in various aspects of reward decision making, we have come up with a schematic illustration to summarize the distributed representations of valuation and valence in reward processing (Figure 3). We tentatively group different brain regions based on their roles in different processes, although each region may serve multiple functions and interact with other brain areas in a far more complex way. When facing alternative choices, each of which has distinctive characteristics such as magnitude and probability, these properties need to be converted into comparable value-based information, a “common currency”. Not only do we compare the values of these alternative choices, but we also compare the factual and projected values as well as the fictional values associated with the un-chosen choice (e.g., the prediction error signal). The ventral striatum and medial OFC have been implicated in this value-based representation. The inferior parietal lobule has also been found to be involved in representing and comparing numerical information. In addition, value-based decision making inevitably results in evaluation of the choices, based on the valence of the outcomes and associated emotional responses. While the ventral striatum and medial OFC are also involved in detecting the positive reward valence, the lateral OFC, anterior insula, ACC and amygdala are mostly implicated in processing of the negative reward valence, most likely linked to their evaluative roles in negative emotional responses. Because of the negative affect usually associated with risk, the anterior insula and ACC are also involved in reward anticipation of risky decisions, especially for uncertainty-averse responses in anticipation of loss. Finally, the frontoparietal regions serve to integrate and act upon these signals in order to produce optimal decisions (e.g., win-stay-loss-switch).

Figure 3 

A schematic framework illustrates the roles of core brain areas involved in different aspects of reward-related decision making. The grid pattern denotes the medial orbitofrontal cortex encoding positive valence; the dash pattern denotes the anterior

4.5 Caveats

A couple of methodological caveats need to be noted. The first is related to the bias in reporting the results in different studies. Some studies are purely ROI-based, which were excluded from the current study. Still, others singled out or put more emphasis on a prior regions by reporting more coordinates or contrasts related to those regions. They could bias the results toward confirming the “hotspots”. Secondly, we want to caution about conceptual distinction of different aspects of reward processing. We classified various contrasts into different categories of theoretical interest. However, with real life decisions or in many experimental tasks, these aspects do not necessarily have clear divisions. For example, evaluation of the previous choice and reward outcome may intermingle with upcoming reward anticipation and decision making. There is no clear boundary across different stages of reward processing, leaving our current classification open for discussion. Nonetheless, this hypothesis-driven approach is greatly needed (Caspers et al., 2010; Mohr et al., 2010; Richlan et al., 2009), which complements the data-driven nature of meta-analysis. Many factors related to reward decision making, such as risk assessment and types of reward (e.g., primary vs. secondary, monetary vs. social), call for additional meta-analyses.

Research Highlights

  • We conducted two sets of coordinate-based meta-analyses on 142 fMRI studies of reward.
  • The core reward circuitry included the nucleus accumbens, insula, orbitofrontal, cingulate, and frontoparietal regions.
  • The nucleus accumbens was activated by both positive and negative rewards across various reward processing stages.
  • Other regions showed preferential responses toward positive or negative rewards, or during anticipation or outcome.

Supplementary Material

02

03

Acknowledgments

This study is supported by the Hundred-Talent Project of the Chinese Academy of Sciences, NARSAD Young Investigator Award (XL), and NIH Grant R21MH083164 (JF). The authors wish to thank the development team of BrainMap and Sergi G. Costafreda for providing excellent tools for this study.

Appendix

List of articles included in the meta-analyses of the current study.

1. Abler B, Walter H, Erk S, Kammerer H, Spitzer M. Prediction error as a linear function of reward probability is coded in human nucleus accumbens. Neuroimage. 2006;31:790–795. [PubMed]
2. Adcock RA, Thangavel A, Whitfield-Gabrieli S, Knutson B, Gabrieli JD. Reward-motivated learning: mesolimbic activation precedes memory formation. Neuron. 2006;50:507–517. [PubMed]
3. Akitsuki Y, Sugiura M, Watanabe J, Yamashita K, Sassa Y, Awata S, Matsuoka H, Maeda Y, Matsue Y, Fukuda H, Kawashima R. Context-dependent cortical activation in response to financial reward and penalty: an event-related fMRI study. Neuroimage. 2003;19:1674–1685. [PubMed]
4. Ballard K, Knutson B. Dissociable neural representations of future reward magnitude and delay during temporal discounting. Neuroimage. 2009;45:143–150. [PMC free article] [PubMed]
5. Behrens TE, Woolrich MW, Walton ME, Rushworth MF. Learning the value of information in an uncertain world. Nature Neuroscience. 2007;10:1214–1221. [PubMed]
6. Bjork JM, Hommer DW. Anticipating instrumentally obtained and passively-received rewards: a factorial fMRI investigation. Behavioural Brain Research. 2007;177:165–170. [PMC free article] [PubMed]
7. Bjork JM, Knutson B, Fong GW, Caggiano DM, Bennett SM, Hommer DW. Incentive-elicited brain activation in adolescents: similarities and differences from young adults. Journal of Neuroscience. 2004;24:1793–1802. [PubMed]
8. Bjork JM, Momenan R, Smith AR, Hommer DW. Reduced posterior mesofrontal cortex activation by risky rewards in substance-dependent patients. Drug and Alcohol Dependence. 2008;95:115–128. [PMC free article] [PubMed]
9. Bjork JM, Smith AR, Danube CL, Hommer DW. Developmental differences in posterior mesofrontal cortex recruitment by risky rewards. Journal of Neuroscience. 2007;27:4839–4849. [PubMed]
10. Blair K, Marsh AA, Morton J, Vythilingam M, Jones M, Mondillo K, Pine DC, Drevets WC, Blair JR. Choosing the lesser of two evils, the better of two goods: specifying the roles of ventromedial prefrontal cortex and dorsal anterior cingulate in object choice. Journal of Neuroscience. 2006;26:11379–11386. [PubMed]
11. Breiter HC, Aharon I, Kahneman D, Dale A, Shizgal P. Functional imaging of neural responses to expectancy and experience of monetary gains and losses. Neuron. 2001;30:619–639. [PubMed]
12. Budhani S, Marsh AA, Pine DS, Blair RJ. Neural correlates of response reversal: considering acquisition. Neuroimage. 2007;34:1754–1765. [PubMed]
13. Bush G, Vogt BA, Holmes J, Dale AM, Greve D, Jenike MA, Rosen BR. Dorsal anterior cingulate cortex: a role in reward-based decision making. Proceedings of the National Academy of Sciences of the United States of America. 2002;99:523–528. [PMC free article] [PubMed]
14. Camara E, Rodriguez-Fornells A, Munte TF. Functional connectivity of reward processing in the brain. Frontiers in Human Neuroscience. 2008;2:19. [PMC free article] [PubMed]
15. Campbell-Meiklejohn DK, Woolrich MW, Passingham RE, Rogers RD. Knowing when to stop: the brain mechanisms of chasing losses. Biological Psychiatry. 2008;63:293–300. [PubMed]
16. Chiu PH, Lohrenz TM, Montague PR. Smokers’ brains compute, but ignore, a fictive error signal in a sequential investment task. Nature Neuroscience. 2008;11:514–520. [PubMed]
17. Clithero JA, Carter RM, Huettel SA. Local pattern classification differentiates processes of economic valuation. Neuroimage. 2009;45:1329–1338. [PMC free article] [PubMed]
18. Cohen MX. Individual differences and the neural representations of reward expectation and reward prediction error. Social Cognitive and Affective Neuroscience. 2007;2:20–30. [PMC free article] [PubMed]
19. Cohen MX, Elger CE, Weber B. Amygdala tractography predicts functional connectivity and learning during feedback-guided decision-making. Neuroimage. 2008;39:1396–1407. [PubMed]
20. Cohen MX, Heller AS, Ranganath C. Functional connectivity with anterior cingulate and orbitofrontal cortices during decision-making. Brain Research Cognitive Brain Research. 2005a;23:61–70. [PubMed]
21. Cohen MX, Ranganath C. Behavioral and neural predictors of upcoming decisions. Cognitive, Affective, & Behavioral Neuroscience. 2005;5:117–126. [PubMed]
22. Cohen MX, Young J, Baek JM, Kessler C, Ranganath C. Individual differences in extraversion and dopamine genetics predict neural reward responses. Brain Research Cognitive Brain Research. 2005b;25:851–861. [PubMed]
23. Cooper JC, Knutson B. Valence and salience contribute to nucleus accumbens activation. Neuroimage. 2008;39:538–547. [PMC free article] [PubMed]
24. Coricelli G, Critchley HD, Joffily M, O’Doherty JP, Sirigu A, Dolan RJ. Regret and its avoidance: a neuroimaging study of choice behavior. Nature Neuroscience. 2005;8:1255–1262. [PubMed]
25. Cox SM, Andrade A, Johnsrude IS. Learning to like: a role for human orbitofrontal cortex in conditioned reward. Journal of Neuroscience. 2005;25:2733–2740. [PubMed]
26. Critchley HD, Mathias CJ, Dolan RJ. Neural activity in the human brain relating to uncertainty and arousal during anticipation. Neuron. 2001;29:537–545. [PubMed]
27. D’Ardenne K, McClure SM, Nystrom LE, Cohen JD. BOLD responses reflecting dopaminergic signals in the human ventral tegmental area. Science. 2008;319:1264–1267. [PubMed]
28. Daw ND, O’Doherty JP, Dayan P, Seymour B, Dolan RJ. Cortical substrates for exploratory decisions in humans. Nature. 2006;441:876–879. [PMC free article] [PubMed]
29. Delgado MR, Locke HM, Stenger VA, Fiez JA. Dorsal striatum responses to reward and punishment: effects of valence and magnitude manipulations. Cognitive, Affective, & Behavioral Neuroscience. 2003;3:27–38. [PubMed]
30. Delgado MR, Miller MM, Inati S, Phelps EA. An fMRI study of reward-related probability learning. Neuroimage. 2005;24:862–873. [PubMed]
31. Delgado MR, Nystrom LE, Fissell C, Noll DC, Fiez JA. Tracking the hemodynamic responses to reward and punishment in the striatum. Journal of Neurophysiology. 2000;84:3072–3077. [PubMed]
32. Delgado MR, Schotter A, Ozbay EY, Phelps EA. Understanding overbidding: using the neural circuitry of reward to design economic auctions. Science. 2008;321:1849–1852. [PMC free article] [PubMed]
33. Dillon DG, Holmes AJ, Jahn AL, Bogdan R, Wald LL, Pizzagalli DA. Dissociation of neural regions associated with anticipatory versus consummatory phases of incentive processing. Psychophysiology. 2008;45:36–49. [PMC free article] [PubMed]
34. Elliott R, Agnew Z, Deakin JF. Medial orbitofrontal cortex codes relative rather than absolute value of financial rewards in humans. European Journal of Neuroscience. 2008;27:2213–2218. [PubMed]
35. Elliott R, Agnew Z, Deakin JF. Hedonic and Informational Functions of the Human Orbitofrontal Cortex. Cerebral Cortex 2009 [PubMed]
36. Elliott R, Friston KJ, Dolan RJ. Dissociable neural responses in human reward systems. Journal of Neuroscience. 2000;20:6159–6165. [PubMed]
37. Elliott R, Newman JL, Longe OA, Deakin JF. Differential response patterns in the striatum and orbitofrontal cortex to financial reward in humans: a parametric functional magnetic resonance imaging study. Journal of Neuroscience. 2003;23:303–307. [PubMed]
38. Elliott R, Newman JL, Longe OA, William Deakin JF. Instrumental responding for rewards is associated with enhanced neuronal response in subcortical reward systems. Neuroimage. 2004;21:984–990. [PubMed]
39. Ernst M, Nelson EE, Jazbec S, McClure EB, Monk CS, Leibenluft E, Blair J, Pine DS. Amygdala and nucleus accumbens in responses to receipt and omission of gains in adults and adolescents. Neuroimage. 2005;25:1279–1291. [PubMed]
40. Ernst M, Nelson EE, McClure EB, Monk CS, Munson S, Eshel N, Zarahn E, Leibenluft E, Zametkin A, Towbin K, Blair J, Charney D, Pine DS. Choice selection and reward anticipation: an fMRI study. Neuropsychologia. 2004;42:1585–1597. [PubMed]
41. Fujiwara J, Tobler PN, Taira M, Iijima T, Tsutsui K. A parametric relief signal in human ventrolateral prefrontal cortex. Neuroimage. 2009a;44:1163–1170. [PubMed]
42. Fujiwara J, Tobler PN, Taira M, Iijima T, Tsutsui K. Segregated and integrated coding of reward and punishment in the cingulate cortex. Journal of Neurophysiology. 2009b;101:3284–3293. [PubMed]
43. Fukui H, Murai T, Fukuyama H, Hayashi T, Hanakawa T. Functional activity related to risk anticipation during performance of the Iowa Gambling Task. Neuroimage. 2005;24:253–259. [PubMed]
44. Glascher J, Hampton AN, O’Doherty JP. Determining a role for ventromedial prefrontal cortex in encoding action-based value signals during reward-related decision making. Cerebral Cortex. 2009;19:483–495. [PMC free article] [PubMed]
45. Gottfried JA, O’Doherty J, Dolan RJ. Encoding predictive reward value in human amygdala and orbitofrontal cortex. Science. 2003;301:1104–1107. [PubMed]
46. Hare TA, O’Doherty J, Camerer CF, Schultz W, Rangel A. Dissociating the role of the orbitofrontal cortex and the striatum in the computation of goal values and prediction errors. Journal of Neuroscience. 2008;28:5623–5630. [PubMed]
47. Hariri AR, Brown SM, Williamson DE, Flory JD, de Wit H, Manuck SB. Preference for immediate over delayed rewards is associated with magnitude of ventral striatal activity. Journal of Neuroscience. 2006;26:13213–13217. [PubMed]
48. Haruno M, Kawato M. Different neural correlates of reward expectation and reward expectation error in the putamen and caudate nucleus during stimulus-action-reward association learning. Journal of Neurophysiology. 2006;95:948–959. [PubMed]
49. Haruno M, Kuroda T, Doya K, Toyama K, Kimura M, Samejima K, Imamizu H, Kawato M. A neural correlate of reward-based behavioral learning in caudate nucleus: a functional magnetic resonance imaging study of a stochastic decision task. Journal of Neuroscience. 2004;24:1660–1665. [PubMed]
50. Heekeren HR, Wartenburger I, Marschner A, Mell T, Villringer A, Reischies FM. Role of ventral striatum in reward-based decision making. Neuroreport. 2007;18:951–955. [PubMed]
51. Hewig J, Straube T, Trippe RH, Kretschmer N, Hecht H, Coles MG, Miltner WH. Decision-making under Risk: An fMRI Study. Journal of Cognitive Neuroscience 2008 [PubMed]
52. Hommer DW, Knutson B, Fong GW, Bennett S, Adams CM, Varnera JL. Amygdalar recruitment during anticipation of monetary rewards: an event-related fMRI study. Annals of the New York Academy of Sciences. 2003;985:476–478. [PubMed]
53. Hsu M, Krajbich I, Zhao C, Camerer CF. Neural response to reward anticipation under risk is nonlinear in probabilities. Journal of Neuroscience. 2009;29:2231–2237. [PubMed]
54. Huettel SA. Behavioral, but not reward, risk modulates activation of prefrontal, parietal, and insular cortices. Cognitive, Affective, & Behavioral Neuroscience. 2006;6:141–151. [PubMed]
55. Huettel SA, Stowe CJ, Gordon EM, Warner BT, Platt ML. Neural signatures of economic preferences for risk and ambiguity. Neuron. 2006;49:765–775. [PubMed]
56. Ino T, Nakai R, Azuma T, Kimura T, Fukuyama H. Differential activation of the striatum for decision making and outcomes in a monetary task with gain and loss. Cortex 2009 [PubMed]
57. Izuma K, Saito DN, Sadato N. Processing of social and monetary rewards in the human striatum. Neuron. 2008;58:284–294. [PubMed]
58. Izuma K, Saito DN, Sadato N. Processing of the Incentive for Social Approval in the Ventral Striatum during Charitable Donation. Journal of Cognitive Neuroscience 2009 [PubMed]
59. Juckel G, Schlagenhauf F, Koslowski M, Wustenberg T, Villringer A, Knutson B, Wrase J, Heinz A. Dysfunction of ventral striatal reward prediction in schizophrenia. Neuroimage. 2006;29:409–416. [PubMed]
60. Kable JW, Glimcher PW. The neural correlates of subjective value during intertemporal choice. Nature Neuroscience. 2007;10:1625–1633. [PMC free article] [PubMed]
61. Kahnt T, Park SQ, Cohen MX, Beck A, Heinz A, Wrase J. Dorsal Striatal-midbrain Connectivity in Humans Predicts How Reinforcements Are Used to Guide Decisions. Journal of Cognitive Neuroscience. 2009;21:1332–1345. [PubMed]
62. Kim H, Shimojo S, O’Doherty JP. Is avoiding an aversive outcome rewarding? Neural substrates of avoidance learning in the human brain. PLoS Biology. 2006;4:e233. [PMC free article] [PubMed]
63. Kirsch P, Schienle A, Stark R, Sammer G, Blecker C, Walter B, Ott U, Burkart J, Vaitl D. Anticipation of reward in a nonaversive differential conditioning paradigm and the brain reward system: an event-related fMRI study. Neuroimage. 2003;20:1086–1095. [PubMed]
64. Knutson B, Adams CM, Fong GW, Hommer D. Anticipation of increasing monetary reward selectively recruits nucleus accumbens. Journal of Neuroscience. 2001a;21:RC159. [PubMed]
65. Knutson B, Fong GW, Adams CM, Varner JL, Hommer D. Dissociation of reward anticipation and outcome with event-related fMRI. Neuroreport. 2001b;12:3683–3687. [PubMed]
66. Knutson B, Fong GW, Bennett SM, Adams CM, Hommer D. A region of mesial prefrontal cortex tracks monetarily rewarding outcomes: characterization with rapid event-related fMRI. Neuroimage. 2003;18:263–272. [PubMed]
67. Knutson B, Rick S, Wimmer GE, Prelec D, Loewenstein G. Neural predictors of purchases. Neuron. 2007;53:147–156. [PMC free article] [PubMed]
68. Knutson B, Taylor J, Kaufman M, Peterson R, Glover G. Distributed neural representation of expected value. Journal of Neuroscience. 2005;25:4806–4812. [PubMed]
69. Knutson B, Westdorp A, Kaiser E, Hommer D. FMRI visualization of brain activity during a monetary incentive delay task. Neuroimage. 2000;12:20–27. [PubMed]
70. Knutson B, Wimmer GE. Splitting the difference: how does the brain code reward episodes? Annals of the New York Academy of Sciences. 2007;1104:54–69. [PubMed]
71. Knutson B, Wimmer GE, Rick S, Hollon NG, Prelec D, Loewenstein G. Neural antecedents of the endowment effect. Neuron. 2008;58:814–822. [PubMed]
72. Koch K, Schachtzabel C, Wagner G, Reichenbach JR, Sauer H, Schlosser R. The neural correlates of reward-related trial-and-error learning: An fMRI study with a probabilistic learning task. Learning & Memory. 2008;15:728–732. [PubMed]
73. Kuhnen CM, Knutson B. The neural basis of financial risk taking. Neuron. 2005;47:763–770. [PubMed]
74. Landmann C, Dehaene S, Pappata S, Jobert A, Bottlaender M, Roumenov D, Le Bihan D. Dynamics of prefrontal and cingulate activity during a reward-based logical deduction task. Cerebral Cortex. 2007;17:749–759. [PubMed]
75. Liu X, Powell DK, Wang H, Gold BT, Corbly CR, Joseph JE. Functional dissociation in frontal and striatal areas for processing of positive and negative reward information. Journal of Neuroscience. 2007;27:4587–4597. [PubMed]
76. Lohrenz T, McCabe K, Camerer CF, Montague PR. Neural signature of fictive learning signals in a sequential investment task. Proceedings of the National Academy of Sciences of the United States of America. 2007;104:9493–9498. [PMC free article] [PubMed]
77. Longe O, Senior C, Rippon G. The lateral and ventromedial prefrontal cortex work as a dynamic integrated system: evidence from FMRI connectivity analysis. Journal of Cognitive Neuroscience. 2009;21:141–154. [PubMed]
78. Luhmann CC, Chun MM, Yi DJ, Lee D, Wang XJ. Neural dissociation of delay and uncertainty in intertemporal choice. Journal of Neuroscience. 2008;28:14459–14466. [PMC free article] [PubMed]
79. Marsh AA, Blair KS, Vythilingam M, Busis S, Blair RJ. Response options and expectations of reward in decision-making: the differential roles of dorsal and rostral anterior cingulate cortex. Neuroimage. 2007;35:979–988. [PMC free article] [PubMed]
80. McClure SM, Berns GS, Montague PR. Temporal prediction errors in a passive learning task activate human striatum. Neuron. 2003;38:339–346. [PubMed]
81. McLean J, Brennan D, Wyper D, Condon B, Hadley D, Cavanagh J. Localisation of regions of intense pleasure response evoked by soccer goals. Psychiatry Research. 2009;171:33–43. [PubMed]
82. Nieuwenhuis S, Heslenfeld DJ, von Geusau NJ, Mars RB, Holroyd CB, Yeung N. Activity in human reward-sensitive brain areas is strongly context dependent. Neuroimage. 2005a;25:1302–1309. [PubMed]
83. Nieuwenhuis S, Slagter HA, von Geusau NJ, Heslenfeld DJ, Holroyd CB. Knowing good from bad: differential activation of human cortical areas by positive and negative outcomes. European Journal of Neuroscience. 2005b;21:3161–3168. [PubMed]
84. O’Doherty J, Critchley H, Deichmann R, Dolan RJ. Dissociating valence of outcome from behavioral control in human orbital and ventral prefrontal cortices. Journal of Neuroscience. 2003a;23:7931–7939. [PubMed]
85. O’Doherty J, Dayan P, Schultz J, Deichmann R, Friston K, Dolan RJ. Dissociable roles of ventral and dorsal striatum in instrumental conditioning. Science. 2004;304:452–454. [PubMed]
86. O’Doherty J, Kringelbach ML, Rolls ET, Hornak J, Andrews C. Abstract reward and punishment representations in the human orbitofrontal cortex. Nature Neuroscience. 2001;4:95–102. [PubMed]
87. O’Doherty JP, Dayan P, Friston K, Critchley H, Dolan RJ. Temporal difference models and reward-related learning in the human brain. Neuron. 2003b;38:329–337. [PubMed]
88. O’Doherty JP, Deichmann R, Critchley HD, Dolan RJ. Neural responses during anticipation of a primary taste reward. Neuron. 2002;33:815–826. [PubMed]
89. Paulus MP, Frank LR. Anterior cingulate activity modulates nonlinear decision weight function of uncertain prospects. Neuroimage. 2006;30:668–677. [PubMed]
90. Paulus MP, Rogalsky C, Simmons A, Feinstein JS, Stein MB. Increased activation in the right insula during risk-taking decision making is related to harm avoidance and neuroticism. Neuroimage. 2003;19:1439–1448. [PubMed]
91. Pessiglione M, Petrovic P, Daunizeau J, Palminteri S, Dolan RJ, Frith CD. Subliminal instrumental conditioning demonstrated in the human brain. Neuron. 2008;59:561–567. [PMC free article] [PubMed]
92. Petrovic P, Pleger B, Seymour B, Kloppel S, De Martino B, Critchley H, Dolan RJ. Blocking central opiate function modulates hedonic impact and anterior cingulate response to rewards and losses. Journal of Neuroscience. 2008;28:10509–10516. [PMC free article] [PubMed]
93. Plassmann H, O’Doherty J, Rangel A. Orbitofrontal cortex encodes willingness to pay in everyday economic transactions. Journal of Neuroscience. 2007;27:9984–9988. [PubMed]
94. Pleger B, Blankenburg F, Ruff CC, Driver J, Dolan RJ. Reward facilitates tactile judgments and modulates hemodynamic responses in human primary somatosensory cortex. Journal of Neuroscience. 2008;28:8161–8168. [PMC free article] [PubMed]
95. Pochon JB, Levy R, Fossati P, Lehericy S, Poline JB, Pillon B, Le Bihan D, Dubois B. The neural system that bridges reward and cognition in humans: an fMRI study. Proceedings of the National Academy of Sciences of the United States of America. 2002;99:5669–5674. [PMC free article] [PubMed]
96. Preuschoff K, Bossaerts P, Quartz SR. Neural differentiation of expected reward and risk in human subcortical structures. Neuron. 2006;51:381–390. [PubMed]
97. Preuschoff K, Quartz SR, Bossaerts P. Human insula activation reflects risk prediction errors as well as risk. Journal of Neuroscience. 2008;28:2745–2752. [PubMed]
98. Ramnani N, Elliott R, Athwal BS, Passingham RE. Prediction error for free monetary reward in the human prefrontal cortex. Neuroimage. 2004;23:777–786. [PubMed]
99. Ramnani N, Miall RC. Instructed delay activity in the human prefrontal cortex is modulated by monetary reward expectation. Cerebral Cortex. 2003;13:318–327. [PubMed]
100. Remijnse PL, Nielen MM, Uylings HB, Veltman DJ. Neural correlates of a reversal learning task with an affectively neutral baseline: an event-related fMRI study. Neuroimage. 2005;26:609–618. [PubMed]
101. Remijnse PL, Nielen MM, van Balkom AJ, Cath DC, van Oppen P, Uylings HB, Veltman DJ. Reduced orbitofrontal-striatal activity on a reversal learning task in obsessive-compulsive disorder. Archives of General Psychiatry. 2006;63:1225–1236. [PubMed]
102. Reuter J, Raedler T, Rose M, Hand I, Glascher J, Buchel C. Pathological gambling is linked to reduced activation of the mesolimbic reward system. Nature Neuroscience. 2005;8:147–148. [PubMed]
103. Rilling JK, Gutman D, Zeh T, Pagnoni G, Berns G, Kilts C. A neural basis for social cooperation. Neuron. 2002;35:395–405. [PubMed]
104. Rilling JK, Sanfey AG, Aronson JA, Nystrom LE, Cohen JD. Opposing BOLD responses to reciprocated and unreciprocated altruism in putative reward pathways. Neuroreport. 2004;15:2539–2543. [PubMed]
105. Rogers RD, Ramnani N, Mackay C, Wilson JL, Jezzard P, Carter CS, Smith SM. Distinct portions of anterior cingulate cortex and medial prefrontal cortex are activated by reward processing in separable phases of decision-making cognition. Biological Psychiatry. 2004;55:594–602. [PubMed]
106. Rolls ET, McCabe C, Redoute J. Expected value, reward outcome, and temporal difference error representations in a probabilistic decision task. Cerebral Cortex. 2008;18:652–663. [PubMed]
107. Sailer U, Robinson S, Fischmeister FP, Konig D, Oppenauer C, Lueger-Schuster B, Moser E, Kryspin-Exner I, Bauer H. Altered reward processing in the nucleus accumbens and mesial prefrontal cortex of patients with posttraumatic stress disorder. Neuropsychologia. 2008;46:2836–2844. [PubMed]
108. Sailer U, Robinson S, Fischmeister FP, Moser E, Kryspin-Exner I, Bauer H. Imaging the changing role of feedback during learning in decision-making. Neuroimage. 2007;37:1474–1486. [PubMed]
109. Samanez-Larkin GR, Gibbs SE, Khanna K, Nielsen L, Carstensen LL, Knutson B. Anticipation of monetary gain but not loss in healthy older adults. Nature Neuroscience. 2007;10:787–791. [PMC free article] [PubMed]
110. Sanfey AG, Rilling JK, Aronson JA, Nystrom LE, Cohen JD. The neural basis of economic decision-making in the Ultimatum Game. Science. 2003;300:1755–1758. [PubMed]
111. Satterthwaite TD, Green L, Myerson J, Parker J, Ramaratnam M, Buckner RL. Dissociable but inter-related systems of cognitive control and reward during decision making: evidence from pupillometry and event-related fMRI. Neuroimage. 2007;37:1017–1031. [PubMed]
112. Schonberg T, Daw ND, Joel D, O’Doherty JP. Reinforcement learning signals in the human striatum distinguish learners from nonlearners during reward-based decision making. Journal of Neuroscience. 2007;27:12860–12867. [PubMed]
113. Smith BW, Mitchell DG, Hardin MG, Jazbec S, Fridberg D, Blair RJ, Ernst M. Neural substrates of reward magnitude, probability, and risk during a wheel of fortune decision-making task. Neuroimage. 2009;44:600–609. [PMC free article] [PubMed]
114. Spicer J, Galvan A, Hare TA, Voss H, Glover G, Casey B. Sensitivity of the nucleus accumbens to violations in expectation of reward. Neuroimage. 2007;34:455–461. [PMC free article] [PubMed]
115. Staudinger MR, Erk S, Abler B, Walter H. Cognitive reappraisal modulates expected value and prediction error encoding in the ventral striatum. Neuroimage. 2009;47:713–721. [PubMed]
116. Strohle A, Stoy M, Wrase J, Schwarzer S, Schlagenhauf F, Huss M, Hein J, Nedderhut A, Neumann B, Gregor A, Juckel G, Knutson B, Lehmkuhl U, Bauer M, Heinz A. Reward anticipation and outcomes in adult males with attention-deficit/hyperactivity disorder. Neuroimage. 2008;39:966–972. [PubMed]
117. Tanaka SC, Doya K, Okada G, Ueda K, Okamoto Y, Yamawaki S. Prediction of immediate and future rewards differentially recruits cortico-basal ganglia loops. Nature Neuroscience. 2004;7:887–893. [PubMed]
118. Tanaka SC, Samejima K, Okada G, Ueda K, Okamoto Y, Yamawaki S, Doya K. Brain mechanism of reward prediction under predictable and unpredictable environmental dynamics. Neural Networks. 2006;19:1233–1241. [PubMed]
119. Taylor SF, Welsh RC, Wager TD, Phan KL, Fitzgerald KD, Gehring WJ. A functional neuroimaging study of motivation and executive function. Neuroimage. 2004;21:1045–1054. [PubMed]
120. Tobler PN, Christopoulos GI, O’Doherty JP, Dolan RJ, Schultz W. Neuronal distortions of reward probability without choice. Journal of Neuroscience. 2008;28:11703–11711. [PMC free article] [PubMed]
121. Tobler PN, O’Doherty JP, Dolan RJ, Schultz W. Reward value coding distinct from risk attitude-related uncertainty coding in human reward systems. Journal of Neurophysiology. 2007;97:1621–1632. [PMC free article] [PubMed]
122. Tom SM, Fox CR, Trepel C, Poldrack RA. The neural basis of loss aversion in decision-making under risk. Science. 2007;315:515–518. [PubMed]
123. Tricomi EM, Delgado MR, Fiez JA. Modulation of caudate activity by action contingency. Neuron. 2004;41:281–292. [PubMed]
124. Ullsperger M, von Cramon DY. Error monitoring using external feedback: specific roles of the habenular complex, the reward system, and the cingulate motor area revealed by functional magnetic resonance imaging. Journal of Neuroscience. 2003;23:4308–4314. [PubMed]
125. Ursu S, Carter CS. Outcome representations, counterfactual comparisons and the human orbitofrontal cortex: implications for neuroimaging studies of decision-making. Brain Research Cognitive Brain Research. 2005;23:51–60. [PubMed]
126. van den Bos W, McClure SM, Harris LT, Fiske ST, Cohen JD. Dissociating affective evaluation and social cognitive processes in the ventral medial prefrontal cortex. Cognitive, Affective, & Behavioral Neuroscience. 2007;7:337–346. [PubMed]
127. Vollm B, Richardson P, McKie S, Elliott R, Dolan M, Deakin B. Neuronal correlates of reward and loss in Cluster B personality disorders: a functional magnetic resonance imaging study. Psychiatry Research. 2007;156:151–167. [PubMed]
128. Volz KG, Schubotz RI, von Cramon DY. Predicting events of varying probability: uncertainty investigated by fMRI. Neuroimage. 2003;19:271–280. [PubMed]
129. Volz KG, Schubotz RI, von Cramon DY. Variants of uncertainty in decision-making and their neural correlates. Brain Research Bulletin. 2005;67:403–412. [PubMed]
130. Weber B, Aholt A, Neuhaus C, Trautner P, Elger CE, Teichert T. Neural evidence for Reference-dependence in real-market-transactions. Neuroimage. 2007;35:441–447. [PubMed]
131. Weber BJ, Huettel SA. The neural substrates of probabilistic and intertemporal decision making. Brain Research. 2008;1234:104–115. [PMC free article] [PubMed]
132. Wittmann M, Leland DS, Paulus MP. Time and decision making: differential contribution of the posterior insular cortex and the striatum during a delay discounting task. Experimental Brain Research. 2007;179:643–653. [PubMed]
133. Wrase J, Kahnt T, Schlagenhauf F, Beck A, Cohen MX, Knutson B, Heinz A. Different neural systems adjust motor behavior in response to reward and punishment. Neuroimage. 2007a;36:1253–1262. [PubMed]
134. Wrase J, Schlagenhauf F, Kienast T, Wustenberg T, Bermpohl F, Kahnt T, Beck A, Strohle A, Juckel G, Knutson B, Heinz A. Dysfunction of reward processing correlates with alcohol craving in detoxified alcoholics. Neuroimage. 2007b;35:787–794. [PubMed]
135. Xu L, Liang ZY, Wang K, Li S, Jiang T. Neural mechanism of intertemporal choice: from discounting future gains to future losses. Brain Research. 2009;1261:65–74. [PubMed]
136. Xue G, Lu Z, Levin IP, Weller JA, Li X, Bechara A. Functional dissociations of risk and reward processing in the medial prefrontal cortex. Cerebral Cortex. 2009;19:1019–1027. [PMC free article] [PubMed]
137. Yacubian J, Glascher J, Schroeder K, Sommer T, Braus DF, Buchel C. Dissociable systems for gain- and loss-related value predictions and errors of prediction in the human brain. Journal of Neuroscience. 2006;26:9530–9537. [PubMed]
138. Yacubian J, Sommer T, Schroeder K, Glascher J, Kalisch R, Leuenberger B, Braus DF, Buchel C. Gene-gene interaction associated with neural reward sensitivity. Proceedings of the National Academy of Sciences of the United States of America. 2007;104:8125–8130. [PMC free article] [PubMed]
139. Yarkoni T, Braver TS, Gray JR, Green L. Prefrontal brain activity predicts temporally extended decision-making behavior. Journal of the Experimental Analysis of Behavior. 2005a;84:537–554. [PMC free article] [PubMed]
140. Yarkoni T, Gray JR, Chrastil ER, Barch DM, Green L, Braver TS. Sustained neural activity associated with cognitive control during temporally extended decision making. Brain Research Cognitive Brain Research. 2005b;23:71–84. [PubMed]
141. Zalla T, Koechlin E, Pietrini P, Basso G, Aquino P, Sirigu A, Grafman J. Differential amygdala responses to winning and losing: a functional magnetic resonance imaging study in humans. European Journal of Neuroscience. 2000;12:1764–1770. [PubMed]
142. Zink CF, Pagnoni G, Martin-Skurski ME, Chappelow JC, Berns GS. Human striatal responses to monetary reward depend on saliency. Neuron. 2004;42:509–517. [PubMed]

Footnotes

Author contributions: XL designed and supervised the whole study. JH and MS made equal contributions to this study, performing literature search, data extraction and organization. JF participated in discussion and manuscript preparation.

Publisher’s Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

References

1. Andersen RA, Cui H. Intention, action planning, and decision making in parietal-frontal circuits. Neuron. 2009;63:568–583. [PubMed]
2. Arana FS, Parkinson JA, Hinton E, Holland AJ, Owen AM, Roberts AC. Dissociable contributions of the human amygdala and orbitofrontal cortex to incentive motivation and goal selection. J Neurosci. 2003;23:9632–9638. [PubMed]
3. Bach DR, Seymour B, Dolan RJ. Neural activity associated with the passive prediction of ambiguity and risk for aversive events. J Neurosci. 2009;29:1648–1656. [PMC free article] [PubMed]
4. Bechara A. Decision making, impulse control and loss of willpower to resist drugs: a neurocognitive perspective. Nat Neurosci. 2005;8:1458–1463. [PubMed]
5. Bjork JM, Knutson B, Fong GW, Caggiano DM, Bennett SM, Hommer DW. Incentive-elicited brain activation in adolescents: similarities and differences from young adults. J Neurosci. 2004;24:1793–1802. [PubMed]
6. Breiter HC, Aharon I, Kahneman D, Dale A, Shizgal P. Functional imaging of neural responses to expectancy and experience of monetary gains and losses. Neuron. 2001;30:619–639. [PubMed]
7. Brett M, Johnsrude IS, Owen AM. The problem of functional localization in the human brain. Nat Rev Neurosci. 2002;3:243–249. [PubMed]
8. Caspers S, Zilles K, Laird AR, Eickhoff SB. ALE meta-analysis of action observation and imitation in the human brain. Neuroimage. 2010;50:1148–1167. [PubMed]
9. Cohen Kadosh R, Henik A, Rubinsten O, Mohr H, Dori H, van de Ven V, Zorzi M, Hendler T, Goebel R, Linden DE. Are numbers special? The comparison systems of the human brain investigated by fMRI. Neuropsychologia. 2005;43:1238–1248. [PubMed]
10. Cohen MX, Ranganath C. Behavioral and neural predictors of upcoming decisions. Cogn Affect Behav Neurosci. 2005;5:117–126. [PubMed]
11. Costafreda SG, David AS, Brammer MJ. A parametric approach to voxel-based meta-analysis. Neuroimage. 2009;46:115–122. [PubMed]
12. Cox CL, Gotimer K, Roy AK, Castellanos FX, Milham MP, Kelly C. Your resting brain CAREs about your risky behavior. PLoS One. 2010;5:e12296. [PMC free article] [PubMed]
13. Cox SM, Andrade A, Johnsrude IS. Learning to like: a role for human orbitofrontal cortex in conditioned reward. J Neurosci. 2005;25:2733–2740. [PubMed]
14. Craig AD. How do you feel? Interoception: the sense of the physiological condition of the body. Nat Rev Neurosci. 2002;3:655–666. [PubMed]
15. Craig AD. How do you feel–now? The anterior insula and human awareness. Nat Rev Neurosci. 2009;10:59–70. [PubMed]
16. Critchley HD, Mathias CJ, Dolan RJ. Neural activity in the human brain relating to uncertainty and arousal during anticipation. Neuron. 2001;29:537–545. [PubMed]
17. d’Acremont M, Bossaerts P. Neurobiological studies of risk assessment: a comparison of expected utility and mean-variance approaches. Cogn Affect Behav Neurosci. 2008;8:363–374. [PubMed]
18. Delgado MR, Miller MM, Inati S, Phelps EA. An fMRI study of reward-related probability learning. Neuroimage. 2005;24:862–873. [PubMed]
19. Drevets WC. Neuroimaging and neuropathological studies of depression: implications for the cognitive-emotional features of mood disorders. Curr Opin Neurobiol. 2001;11:240–249. [PubMed]
20. Eickhoff SB, Laird AR, Grefkes C, Wang LE, Zilles K, Fox PT. Coordinate-based activation likelihood estimation meta-analysis of neuroimaging data: a random-effects approach based on empirical estimates of spatial uncertainty. Hum Brain Mapp. 2009;30:2907–2926. [PMC free article] [PubMed]
21. Elliott R, Agnew Z, Deakin JF. Hedonic and Informational Functions of the Human Orbitofrontal Cortex. Cereb Cortex. 2010;20:198–204. [PubMed]
22. Elliott R, Friston KJ, Dolan RJ. Dissociable neural responses in human reward systems. J Neurosci. 2000;20:6159–6165. [PubMed]
23. Ernst M, Nelson EE, Jazbec S, McClure EB, Monk CS, Leibenluft E, Blair J, Pine DS. Amygdala and nucleus accumbens in responses to receipt and omission of gains in adults and adolescents. Neuroimage. 2005;25:1279–1291. [PubMed]
24. Fellows LK. The cognitive neuroscience of human decision making: a review and conceptual framework. Behav Cogn Neurosci Rev. 2004;3:159–172. [PubMed]
25. FitzGerald TH, Seymour B, Dolan RJ. The role of human orbitofrontal cortex in value comparison for incommensurable objects. J Neurosci. 2009;29:8388–8395. [PMC free article] [PubMed]
26. Fox MD, Snyder AZ, Vincent JL, Corbetta M, Van Essen DC, Raichle ME. The human brain is intrinsically organized into dynamic, anticorrelated functional networks. Proc Natl Acad Sci USA. 2005;102:9673–9678. [PMC free article] [PubMed]
27. Frank MJ, Claus ED. Anatomy of a decision: striato-orbitofrontal interactions in reinforcement learning, decision making, and reversal. Psychol Rev. 2006;113:300–326. [PubMed]
28. Garavan H, Stout JC. Neurocognitive insights into substance abuse. Trends Cogn Sci. 2005;9:195–201. [PubMed]
29. Gold JI, Shadlen MN. The neural basis of decision making. Annu Rev Neurosci. 2007;30:535–574. [PubMed]
30. Gottfried JA, O’Doherty J, Dolan RJ. Encoding predictive reward value in human amygdala and orbitofrontal cortex. Science. 2003;301:1104–1107. [PubMed]
31. Gu X, Liu X, Guise KG, Naidich TP, Hof PR, Fan J. Functional dissociation of the frontoinsular and anterior cingulate cortices in empathy for pain. J Neurosci. 2010;30:3739–3744. [PMC free article] [PubMed]
32. Hare TA, O’Doherty J, Camerer CF, Schultz W, Rangel A. Dissociating the role of the orbitofrontal cortex and the striatum in the computation of goal values and prediction errors. J Neurosci. 2008;28:5623–5630. [PubMed]
33. Hsu M, Krajbich I, Zhao C, Camerer CF. Neural Response to Reward Anticipation under Risk Is Nonlinear in Probabilities. J Neurosci. 2009;29:2231–2237. [PubMed]
34. Hubbard EM, Piazza M, Pinel P, Dehaene S. Interactions between number and space in parietal cortex. Nat Rev Neurosci. 2005;6:435–448. [PubMed]
35. Huettel SA. Behavioral, but not reward, risk modulates activation of prefrontal, parietal, and insular cortices. Cogn Affect Behav Neurosci. 2006;6:141–151. [PubMed]
36. Kim H, Shimojo S, O’Doherty JP. Is avoiding an aversive outcome rewarding? Neural substrates of avoidance learning in the human brain. PLoS Biol. 2006;4:e233. [PMC free article] [PubMed]
37. Knutson B, Adams CM, Fong GW, Hommer D. Anticipation of increasing monetary reward selectively recruits nucleus accumbens. J Neurosci. 2001a;21:RC159. [PubMed]
38. Knutson B, Fong GW, Adams CM, Varner JL, Hommer D. Dissociation of reward anticipation and outcome with event-related fMRI. Neuroreport. 2001b;12:3683–3687. [PubMed]
39. Knutson B, Fong GW, Bennett SM, Adams CM, Hommer D. A region of mesial prefrontal cortex tracks monetarily rewarding outcomes: characterization with rapid event-related fMRI. Neuroimage. 2003;18:263–272. [PubMed]
40. Kringelbach ML. The human orbitofrontal cortex: linking reward to hedonic experience. Nat Rev Neurosci. 2005;6:691–702. [PubMed]
41. Kringelbach ML, O’Doherty J, Rolls ET, Andrews C. Activation of the human orbitofrontal cortex to a liquid food stimulus is correlated with its subjective pleasantness. Cereb Cortex. 2003;13:1064–1071. [PubMed]
42. Kringelbach ML, Rolls ET. The functional neuroanatomy of the human orbitofrontal cortex: evidence from neuroimaging and neuropsychology. Prog Neurobiol. 2004;72:341–372. [PubMed]
43. Kuhnen CM, Knutson B. The neural basis of financial risk taking. Neuron. 2005;47:763–770. [PubMed]
44. Laird AR, Eickhoff SB, Li K, Robin DA, Glahn DC, Fox PT. Investigating the functional heterogeneity of the default mode network using coordinate-based meta-analytic modeling. J Neurosci. 2009;29:14496–14505. [PMC free article] [PubMed]
45. Laird AR, Fox PM, Price CJ, Glahn DC, Uecker AM, Lancaster JL, Turkeltaub PE, Kochunov P, Fox PT. ALE meta-analysis: controlling the false discovery rate and performing statistical contrasts. Hum Brain Mapp. 2005;25:155–164. [PubMed]
46. Lau HC, Rogers RD, Haggard P, Passingham RE. Attention to intention. Science. 2004a;303:1208–1210. [PubMed]
47. Lau HC, Rogers RD, Ramnani N, Passingham RE. Willed action and attention to the selection of action. Neuroimage. 2004b;21:1407–1415. [PubMed]
48. Liu X, Powell DK, Wang H, Gold BT, Corbly CR, Joseph JE. Functional dissociation in frontal and striatal areas for processing of positive and negative reward information. J Neurosci. 2007;27:4587–4597. [PubMed]
49. McClure SM, Berns GS, Montague PR. Temporal prediction errors in a passive learning task activate human striatum. Neuron. 2003;38:339–346. [PubMed]
50. Mohr PN, Biele G, Heekeren HR. Neural processing of risk. J Neurosci. 2010;30:6613–6619. [PubMed]
51. Nieuwenhuis S, Slagter HA, von Geusau NJ, Heslenfeld DJ, Holroyd CB. Knowing good from bad: differential activation of human cortical areas by positive and negative outcomes. Eur J Neurosci. 2005;21:3161–3168. [PubMed]
52. O’Doherty JP, Critchley H, Deichmann R, Dolan RJ. Dissociating valence of outcome from behavioral control in human orbital and ventral prefrontal cortices. J Neurosci. 2003a;23:7931–7939. [PubMed]
53. O’Doherty JP, Dayan P, Friston K, Critchley H, Dolan RJ. Temporal difference models and reward-related learning in the human brain. Neuron. 2003b;38:329–337. [PubMed]
54. O’Doherty JP, Deichmann R, Critchley HD, Dolan RJ. Neural responses during anticipation of a primary taste reward. Neuron. 2002;33:815–826. [PubMed]
55. O’Doherty JP, Kringelbach ML, Rolls ET, Hornak J, Andrews C. Abstract reward and punishment representations in the human orbitofrontal cortex. Nat Neurosci. 2001;4:95–102. [PubMed]
56. Pagnoni G, Zink CF, Montague PR, Berns GS. Activity in human ventral striatum locked to errors of reward prediction. Nat Neurosci. 2002;5:97–98. [PubMed]
57. Paulus MP, Rogalsky C, Simmons A, Feinstein JS, Stein MB. Increased activation in the right insula during risk-taking decision making is related to harm avoidance and neuroticism. Neuroimage. 2003;19:1439–1448. [PubMed]
58. Phan KL, Wager T, Taylor SF, Liberzon I. Functional neuroanatomy of emotion: a meta-analysis of emotion activation studies in PET and fMRI. Neuroimage. 2002;16:331–348. [PubMed]
59. Plassmann H, O’Doherty J, Rangel A. Orbitofrontal cortex encodes willingness to pay in everyday economic transactions. J Neurosci. 2007;27:9984–9988. [PubMed]
60. Raichle ME, MacLeod AM, Snyder AZ, Powers WJ, Gusnard DA, Shulman GL. A default mode of brain function. Proc Natl Acad Sci USA. 2001;98:676–682. [PMC free article] [PubMed]
61. Raichle ME, Snyder AZ. A default mode of brain function: A brief history of an evolving idea. Neuroimage. 2007;37:1083–1090. [PubMed]
62. Ramnani N, Elliott R, Athwal BS, Passingham RE. Prediction error for free monetary reward in the human prefrontal cortex. Neuroimage. 2004;23:777–786. [PubMed]
63. Richlan F, Kronbichler M, Wimmer H. Functional abnormalities in the dyslexic brain: a quantitative meta-analysis of neuroimaging studies. Hum Brain Mapp. 2009;30:3299–3308. [PMC free article] [PubMed]
64. Rogers RD, Ramnani N, Mackay C, Wilson JL, Jezzard P, Carter CS, Smith SM. Distinct portions of anterior cingulate cortex and medial prefrontal cortex are activated by reward processing in separable phases of decision-making cognition. Biol Psychiatry. 2004;55:594–602. [PubMed]
65. Salimi-Khorshidi G, Smith SM, Keltner JR, Wager TD, Nichols TE. Meta-analysis of neuroimaging data: a comparison of image-based and coordinate-based pooling of studies. Neuroimage. 2009;45:810–823. [PubMed]
66. Schoenbaum G, Roesch M. Orbitofrontal cortex, associative learning, and expectancies. Neuron. 2005;47:633–636. [PMC free article] [PubMed]
67. Schoenbaum G, Roesch MR, Stalnaker TA, Takahashi YK. A new perspective on the role of the orbitofrontal cortex in adaptive behaviour. Nat Rev Neurosci 2009 [PMC free article] [PubMed]
68. Schoenbaum G, Setlow B, Saddoris MP, Gallagher M. Encoding predicted outcome and acquired value in orbitofrontal cortex during cue sampling depends upon input from basolateral amygdala. Neuron. 2003;39:855–867. [PubMed]
69. Schultz W, Tremblay L, Hollerman JR. Reward processing in primate orbitofrontal cortex and basal ganglia. Cereb Cortex. 2000;10:272–284. [PubMed]
70. Sugrue LP, Corrado GS, Newsome WT. Choosing the greater of two goods: neural currencies for valuation and decision making. Nat Rev Neurosci. 2005;6:363–375. [PubMed]
71. Tom SM, Fox CR, Trepel C, Poldrack RA. The neural basis of loss aversion in decision-making under risk. Science. 2007;315:515–518. [PubMed]
72. Toro R, Fox PT, Paus T. Functional coactivation map of the human brain. Cereb Cortex. 2008;18:2553–2559. [PMC free article] [PubMed]
73. Tremblay L, Schultz W. Relative reward preference in primate orbitofrontal cortex. Nature. 1999;398:704–708. [PubMed]
74. Tremblay L, Schultz W. Modifications of reward expectation-related neuronal activity during learning in primate orbitofrontal cortex. J Neurophysiol. 2000;83:1877–1885. [PubMed]
75. Trepel C, Fox CR, Poldrack RA. Prospect theory on the brain? Toward a cognitive neuroscience of decision under risk. Brain Res Cogn Brain Res. 2005;23:34–50. [PubMed]
76. Turkeltaub PE, Eden GF, Jones KM, Zeffiro TA. Meta-analysis of the functional neuroanatomy of single-word reading: method and validation. Neuroimage. 2002;16:765–780. [PubMed]
77. Ullsperger M, von Cramon DY. Error monitoring using external feedback: specific roles of the habenular complex, the reward system, and the cingulate motor area revealed by functional magnetic resonance imaging. J Neurosci. 2003;23:4308–4314. [PubMed]
78. Volkow ND, Fowler JS, Wang GJ. The addicted human brain: insights from imaging studies. J Clin Invest. 2003;111:1444–1451. [PMC free article] [PubMed]
79. Volkow ND, Wise RA. How can drug addiction help us understand obesity? Nat Neurosci. 2005;8:555–560. [PubMed]
80. Wallis JD. Orbitofrontal cortex and its contribution to decision-making. Annu Rev Neurosci. 2007;30:31–56. [PubMed]
81. Xiong J, Gao JH, Lancaster JL, Fox PT. Clustered pixels analysis for functional MRI activation studies of the human brain. Hum Brain Mapp. 1995;3:287–301.
82. Yang T, Shadlen MN. Probabilistic reasoning by neurons. Nature. 2007;447:1075–1080. [PubMed]
83. Zalla T, Koechlin E, Pietrini P, Basso G, Aquino P, Sirigu A, Grafman J. Differential amygdala responses to winning and losing: a functional magnetic resonance imaging study in humans. Eur J Neurosci. 2000;12:1764–1770. [PubMed]