Effects of lesions of the nucleus accumbens core on choice between small certain rewards and large uncertain rewards in rats
© Cardinal and Howes; licensee BioMed Central Ltd. 2005
Received: 18 April 2005
Accepted: 28 May 2005
Published: 28 May 2005
Animals must frequently make choices between alternative courses of action, seeking to maximize the benefit obtained. They must therefore evaluate the magnitude and the likelihood of the available outcomes. Little is known of the neural basis of this process, or what might predispose individuals to be overly conservative or to take risks excessively (avoiding or preferring uncertainty, respectively). The nucleus accumbens core (AcbC) is known to contribute to rats' ability to choose large, delayed rewards over small, immediate rewards; AcbC lesions cause impulsive choice and an impairment in learning with delayed reinforcement. However, it is not known how the AcbC contributes to choice involving probabilistic reinforcement, such as between a large, uncertain reward and a small, certain reward. We examined the effects of excitotoxic lesions of the AcbC on probabilistic choice in rats.
Rats chose between a single food pellet delivered with certainty (p = 1) and four food pellets delivered with varying degrees of uncertainty (p = 1, 0.5, 0.25, 0.125, and 0.0625) in a discrete-trial task, with the large-reinforcer probability decreasing or increasing across the session. Subjects were trained on this task and then received excitotoxic or sham lesions of the AcbC before being retested. After a transient period during which AcbC-lesioned rats exhibited relative indifference between the two alternatives compared to controls, AcbC-lesioned rats came to exhibit risk-averse choice, choosing the large reinforcer less often than controls when it was uncertain, to the extent that they obtained less food as a result. Rats behaved as if indifferent between a single certain pellet and four pellets at p = 0.32 (sham-operated) or at p = 0.70 (AcbC-lesioned) by the end of testing. When the probabilities did not vary across the session, AcbC-lesioned rats and controls strongly preferred the large reinforcer when it was certain, and strongly preferred the small reinforcer when the large reinforcer was very unlikely (p = 0.0625), with no differences between AcbC-lesioned and sham-operated groups.
These results support the view that the AcbC contributes to action selection by promoting the choice of uncertain, as well as delayed, reinforcement.
Animals often need to choose between different courses of action on the basis of the eventual rewarding or reinforcing outcomes of those actions. However, the relationship between an action and an outcome is frequently uncertain: animals do not always obtain that for which they work. Therefore, animals must incorporate information on the probability of obtaining different rewards when making decisions about what to do. Little is known of the neural basis of this process. Furthermore, when making decisions under conditions of uncertainty, individuals vary as to how much uncertainty or risk they are willing to tolerate. Formally, individuals differ in how much they 'discount' the value of reinforcers as the uncertainty of the reinforcer increases (i.e. as the probability of the reinforcer declines, or the odds against obtaining the reinforcer increase) . Risk taking is one aspect of the personality trait of impulsivity [2–4] and is a feature of a number of psychiatric disorders, including pathological gambling and certain personality disorders [5–8]. The term 'risk' implies exposure to the possibility of an aversive consequence , which may include the possibility of not obtaining an anticipated reward. In the appetitive domain, risk taking is exemplified by the tendency to choose large rewards that are very uncertain, in preference to smaller, certain rewards. Abnormal risk taking may reflect dysfunction of reinforcement learning systems that mediate the effects of uncertain reward or punishment.
The nucleus accumbens (Acb) is one candidate structure that may influence choice involving uncertainty. The Acb responds to anticipated rewards in humans, other primates, and rats [10–17], and is innervated by dopamine (DA) neurons that respond to errors in reward prediction in a manner appropriate for a teaching signal [18–21]. There is clear evidence that the Acb is involved in the processing of delayed reinforcement and its influence upon choice. Damage to the nucleus accumbens core (AcbC) produces impulsive choice in rats [22, 23], reducing their ability to choose large, delayed rewards in preference to small, immediate rewards, yet these and other similar lesions do not appear to impair rats' ability to discriminate reward size [23–31]. Furthermore, AcbC lesions impair rats' ability to learn instrumental actions when the outcomes of those actions are delayed . The Acb may also be involved in the processing of uncertain or probabilistic reinforcement. DA neurons that innervate the Acb may fire in a manner related to reward probability [32–34] and the midbrain, the site of the cell bodies of these neurons, responds to stimulus uncertainty in humans . A greater blood flow response is observed in the human Acb during the selection of high-reward/high-risk options, compared to low-reward/low-risk outcomes, in a task where the risk is of not winning , with similar activation to high-reward/high-risk option selection in a task where the risk is of losing ; this latter activation was correlated with personality measures of harm avoidance. However, these studies are correlative, and it is not known whether the AcbC is causally involved in regulating choice involving uncertain reinforcement.
Testing schedule for probabilistic choice task. Subjects were trained and tested according to the schedule shown here. Initial pre- and postoperative testing was conducted with the probability of large reinforcer delivery declining across trial blocks from 1 to 0.0625 (the steps were p = 1, 0.5, 0.25, 0.125, and 0.0625). Subsequently, subjects were tested alternating between the hungry and sated state (as described in the Methods), before the reinforcement probabilities were manipulated further, as indicated.
Probability (small reward of 1 pellet)
Probability (large reward of 4 pellets)
Preoperative training (12 sessions)
p = 1
p = 1 to 0.0625, decreasing
Surgery and recovery
Postoperative baseline testing (12 sessions)
p = 1
p = 1 to 0.0625, decreasing
Satiety tests (4 sessions)
p = 1
p = 1 to 0.0625, decreasing
Equal probabilities, certain (6 sessions)
p = 1
p = 1
Increasing probabilities (12 sessions)
p = 1
p = 0.0625 to 1, increasing
Large reinforcer always very unlikely (6 sessions)
p = 1
p = 0.0625
Early postoperative choice
In the initial postoperative period, AcbC-lesioned rats exhibited relative indifference between the two alternatives, choosing the large reinforcer close to 50% of the time at all large-reinforcer probabilities; as a result, AcbC-lesioned rats were more likely than shams to choose the large reinforcer when it was most uncertain (Figure 3b). An analysis of choice ratios calculated across sessions 13–15 was performed using the ANOVA model lesion2 × (probability5 × S). This revealed a lesion × probability interaction (F3.3,53.5 = 5.22, = .836, p = .002). Comparison of the two groups at individual large-reinforcer probabilities demonstrated that AcbC-lesioned rats chose the large/uncertain reinforcer more than shams at preinforcer = 0.0625 (pstatistical = .02), and at preinforcer = 0.125 (pstatistical = .009), but did not differ from shams at reinforcer probabilities of 0.25–1 (pstatistical ≥ .158). Nevertheless, simple effects of probability persisted both in shams (F2.8,30.9 = 32.3, = .702, p < .001) and in AcbC-lesioned rats (F4,20 = 5.37, p = .004). Choice at each preinforcer was compared to 50% (indifference) using post hoc two-tailed one-sample t tests, correcting pstatistical values using the Šidák correction for 5 comparisons. For shams, choice differed significantly from 50% at large-reinforcer probabilities of 0.0625 (when choice of the large reinforcer was less than 50%), 0.125 (less than 50%), and 1 (greater than 50%) (corrected pstatistical ≤ 0.007), but for AcbC-lesioned rats, choice did not differ significantly from 50% at any large-reinforcer probability (corrected pstatistical ≥ 0.81).
Final postoperative choice
By the final three sessions of the basic task (sessions 22–24; see Table 1), the pattern of choice in AcbC-lesioned rats had changed (Figure 3c). Once more, an analysis of choice ratios using the model lesion2 × (probability5 × S) revealed a lesion × probability interaction (F2.9,46.4 = 5.78, = .726, p = .002). By now, however, AcbC-lesioned rats did not differ from shams with reinforcer probabilities of 0.0625-0.25 (pstatistical ≥ .386) but chose the large reinforcer less than shams when its probability was 0.5 (pstatistical = .037) and 1 (pstatistical = .015). As before, effects of probability persisted both in shams (F2.3,24.9 = 49.5, = .565, p < .001) and in AcbC-lesioned rats (F4,20 = 9.45, p < .001).
Choice when both reinforcers were certain, or both uncertain
When the large and small reinforcers were both delivered with certainty, AcbC-lesioned and sham-operated rats strongly preferred the large reinforcer; when the small reinforcer was certain and the large reinforcer was consistently unlikely (preinforcer = 0.0625), all rats strongly preferred the small reinforcer (Figure 3d). There were no group differences in either case. This indicates that both AcbC-lesioned and sham-operated rats successfully discriminated the large reinforcer from the small reinforcer, and discriminated the certain large reinforcer from the uncertain large reinforcer. Choice ratios from the final sessions of training in these two conditions (sessions 34 and 52; see Table 1) were analysed using the model lesion2 × (trial block5 × S). In the 'certain' condition (session 34), there was no effect of lesion (F1,15 = 2.54, p = .132), no lesion × block interaction (F = 1.42, NS), and no effect of trial block (F1.5,21.9 = 2.12, = .365, p = .154). Similarly, in the 'uncertain' condition (session 52), there was no effect of lesion (F = 1.35, NS), no lesion × block interaction (F = 1.31, NS), and no effect of trial block (F < 1, NS).
Choice with ascending probabilities
After rats had been trained with the large-reinforcer probability increasing across the session, choice behaviour was similar to that with the decreasing-probability version of the task used initially, with AcbC-lesioned rats choosing the large/uncertain reinforcer less often than shams (Figure 3e; compare Figure 3c). Choice ratios from sessions 44–46 (see Table 1) were analysed using the model lesion2 × (probability5 × S). As before, there was a lesion × probability interaction (F4,64 = 9.29, p < .001), in addition to main effects of lesion (F1,16 = 19.5, p < .001) and probability (F4,64 = 95.6, p < .001), and there were strong effects of probability for both AcbC-lesioned rats (F1,20 = 20.7, p < .001) and shams (F3.1,34.4 = 119.6, = .781, p < .001). AcbC-lesioned rats differed from shams at reinforcer probabilities of 0.125 (pstatistical = .033), 0.25 (pstatistical < .001), 0.5 (pstatistical < .001), and 1 (pstatistical = .013), but not at preinforcer = 0.0625 (pstatistical = .881).
Postoperative choice: analysis by experienced probability
These analyses confirmed the pattern of results obtained on the basis of programmed probabilities. For the preoperative sessions, as expected, there was a main effect of experienced probability (F1,54 = 319.1, p < .001) but no significant terms involving lesion intent (Fs < 1, NS). For the baseline (decreasing-probability) task, both early (sessions 13–15) and late (sessions 22–24) in the postoperative testing, there was a lesion × experienced probability interaction (early: F1,54 = 25.7, p < .001; late: F1,54 = 20.8, p < .001). For the increasing-probability schedule (sessions 44–46), there was no lesion × experienced probability interaction (F1,54 = 1.80, p = .185) but there was a main effect of lesion (F1,16.0 = 9.36, p = .007).
Choice ratios from sham-operated rats on sessions 22–24 (the final 3 postoperative sessions on the basic task; see Table 1) were analysed using four different linear predictors, based either on the probability of delivery of a large reinforcer (given choice of the Large lever), or of the odds against delivery of a large reinforcer, calculated as odds against = (1 - p)/p. This established that choice patterns were predicted best, in linear fashion, by experienced probabilities (within-subject predictor allowing different slopes for each subject, r2 = 0.85) and programmed probabilities (r2 = 0.84), rather than by experienced odds (r2 = 0.61) or programmed odds (r2 = 0.67). Since optimal behaviour would give choice that was a step function of probability (i.e. it is optimal to choose the small/certain lever whenever the 4-pellet reinforcer is delivered with preinforcer < 0.25 and to choose the large/uncertain lever whenever preinforcer > 0.25), a single-parameter continuous function approximating a step function was also used to predict subjects' choice [the logistic function y = 100/e-(x-m)/bwith y as the percentage choice of the large reinforcer, x as the programmed probability, b = 0.01 as an approximation to b = 0 and m as the free parameter], but this gave a poor fit (r2 calculated as SSmodel/SStotal for a nonlinear fit: mean r2 = 0.26; note that individual values of r2 can fall outside the range [0,1] when calculated this way for nonlinear models) . Consequently, since choice was best described as a linear function of probability, indifference probabilities were calculated for sham-operated and AcbC-lesioned rats, namely the probability at which rats were equally likely to choose the small/certain and large/uncertain reinforcers. These were calculated via a linear regression of probability on choice (i.e. a regression in which probability was predicted from choice). This method has the potential to produce nonsensical probabilities for individual rats (if, for example, an individual's curve does not go both above and below the 50% choice point in a given set of sessions) but is nonetheless useful for group comparison. Experienced large-reinforcer probabilities (across all types of trials) were used, rather than programmed probabilities, though the pattern of results presented below was not altered by the use of programmed probabilities instead.
Omissions and latencies
Omissions were infrequent and not influenced by reinforcer probability or the lesion. Omission data from the final postoperative baseline sessions (sessions 22–24) were analysed. Overall, omissions (either failures to initiate a trial or to respond to an initiated trial) across all trial types occurred at a rate of 2.9 ± 0.9 % (sham) and 5.5 ± 1.9 % (AcbC). Omissions on choice trials for the same sessions were analysed using the model lesion2 × (probability5 × S). There were no effects of lesion (F1,16 = 1.95, NS) or probability (F1.6,25.2 = 2.56, = .394, p = .107), and no interaction (F = 1.04, NS). Almost all omissions were failures to initiate a trial (shams 0.9% of choice trials, AcbC 4.4%) rather than failures to respond once a trial had been initiated (shams 0.06% of choice trials, AcbC 0%).
Initiation latencies on choice trials for sessions 22–24 were analysed in the same manner. They were not affected by the lesion (F < 1, NS), nor by the large-reinforcer probability (F4,64 = 1.41, NS), and there was no lesion × probability interaction (F < 1, NS).
Response latencies were not affected by the lesion, but were affected both by the time in the session, with responding tending to get slower as the session progressed, and by the likelihood of obtaining a large reinforcer, with responding tending to get faster as large-reinforcer delivery became more likely. Response latencies on choice trials for sessions 22–24 were analysed using the model lesion2 × (trial block5 × choice2 × S). Response latencies varied across trial blocks: response latencies were initially 0.82 s (in the first trial block, when the large-reinforcer probability was 1) and slowed to 1.1 s (in the last trial block, when the large-reinforcer probability was 0.0625) (F3.1,25.1 = 2.97, = .785, p = .049). Latencies were not affected by the lesion, or the lever being chosen, and there were no interactions (maximum F was for response: F1,8 = 2.96, p = .124). To establish whether these effects were due to the large-reinforcer probability, or to progressive satiation or the passage of time, data from sessions 44–46 were also analysed, because in these sessions the large-reinforcer probability increased within the session. This time, there was a response × trial block interaction (F4,28 = 6.44, p = .001), with no other terms significant (Fs < 1, NS). Responding on the small/certain lever initially took 0.71 s in the first trial block and slowed to 0.95 s in the last trial block (F2.3,24.8 = 3.58, = .564, p = .038), but responding on the large/uncertain lever initially took 0.97 s (in the first trial block, when the large-reinforcer probability was 0.0625) and speeded up to 0.79 s (in the last trial block, when the large-reinforcer probability was 1) (F3.5,38.8 = 3.222, = .883, p = .027).
The lesion did not affect the latency to collect reward. Food collection latencies on rewarded trials were analysed across sessions 22–24, this time including both forced and choice trials to enable an analysis by response and probability. The model lesion2 × (probability5 × response2 × S) was used; this revealed main effects of response (F1,13 = 13.8, p = .003) and probability (F2.5,32.8 = 3.53, = .631, p = .031), but no other significant terms (maximum F was for lesion × response, F1,13 = 3.94, p = .069). Collection was faster following delivery of the large reinforcer than the small (4.1 versus 5.3 s, respectively), and got slightly slower across the session (4.4 s in the first trial block and 4.9 s in the last).
Amount of food obtained
Effects of hunger and satiety on choice
Alternating between hunger and satiety had no substantial effects on choice (Figure 6b). Choice ratios for sessions 25–28 were analysed using the model lesion2 × (hunger2 × probability5 × S). As before, a main effect of probability (F2.0,32.6 = 38.4, = .510, p < .001) and a lesion × probability interaction (F2.0,32.6 = 4.29, = .510, p = .022) were present; in addition, there was a marginally significant lesion × hunger × probability interaction (F4,64 = 2.51, p = .05). However, an effect of hunger was not detectable in either group alone, either for shams (hunger: F1,11 = 2.45, NS; hunger × probability: F4,44 = 2.18, NS) or for AcbC-lesioned rats (hunger: F < 1, NS; hunger × probability: F4,20 = 1.79, NS). Similarly, the differences between groups persisted both in the hungry (lesion × probability: F2.6,41.7 = 3.66, = .652, p = .024) and the sated (lesion × probability: F2.5,39.3 = 4.24, = .615, p = .016) conditions.
Locomotor activity and body mass
These results suggest that the AcbC contributes to the selection of uncertain rewards. AcbC-lesioned rats exhibited risk-averse choice: they chose large, uncertain rewards less than sham-operated controls when offered a smaller, certain alternative, even though they showed a strong and unaltered preference for large rewards over small rewards, and for certain rewards over uncertain rewards. By the end of testing, the control group behaved as if indifferent between a single certain food pellet and four pellets delivered with p = 0.32 (close to the probability of 0.25 that would represent rational indifference), while the AcbC-lesioned group behaved as if indifferent between a single certain pellet and four pellets delivered with p = 0.70.
Though these results establish that the lesions used in this study caused this pattern of behaviour, the precise mechanism by which this occurs is unknown: for example, it is possible that the damage caused to structures adjacent to the AcbC, though limited, played a role in this pattern of choice, or that adaptations in other structures consequent upon the lesion were important in the behavioural effects (particularly given that risk aversion was not apparent immediately but emerged with further time and postoperative experience with the task).
Choice in normal subjects
The dominant model of uncertainty or probability discounting [1, 44–46] suggests that subjects calculate a value for each reinforcer, according to its size and other parameters, and discount this by multiplying it by 1/(1+Hθ), where θ represents the odds against obtaining the reinforcer, θ = (1 - p)/p, and H represents an odds discounting parameter that is specific to the individual subject but stable over time for that subject. In this model, value is a hyperbolic function of the odds θ; such a hyperbolic function is supported by empirical research, at least in humans [44, 45, 47–50]. The present task is not well suited to evaluating such a quantitative model, since in discrete-trial schedules it is often the case that animals maximize, or allocate most of their choices to whichever option is the more favourable . However, the behaviour of normal subjects here can be evaluated as to its optimality. In the present task, neither risk aversion nor risk taking is optimal if carried to extremes. Optimal behaviour, to maximize the expected amount of food, is to choose the small/certain lever when the large (4-pellet) reinforcer probability is less than 0.25, to choose the large/uncertain lever when the probability exceeds 0.25, and to be indifferent at p = 0.25 (i.e. to exhibit a step function in choice). Shams' choice of the large reinforcer behaviour was better described by a linear function of the large-reinforcer probability than by such a step function. Nevertheless, shams' behaviour was reasonably close to the optimal in the most obvious way to measure optimality, namely the amount of food obtained (Figure 6b).
Effects of AcbC lesions in terms of conditioning processes
AcbC-lesioned rats chose the large, uncertain reinforcer less often than shams did, but only when a smaller certain reinforcer was available as an alternative; that is, they exhibited risk-averse choice. A number of simple explanations of the present results may be ruled out. For example, it is unlikely that the pattern of choice exhibited by AcbC-lesioned rats can be explained in terms of perseveration, within a session, on the initially-optimal lever. It might be that animals that perseverated on the lever delivering the small, certain reinforcer, because that lever was initially optimal, would appear to exhibit risk-averse choice in sessions in which the large-reinforcer probability increased across the session (Figure 3e), but this could not explain the same pattern of choice in sessions in which the same lever was initially suboptimal, i.e. when the large-reinforcer probability decreased across the session (Figure 3c). Furthermore, although AcbC lesions are known to affect processes through which Pavlovian conditioned stimuli (CSs) affect behaviour, including Pavlovian-instrumental transfer (PIT), autoshaping, and conditioned reinforcement [38, 52–57], there was no Pavlovian CS that was differentially associated with uncertain as opposed to certain reinforcement in this task, so these effects cannot explain the present results. It might be that the AcbC lesion impaired subjects' knowledge of the instrumental action-outcome contingency specifically for the uncertain outcome. There is some debate about the role of the AcbC in instrumental conditioning (see [43, 58, 59]) and goal-directed action, a subset of instrumental conditioning [58, 60, 61]. Manipulation of the AcbC can certainly affect instrumental learning [62–65]. However, the AcbC is not required for simple instrumental conditioning: rats with AcbC lesions acquire lever-press responses on fixed-ratio-1 schedules at supernormal levels , and rats with Acb or AcbC lesions are fully sensitive to changes in the action-outcome contingency [25, 53, 66]. However, when acquiring a sequence of random ratio schedules, AcbC-lesioned rats respond somewhat less than sham-operated controls , while lesions of the whole Acb made rats respond slightly, though not significantly, less on a similar sequence of random ratio schedules . Random ratio schedules clearly involve probabilistic reinforcement, so these results are consistent with the possibility that the present impairment shown by AcbC-lesioned rats in choosing large, unlikely rewards is due to impaired instrumental conditioning when the outcome is uncertain – and, conversely, that the impairment in simple instrumental learning seen previously  was specifically a result of the reward uncertainty inherent in a random ratio schedule, given that AcbC-lesioned rats learn instrumental responses normally or supernormally with certain immediate reinforcement . It is also possible that AcbC-lesioned rats represent the instrumental contingency normally with uncertain reward, but simply value the uncertain outcome less and respond less for it accordingly, as discussed next.
Effects of AcbC lesions in terms of probability discounting and reinforcer magnitude sensitivity
Since the present study required rats to choose between small, certain and large, uncertain rewards, an effect of the lesion to alter the perception of relative reward magnitude might affect choice, just as an alteration in the perception of reward probability might. For example, altering the absolute magnitudes of the reinforcers can affect choice involving probabilistic reinforcement [67, 68], as would be predicted if reinforcer 'value' is not simply a linear function of physical magnitude . Specifically, the present results (a tendency for AcbC-lesioned rats to choose the small, certain reinforcer more than shams) could be explained by 'risk aversion' (increased or steeper uncertainty/odds/probability discounting), or if the difference between 1 and 4 pellets was perceived to be smaller by AcbC-lesioned subjects than by shams (due to reduced discrimination between the two reinforcer magnitudes, or perhaps with a normal ability to tell the two apart but with an altered perception of relative value). For example, if a normal subject assigned values of 1 and 4 to the reinforcers, and a lesioned subject assigned values of 1 and 3 to the same reinforcers, then the lesioned subject would be less likely than the sham to choose the large reinforcer when it was made uncertain, even without any primary abnormality in the processing of probability. At first glance, this interpretation would appear to be supported by the observation that AcbC-lesioned rats chose the large reinforcer somewhat less often than shams when it was certain, as well as when it was uncertain. However, several lines of evidence suggest this explanation is not the correct one. When the large and the small reinforcers were both made consistently certain, there were no differences between AcbC-lesioned rats and controls (Figure 3d). Furthermore, other evidence indicates that AcbC lesions do not impair reinforcer magnitude discrimination or the perception of relative reinforcer value. Excitotoxic lesions of the whole Acb do not prevent rats from detecting changes in reward value (induced either by altering the concentration of a sucrose reward or by changing the deprivational state of the subject) . Such lesions also do not impair rats' ability to respond faster when environmental cues predict the availability of larger rewards , and nor does inactivation of the Acb with local anaesthetic or blockade of AMPA glutamate receptors in the Acb [27, 69]; the effects of intra-Acb NMDA receptor antagonists have varied [69–71]. AcbC-lesioned rats can still discriminate large from small rewards [23, 28]. Similarly, DA depletion of the Acb does not affect the ability to discriminate large from small reinforcers [29–31], and systemic DA antagonists do not affect the perceived quantity of food as assessed in a psychophysical procedure . Furthermore, a recent study found evidence that AcbC-lesioned rats may even show somewhat enhanced reinforcer magnitude discrimination (or an exaggerated perception of relative value) . Given that reinforcer magnitude discrimination appears to be unimpaired, at worst, by AcbC lesions, the observation in the present study that AcbC-lesioned rats chose the large reinforcer somewhat less often than controls in the task in which large-reinforcer probabilities changed throughout the session is more likely to be explained by within-session generalization [23, 73, 74] – i.e. that avoidance of the large reinforcer during trial blocks when it was uncertain generalized to trial blocks when it was certain. Together, these findings suggest that the present results are best explained as an effect of AcbC lesions to increase the rate of uncertainty/odds/probability discounting – effectively, a tendency to behave as if an uncertain outcome were less likely than it really is.
Probability versus delay discounting
It is known that AcbC lesions affect choice and learning involving delayed reinforcement [22–24]. It has been suggested that delay (or temporal) discounting, the process by which delayed reinforcers lose value, and probability (or odds) discounting, the process by which uncertain reinforcers lose value, reflect the same underlying process [44, 45, 75–81]. For example, in the present task, choosing the uncertain large reinforcer five times but only obtaining it on the fifth response might be seen as equivalent to a very long delay, on average, between choice of the large reinforcer and its eventual delivery. Alternatively, delays may be seen as entailing the ecological risk of losing the reward during the delay. The failure of AcbC-lesioned rats to choose an uncertain reinforcer (risk aversion, as seen in the stable phase of the present results) and their failure to choose a delayed reinforcer may therefore be explained in the same way. However, there is evidence that time and probability discounting are different and dissociable processes [1, 46, 82]. Most simply, it is not surprising that currency inflation affects human decisions involving delayed but not probabilistic financial reward . Moreover, the absolute magnitude of rewards can have different effects on delayed and probabilistic discounting [46, 84, 85]. A study looking at human choices in a gambling task found that individuals' propensity to choose rapidly (one, perhaps motoric, measure of delay aversion) and their propensity to bet large amounts of money on uncertain outcomes (a measure of risk taking) represented independent factors . Some studies have found abnormal delay discounting, but not uncertainty discounting, in drug addicts [82, 87–89], while gamblers have been observed to discount probabilistic rewards less steeply than controls (i.e. to take risks) without showing differences in delay discounting .
Implications for AcbC function and impulsivity
Impulsivity is multifaceted, reflecting – at the least – individual differences in distinct and dissociable processes involving information gathering, the selection of outcomes, and the inhibition of motor actions . Furthermore, as discussed above, delay discounting and probability discounting may also reflect separate processes. Damage to the AcbC can produce impulsive choice in the sense of an impaired ability to choose delayed rewards , in addition to hyperactivity [22, 24, 38, 41], though without impairments in attentional function  and without motoric impulsivity as assessed by the stop-signal task . In the context of choice involving uncertain appetitive reinforcement, 'impulsivity' would equate to risk taking (less steep uncertainty discounting or greater willingness to choose unlikely rewards). AcbC lesions, however, produced a risk-averse or conservative pattern of choice in the present study. Clearly, then, AcbC-lesioned rats cannot be characterized as impulsive in all senses. A more appropriate unifying concept would seem to be that the AcbC promotes the selection, and perhaps the salience, of uncertain and delayed rewards – perhaps, in general, of rewards that are not certain, imminent, or present . The AcbC promotes choice of  and learning with  delayed rewards. It appears to promote the selection of uncertain reinforcers (present results), and this is compatible with human imaging studies showing increased Acb blood flow during the selection of high-risk options [36, 37]. The Acb is required for PIT, the process by which Pavlovian CSs signalling reward enhance instrumental responding for those rewards [52, 53]. It is also required for autoshaping, or locomotor approach to appetitive Pavlovian CSs [38, 54–57], and it influences conditioned reinforcement, the process of working for CSs previously paired with reinforcement [38, 93–95]. Acb DA also contributes to subjects' motivation to work hard [96–100].
It is not known whether AcbC lesions would produce similar effects on choice involving uncertain aversive events. It would be expected that increased odds/uncertainty/probability discounting – effectively, a tendency to behave as if an uncertain outcome were less likely than it really is – would produce risk aversion for appetitive outcomes (reduced willingness to choose large, unlikely rewards) but risk proneness for aversive outcomes (increased willingness to choose large, uncertain punishments over small, certain punishments) . In humans, at least, the delay and probability discounting processes appear similar for rewards and losses [46, 101].
Relationship to structures and neuromodulator systems innervating the AcbC
The prefrontal cortex (PFC), which projects heavily to the AcbC , is also involved in decision-making under conditions of uncertainty. Humans with orbitofrontal cortex (OFC) or ventromedial PFC damage are impaired on the Iowa gambling task [103–105], in which subjects must learn to differentiate between low-reward, low-risk card decks that yield a net positive outcome and high-reward, high-risk decks that yield a net negative outcome, though the precise locus and nature of the deficit seen on this task is debated [106–108]. Choice between small, likely rewards and large, unlikely rewards increases cerebral blood flow in orbital and inferior PFC [109, 110], and OFC damage also impairs performance of a task requiring human subjects to choose between two possible outcomes and to bet on their choice, with lesioned subjects deciding slowly and failing to choose the optimal, most likely outcome . Excitotoxic lesions of the OFC make rats less likely than sham-operated controls to choose a large, uncertain reward over a small, certain reward ; OFC-lesioned rats had lower indifference odds (higher indifference probabilities; steeper uncertainty discounting) and exhibited risk-averse choice, just like the AcbC-lesioned subjects in the present study. There is direct evidence that OFC lesions do alter sensitivity to the relative magnitudes of the two rewards , as does OFC DA depletion , but the effects on uncertainty discounting are present in addition to those on reinforcer magnitude sensitivity .
The Acb is also innervated by the a number of neuromodulator systems, including the serotonin (5-hydroxytryptamine; 5-HT) system . Although manipulations of 5-HT influence choice involving delayed reinforcement, there is less evidence that they influence choice involving uncertainty and risk. Correlational studies have indicated that low cerebrospinal fluid (CSF) levels of the 5-HT metabolite 5-hydroxyindoleacetic acid (5-HIAA) are associated with risk taking in monkeys  and impulsive aggression, violence, and suicide in humans [118–122]. Forebrain 5-HT depletion tends to steepen temporal (delay) discounting (reviewed briefly by ); however, it does not appear to influence choice involving probabilistic reinforcement. Dietary tryptophan depletion [123–125] decreases levels of 5-HT metabolites in CSF, an indirect indicator of brain 5-HT levels, but has not been shown to affect probability discounting in humans [126, 127]; similarly, forebrain 5-HT depletion in rats does not affect choice between small, certain rewards and large, uncertain rewards . The AcbC also receives a substantial DA innervation, and DA neurons respond to reward prediction errors [18–21]. Although systemic D2-type DA receptor antagonists can induce impulsive choice involving delayed reinforcement , this effect may not occur in the Acb , the response of DA neurons specifically to reward uncertainty is debated [32–34], and little is known of the role of DA in choice involving uncertain rewards. Systemic noradrenergic (NA) blockade has also been shown to affect decision-making under uncertainty in humans, by reducing the discrimination between magnitudes of different losses when the probability of losing was high , though NA reuptake inhibition has not been shown to affect the Iowa gambling task . However, the Acb does not receive a substantial NA innervation .
We have shown that excitotoxic lesions of the AcbC induce risk-averse choice in rats. AcbC lesions did not prevent rats from discriminating a large reward from a small reward, or a certain reward from an uncertain reward. However, when offered the choice between a small/certain reward and a large/uncertain reward, AcbC-lesioned rats showed a reduced preference for the large/uncertain reward (compared to sham-operated controls) in their final pattern of postoperative choice. AcbC-lesioned rats exhibited a tendency to behave as if an uncertain outcome were less likely than was really the case. Together with previous studies, these results suggest that the AcbC contributes to reinforcement and choice particularly when the reinforcer is temporally distant or uncertain.
Subjects and housing conditions
The subjects were 24 male Lister hooded rats (Harlan-Olac UK Ltd) housed in a temperature-controlled room (minimum 22°C) under a 12:12 h reversed light-dark cycle (lights off 07:30 to 19:30). Subjects were approximately 15 weeks old on arrival at the laboratory and were given a minimum of a week to acclimatize, with free access to food, before experiments began. Preoperatively, subjects were housed in pairs; postoperatively, they were housed individually. Experiments took place between 09:00 and 21:00, with individual subjects being tested at a consistent time of day. Subjects had free access to water. During behavioural testing, subjects were fed ~15–16 g/day, an amount that maintains ~85–90% of free-feeding mass in normal male Lister hooded rats (the free-feeding mass being a steadily-increasing quantity at this age). Feeding occurred in the home cages at the end of the experimental day. As it was possible for subjects to earn substantial amounts of food in the behavioural tasks, the amount of food actually earned was taken into account when feeding with the maintenance diet in the home cages. All procedures were subject to UK Home Office approval (Project Licence 80/1767) under the Animals (Scientific Procedures) Act 1986.
Behavioural testing was conducted in one of two types of operant chamber of identical configuration (from Med Associates Inc., Georgia, Vermont, USA, or Paul Fray Ltd, Cambridge, UK). Each chamber was fitted with a 2.8 W overhead house light and two retractable levers on either side of an alcove fitted with an infrared photodiode to detect head entry and a 2.8 W lightbulb ('traylight'). Sucrose pellets (45 mg, Rodent Diet Formula P, Noyes, Lancaster, New Hampshire, USA) could be delivered into the alcove. The chambers were enclosed within sound-attenuating boxes fitted with fans to provide air circulation. The apparatus was controlled by software written by RNC in C++  using the Whisker control system . Equal numbers of subjects were trained in the two brands of operant chamber (12 in each type). Individual subjects were always tested in the same operant chamber.
Rats were first trained to press the left lever for single pellets on a fixed-ratio-1 schedule, in 30-min sessions, until they had obtained a total of 100 pellets. This procedure was repeated for the right lever. They were then trained to nosepoke to initiate presentation of a lever in discrete trials. Each session began with the levers retracted and the operant chamber in darkness. Every 40 s, a trial began with illumination of the houselight and the traylight. The subject was required to make a nosepoke response within 10 s, or the current trial was aborted and the chamber returned to darkness. If the subject nosepoked within this time limit, the traylight was extinguished and a single lever presented. If the rat failed to respond on the lever within 10 s, the lever was retracted and the chamber darkened, but if it responded, the houselight was switched off, a single pellet was delivered immediately and the traylight was illuminated until the rat collected the pellet (or a 10-s collection time limit elapsed, whereupon the chamber was darkened). In every pair of trials, the left lever was presented once and the right lever once, though the order within the pair of trials was random. Rats were trained to a criterion of 60 successful trials in one hour (the maximum possible with a 40-s period being 90). They then proceeded to the full task.
Probabilistic choice task
The task was based on delayed reinforcement choice tasks that have been described before [73, 74]. The session began in darkness with the levers retracted; this was designated the intertrial state. Trials began at 40-s intervals; the format of a single trial is shown in Figure 2. Each trial began with the illumination of the houselight and the traylight. The rat was required to make a nosepoke response, ensuring that it was centrally located at the start of the trial (latency to poke was designated the initiation latency). If the rat did not respond within 10 s of the start of the trial, the operant chamber was reset to the intertrial state until the next trial began and the trial was scored as an omission. If the rat was already nosepoking when the trial began, the next stage followed immediately. Upon a successful nosepoke, the traylight was extinguished and one or both levers were extended. One lever was designated the Large/Uncertain lever, the other the Small/Certain lever (counterbalanced left/right). The latency to choose a lever was recorded. (If the rat did not respond within 10 s of lever presentation, the chamber was reset to the intertrial state until the next trial and the trial was scored as an omission.) When a lever was chosen, both levers were retracted and the houselight was switched off. Choice of the Small lever caused the certain delivery of one pellet; choice of the Large lever caused the delivery of 4 pellets with a particular probability (see below). When reinforcement was delivered, the traylight was switched on. Multiple pellets were delivered 0.5 s apart. If the rat collected the pellets before the next trial began, then the traylight was switched off and the time from delivery of the first pellet until a nosepoke occurred was recorded as the collection latency. If the rat did not collect the food within 10 s of its delivery, the operant chamber entered the intertrial state, though collection latencies were still recorded up to the start of the next trial. The chamber was then in the intertrial state and remained so until the next trial. There was no mechanism to remove uneaten pellets, but failure to collect the reward was an extremely rare event. The large-reinforcer probability was varied systematically across the session as follows. A session consisted of 5 blocks, each comprising 16 trials in which only one lever was presented (8 trials for each lever, randomized in pairs) followed by 10 free-choice trials. The probability that the large reinforcer was delivered, given that the Large lever had been chosen (preinforcer), varied across blocks: it was initially 1, 0.5, 0.25, 0.125, and 0.0625, respectively, for each block. As trials began every 40 s and there were 130 trials per session, the total session length was ~87 minutes; subjects received one session per day. Choice ratios (percentage choice of the large reinforcer, for each trial block) were calculated using only choice trials on which the subject responded.
Excitotoxic lesions of the AcbC
Subjects were anaesthetized with Avertin (2% w/v 2,2,2-tribromoethanol, 1% w/v 2-methylbutan-2-ol, and 8% v/v ethanol in phosphate-buffered saline, sterilized by filtration, 10 ml/kg intraperitoneally) and placed in a Kopf or Stoelting stereotaxic frame (David Kopf Instruments, Tujunga, California, USA; Stoelting Co., Wood Dale, Illinois, USA) fitted with atraumatic ear bars. The skull was exposed and a dental drill was used to remove the bone directly above the injection sites. The dura mater was broken with the tip of a hypodermic needle, avoiding damage to underlying venous sinuses. Excitotoxic lesions of the AcbC were made by injecting 0.5 μl of 0.09 M quinolinic acid (Sigma, UK) per side through a glass micropipette at coordinates 1.2 mm anterior to bregma, ± 1.8 mm from the midline, and 7.1 mm below the skull surface at bregma; the incisor bar was 3.3 mm below the interaural line . The toxin had been dissolved in 0.1 M phosphate buffer (composition 0.07 M Na2HPO4, 0.028 M NaH2PO4 in double-distilled water, sterilized by filtration) and adjusted with NaOH to a final pH of 7.2–7.4. Toxin was injected over 3 min and the micropipette was left in place for 2 min following injections. Sham lesions were made in the same manner except that vehicle was infused. At the end of the operation, animals were given 15 ml/kg of sterile 5% w/v glucose, 0.9% w/v sodium chloride intraperitoneally. They were given a week to recover, with free access to food, and were handled regularly. Any instances of postoperative constipation were treated with liquid paraffin orally and rectally. At the end of this period, food restriction commenced or was resumed.
Subjects were trained preoperatively and tested postoperatively according to the schedule shown in Table 1. In the basic task, used for preoperative training, the probability of large reinforcer delivery declined across trial blocks from 1 to 0.0625 (in the order 1, 0.5, 0.25, 0.125, 0.0625). After subjects had been tested postoperatively for 12 sessions on this schedule, satiety tests were given, to establish the effect of varying primary motivational state on preference for probabilistic reinforcement. Subjects were tested for 4 sessions while alternating between hungry and sated states on consecutive days in counterbalanced fashion (half the subjects experienced hungry and sated days in the order HSHS, and half in the order SHSH). Following a 'hungry' session, animals were placed on free food (maintenance diet) until the start of the next day's 'sated' session, at which time the food was again removed for the 'hungry' session to follow. The comparison was therefore between food deprivation for ~22 h and satiety. Next, subjects were returned to the hungry state and tested for 6 sessions on a schedule in which both the large and small reinforcer were delivered with certainty. Next, the element of uncertainty was reintroduced for another 12 sessions, but this time the probability of large reinforcer delivery (given that the Large lever had been chosen) increased across blocks from 0.0625 to 1 (in the order 0.0625, 0.125, 0.25, 0.5, 1). Finally, subjects were tested for 6 sessions with the large reinforcer always being very unlikely (p = 0.0625), with the small reinforcer remaining certain.
Locomotor activity in a novel environment
Locomotor activity was measured in wire mesh cages, 25 (W) × 40 (D) × 18 (H) cm, each equipped with a water bottle and two horizontal photocell beams situated 1 cm from the floor that enabled movements along the long axis of the cage to be registered. Subjects were placed in these cages, which were initially unfamiliar to them, and their activity was recorded for 2 h. All animals were tested in the food-deprived state. Locomotor hyperactivity and reduced body mass gain have previously been part of the phenotype of AcbC-lesioned rats, though without alterations in the consumption of the reinforcer used in the present experiments [22, 24, 38, 39, 41].
Rats were deeply anaesthetized with pentobarbitone sodium (200 mg/ml, minimum of 1.5 ml i.p.) and perfused transcardially with 0.01 M phosphate-buffered saline (PBS) followed by 4% paraformaldehyde in PBS. Their brains were removed and postfixed in paraformaldehyde before being dehydrated in 20% sucrose for cryoprotection. The brains were sectioned coronally at 60 μm thickness on a freezing microtome and every third section mounted on chromium potassium sulphate/gelatin-coated glass microscope slides and allowed to dry. Sections were passed through a series of ethanol solutions of descending concentration (3 minutes in each of 100%, 95%, and 70% v/v ethanol in water) and stained for ~5 min with cresyl violet. The stain comprises 0.05% w/v aqueous cresyl violet (Raymond A. Lamb Ltd, Eastbourne, UK), 2 mM acetic acid, and 5 mM formic acid in water. Following staining, sections were rinsed in water and 70% ethanol before being differentiated in 95% ethanol. Finally, they were dehydrated and delipidated in 100% ethanol and Histoclear (National Diagnostics, UK) before being cover-slipped using DePeX mounting medium (BDH, UK) and allowed to dry. The sections were used to verify lesion placement and assess the extent of lesion-induced neuronal loss. Lesions were detectable as the absence of visible neurons (cell bodies of the order of 100 μm in diameter with a characteristic shape and appearance), often associated with a degree of tissue collapse (sometimes with consequent ventricular expansion when the lesion was adjacent to a ventricle) and gliosis (visible as the presence of smaller, densely-staining cells).
Data collected by the chamber control programs were imported into a relational database (Microsoft Access 97) for case selection and analysed with SPSS 11. Figures were created with SigmaPlot 2001/v7 and Adobe Illustrator 8. All graphs show group means and error bars are ±1 standard error of the mean (SEM) unless otherwise stated. Count data (e.g. locomotor activity counts), for which variance increases with the mean, were subjected to a square-root transformation prior to any analysis . Homogeneity of variance was verified using Levene's test . General linear models are described as dependent variable = A2 × B cov × (C5 × D cov × S) where A is a between-subjects factor with two levels, B is a between-subjects covariate, C is a within-subjects factor with five levels, and D is a within-subjects covariate; S denotes subjects in designs involving within-subjects factors . For repeated measures analyses, Mauchly's test of sphericity of the covariance matrix was applied  and the degrees of freedom corrected to more conservative values by multiplying them by the Huynh-Feldt epsilon for any terms involving factors in which the sphericity assumption was violated . Where multiple comparisons were conducted post hoc following a significant overall ANOVA effect for a factor with more than three levels, p values were corrected using the Šidák correction , in which pcorrected = 1 - (1 - puncorrected) n for n comparisons.
List of abbreviations used
nucleus accumbens core
nucleus accumbens shell
analysis of variance
- p :
- p reinforcer :
probability of delivery of the large reinforcer after it had been chosen
- p statistical :
statistical p value (probability of obtaining the observed data, or results more extreme, were the null hypothesis to be true)
- r 2 :
proportion of variance explained
standard error of the difference between means
standard error of the mean
sum of squares (sum of squared deviations)
volume per unit volume
weight per unit volume
- [x, y):
a range that includes x but not y
Supported by a Wellcome Trust programme grant (to Trevor W. Robbins, Barry J. Everitt, Angela C. Roberts, and Barbara J. Sahakian); conducted within the UK Medical Research Council (MRC) Behavioural and Clinical Neuroscience Centre, Cambridge. We thank three anonymous referees for their helpful comments. Competing interests: none declared.
- Ho MY, Mobini S, Chiang TJ, Bradshaw CM, Szabadi E: Theory and method in the quantitative analysis of "impulsive choice" behaviour: implications for psychopharmacology. Psychopharmacology. 1999, 146: 362-372.PubMedGoogle Scholar
- Evenden JL: Impulsivity: a discussion of clinical and experimental findings. Journal of Psychopharmacology. 1999, 13: 180-192.PubMedGoogle Scholar
- Eysenck SGB: The I7: development of a measure of impulsivity and its relationship to the superfactors of personality. The impulsive client: theory, research and treatment. Edited by: McCown WG, Johnson JL and Shure MB. 1993, Washington DC, American Psychological AssociationGoogle Scholar
- Daruna JH, Barnes PA: A neurodevelopmental view of impulsivity. The impulsive client: theory, research and treatment. Edited by: McCown WG, Johnson JL and Shure MB. 1993, Washington, DC, American Psychological AssociationGoogle Scholar
- APA: Diagnostic and Statistical Manual of Mental Disorders, fourth edition, text revision (DSM-IV-TR). 2000, Washington DC, American Psychiatric AssociationGoogle Scholar
- Roy A, DeJong J, Linnoila M: Extraversion in pathological gamblers: correlates with indices of noradrenergic function. Archives of General Psychiatry. 1989, 46: 679-681.PubMedGoogle Scholar
- Coccaro EF, Siever LJ: The neuropsychopharmacology of personality disorders. Psychopharmacology: The Fouth Generation of Progress. Edited by: Bloom FE and Kupfer DJ. 1995, New York, American College of Neuropsychopharmacology / Raven Press [ http://www.acnp.org/content-32.html ], 1567-1579.Google Scholar
- Holt DD, Green L, Myerson J: Is discounting impulsive?. Evidence from temporal and probability discounting in gambling and non-gambling college students. Behav Processes. 2003, 64: 355-367. 10.1016/S0376-6357(03)00141-4.PubMedGoogle Scholar
- OUP: New Shorter Oxford English Dictionary. 1997, Oxford, UK, Oxford University PressGoogle Scholar
- Knutson B, Adams CM, Fong GW, Hommer D: Anticipation of increasing monetary reward selectively recruits nucleus accumbens. J Neurosci. 2001, 21: RC159.PubMedGoogle Scholar
- Breiter HC, Aharon I, Kahneman D, Dale A, Shizgal P: Functional imaging of neural responses to expectancy and experience of monetary gains and losses. Neuron. 2001, 30: 619-639. 10.1016/S0896-6273(01)00303-8.PubMedGoogle Scholar
- Schultz W, Tremblay L, Hollerman JR: Reward processing in primate orbitofrontal cortex and basal ganglia. Cereb Cortex. 2000, 10: 272-284. 10.1093/cercor/10.3.272.PubMedGoogle Scholar
- Miyazaki K, Mogi E, Araki N, Matsumoto G: Reward-quality dependent anticipation in rat nucleus accumbens. Neuroreport. 1998, 9: 3943-3948.PubMedGoogle Scholar
- Schultz W, Apicella P, Scarnati E, Ljungberg T: Neuronal activity in monkey ventral striatum related to the expectation of reward. Journal of Neuroscience. 1992, 12: 4595-4610.PubMedGoogle Scholar
- Bjork JM, Knutson B, Fong GW, Caggiano DM, Bennett SM, Hommer DW: Incentive-elicited brain activation in adolescents: similarities and differences from young adults. J Neurosci. 2004, 24: 1793-1802. 10.1523/JNEUROSCI.4862-03.2004.PubMedGoogle Scholar
- Martin PD, Ono T: Effects of reward anticipation, reward presentation, and spatial parameters on the firing of single neurons recorded in the subiculum and nucleus accumbens of freely moving rats. Behav Brain Res. 2000, 116: 23-38. 10.1016/S0166-4328(00)00249-7.PubMedGoogle Scholar
- Cromwell HC, Schultz W: Effects of expectations for different reward magnitudes on neuronal activity in primate striatum. J Neurophysiol. 2003, 89: 2823-2838.PubMedGoogle Scholar
- Schultz W, Dickinson A: Neuronal coding of prediction errors. Annual Review of Neuroscience. 2000, 23: 473-500. 10.1146/annurev.neuro.23.1.473.PubMedGoogle Scholar
- Schultz W, Dayan P, Montague PR: A neural substrate of prediction and reward. Science. 1997, 275: 1593-1599. 10.1126/science.275.5306.1593.PubMedGoogle Scholar
- Schultz W: Predictive reward signal of dopamine neurons. J Neurophysiol. 1998, 80: 1-27.PubMedGoogle Scholar
- Schultz W, Tremblay L, Hollerman JR: Reward prediction in primate basal ganglia and frontal cortex. Neuropharmacology. 1998, 37: 421-429. 10.1016/S0028-3908(98)00071-9.PubMedGoogle Scholar
- Cardinal RN, Pennicott DR, Sugathapala CL, Robbins TW, Everitt BJ: Impulsive choice induced in rats by lesions of the nucleus accumbens core. Science. 2001, 292: 2499-2501. 10.1126/science.1060818.PubMedGoogle Scholar
- Cardinal RN, Robbins TW, Everitt BJ: Choosing delayed rewards: perspectives from learning theory, neurochemistry, and neuroanatomy. Choice, Behavioral Economics and Addiction. Edited by: Heather N and Vuchinich RE. 2003, Oxford, Elsevier, 183-213, 217-218.Google Scholar
- Cardinal RN, Cheung THC: Nucleus accumbens core lesions retard instrumental learning and performance with delayed reinforcement in the rat. BMC Neuroscience. 2005, 6: 9-10.1186/1471-2202-6-9.PubMed CentralPubMedGoogle Scholar
- Balleine B, Killcross S: Effects of ibotenic acid lesions of the nucleus accumbens on instrumental action. Behavioural Brain Research. 1994, 65: 181-193. 10.1016/0166-4328(94)90104-X.PubMedGoogle Scholar
- Brown VJ, Bowman EM: Discriminative cues indicating reward magnitude continue to determine reaction time of rats following lesions of the nucleus accumbens. European Journal of Neuroscience. 1995, 7: 2479-2485.PubMedGoogle Scholar
- Giertler C, Bohn I, Hauber W: Transient inactivation of the rat nucleus accumbens does not impair guidance of instrumental behaviour by stimuli predicting reward magnitude. Behav Pharmacol. 2004, 15: 55-63. 10.1097/00008877-200402000-00007.PubMedGoogle Scholar
- Cardinal RN, Winstanley CA, Robbins TW, Everitt BJ: Limbic corticostriatal systems and delayed reinforcement. Annals of the New York Academy of Sciences. 2004, 1021: 33-50. 10.1196/annals.1308.004.PubMedGoogle Scholar
- Salamone JD, Cousins MS, Bucher S: Anhedonia or anergia? Effects of haloperidol and nucleus accumbens dopamine depletion on instrumental response selection in a T-maze cost/benefit procedure. Behavioural Brain Research. 1994, 65: 221-229. 10.1016/0166-4328(94)90108-2.PubMedGoogle Scholar
- Salamone JD, Wisniecki A, Carlson BB, Correa M: Nucleus accumbens dopamine depletions make animals highly sensitive to high fixed ratio requirements but do not impair primary food reinforcement. Neuroscience. 2001, 105: 863-870. 10.1016/S0306-4522(01)00249-4.PubMedGoogle Scholar
- Cousins MS, Atherton A, Turner L, Salamone JD: Nucleus accumbens dopamine depletions alter relative response allocation in a T-maze cost/benefit task. Behavioural Brain Research. 1996, 74: 189-197. 10.1016/0166-4328(95)00151-4.PubMedGoogle Scholar
- Fiorillo CD, Tobler PN, Schultz W: Discrete coding of reward probability and uncertainty by dopamine neurons. Science. 2003, 299: 1898-1902. 10.1126/science.1077349.PubMedGoogle Scholar
- Tobler PN, Fiorillo CD, Schultz W: Adaptive coding of reward value by dopamine neurons. Science. 2005, 307: 1642-1645. 10.1126/science.1105370.PubMedGoogle Scholar
- Niv Y, Duff MO, Dayan P: Dopamine, uncertainty and TD learning. Behavioral and Brain Functions. 2005, 1: 6-10.1186/1744-9081-1-6.PubMed CentralPubMedGoogle Scholar
- Aron AR, Shohamy D, Clark J, Myers C, Gluck MA, Poldrack RA: Human midbrain sensitivity to cognitive feedback and uncertainty during classification learning. J Neurophysiol. 2004, 92: 1144-1152. 10.1152/jn.01209.2003.PubMedGoogle Scholar
- Ernst M, Nelson EE, McClure EB, Monk CS, Munson S, Eshel N, Zarahn E, Leibenluft E, Zametkin A, Towbin K, Blair J, Charney D, Pine DS: Choice selection and reward anticipation: an fMRI study. Neuropsychologia. 2004, 42: 1585-1597. 10.1016/j.neuropsychologia.2004.05.011.PubMedGoogle Scholar
- Matthews SC, Simmons AN, Lane SD, Paulus MP: Selective activation of the nucleus accumbens during risk-taking decision making. Neuroreport. 2004, 15: 2123-2127. 10.1097/00001756-200409150-00025.PubMedGoogle Scholar
- Parkinson JA, Olmstead MC, Burns LH, Robbins TW, Everitt BJ: Dissociation in effects of lesions of the nucleus accumbens core and shell on appetitive Pavlovian approach behavior and the potentiation of conditioned reinforcement and locomotor activity by d-amphetamine. Journal of Neuroscience. 1999, 19: 2401-2411.PubMedGoogle Scholar
- Cardinal RN: Neuropsychology of reinforcement processes in the rat [unpublished PhD thesis; www.pobox.com/~rudolf/publications/2001/PhD ]. 2001, Cambridge, UK, University of CambridgeGoogle Scholar
- Cameron AC, Windmeijer FAG: An R-squared measure of goodness of fit for some common nonlinear regression models. Journal of Econometrics. 1997, 77: 329-342. 10.1016/S0304-4076(96)01818-0.Google Scholar
- Maldonado-Irizarry CS, Kelley AE: Excitotoxic lesions of the core and shell subregions of the nucleus accumbens differentially disrupt body-weight regulation and motor activity in the rat. Brain Research Bulletin. 1995, 38: 551-559. 10.1016/0361-9230(95)02030-2.PubMedGoogle Scholar
- Parkinson JA: Limbic corticostriatal circuitry underlying Pavlovian associative learning [unpublished PhD thesis]. 1998, Cambridge, UK, University of CambridgeGoogle Scholar
- Kelley AE: Ventral striatal control of appetitive motivation: role in ingestive behavior and reward-related learning. Neurosci Biobehav Rev. 2004, 27: 765-776. 10.1016/j.neubiorev.2003.11.015.PubMedGoogle Scholar
- Rachlin H, Raineri A, Cross D: Subjective probability and delay. Journal of the Experimental Analysis of Behavior. 1991, 55: 233-244.PubMed CentralPubMedGoogle Scholar
- Rachlin H, Logue AW, Gibbon J, Frankel M: Cognition and behavior in studies of choice. Psychological Review. 1986, 93: 33-45. 10.1037//0033-295X.93.1.33.Google Scholar
- Green L, Myerson J: A discounting framework for choice with delayed and probabilistic rewards. Psychol Bull. 2004, 130: 769-792. 10.1037/0033-2909.130.5.769.PubMed CentralPubMedGoogle Scholar
- Rachlin H, Siegel E: Temporal patterning in probabilistic choice. Organizational Behavior and Human Decision Processes. 1994, 59: 161-176. 10.1006/obhd.1994.1054.Google Scholar
- Rachlin H, Brown J, Cross D: Discounting in judgments of delay and probability. Journal of Behavioral Decision Making. 2000, 13: 145-149. 10.1002/(SICI)1099-0771(200004/06)13:2<145::AID-BDM320>3.0.CO;2-4.Google Scholar
- Richards JB, Zhang L, Mitchell SH, de Wit H: Delay or probability discounting in a model of impulsive behavior: effect of alcohol. J Exp Anal Behav. 1999, 71: 121-143.PubMed CentralPubMedGoogle Scholar
- Kacelnik A: Normative and descriptive models of decision making: time discounting and risk sensitivity. Ciba Found Symp. 1997, 208: 51-67; discussion 67-70.PubMedGoogle Scholar
- Mackintosh NJ: The Psychology of Animal Learning. 1974, London, Academic PressGoogle Scholar
- Hall J, Parkinson JA, Connor TM, Dickinson A, Everitt BJ: Involvement of the central nucleus of the amygdala and nucleus accumbens core in mediating Pavlovian influences on instrumental behaviour. European Journal of Neuroscience. 2001, 13: 1984-1992. 10.1046/j.0953-816x.2001.01577.x.PubMedGoogle Scholar
- de Borchgrave R, Rawlins JN, Dickinson A, Balleine BW: Effects of cytotoxic nucleus accumbens lesions on instrumental conditioning in rats. Exp Brain Res. 2002, 144: 50-68. 10.1007/s00221-002-1031-y.PubMedGoogle Scholar
- Parkinson JA, Robbins TW, Everitt BJ: Selective excitotoxic lesions of the nucleus accumbens core and shell differentially affect aversive Pavlovian conditioning to discrete and contextual cues. Psychobiology. 1999, 27: 256-266.Google Scholar
- Parkinson JA, Willoughby PJ, Robbins TW, Everitt BJ: Disconnection of the anterior cingulate cortex and nucleus accumbens core impairs Pavlovian approach behavior: Further evidence for limbic cortical-ventral striatopallidal systems. Behavioral Neuroscience. 2000, 114: 42-63. 10.1037//0735-7044.114.1.42.PubMedGoogle Scholar
- Parkinson JA, Dalley JW, Cardinal RN, Bamford A, Fehnert B, Lachenal G, Rudarakanchana N, Halkerston KM, Robbins TW, Everitt BJ: Nucleus accumbens dopamine depletion impairs both acquisition and performance of appetitive Pavlovian approach behaviour: implications for mesoaccumbens dopamine function. Behavioural Brain Research. 2002, 137: 149-163. 10.1016/S0166-4328(02)00291-7.PubMedGoogle Scholar
- Cardinal RN, Parkinson JA, Lachenal G, Halkerston KM, Rudarakanchana N, Hall J, Morrison CH, Howes SR, Robbins TW, Everitt BJ: Effects of lesions of the nucleus accumbens core, anterior cingulate cortex, and central nucleus of the amygdala on autoshaping performance in rats. Behavioral Neuroscience. 2002, 116: 553-567. 10.1037//0735-7044.116.4.553.PubMedGoogle Scholar
- Cardinal RN, Parkinson JA, Hall J, Everitt BJ: Emotion and motivation: the role of the amygdala, ventral striatum, and prefrontal cortex. Neurosci Biobehav Rev. 2002, 26: 321-352. 10.1016/S0149-7634(02)00007-6.PubMedGoogle Scholar
- Cardinal RN, Everitt BJ: Neural and psychological mechanisms underlying appetitive learning: links to drug addiction. Curr Opin Neurobiol. 2004, 14: 156-162. 10.1016/j.conb.2004.03.004.PubMedGoogle Scholar
- Dickinson A, Balleine B: Motivational control of goal-directed action. Animal Learning & Behavior. 1994, 22: 1-18.Google Scholar
- Dickinson A: Instrumental conditioning. Animal Learning and Cognition. Edited by: Mackintosh NJ. 1994, San Diego, Academic Press, 45-79.Google Scholar
- Kelley AE, Smith-Roe SL, Holahan MR: Response-reinforcement learning is dependent on N-methyl-D-aspartate receptor activation in the nucleus accumbens core. Proc Natl Acad Sci U S A. 1997, 94: 12174-12179. 10.1073/pnas.94.22.12174.PubMed CentralPubMedGoogle Scholar
- Baldwin AE, Sadeghian K, Holahan MR, Kelley AE: Appetitive instrumental learning is impaired by inhibition of cAMP-dependent protein kinase within the nucleus accumbens. Neurobiol Learn Mem. 2002, 77: 44-62. 10.1006/nlme.2000.4002.PubMedGoogle Scholar
- Smith-Roe SL, Kelley AE: Coincident activation of NMDA and dopamine D1 receptors within the nucleus accumbens core is required for appetitive instrumental learning. J Neurosci. 2000, 20: 7737-7742.PubMedGoogle Scholar
- Hernandez PJ, Sadeghian K, Kelley AE: Early consolidation of instrumental learning requires protein synthesis in the nucleus accumbens. Nat Neurosci. 2002, 5: 1327-1331. 10.1038/nn973.PubMedGoogle Scholar
- Corbit LH, Muir JL, Balleine BW: The role of the nucleus accumbens in instrumental conditioning: evidence of a functional dissociation between accumbens core and shell. Journal of Neuroscience. 2001, 21: 3251-3260.PubMedGoogle Scholar
- Mazur JE: Choice between small certain and large uncertain reinforcers. Animal Learning & Behavior. 1988, 16: 199-205.Google Scholar
- Kirby KN, Marakovic NN: Delay-discounting probabilistic rewards: Rates decrease as amounts increase. Psychonomic Bulletin & Review. 1996Google Scholar
- Giertler C, Bohn I, Hauber W: Involvement of NMDA and AMPA/KA receptors in the nucleus accumbens core in instrumental learning guided by reward-predictive cues. Eur J Neurosci. 2005, 21: 1689-1702.PubMedGoogle Scholar
- Hauber W, Bohn I, Giertler C: NMDA, but not dopamine D(2), receptors in the rat nucleus accumbens are involved in guidance of instrumental behavior by stimuli predicting reward magnitude. Journal of Neuroscience. 2000, 20: 6282-6288.PubMedGoogle Scholar
- Giertler C, Bohn I, Hauber W: The rat nucleus accumbens is involved in guiding of instrumental responses by stimuli predicting reward magnitude. Eur J Neurosci. 2003, 18: 1993-1996. 10.1046/j.1460-9568.2003.02904.x.PubMedGoogle Scholar
- Martin-Iverson MT, Wilkie D, Fibiger HC: Effects of haloperidol and d-amphetamine on perceived quantity of food and tones. Psychopharmacology. 1987, 93: 374-381. 10.1007/BF00187260.PubMedGoogle Scholar
- Cardinal RN, Robbins TW, Everitt BJ: The effects of d-amphetamine, chlordiazepoxide, alpha-flupenthixol and behavioural manipulations on choice of signalled and unsignalled delayed reinforcement in rats. Psychopharmacology. 2000, 152: 362-375. 10.1007/s002130000536.PubMedGoogle Scholar
- Evenden JL, Ryan CN: The pharmacology of impulsive behaviour in rats: the effects of drugs on response choice with varying delays of reinforcement. Psychopharmacology. 1996, 128: 161-170. 10.1007/s002130050121.PubMedGoogle Scholar
- Sozou PD: On hyperbolic discounting and uncertain hazard rates [ DOI: 10.1098/rspb.1998.0534 ]. Proceedings of the Royal Society of London Series B: Biological Sciences. 1998, 265: 2015-2020. 10.1098/rspb.1998.0534.PubMed CentralGoogle Scholar
- Green L, Myerson J: Exponential versus hyperbolic discounting of delayed outcomes: Risk and waiting time. American Zoologist. 1996, 36: 496-505.Google Scholar
- Stevenson MK: A discounting model for decisions with delayed positive and negative outcomes. Journal of Experimental Psychology: General. 1986, 115: 131-154. 10.1037//0096-3422.214.171.124.Google Scholar
- Rachlin H, Castrogiovanni A, Cross D: Probability and delay in commitment. Journal of the Experimental Analysis of Behavior. 1987, 48: 347-353.PubMed CentralPubMedGoogle Scholar
- Mazur JE: Theories of probabilistic reinforcement. Journal of the Experimental Analysis of Behavior. 1989, 51: 87-99.PubMed CentralPubMedGoogle Scholar
- Mazur JE: Conditioned reinforcement and choice with delayed and uncertain primary reinforcers. Journal of the Experimental Analysis of Behavior. 1995, 63: 139-150.PubMed CentralPubMedGoogle Scholar
- Mazur JE: Choice, delay, probability, and conditioned reinforcement. Animal Learning & Behavior. 1997, 25: 131-147.Google Scholar
- Mitchell SH: Discounting the value of commodities according to different types of cost. Choice, Behavioral Economics and Addiction. Edited by: Heather N and Vuchinich RE. 2003, Oxford, Elsevier, 339-357.Google Scholar
- Ostaszewski P, Green L, Myerson J: Effects of inflation on the subjective value of delayed and probabilistic rewards. Psychonomic Bulletin & Review. 1998, 5: 324-333.Google Scholar
- Green L, Myerson J, Ostaszewski P: Amount of reward has opposite effects on the discounting of delayed and probabilistic outcomes. J Exp Psychol Learn Mem Cogn. 1999, 25: 418-427. 10.1037//0278-73126.96.36.1998.PubMedGoogle Scholar
- Myerson J, Green L, Hanson JS, Hold DD, Estle SJ: Discounting delayed and probabilistic rewards: Processes and traits. Journal of Economic Psychology. 2003, 24: 619-635. 10.1016/S0167-4870(03)00005-9.Google Scholar
- Deakin J, Aitken M, Robbins T, Sahakian BJ: Risk taking during decision-making in normal volunteers changes with age. J Int Neuropsychol Soc. 2004, 10: 590-598. 10.1017/S1355617704104104.PubMedGoogle Scholar
- Reynolds B, Richards JB, Horn K, Karraker K: Delay discounting and probability discounting as related to cigarette smoking status in adults. Behav Processes. 2004, 65: 35-42. 10.1016/S0376-6357(03)00109-8.PubMedGoogle Scholar
- Vuchinich RE, Calamas ML: Does the repeated gambles procedure measure impulsivity in social drinkers?. Exp Clin Psychopharmacol. 1997, 5: 157-162. 10.1037//1064-12188.8.131.52.PubMedGoogle Scholar
- Mitchell SH: Measures of impulsivity in cigarette smokers and non-smokers. Psychopharmacology. 1999, 146: 455-464.PubMedGoogle Scholar
- Evenden JL: Varieties of impulsivity. Psychopharmacology. 1999, 146: 348-361.PubMedGoogle Scholar
- Christakou A, Robbins TW, Everitt BJ: Prefrontal cortical-ventral striatal interactions involved in affective modulation of attentional performance: implications for corticostriatal circuit function. J Neurosci. 2004, 24: 773-780. 10.1523/JNEUROSCI.0949-03.2004.PubMedGoogle Scholar
- Eagle DM, Robbins TW: Lesions of the medial prefrontal cortex or nucleus accumbens core do not impair inhibitory control in rats performing a stop-signal reaction time task. Behav Brain Res. 2003, 146: 131-144. 10.1016/j.bbr.2003.09.022.PubMedGoogle Scholar
- Taylor JR, Robbins TW: Enhanced behavioural control by conditioned reinforcers following microinjections of d-amphetamine into the nucleus accumbens. Psychopharmacology. 1984, 84: 405-412. 10.1007/BF00555222.PubMedGoogle Scholar
- Taylor JR, Robbins TW: 6-Hydroxydopamine lesions of the nucleus accumbens, but not of the caudate nucleus, attenuate enhanced responding with reward-related stimuli produced by intra-accumbens d-amphetamine. Psychopharmacology. 1986, 90: 390-397. 10.1007/BF00179197.PubMedGoogle Scholar
- Cador M, Taylor JR, Robbins TW: Potentiation of the effects of reward-related stimuli by dopaminergic-dependent mechanisms in the nucleus accumbens. Psychopharmacology. 1991, 104: 377-385.PubMedGoogle Scholar
- Salamone JD, Correa M: Motivational views of reinforcement: implications for understanding the behavioral functions of nucleus accumbens dopamine. Behav Brain Res. 2002, 137: 3-25. 10.1016/S0166-4328(02)00282-6.PubMedGoogle Scholar
- Salamone JD, Correa M, Mingote SM, Weber SM: Nucleus accumbens dopamine and the regulation of effort in food-seeking behavior: implications for studies of natural motivation, psychiatry, and drug abuse. J Pharmacol Exp Ther. 2003, 305: 1-8. 10.1124/jpet.102.035063.PubMedGoogle Scholar
- Ikemoto S, Panksepp J: The role of nucleus accumbens dopamine in motivated behavior: a unifying interpretation with special reference to reward-seeking. 0165-0173. 1999, 31: 6-41.PubMedGoogle Scholar
- Salamone JD, Correa M, Mingote SM, Weber SM: Beyond the reward hypothesis: alternative functions of nucleus accumbens dopamine. Curr Opin Pharmacol. 2005, 5: 34-41. 10.1016/j.coph.2004.09.004.PubMedGoogle Scholar
- Mingote S, Weber SM, Ishiwari K, Correa M, Salamone JD: Ratio and time requirements on operant schedules: effort-related effects of nucleus accumbens dopamine depletions. Eur J Neurosci. 2005, 21: 1749-1757.PubMedGoogle Scholar
- Ostaszewski P, Karzel K: Discounting of delayed and probabilistic losses of different amounts. European Psychologist. 2002, 7: 295-301. 10.1027//1016-9040.7.4.295.Google Scholar
- Brog JS, Salyapongse A, Deutch AY, Zahm DS: The patterns of afferent innervation of the core and shell in the "accumbens" part of the rat ventral striatum: immunohistochemical detection of retrogradely transported fluoro-gold. Journal of Comparative Neurology. 1993, 338: 255-278. 10.1002/cne.903380209.PubMedGoogle Scholar
- Bechara A, Damasio AR, Damasio H, Anderson SW: Insensitivity to future consequences following damage to human prefrontal cortex. Cognition. 1994, 50: 7-15. 10.1016/0010-0277(94)90018-3.PubMedGoogle Scholar
- Bechara A, Tranel D, Damasio H, Damasio AR: Failure to respond autonomically to anticipated future outcomes following damage to prefrontal cortex. Cerebral Cortex. 1996, 6: 215-225.PubMedGoogle Scholar
- Bechara A, Damasio H, Tranel D, Damasio AR: Deciding advantageously before knowing the advantageous strategy. Science. 1997, 275: 1293-1295. 10.1126/science.275.5304.1293.PubMedGoogle Scholar
- Manes F, Sahakian B, Clark L, Rogers R, Antoun N, Aitken M, Robbins T: Decision-making processes following damage to the prefrontal cortex. Brain. 2002, 125: 624-639. 10.1093/brain/awf049.PubMedGoogle Scholar
- Clark L, Manes F, Antoun N, Sahakian BJ, Robbins TW: The contributions of lesion laterality and lesion volume to decision-making impairment following frontal lobe damage. Neuropsychologia. 2003, 41: 1474-1483. 10.1016/S0028-3932(03)00081-2.PubMedGoogle Scholar
- Fellows LK, Farah MJ: Different underlying impairments in decision-making following ventromedial and dorsolateral frontal lobe damage in humans. Cereb Cortex. 2005, 15: 58-63. 10.1093/cercor/bhh108.PubMedGoogle Scholar
- Rogers RD, Owen AM, Middleton HC, Williams EJ, Pickard JD, Sahakian BJ, Robbins TW: Choosing between small, likely rewards and large, unlikely rewards activates inferior and orbital prefrontal cortex. Journal of Neuroscience. 1999, 19: 9029-9038.PubMedGoogle Scholar
- Rogers RD, Ramnani N, Mackay C, Wilson JL, Jezzard P, Carter CS, Smith SM: Distinct portions of anterior cingulate cortex and medial prefrontal cortex are activated by reward processing in separable phases of decision-making cognition. Biol Psychiatry. 2004, 55: 594-602. 10.1016/j.biopsych.2003.11.012.PubMedGoogle Scholar
- Rogers RD, Everitt BJ, Baldacchino A, Blackshaw AJ, Swainson R, Wynne K, Baker NB, Hunter J, Carthy T, Booker E, London M, Deakin JF, Sahakian BJ, Robbins TW: Dissociable deficits in the decision-making cognition of chronic amphetamine abusers, opiate abusers, patients with focal damage to prefrontal cortex, and tryptophan-depleted normal volunteers: evidence for monoaminergic mechanisms. Neuropsychopharmacology. 1999, 20: 322-339. 10.1016/S0893-133X(98)00091-8.PubMedGoogle Scholar
- Mobini S, Body S, Ho MY, Bradshaw CM, Szabadi E, Deakin JF, Anderson IM: Effects of lesions of the orbitofrontal cortex on sensitivity to delayed and probabilistic reinforcement. Psychopharmacology. 2002, 160: 290-298. 10.1007/s00213-001-0983-0.PubMedGoogle Scholar
- Kheramin S, Body S, Herrera FM, Bradshaw CM, Szabadi E, Deakin JF, Anderson IM: The effect of orbital prefrontal cortex lesions on performance on a progressive ratio schedule: implications for models of inter-temporal choice. Behav Brain Res. 2005, 156: 145-152. 10.1016/j.bbr.2004.05.017.PubMedGoogle Scholar
- Kheramin S, Body S, Ho MY, Velazquez-Martinez DN, Bradshaw CM, Szabadi E, Deakin JF, Anderson IM: Effects of orbital prefrontal cortex dopamine depletion on inter-temporal choice: a quantitative analysis. Psychopharmacology (Berl). 2004, 175: 206-214.Google Scholar
- Kheramin S, Body S, Ho M, Velazquez-Martinez DN, Bradshaw CM, Szabadi E, Deakin JF, Anderson IM: Role of the orbital prefrontal cortex in choice between delayed and uncertain reinforcers: a quantitative analysis. Behav Processes. 2003, 64: 239-250. 10.1016/S0376-6357(03)00142-6.PubMedGoogle Scholar
- Halliday G, Harding A, Paxinos G: Serotonin and tachykinin systems. The Rat Nervous System. Edited by: Paxinos G. 1995, London, Academic Press, 929-974.Google Scholar
- Mehlman PT, Higley JD, Faucher I, Lilly AA, Taub DM, Vickers J, Suomi SJ, Linnoila M: Low CSF 5-HIAA concentrations and severe aggression and impaired impulse control in nonhuman primates. American Journal of Psychiatry. 1994, 151: 1485-1491.PubMedGoogle Scholar
- Åsberg M, Träskman L, Thorén P: 5-HIAA in the cerebrospinal fluid: a biochemical suicide predictor. Archives of General Psychiatry. 1976, 33: 1193-1197.PubMedGoogle Scholar
- Linnoila M, Virkkunen M, Scheinin M, Nuutila A, Rimon R, Goodwin FK: Low cerebrospinal fluid 5-hydroxyindoleacetic acid concentration differentiates impulsive from nonimpulsive violent behavior. Life Sciences. 1983, 33: 2609-2614. 10.1016/0024-3205(83)90344-2.PubMedGoogle Scholar
- Brown GL, Linnoila M: CSF serotonin metabolite (5HIAA) studies in depression, impulsivity and violence. Journal of Clinical Psychiatry. 1990, 51 (supplement 4): 31-41.Google Scholar
- Linnoila M, Virkkunen M, George T, Higley D: Impulse control disorders. International Clinical Psychopharmacology. 1993, 8 (Supplement 1): 53-56.Google Scholar
- Mann JJ: Neurobiology of suicidal behaviour. Nat Rev Neurosci. 2003, 4: 819-828. 10.1038/nrn1220.PubMedGoogle Scholar
- Biggio G, Fadda F, Fanni P, Tagliamonte A, Gessa G: Rapid depletion of serum tryptophan, brain tryptophan, serotonin and 5-hydroxyindoleacetic acid by a trytophan-free diet. Life Sciences. 1974, 14: 1321-1329. 10.1016/0024-3205(74)90440-8.PubMedGoogle Scholar
- Clemens JA, Bennett DR, Fuller RW: The effect of a tryptophan-free diet on prolactin and corticosterone release by serotonergic stimuli. Horm Metab Res. 1980, 12: 35-38.PubMedGoogle Scholar
- Delgado PL, Charney DS, Price LH, Landis H, Heninger GR: Neuroendocrine and behavioral effects of dietary tryptophan restriction in healthy subjects. Life Sci. 1989, 45: 2323-2332. 10.1016/0024-3205(89)90114-8.PubMedGoogle Scholar
- Anderson IM, Richell RA, Bradshaw CM: The effect of acute tryptophan depletion on probabilistic choice. J Psychopharmacol. 2003, 17: 3-7. 10.1177/0269881103017001687.PubMedGoogle Scholar
- Rogers RD, Tunbridge EM, Bhagwagar Z, Drevets WC, Sahakian BJ, Carter CS: Tryptophan depletion alters the decision-making of healthy volunteers through altered processing of reward cues. Neuropsychopharmacology. 2003, 28: 153-162. 10.1038/sj.npp.1300001.PubMedGoogle Scholar
- Mobini S, Chiang TJ, Ho MY, Bradshaw CM, Szabadi E: Effects of central 5-hydroxytryptamine depletion on sensitivity to delayed and probabilistic reinforcement. Psychopharmacology (Berl). 2000, 152: 390-397. 10.1007/s002130000542.Google Scholar
- Wade TR, de Wit H, Richards JB: Effects of dopaminergic drugs on delayed reward as a measure of impulsive behavior in rats. Psychopharmacology. 2000, 150: 90-101. 10.1007/s002130000402.PubMedGoogle Scholar
- Winstanley CA, Theobald DE, Dalley JW, Robbins TW: Interactions between serotonin and dopamine in the control of impulsive choice in rats: therapeutic implications for impulse control disorders. Neuropsychopharmacology. 2005, 30: 669-682.PubMedGoogle Scholar
- Rogers RD, Lancaster M, Wakeley J, Bhagwagar Z: Effects of beta-adrenoceptor blockade on components of human decision-making. Psychopharmacology (Berl). 2004, 172: 157-164. 10.1007/s00213-003-1641-5.Google Scholar
- O'Carroll RE, Papps BP: Decision making in humans: the effect of manipulating the central noradrenergic system. J Neurol Neurosurg Psychiatry. 2003, 74: 376-378. 10.1136/jnnp.74.3.376.PubMed CentralPubMedGoogle Scholar
- Aston-Jones G, Shipley MT, Grzanna R: The locus coeruleus, A5 and A7 noradrenergic cell groups. The Rat Nervous System. Edited by: Paxinos G. 1995, London, Academic Press, 183-213.Google Scholar
- Stroustrup B: The C++ Programming Language. 1986, Reading, Massachusetts, Addison-WesleyGoogle Scholar
- Cardinal RN, Aitken MRF: Whisker (version 2) [ www.whiskercontrol.com ]. 2001, Cambridge, UK, Cambridge University Technical Services LtdGoogle Scholar
- Paxinos G, Watson C: The Rat Brain in Stereotaxic Coordinates. 1998, London, Academic Press, FourthGoogle Scholar
- Howell DC: Statistical Methods for Psychology. 1997, Belmont, California, Wadsworth, FourthGoogle Scholar
- Levene H: Robust tests for the equality of variance. Contributions to probability and statistics. Edited by: Oklin I. 1960, Palo Alto, California, Stanford University PressGoogle Scholar
- Keppel G: Design and analysis: a researcher's handbook. 1982, London, Englewood Cliffs: Prentice-Hall, SecondGoogle Scholar
- Mauchly JW: Significance test for sphericity of a normal n-variate distribution. Annals of Mathematical Statistics. 1940, 11: 204-209.Google Scholar
- Huynh H, Feldt LS: Conditions under which mean square ratios in repeated measures designs have exact F-distributions. Journal of the American Statistical Association. 1970, 65: 1582-1589.Google Scholar
- Šidák Z: Rectangular confidence regions for the means of multivariate normal distributions. Journal of the American Statistical Association. 1967, 81: 826-831.Google Scholar