Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 22
Filter
Add more filters










Publication year range
1.
Neuron ; 110(13): 2046-2048, 2022 07 06.
Article in English | MEDLINE | ID: mdl-35797959

ABSTRACT

Corticostriatal circuits represent value and choice during value-guided decision making. In this issue of Neuron, Balewski et al. (2022) show that caudate nucleus and orbitofrontal cortex use distinct value signals during choice, which are consistent with two parallel valuation mechanisms, one fast, one slow.


Subject(s)
Caudate Nucleus , Prefrontal Cortex , Choice Behavior/physiology , Decision Making/physiology , Neurons/physiology , Prefrontal Cortex/physiology , Reward
2.
Neuroimage ; 246: 118780, 2022 02 01.
Article in English | MEDLINE | ID: mdl-34875383

ABSTRACT

Learning how to reach a reward over long series of actions is a remarkable capability of humans, and potentially guided by multiple parallel learning modules. Current brain imaging of learning modules is limited by (i) simple experimental paradigms, (ii) entanglement of brain signals of different learning modules, and (iii) a limited number of computational models considered as candidates for explaining behavior. Here, we address these three limitations and (i) introduce a complex sequential decision making task with surprising events that allows us to (ii) dissociate correlates of reward prediction errors from those of surprise in functional magnetic resonance imaging (fMRI); and (iii) we test behavior against a large repertoire of model-free, model-based, and hybrid reinforcement learning algorithms, including a novel surprise-modulated actor-critic algorithm. Surprise, derived from an approximate Bayesian approach for learning the world-model, is extracted in our algorithm from a state prediction error. Surprise is then used to modulate the learning rate of a model-free actor, which itself learns via the reward prediction error from model-free value estimation by the critic. We find that action choices are well explained by pure model-free policy gradient, but reaction times and neural data are not. We identify signatures of both model-free and surprise-based learning signals in blood oxygen level dependent (BOLD) responses, supporting the existence of multiple parallel learning modules in the brain. Our results extend previous fMRI findings to a multi-step setting and emphasize the role of policy gradient and surprise signalling in human learning.


Subject(s)
Brain/physiology , Decision Making/physiology , Functional Neuroimaging/methods , Learning/physiology , Magnetic Resonance Imaging/methods , Adult , Brain/diagnostic imaging , Female , Humans , Male , Models, Biological , Reinforcement, Psychology , Young Adult
3.
Proc Natl Acad Sci U S A ; 118(43)2021 10 26.
Article in English | MEDLINE | ID: mdl-34686596

ABSTRACT

Decisions are based on the subjective values of choice options. However, subjective value is a theoretical construct and not directly observable. Strikingly, distinct theoretical models competing to explain how subjective values are assigned to choice options often make very similar behavioral predictions, which poses a major difficulty for establishing a mechanistic, biologically plausible explanation of decision-making based on behavior alone. Here, we demonstrate that model comparison at the neural level provides insights into model implementation during subjective value computation even though the distinct models parametrically identify common brain regions as computing subjective value. We show that frontal cortical regions implement a model based on the statistical distributions of available rewards, whereas intraparietal cortex and striatum compute subjective value signals according to a model based on distortions in the representations of probabilities. Thus, better mechanistic understanding of how cognitive processes are implemented arises from model comparisons at the neural level, over and above the traditional approach of comparing models at the behavioral level alone.


Subject(s)
Brain/physiology , Choice Behavior/physiology , Adult , Bayes Theorem , Brain/diagnostic imaging , Brain Mapping , Decision Making/physiology , Female , Humans , Male , Models, Neurological , Models, Psychological , Perceptual Masking/physiology , Prefrontal Cortex/diagnostic imaging , Prefrontal Cortex/physiology , Young Adult
4.
Neuropsychologia ; 149: 107654, 2020 12.
Article in English | MEDLINE | ID: mdl-33069790

ABSTRACT

The temporo-parietal junction (TPJ) consistently emerges in other-regarding behavior, including tasks probing affective phenomena such as morality and empathy. Yet the TPJ is also recruited in processes with no affective or social component, such as visuo-spatial processing and mathematical cognition. We present serendipitous findings from a perceptual decision-making task on a bistable stimulus, the Necker Cube, performed in an MRI scanner. The stimulus in question is a transparent, wire-frame cube that evokes spontaneous switches in perception. Individuals can view the cube from below or from above, though a consistent bias is shown towards seeing the cube from above. We replicate this bias, finding participants spend more time in the from-above percept. However, in testing for BOLD differences between percept orientations, we found robust responses in bilateral TPJ for the from-above > from-below perceptual state. We speculate that this neural response comes from the sensory incongruence of viewing an object from above while lying supine in the scanner. We further speculate that the TPJ resolves this incongruence by facilitating an egocentric projection. Such a function would explain the TPJ's ubiquitous response to other-regarding, visuo-spatial and mathematical cognition, as all these phenomena demand an ability to ambulate through the coordinate space. Our findings suggest the TPJ may not play a specific role in social or moral components of other-regarding behavior, such as altruism, and further indirectly suggest that "pure", allocentric altruism may not correlate with the TPJ. Results further have implications on how the TPJ may be modulated by activities such as flight or drone operation. Finally, this study highlights the possible need for congruence between stimuli and body position in neuroimaging studies.


Subject(s)
Magnetic Resonance Imaging , Orientation , Bias , Humans , Parietal Lobe/diagnostic imaging
5.
Neuroimage ; 214: 116766, 2020 07 01.
Article in English | MEDLINE | ID: mdl-32247756

ABSTRACT

Organisms use rewards to navigate and adapt to (uncertain) environments. Error-based learning about rewards is supported by the dopaminergic system, which is thought to signal reward prediction errors to make adjustments to past predictions. More recently, the phasic dopamine response was suggested to have two components: the first rapid component is thought to signal the detection of a potentially rewarding stimulus; the second, slightly later component characterizes the stimulus by its reward prediction error. Error-based learning signals have also been found for risk. However, whether the neural generators of these signals employ a two-component coding scheme like the dopaminergic system is unknown. Here, using human high density EEG, we ask whether risk learning, or more generally speaking surprise-based learning under uncertainty, is similarly comprised of two temporally dissociable components. Using a simple card game, we show that the risk prediction error is reflected in the amplitude of the P3b component. This P3b modulation is preceded by an earlier component, that is modulated by the stimulus salience. Source analyses are compatible with the idea that both the early salience signal and the later risk prediction error signal are generated in insular, frontal, and temporal cortex. The identified sources are parts of the risk processing network that receives input from noradrenergic cells in the locus coeruleus. Finally, the P3b amplitude modulation is mirrored by an analogous modulation of pupil size, which is consistent with the idea that both the P3b and pupil size indirectly reflect locus coeruleus activity.


Subject(s)
Brain/physiology , Dopaminergic Neurons/physiology , Event-Related Potentials, P300/physiology , Learning/physiology , Reward , Adolescent , Adult , Brain Mapping , Electroencephalography , Female , Humans , Male , Reinforcement, Psychology , Uncertainty , Young Adult
6.
Neuroimage ; 210: 116549, 2020 04 15.
Article in English | MEDLINE | ID: mdl-31954844

ABSTRACT

The brain has been theorized to employ inferential processes to overcome the problem of uncertainty. Inference is thought to underlie neural processes, including in disparate domains such as value-based decision-making and perception. Value-based decision-making commonly involves deliberation, a time-consuming process that requires conscious consideration of decision variables. Perception, by contrast, is thought to be automatic and effortless. Both processes may call on a general neural system to resolve for uncertainty however. We addressed this question by directly comparing uncertainty signals in visual perception and an economic task using fMRI. We presented the same individuals with different versions of a bi-stable figure (Necker's cube) and with a gambling task during fMRI acquisition. We experimentally varied uncertainty, either on perceptual state or financial outcome. We found that inferential errors indexed by a formal account of surprise in the gambling task yielded BOLD responses in the anterior insula, in line with earlier findings. Moreover, we found perceptual uncertainty and surprise in the Necker Cube task yielded similar responses in the anterior insula. These results suggest that uncertainty, irrespective of domain, correlates to a common brain region, the anterior insula. These findings provide empirical evidence that the brain interacts with its environment through inferential processes.


Subject(s)
Brain Mapping , Cerebral Cortex/physiology , Decision Making/physiology , Pattern Recognition, Visual/physiology , Uncertainty , Adult , Cerebral Cortex/diagnostic imaging , Female , Humans , Magnetic Resonance Imaging , Male , Young Adult
7.
Front Artif Intell ; 3: 5, 2020.
Article in English | MEDLINE | ID: mdl-33733125

ABSTRACT

Uncertainty presents a problem for both human and machine decision-making. While utility maximization has traditionally been viewed as the motive force behind choice behavior, it has been theorized that uncertainty minimization may supersede reward motivation. Beyond reward, decisions are guided by belief, i.e., confidence-weighted expectations. Evidence challenging a belief evokes surprise, which signals a deviation from expectation (stimulus-bound surprise) but also provides an information gain. To support the theory that uncertainty minimization is an essential drive for the brain, we probe the neural trace of uncertainty-related decision variables, namely confidence, surprise, and information gain, in a discrete decision with a deterministic outcome. Confidence and surprise were elicited with a gambling task administered in a functional magnetic resonance imaging experiment, where agents start with a uniform probability distribution, transition to a non-uniform probabilistic state, and end in a fully certain state. After controlling for reward expectation, we find confidence, taken as the negative entropy of a trial, correlates with a response in the hippocampus and temporal lobe. Stimulus-bound surprise, taken as Shannon information, correlates with responses in the insula and striatum. In addition, we also find a neural response to a measure of information gain captured by a confidence error, a quantity we dub accuracy. BOLD responses to accuracy were found in the cerebellum and precuneus, after controlling for reward prediction errors and stimulus-bound surprise at the same time point. Our results suggest that, even absent an overt need for learning, the human brain expends energy on information gain and uncertainty minimization.

8.
Elife ; 82019 11 11.
Article in English | MEDLINE | ID: mdl-31709980

ABSTRACT

In many daily tasks, we make multiple decisions before reaching a goal. In order to learn such sequences of decisions, a mechanism to link earlier actions to later reward is necessary. Reinforcement learning (RL) theory suggests two classes of algorithms solving this credit assignment problem: In classic temporal-difference learning, earlier actions receive reward information only after multiple repetitions of the task, whereas models with eligibility traces reinforce entire sequences of actions from a single experience (one-shot). Here, we show one-shot learning of sequences. We developed a novel paradigm to directly observe which actions and states along a multi-step sequence are reinforced after a single reward. By focusing our analysis on those states for which RL with and without eligibility trace make qualitatively distinct predictions, we find direct behavioral (choice probability) and physiological (pupil dilation) signatures of reinforcement learning with eligibility trace across multiple sensory modalities.


Subject(s)
Cognition/physiology , Decision Making/physiology , Learning/physiology , Memory/physiology , Pupil/physiology , Reinforcement, Psychology , Reward , Algorithms , Humans , Markov Chains , Models, Neurological , Psychomotor Performance/physiology
9.
Neural Comput ; 30(1): 34-83, 2018 01.
Article in English | MEDLINE | ID: mdl-29064784

ABSTRACT

Surprise describes a range of phenomena from unexpected events to behavioral responses. We propose a novel measure of surprise and use it for surprise-driven learning. Our surprise measure takes into account data likelihood as well as the degree of commitment to a belief via the entropy of the belief distribution. We find that surprise-minimizing learning dynamically adjusts the balance between new and old information without the need of knowledge about the temporal statistics of the environment. We apply our framework to a dynamic decision-making task and a maze exploration task. Our surprise-minimizing framework is suitable for learning in complex environments, even if the environment undergoes gradual or sudden changes, and it could eventually provide a framework to study the behavior of humans and animals as they encounter surprising events.


Subject(s)
Adaptation, Psychological/physiology , Algorithms , Decision Making , Learning/physiology , Neurons/physiology , Environment , Female , Humans , Male , Models, Neurological
10.
Curr Opin Neurol ; 28(4): 344-50, 2015 Aug.
Article in English | MEDLINE | ID: mdl-26110801

ABSTRACT

PURPOSE OF REVIEW: Mild cognitive impairment (MCI) is a comorbid factor in Parkinson's disease. The aim of this review is to examine the recent neuroimaging findings in the search for Parkinson's disease MCI (PD-MCI) biomarkers to gain insight on whether MCI and specific cognitive deficits in Parkinson's disease implicate striatal dopamine or another system. RECENT FINDINGS: The evidence implicates a diffuse pathophysiology in PD-MCI rather than acute dopaminergic involvement. On the one hand, performance in specific cognitive domains, notably in set-shifting and learning, appears to vary with dopaminergic status. On the other hand, motivational states in Parkinson's disease along with their behavioral and physiological indices suggest a noradrenergic contribution to cognitive deficits in Parkinson's disease. Finally, Parkinson's disease's pattern of neurodegeneration offers an avenue for continued research in nigrostriatal dopamine's role in distinct behaviors, as well as the specification of dorsal and ventral striatal functions. SUMMARY: The search for PD-MCI biomarkers has employed an array of neuroimaging techniques, but still yields divergent findings. This may be due in part to MCI's broad definition, encompassing heterogeneous cognitive domains, only some of which are affected in Parkinson's disease. Most domains falling under the MCI umbrella include fronto-dependent executive functions, whereas others, notably learning, rely on the basal ganglia. Given the deterioration of the nigrostriatal dopaminergic system in Parkinson's disease, it has been the prime target of PD-MCI investigation. By testing well defined cognitive deficits in Parkinson's disease, distinct functions can be attributed to specific neural systems, overcoming conflicting results on PD-MCI. Apart from dopamine, other systems such as the neurovascular or noradrenergic systems are affected in Parkinson's disease. These factors may be at the basis of specific facets of PD-MCI for which dopaminergic involvement has not been conclusive. Finally, the impact of both dopaminergic and noradrenergic deficiency on motivational states in Parkinson's disease is examined in light of a plausible link between apathy and cognitive deficits.


Subject(s)
Apathy/physiology , Brain/metabolism , Cognitive Dysfunction/complications , Norepinephrine/metabolism , Parkinson Disease/complications , Cognitive Dysfunction/metabolism , Cognitive Dysfunction/psychology , Humans , Parkinson Disease/metabolism , Parkinson Disease/psychology
11.
Front Neurosci ; 7: 218, 2013.
Article in English | MEDLINE | ID: mdl-24311997
12.
J Neurosci ; 32(47): 16683-92, 2012 Nov 21.
Article in English | MEDLINE | ID: mdl-23175822

ABSTRACT

Individual risk preferences have a large influence on decisions, such as financial investments, career and health choices, or gambling. Decision making under risk has been studied both behaviorally and on a neural level. It remains unclear, however, how risk attitudes are encoded and integrated with choice. Here, we investigate how risk preferences are reflected in neural regions known to process risk. We collected functional magnetic resonance images of 56 human subjects during a gambling task (Preuschoff et al., 2006). Subjects were grouped into risk averters and risk seekers according to the risk preferences they revealed in a separate lottery task. We found that during the anticipation of high-risk gambles, risk averters show stronger responses in ventral striatum and anterior insula compared to risk seekers. In addition, risk prediction error signals in anterior insula, inferior frontal gyrus, and anterior cingulate indicate that risk averters do not dissociate properly between gambles that are more or less risky than expected. We suggest this may result in a general overestimation of prospective risk and lead to risk avoidance behavior. This is the first study to show that behavioral risk preferences are reflected in the passive evaluation of risky situations. The results have implications on public policies in the financial and health domain.


Subject(s)
Anticipation, Psychological/physiology , Gambling/psychology , Risk-Taking , Adult , Cerebral Cortex/physiology , Computer Simulation , Female , Frontal Lobe/physiology , Gyrus Cinguli/physiology , Humans , Image Processing, Computer-Assisted , Learning , Linear Models , Magnetic Resonance Imaging , Male , Neostriatum/physiology , Neuroimaging , Oxygen/blood , Reward , Young Adult
13.
PLoS Comput Biol ; 7(11): e1002280, 2011 Nov.
Article in English | MEDLINE | ID: mdl-22125485

ABSTRACT

This article presents the first attempt to formalize the optimization of experimental design with the aim of comparing models of brain function based on neuroimaging data. We demonstrate our approach in the context of Dynamic Causal Modelling (DCM), which relates experimental manipulations to observed network dynamics (via hidden neuronal states) and provides an inference framework for selecting among candidate models. Here, we show how to optimize the sensitivity of model selection by choosing among experimental designs according to their respective model selection accuracy. Using Bayesian decision theory, we (i) derive the Laplace-Chernoff risk for model selection, (ii) disclose its relationship with classical design optimality criteria and (iii) assess its sensitivity to basic modelling assumptions. We then evaluate the approach when identifying brain networks using DCM. Monte-Carlo simulations and empirical analyses of fMRI data from a simple bimanual motor task in humans serve to demonstrate the relationship between network identification and the optimal experimental design. For example, we show that deciding whether there is a feedback connection requires shorter epoch durations, relative to asking whether there is experimentally induced change in a connection that is known to be present. Finally, we discuss limitations and potential extensions of this work.


Subject(s)
Brain Mapping/methods , Brain/physiology , Models, Neurological , Bayes Theorem , Brain/anatomy & histology , Computational Biology , Computer Simulation , Feedback, Physiological , Humans , Magnetic Resonance Imaging , Monte Carlo Method , Psychomotor Performance/physiology , Reproducibility of Results , Research Design/standards
14.
Front Neurosci ; 5: 115, 2011.
Article in English | MEDLINE | ID: mdl-21994487

ABSTRACT

Our decisions are guided by the rewards we expect. These expectations are often based on incomplete knowledge and are thus subject to uncertainty. While the neurophysiology of expected rewards is well understood, less is known about the physiology of uncertainty. We hypothesize that uncertainty, or more specifically errors in judging uncertainty, are reflected in pupil dilation, a marker that has frequently been associated with decision making, but so far has remained largely elusive to quantitative models. To test this hypothesis, we measure pupil dilation while observers perform an auditory gambling task. This task dissociates two key decision variables - uncertainty and reward - and their errors from each other and from the act of the decision itself. We first demonstrate that the pupil does not signal expected reward or uncertainty per se, but instead signals surprise, that is, errors in judging uncertainty. While this general finding is independent of the precise quantification of these decision variables, we then analyze this effect with respect to a specific mathematical model of uncertainty and surprise, namely risk and risk prediction error. Using this quantification, we find that pupil dilation and risk prediction error are indeed highly correlated. Under the assumption of a tight link between noradrenaline (NA) and pupil size under constant illumination, our data may be interpreted as empirical evidence for the hypothesis that NA plays a similar role for uncertainty as dopamine does for reward, namely the encoding of error signals.

15.
Neuron ; 68(1): 149-60, 2010 Oct 06.
Article in English | MEDLINE | ID: mdl-20920798

ABSTRACT

Little is known about the neurobiological mechanisms underlying prosocial decisions and how they are modulated by social factors such as perceived group membership. The present study investigates the neural processes preceding the willingness to engage in costly helping toward ingroup and outgroup members. Soccer fans witnessed a fan of their favorite team (ingroup member) or of a rival team (outgroup member) experience pain. They were subsequently able to choose to help the other by enduring physical pain themselves to reduce the other's pain. Helping the ingroup member was best predicted by anterior insula activation when seeing him suffer and by associated self-reports of empathic concern. In contrast, not helping the outgroup member was best predicted by nucleus accumbens activation and the degree of negative evaluation of the other. We conclude that empathy-related insula activation can motivate costly helping, whereas an antagonistic signal in nucleus accumbens reduces the propensity to help.


Subject(s)
Brain Mapping , Group Processes , Helping Behavior , Individuality , Interpersonal Relations , Stress, Psychological/pathology , Adult , Brain/blood supply , Brain/physiopathology , Empathy/physiology , Humans , Image Processing, Computer-Assisted , Magnetic Resonance Imaging/methods , Male , Oxygen/blood , Pain/psychology , Predictive Value of Tests , Social Identification , Stress, Psychological/physiopathology , Stress, Psychological/psychology , Surveys and Questionnaires , Young Adult
16.
Trends Cogn Sci ; 13(8): 334-40, 2009 Aug.
Article in English | MEDLINE | ID: mdl-19643659

ABSTRACT

Although accumulating evidence highlights a crucial role of the insular cortex in feelings, empathy and processing uncertainty in the context of decision making, neuroscientific models of affective learning and decision making have mostly focused on structures such as the amygdala and the striatum. Here, we propose a unifying model in which insula cortex supports different levels of representation of current and predictive states allowing for error-based learning of both feeling states and uncertainty. This information is then integrated in a general subjective feeling state which is modulated by individual preferences such as risk aversion and contextual appraisal. Such mechanisms could facilitate affective learning and regulation of body homeostasis, and could also guide decision making in complex and uncertain environments.


Subject(s)
Cerebral Cortex/physiology , Emotions/physiology , Empathy , Uncertainty , Cerebral Cortex/anatomy & histology , Humans , Mental Processes/physiology , Models, Neurological , Models, Psychological , Neural Pathways/physiology
17.
Philos Trans R Soc Lond B Biol Sci ; 363(1511): 3801-11, 2008 Dec 12.
Article in English | MEDLINE | ID: mdl-18829433

ABSTRACT

The acknowledged importance of uncertainty in economic decision making has stimulated the search for neural signals that could influence learning and inform decision mechanisms. Current views distinguish two forms of uncertainty, namely risk and ambiguity, depending on whether the probability distributions of outcomes are known or unknown. Behavioural neurophysiological studies on dopamine neurons revealed a risk signal, which covaried with the standard deviation or variance of the magnitude of juice rewards and occurred separately from reward value coding. Human imaging studies identified similarly distinct risk signals for monetary rewards in the striatum and orbitofrontal cortex (OFC), thus fulfilling a requirement for the mean variance approach of economic decision theory. The orbitofrontal risk signal covaried with individual risk attitudes, possibly explaining individual differences in risk perception and risky decision making. Ambiguous gambles with incomplete probabilistic information induced stronger brain signals than risky gambles in OFC and amygdala, suggesting that the brain's reward system signals the partial lack of information. The brain can use the uncertainty signals to assess the uncertainty of rewards, influence learning, modulate the value of uncertain rewards and make appropriate behavioural choices between only partly known options.


Subject(s)
Neurons/physiology , Reward , Signal Transduction/physiology , Decision Making , Humans , Risk
18.
Neuroimage ; 41(1): 35-44, 2008 May 15.
Article in English | MEDLINE | ID: mdl-18375146

ABSTRACT

How the brain integrates signals from specific areas has been a longstanding critical question for neurobiologists. Two recent observations suggest a new approach to fMRI data analysis of this question. First, in many instances, the brain analyzes inputs by decomposing the information along several salient dimensions. For example, earlier work demonstrated that the brain splits a monetary gamble in terms of expected reward (ER) and variance of the reward (risk) [Preuschoff, K., Bossaerts, P., Quartz, S., 2006. Neural differentiation of expected reward and risk in human subcortical structures. Neuron 51, 381-390]. However, since ER and risk activate separate brain regions, the brain needs to integrate these activations to obtain an overall evaluation of the gamble. Second, recent evidence suggests that the correlation of the activity between neurons may serve a specific organizational purpose [Romo, R., Hernandez, A., Zainos, A., Salinas, E., 2003. Correlated neuronal discharges that increase coding efficiency during perceptual discrimination. Neuron 38, 649-657; Salinas, E., Sejnowski, T.J., 2001. Correlated neuronal activity and the flow of neural information. Nat. Rev. Neurosci. 2, 539]. Specifically, it is hypothesized that correlations allow brain regions to integrate several signals in a way that minimizes noise. Under this hypothesis, we show here that canonical correlation analysis of fMRI data identifies how the signals from several regions are combined. A general linear model then verifies whether the identified combination indeed activates a projection area in the brain. We illustrate the proposed procedure on data recorded while human subjects played a simple card game. We show that the brain adds the signals of ER and risk to form a measure that activates the medial prefrontal cortex, consistent with the role of this brain structure in the evaluation of monetary gambles.


Subject(s)
Brain/physiology , Image Processing, Computer-Assisted/statistics & numerical data , Magnetic Resonance Imaging/statistics & numerical data , Algorithms , Gambling/psychology , Humans , Linear Models , Neurons/physiology , Oxygen/blood , Prefrontal Cortex/physiology , Reproducibility of Results , Reward , Risk-Taking
19.
J Neurosci ; 28(11): 2745-52, 2008 Mar 12.
Article in English | MEDLINE | ID: mdl-18337404

ABSTRACT

Understanding how organisms deal with probabilistic stimulus-reward associations has been advanced by a convergence between reinforcement learning models and primate physiology, which demonstrated that the brain encodes a reward prediction error signal. However, organisms must also predict the level of risk associated with reward forecasts, monitor the errors in those risk predictions, and update these in light of new information. Risk prediction serves a dual purpose: (1) to guide choice in risk-sensitive organisms and (2) to modulate learning of uncertain rewards. To date, it is not known whether or how the brain accomplishes risk prediction. Using functional imaging during a simple gambling task in which we constantly changed risk, we show that an early-onset activation in the human insula correlates significantly with risk prediction error and that its time course is consistent with a role in rapid updating. Additionally, we show that activation previously associated with general uncertainty emerges with a delay consistent with a role in risk prediction. The activations correlating with risk prediction and risk prediction errors are the analogy for risk of activations correlating with reward prediction and reward prediction errors for reward expectation. As such, our findings indicate that our understanding of the neural basis of reward anticipation under uncertainty needs to be expanded to include risk prediction.


Subject(s)
Cerebral Cortex/metabolism , Risk-Taking , Adolescent , Adult , Choice Behavior/physiology , Female , Humans , Magnetic Resonance Imaging/methods , Male , Predictive Value of Tests , Psychomotor Performance/physiology , Reward
20.
Ann N Y Acad Sci ; 1104: 135-46, 2007 May.
Article in English | MEDLINE | ID: mdl-17344526

ABSTRACT

This article analyzes the simple Rescorla-Wagner learning rule from the vantage point of least squares learning theory. In particular, it suggests how measures of risk, such as prediction risk, can be used to adjust the learning constant in reinforcement learning. It argues that prediction risk is most effectively incorporated by scaling the prediction errors. This way, the learning rate needs adjusting only when the covariance between optimal predictions and past (scaled) prediction errors changes. Evidence is discussed that suggests that the dopaminergic system in the (human and nonhuman) primate brain encodes prediction risk, and that prediction errors are indeed scaled with prediction risk (adaptive encoding).


Subject(s)
Brain Mapping , Brain/anatomy & histology , Conditioning, Classical , Learning , Animals , Association Learning , Brain/physiology , Dopamine/metabolism , Humans , Least-Squares Analysis , Neurons/metabolism , Probability , Reinforcement, Psychology , Reward , Risk
SELECTION OF CITATIONS
SEARCH DETAIL
...