Deutsch
 
Hilfe Datenschutzhinweis Impressum
  DetailsucheBrowse

Datensatz

 
 
DownloadE-Mail
  Temporal dynamics of prediction error processing reward-based decision making

Philiastides, M. G., Biele, G., Vavatzanidis, N., Kazzer, P., & Heekeren, H. R. (2010). Temporal dynamics of prediction error processing reward-based decision making. NeuroImage, 53(1), 221-232. doi:10.1016/j.neuroimage.2010.05.052.

Item is

Basisdaten

einblenden: ausblenden:
Genre: Zeitschriftenartikel

Externe Referenzen

einblenden:

Urheber

einblenden:
ausblenden:
 Urheber:
Philiastides, Marios G.1, 2, Autor
Biele, Guido1, 2, 3, Autor
Vavatzanidis, Niki1, Autor           
Kazzer, Philipp1, Autor
Heekeren, Hauke R.1, 2, 3, Autor           
Affiliations:
1Max Planck Institute for Human Development, Berlin, Germany, ou_persistent22              
2MPI for Human Cognitive and Brain Sciences, Max Planck Society, ou_634548              
3Department of Education & Psychology, Freie Universität Berlin, Germany, ou_persistent22              

Inhalt

einblenden:
ausblenden:
Schlagwörter: Decision making; Reward; Reinforcement learning; Prediction error; Single-trial; Model; EEG
 Zusammenfassung: Adaptive decision making depends on the accurate representation of rewards associated with potential choices. These representations can be acquired with reinforcement learning (RL) mechanisms, which use the prediction error (PE, the difference between expected and received rewards) as a learning signal to update reward expectations. While EEG experiments have highlighted the role of feedback-related potentials during performance monitoring, important questions about the temporal sequence of feedback processing and the specific function of feedback-related potentials during reward-based decision making remain. Here, we hypothesized that feedback processing starts with a qualitative evaluation of outcome-valence, which is subsequently complemented by a quantitative representation of PE magnitude. Results of a model-based single-trial analysis of EEG data collected during a reversal learning task showed that around 220 ms after feedback outcomes are initially evaluated categorically with respect to their valence (positive vs. negative). Around 300 ms, and parallel to the maintained valence-evaluation, the brain also represents quantitative information about PE magnitude, thus providing the complete information needed to update reward expectations and to guide adaptive decision making. Importantly, our single-trial EEG analysis based on PEs from an RL model showed that the feedback-related potentials do not merely reflect error awareness, but rather quantitative information crucial for learning reward contingencies.

Details

einblenden:
ausblenden:
Sprache(n): eng - English
 Datum: 2010-05-062010-02-262010-05-192010-05-252010-10-15
 Publikationsstatus: Erschienen
 Seiten: -
 Ort, Verlag, Ausgabe: -
 Inhaltsverzeichnis: -
 Art der Begutachtung: -
 Identifikatoren: DOI: 10.1016/j.neuroimage.2010.05.052
 Art des Abschluß: -

Veranstaltung

einblenden:

Entscheidung

einblenden:

Projektinformation

einblenden:

Quelle 1

einblenden:
ausblenden:
Titel: NeuroImage
Genre der Quelle: Zeitschrift
 Urheber:
Affiliations:
Ort, Verlag, Ausgabe: Orlando, FL : Academic Press
Seiten: - Band / Heft: 53 (1) Artikelnummer: - Start- / Endseite: 221 - 232 Identifikator: ISSN: 1053-8119
CoNE: https://pure.mpg.de/cone/journals/resource/954922650166