Model-free and model-based reward prediction errors in EEG

Sambrook, Tom D, Hardwick, Ben, Wills, Andy J and Goslin, Jeremy (2018) Model-free and model-based reward prediction errors in EEG. NeuroImage, 178. pp. 162-171. ISSN 1053-8119

[img]
Preview
PDF (Accepted manuscript) - Submitted Version
Available under License Creative Commons Attribution Non-commercial No Derivatives.

Download (928kB) | Preview

Abstract

Learning theorists posit two reinforcement learning systems: model-free and model-based. Model-based learning incorporates knowledge about structure and contingencies in the world to assign candidate actions with an expected value. Model-free learning is ignorant of the world’s structure; instead, actions hold a value based on prior reinforcement, with this value updated by expectancy violation in the form of a reward prediction error. Because they use such different learning mechanisms, it has been previously assumed that model-based and model-free learning are computationally dissociated in the brain. However, recent fMRI evidence suggests that the brain may compute reward prediction errors to both model-free and model-based estimates of value, signalling the possibility that these systems interact. Because of its poor temporal resolution, fMRI risks confounding reward prediction errors with other feedback-related neural activity. In the present study, EEG was used to show the presence of both model-based and model-free reward prediction errors and their place in a temporal sequence of events including state prediction errors and action value updates. This demonstration of model-based prediction errors questions a long-held assumption that model-free and model-based learning are dissociated in the brain.

Item Type: Article
Faculty \ School: Faculty of Social Sciences > School of Psychology
Depositing User: Pure Connector
Date Deposited: 10 May 2018 10:30
Last Modified: 25 Jun 2020 00:21
URI: https://ueaeprints.uea.ac.uk/id/eprint/67007
DOI: 10.1016/j.neuroimage.2018.05.023

Actions (login required)

View Item View Item