Reward-Based Learning, Model-Based and Model-Free

Quentin J.M. Huys, Anthony Cruickshank, Peggy Seriès

Research output: Chapter in Book/Report/Conference proceedingEntry for encyclopedia/dictionary

Abstract / Description of output

Reinforcement learning (RL) techniques are a set of solutions for optimal long-term action choice such that actions take into account both immediate and delayed consequences. They fall into two broad classes. Model-based approaches assume an explicit model of the environment and the agent.
The model describes the consequences of actions and the associated returns. From this, optimal policies can be inferred. Psychologically, model-based descriptions apply to goal-directed decisions, in which choices reflect current preferences over outcomes. Model-free approaches forgo any explicit knowledge of the dynamics of the environment or the consequences of actions and evaluate how good actions are through trial-and-error learning. Model-free values underlie habitual and Pavlovian conditioned responses that are emitted reflexively when faced with certain stimuli. While model-based techniques have substantial computational demands, model-free techniques require extensive experience.
Original languageEnglish
Title of host publicationEncyclopedia of Computational Neuroscience
EditorsDieter Jaeger, Ranu Jung
PublisherSpringer
Pages1-10
Number of pages10
ISBN (Electronic)978-1-4614-7320-6
DOIs
Publication statusPublished - 2014

Fingerprint

Dive into the research topics of 'Reward-Based Learning, Model-Based and Model-Free'. Together they form a unique fingerprint.

Cite this