Inferring relevance in a changing world

Reinforcement learning models of human and animal learning usually concentrate on how we learn the relationship between different stimuli or actions and rewards. However, in real world situations stimuli are ill-defined. On the one hand, our immediate environment is extremely multi-dimensional. On t...

Full description

Bibliographic Details
Main Authors: Robert C Wilson, Yael eNiv
Format: Article
Language:English
Published: Frontiers Media S.A. 2012-01-01
Series:Frontiers in Human Neuroscience
Subjects:
Online Access:http://journal.frontiersin.org/Journal/10.3389/fnhum.2011.00189/full
id doaj-4a5067beda274e3ab0cdc72bdbea8598
record_format Article
spelling doaj-4a5067beda274e3ab0cdc72bdbea85982020-11-25T02:54:04ZengFrontiers Media S.A.Frontiers in Human Neuroscience1662-51612012-01-01510.3389/fnhum.2011.001898805Inferring relevance in a changing worldRobert C Wilson0Yael eNiv1Princeton UniversityPrinceton UniversityReinforcement learning models of human and animal learning usually concentrate on how we learn the relationship between different stimuli or actions and rewards. However, in real world situations stimuli are ill-defined. On the one hand, our immediate environment is extremely multi-dimensional. On the other hand, in every decision-making scenario only a few aspects of the environment are relevant for obtaining reward, while most are irrelevant. Thus a key question is how do we learn these relevant dimensions, that is, how do we learn what to learn about? We investigated this process of representation learning experimentally, using a task in which one stimulus dimension was relevant for determining reward at each point in time. As in real life situations, in our task the relevant dimension can change without warning, adding ever-present uncertainty engendered by a constantly changing environment. We show that human performance on this task is better described by a suboptimal strategy based on selective attention and serial hypothesis testing rather than a normative strategy based on probabilistic inference. From this, we conjecture that the problem of inferring relevance in general scenarios is too computationally demanding for the brain to solve optimally. As a result the brain utilizes approximations, employing these even in simplified scenarios in which optimal representation learning is tractable, such as the one in our experiment.http://journal.frontiersin.org/Journal/10.3389/fnhum.2011.00189/fullDecision Makingreinforcement learningselective attentionBayesian inferencerepresentation learning
collection DOAJ
language English
format Article
sources DOAJ
author Robert C Wilson
Yael eNiv
spellingShingle Robert C Wilson
Yael eNiv
Inferring relevance in a changing world
Frontiers in Human Neuroscience
Decision Making
reinforcement learning
selective attention
Bayesian inference
representation learning
author_facet Robert C Wilson
Yael eNiv
author_sort Robert C Wilson
title Inferring relevance in a changing world
title_short Inferring relevance in a changing world
title_full Inferring relevance in a changing world
title_fullStr Inferring relevance in a changing world
title_full_unstemmed Inferring relevance in a changing world
title_sort inferring relevance in a changing world
publisher Frontiers Media S.A.
series Frontiers in Human Neuroscience
issn 1662-5161
publishDate 2012-01-01
description Reinforcement learning models of human and animal learning usually concentrate on how we learn the relationship between different stimuli or actions and rewards. However, in real world situations stimuli are ill-defined. On the one hand, our immediate environment is extremely multi-dimensional. On the other hand, in every decision-making scenario only a few aspects of the environment are relevant for obtaining reward, while most are irrelevant. Thus a key question is how do we learn these relevant dimensions, that is, how do we learn what to learn about? We investigated this process of representation learning experimentally, using a task in which one stimulus dimension was relevant for determining reward at each point in time. As in real life situations, in our task the relevant dimension can change without warning, adding ever-present uncertainty engendered by a constantly changing environment. We show that human performance on this task is better described by a suboptimal strategy based on selective attention and serial hypothesis testing rather than a normative strategy based on probabilistic inference. From this, we conjecture that the problem of inferring relevance in general scenarios is too computationally demanding for the brain to solve optimally. As a result the brain utilizes approximations, employing these even in simplified scenarios in which optimal representation learning is tractable, such as the one in our experiment.
topic Decision Making
reinforcement learning
selective attention
Bayesian inference
representation learning
url http://journal.frontiersin.org/Journal/10.3389/fnhum.2011.00189/full
work_keys_str_mv AT robertcwilson inferringrelevanceinachangingworld
AT yaeleniv inferringrelevanceinachangingworld
_version_ 1724722711060545536