Feature Selection for Value Function Approximation
<p>The field of reinforcement learning concerns the question of automated action selection given past experiences. As an agent moves through the state space, it must recognize which state choices are best in terms of allowing it to reach its goal. This is quantified with value functions, whi...
Main Author: | |
---|---|
Other Authors: | |
Published: |
2011
|
Subjects: | |
Online Access: | http://hdl.handle.net/10161/3891 |
id |
ndltd-DUKE-oai-dukespace.lib.duke.edu-10161-3891 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-DUKE-oai-dukespace.lib.duke.edu-10161-38912013-01-07T20:07:45ZFeature Selection for Value Function ApproximationTaylor, GavinArtificial IntelligenceComputer ScienceFeature SelectionReinforcement LearningValue Function Approximation<p>The field of reinforcement learning concerns the question of automated action selection given past experiences. As an agent moves through the state space, it must recognize which state choices are best in terms of allowing it to reach its goal. This is quantified with value functions, which evaluate a state and return the sum of rewards the agent can expect to receive from that state. Given a good value function, the agent can choose the actions which maximize this sum of rewards. Value functions are often chosen from a linear space defined by a set of features; this method offers a concise structure, low computational effort, and resistance to overfitting. However, because the number of features is small, this method depends heavily on these few features being expressive and useful, making the selection of these features a core problem. This document discusses this selection.</p><p>Aside from a review of the field, contributions include a new understanding of the role approximate models play in value function approximation, leading to new methods for analyzing feature sets in an intuitive way, both using the linear and the related kernelized approximation architectures. Additionally, we present a new method for automatically choosing features during value function approximation which has a bounded approximation error and produces superior policies, even in extremely noisy domains.</p>DissertationParr, Ronald2011Dissertationhttp://hdl.handle.net/10161/3891 |
collection |
NDLTD |
sources |
NDLTD |
topic |
Artificial Intelligence Computer Science Feature Selection Reinforcement Learning Value Function Approximation |
spellingShingle |
Artificial Intelligence Computer Science Feature Selection Reinforcement Learning Value Function Approximation Taylor, Gavin Feature Selection for Value Function Approximation |
description |
<p>The field of reinforcement learning concerns the question of automated action selection given past experiences. As an agent moves through the state space, it must recognize which state choices are best in terms of allowing it to reach its goal. This is quantified with value functions, which evaluate a state and return the sum of rewards the agent can expect to receive from that state. Given a good value function, the agent can choose the actions which maximize this sum of rewards. Value functions are often chosen from a linear space defined by a set of features; this method offers a concise structure, low computational effort, and resistance to overfitting. However, because the number of features is small, this method depends heavily on these few features being expressive and useful, making the selection of these features a core problem. This document discusses this selection.</p><p>Aside from a review of the field, contributions include a new understanding of the role approximate models play in value function approximation, leading to new methods for analyzing feature sets in an intuitive way, both using the linear and the related kernelized approximation architectures. Additionally, we present a new method for automatically choosing features during value function approximation which has a bounded approximation error and produces superior policies, even in extremely noisy domains.</p> === Dissertation |
author2 |
Parr, Ronald |
author_facet |
Parr, Ronald Taylor, Gavin |
author |
Taylor, Gavin |
author_sort |
Taylor, Gavin |
title |
Feature Selection for Value Function Approximation |
title_short |
Feature Selection for Value Function Approximation |
title_full |
Feature Selection for Value Function Approximation |
title_fullStr |
Feature Selection for Value Function Approximation |
title_full_unstemmed |
Feature Selection for Value Function Approximation |
title_sort |
feature selection for value function approximation |
publishDate |
2011 |
url |
http://hdl.handle.net/10161/3891 |
work_keys_str_mv |
AT taylorgavin featureselectionforvaluefunctionapproximation |
_version_ |
1716473566537449472 |