Reinforcement learning using the game of soccer

Trial and error learning methods are often ineffective when applied to robots. This is due to certain characteristics found in robotic domains such as large continuous state spaces, noisy sensors and faulty actuators. Learning algorithms work best with small discrete state spaces, discrete determ...

Full description

Bibliographic Details
Main Author: Ford, Roger David
Language:English
Published: 2009
Online Access:http://hdl.handle.net/2429/5465
id ndltd-LACETR-oai-collectionscanada.gc.ca-BVAU.2429-5465
record_format oai_dc
spelling ndltd-LACETR-oai-collectionscanada.gc.ca-BVAU.2429-54652014-03-14T15:40:16Z Reinforcement learning using the game of soccer Ford, Roger David Trial and error learning methods are often ineffective when applied to robots. This is due to certain characteristics found in robotic domains such as large continuous state spaces, noisy sensors and faulty actuators. Learning algorithms work best with small discrete state spaces, discrete deterministic actions, and accurate identification of state. Since trial and error learning requires that an agent learn by trying actions under all possible situations, the large continuous state space is the most problematic of the above characteristics, causing the learning algorithm to become inefficient. There is rarely enough time to explicitly visit every state or enough memory to store the best action for every state. This thesis explores methods for achieving reinforcement learning on large continuous state spaces, where actions are not discrete. This is done by creating abstract states, allowing one state to represent numerous similar states. This saves time since not every state in the abstract state needs to be visited and saves space since only one state needs to be stored. The algorithm tested in this thesis learns which volumes of the state space are similar by recursively subdividing each volume with a KD-tree. Identifying if an abstract state should be split, which dimension should be split, and where that dimension should be split is done by collecting statistics on the previous effects of actions. Continuous actions are dealt with by giving actions inertia, so they can persist past state boundaries if it necessary. 2009-03-04 2009-03-04 1994 2009-03-04 1994-11 Electronic Thesis or Dissertation http://hdl.handle.net/2429/5465 eng UBC Retrospective Theses Digitization Project [http://www.library.ubc.ca/archives/retro_theses/]
collection NDLTD
language English
sources NDLTD
description Trial and error learning methods are often ineffective when applied to robots. This is due to certain characteristics found in robotic domains such as large continuous state spaces, noisy sensors and faulty actuators. Learning algorithms work best with small discrete state spaces, discrete deterministic actions, and accurate identification of state. Since trial and error learning requires that an agent learn by trying actions under all possible situations, the large continuous state space is the most problematic of the above characteristics, causing the learning algorithm to become inefficient. There is rarely enough time to explicitly visit every state or enough memory to store the best action for every state. This thesis explores methods for achieving reinforcement learning on large continuous state spaces, where actions are not discrete. This is done by creating abstract states, allowing one state to represent numerous similar states. This saves time since not every state in the abstract state needs to be visited and saves space since only one state needs to be stored. The algorithm tested in this thesis learns which volumes of the state space are similar by recursively subdividing each volume with a KD-tree. Identifying if an abstract state should be split, which dimension should be split, and where that dimension should be split is done by collecting statistics on the previous effects of actions. Continuous actions are dealt with by giving actions inertia, so they can persist past state boundaries if it necessary.
author Ford, Roger David
spellingShingle Ford, Roger David
Reinforcement learning using the game of soccer
author_facet Ford, Roger David
author_sort Ford, Roger David
title Reinforcement learning using the game of soccer
title_short Reinforcement learning using the game of soccer
title_full Reinforcement learning using the game of soccer
title_fullStr Reinforcement learning using the game of soccer
title_full_unstemmed Reinforcement learning using the game of soccer
title_sort reinforcement learning using the game of soccer
publishDate 2009
url http://hdl.handle.net/2429/5465
work_keys_str_mv AT fordrogerdavid reinforcementlearningusingthegameofsoccer
_version_ 1716650643801767936