Reinforcement learning using the game of soccer
Trial and error learning methods are often ineffective when applied to robots. This is due to certain characteristics found in robotic domains such as large continuous state spaces, noisy sensors and faulty actuators. Learning algorithms work best with small discrete state spaces, discrete determ...
Main Author: | |
---|---|
Language: | English |
Published: |
2009
|
Online Access: | http://hdl.handle.net/2429/5465 |
id |
ndltd-LACETR-oai-collectionscanada.gc.ca-BVAU.2429-5465 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-LACETR-oai-collectionscanada.gc.ca-BVAU.2429-54652014-03-14T15:40:16Z Reinforcement learning using the game of soccer Ford, Roger David Trial and error learning methods are often ineffective when applied to robots. This is due to certain characteristics found in robotic domains such as large continuous state spaces, noisy sensors and faulty actuators. Learning algorithms work best with small discrete state spaces, discrete deterministic actions, and accurate identification of state. Since trial and error learning requires that an agent learn by trying actions under all possible situations, the large continuous state space is the most problematic of the above characteristics, causing the learning algorithm to become inefficient. There is rarely enough time to explicitly visit every state or enough memory to store the best action for every state. This thesis explores methods for achieving reinforcement learning on large continuous state spaces, where actions are not discrete. This is done by creating abstract states, allowing one state to represent numerous similar states. This saves time since not every state in the abstract state needs to be visited and saves space since only one state needs to be stored. The algorithm tested in this thesis learns which volumes of the state space are similar by recursively subdividing each volume with a KD-tree. Identifying if an abstract state should be split, which dimension should be split, and where that dimension should be split is done by collecting statistics on the previous effects of actions. Continuous actions are dealt with by giving actions inertia, so they can persist past state boundaries if it necessary. 2009-03-04 2009-03-04 1994 2009-03-04 1994-11 Electronic Thesis or Dissertation http://hdl.handle.net/2429/5465 eng UBC Retrospective Theses Digitization Project [http://www.library.ubc.ca/archives/retro_theses/] |
collection |
NDLTD |
language |
English |
sources |
NDLTD |
description |
Trial and error learning methods are often ineffective when applied to robots. This is
due to certain characteristics found in robotic domains such as large continuous state
spaces, noisy sensors and faulty actuators. Learning algorithms work best with small
discrete state spaces, discrete deterministic actions, and accurate identification of state.
Since trial and error learning requires that an agent learn by trying actions under all
possible situations, the large continuous state space is the most problematic of the above
characteristics, causing the learning algorithm to become inefficient. There is rarely
enough time to explicitly visit every state or enough memory to store the best action for
every state.
This thesis explores methods for achieving reinforcement learning on large continuous
state spaces, where actions are not discrete. This is done by creating abstract states,
allowing one state to represent numerous similar states. This saves time since not every
state in the abstract state needs to be visited and saves space since only one state needs
to be stored.
The algorithm tested in this thesis learns which volumes of the state space are similar
by recursively subdividing each volume with a KD-tree. Identifying if an abstract state
should be split, which dimension should be split, and where that dimension should be
split is done by collecting statistics on the previous effects of actions. Continuous actions
are dealt with by giving actions inertia, so they can persist past state boundaries if it
necessary. |
author |
Ford, Roger David |
spellingShingle |
Ford, Roger David Reinforcement learning using the game of soccer |
author_facet |
Ford, Roger David |
author_sort |
Ford, Roger David |
title |
Reinforcement learning using the game of soccer |
title_short |
Reinforcement learning using the game of soccer |
title_full |
Reinforcement learning using the game of soccer |
title_fullStr |
Reinforcement learning using the game of soccer |
title_full_unstemmed |
Reinforcement learning using the game of soccer |
title_sort |
reinforcement learning using the game of soccer |
publishDate |
2009 |
url |
http://hdl.handle.net/2429/5465 |
work_keys_str_mv |
AT fordrogerdavid reinforcementlearningusingthegameofsoccer |
_version_ |
1716650643801767936 |