Abstraction and search for decision-theoretic planning
We investigate the use Markov Decision Processes a.s a means of representing worlds in which actions have probabilistic effects. Markov Decision Processes provide many representational advantages over traditional planning representations. As well as being able to represent actions with more than...
Main Author: | |
---|---|
Format: | Others |
Language: | English |
Published: |
2009
|
Online Access: | http://hdl.handle.net/2429/5461 |
id |
ndltd-UBC-oai-circle.library.ubc.ca-2429-5461 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-UBC-oai-circle.library.ubc.ca-2429-54612018-01-05T17:32:35Z Abstraction and search for decision-theoretic planning Dearden, Richard William We investigate the use Markov Decision Processes a.s a means of representing worlds in which actions have probabilistic effects. Markov Decision Processes provide many representational advantages over traditional planning representations. As well as being able to represent actions with more than one possible result, they also provide a much richer way to represent good and bad states of the world. Conventional approaches for finding optimal plans for Markov Decision Processes are computationally expensive and generally impractical for the large domains and real-time requirements of many planning applications. For this reason, we have concentrated on producing approximately optimal plans using a minimal amount of computation. We describe two complementary methods for planning. The first is to generate ap proximately optimal plans using abstraction. By ignoring certain features of a planning problem, we can create a smaller problem for which an optimal plan can be efficiently found by conventional means. The plan for this smaller problem can be directly applied to the original problem, and also provides an estimate of the value of each possible state of the world. Our second technique uses these estimates as a heuristic, and applies game tree search techniques to try to determine a better action to perform in the current state of the system. By repeatedly choosing an action to perform by searching, and executing the action, we provide a planning algorithm which has a complexity that is independent of the number of possible states of the world. Science, Faculty of Computer Science, Department of Graduate 2009-03-04 2009-03-04 1994 1994-11 Text Thesis/Dissertation http://hdl.handle.net/2429/5461 eng For non-commercial purposes only, such as research, private study and education. Additional conditions apply, see Terms of Use https://open.library.ubc.ca/terms_of_use. 1862595 bytes application/pdf |
collection |
NDLTD |
language |
English |
format |
Others
|
sources |
NDLTD |
description |
We investigate the use Markov Decision Processes a.s a means of representing worlds
in which actions have probabilistic effects. Markov Decision Processes provide many
representational advantages over traditional planning representations. As well as being
able to represent actions with more than one possible result, they also provide a much
richer way to represent good and bad states of the world. Conventional approaches for
finding optimal plans for Markov Decision Processes are computationally expensive and
generally impractical for the large domains and real-time requirements of many planning
applications. For this reason, we have concentrated on producing approximately optimal
plans using a minimal amount of computation.
We describe two complementary methods for planning. The first is to generate ap
proximately optimal plans using abstraction. By ignoring certain features of a planning
problem, we can create a smaller problem for which an optimal plan can be efficiently
found by conventional means. The plan for this smaller problem can be directly applied
to the original problem, and also provides an estimate of the value of each possible state
of the world. Our second technique uses these estimates as a heuristic, and applies game
tree search techniques to try to determine a better action to perform in the current state
of the system. By repeatedly choosing an action to perform by searching, and executing
the action, we provide a planning algorithm which has a complexity that is independent
of the number of possible states of the world. === Science, Faculty of === Computer Science, Department of === Graduate |
author |
Dearden, Richard William |
spellingShingle |
Dearden, Richard William Abstraction and search for decision-theoretic planning |
author_facet |
Dearden, Richard William |
author_sort |
Dearden, Richard William |
title |
Abstraction and search for decision-theoretic planning |
title_short |
Abstraction and search for decision-theoretic planning |
title_full |
Abstraction and search for decision-theoretic planning |
title_fullStr |
Abstraction and search for decision-theoretic planning |
title_full_unstemmed |
Abstraction and search for decision-theoretic planning |
title_sort |
abstraction and search for decision-theoretic planning |
publishDate |
2009 |
url |
http://hdl.handle.net/2429/5461 |
work_keys_str_mv |
AT deardenrichardwilliam abstractionandsearchfordecisiontheoreticplanning |
_version_ |
1718587114774331392 |