Abstraction and search for decision-theoretic planning

We investigate the use Markov Decision Processes a.s a means of representing worlds in which actions have probabilistic effects. Markov Decision Processes provide many representational advantages over traditional planning representations. As well as being able to represent actions with more than...

Full description

Bibliographic Details
Main Author: Dearden, Richard William
Format: Others
Language:English
Published: 2009
Online Access:http://hdl.handle.net/2429/5461
id ndltd-UBC-oai-circle.library.ubc.ca-2429-5461
record_format oai_dc
spelling ndltd-UBC-oai-circle.library.ubc.ca-2429-54612018-01-05T17:32:35Z Abstraction and search for decision-theoretic planning Dearden, Richard William We investigate the use Markov Decision Processes a.s a means of representing worlds in which actions have probabilistic effects. Markov Decision Processes provide many representational advantages over traditional planning representations. As well as being able to represent actions with more than one possible result, they also provide a much richer way to represent good and bad states of the world. Conventional approaches for finding optimal plans for Markov Decision Processes are computationally expensive and generally impractical for the large domains and real-time requirements of many planning applications. For this reason, we have concentrated on producing approximately optimal plans using a minimal amount of computation. We describe two complementary methods for planning. The first is to generate ap proximately optimal plans using abstraction. By ignoring certain features of a planning problem, we can create a smaller problem for which an optimal plan can be efficiently found by conventional means. The plan for this smaller problem can be directly applied to the original problem, and also provides an estimate of the value of each possible state of the world. Our second technique uses these estimates as a heuristic, and applies game tree search techniques to try to determine a better action to perform in the current state of the system. By repeatedly choosing an action to perform by searching, and executing the action, we provide a planning algorithm which has a complexity that is independent of the number of possible states of the world. Science, Faculty of Computer Science, Department of Graduate 2009-03-04 2009-03-04 1994 1994-11 Text Thesis/Dissertation http://hdl.handle.net/2429/5461 eng For non-commercial purposes only, such as research, private study and education. Additional conditions apply, see Terms of Use https://open.library.ubc.ca/terms_of_use. 1862595 bytes application/pdf
collection NDLTD
language English
format Others
sources NDLTD
description We investigate the use Markov Decision Processes a.s a means of representing worlds in which actions have probabilistic effects. Markov Decision Processes provide many representational advantages over traditional planning representations. As well as being able to represent actions with more than one possible result, they also provide a much richer way to represent good and bad states of the world. Conventional approaches for finding optimal plans for Markov Decision Processes are computationally expensive and generally impractical for the large domains and real-time requirements of many planning applications. For this reason, we have concentrated on producing approximately optimal plans using a minimal amount of computation. We describe two complementary methods for planning. The first is to generate ap proximately optimal plans using abstraction. By ignoring certain features of a planning problem, we can create a smaller problem for which an optimal plan can be efficiently found by conventional means. The plan for this smaller problem can be directly applied to the original problem, and also provides an estimate of the value of each possible state of the world. Our second technique uses these estimates as a heuristic, and applies game tree search techniques to try to determine a better action to perform in the current state of the system. By repeatedly choosing an action to perform by searching, and executing the action, we provide a planning algorithm which has a complexity that is independent of the number of possible states of the world. === Science, Faculty of === Computer Science, Department of === Graduate
author Dearden, Richard William
spellingShingle Dearden, Richard William
Abstraction and search for decision-theoretic planning
author_facet Dearden, Richard William
author_sort Dearden, Richard William
title Abstraction and search for decision-theoretic planning
title_short Abstraction and search for decision-theoretic planning
title_full Abstraction and search for decision-theoretic planning
title_fullStr Abstraction and search for decision-theoretic planning
title_full_unstemmed Abstraction and search for decision-theoretic planning
title_sort abstraction and search for decision-theoretic planning
publishDate 2009
url http://hdl.handle.net/2429/5461
work_keys_str_mv AT deardenrichardwilliam abstractionandsearchfordecisiontheoreticplanning
_version_ 1718587114774331392