On maximum-reward motion in stochastic environments

Thesis: S.M., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, 2015. === Cataloged from PDF version of thesis. === Includes bibliographical references (pages 75-77). === In this thesis, we consider the problem of an autonomous mobile robot operating in a stochastic...

Full description

Bibliographic Details
Main Author: Ma, Fangchang
Other Authors: Sertac Karaman.
Format: Others
Language:English
Published: Massachusetts Institute of Technology 2015
Subjects:
Online Access:http://hdl.handle.net/1721.1/98696
id ndltd-MIT-oai-dspace.mit.edu-1721.1-98696
record_format oai_dc
spelling ndltd-MIT-oai-dspace.mit.edu-1721.1-986962019-05-02T16:23:35Z On maximum-reward motion in stochastic environments Ma, Fangchang Sertac Karaman. Massachusetts Institute of Technology. Department of Aeronautics and Astronautics. Massachusetts Institute of Technology. Department of Aeronautics and Astronautics. Aeronautics and Astronautics. Thesis: S.M., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, 2015. Cataloged from PDF version of thesis. Includes bibliographical references (pages 75-77). In this thesis, we consider the problem of an autonomous mobile robot operating in a stochastic reward field to maximize total rewards collected in an online setting. This is a generalization of the problem where an unmanned aerial vehicle (UAV) collects data from randomly deployed unattended ground sensors (UGS). Specifically, the rewards are assumed to be generated by a Poisson point process. The robot has a limited perception range, and thus it discovers the reward field on the fly. The robot is assumed to be a dynamical system with substantial drift in one direction, e.g., a high-speed airplane, so it cannot traverse the entire field. The task of the robot is to maximize the total rewards collected during the course of the mission, given above constraints. Under such assumptions, we analyze the performance of a simple receding-horizon planning algorithm with respect to the perception range, robot agility and computational resources available. Firstly, we show that, with highly limited perception range, the robot is able to collect as many rewards as if it could see the entire reward field, if and only if the reward distribution is light-tailed. The second result attained shows that the expected rewards collected scale proportionally to the square root of the robot agility. Finally, we are able to prove that the overall computational workload increases linearly with the mission length, i.e., the distance of travel. We verify our results in simulation examples. At the end, we present one interesting application of our theoretical study to the ground sensor selection problem. For an inference/estimation task, we prove that sensors with randomized quality outperform those with homogeneous precisions, since random sensors yield a higher confidence level of estimation (lower variance), under certain technical assumptions. This finding might have practical implications on the design of UAV-UGS systems. by Fangchang Ma. S.M. 2015-09-17T19:05:31Z 2015-09-17T19:05:31Z 2015 2015 Thesis http://hdl.handle.net/1721.1/98696 920688365 eng M.I.T. theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. See provided URL for inquiries about permission. http://dspace.mit.edu/handle/1721.1/7582 77 pages application/pdf Massachusetts Institute of Technology
collection NDLTD
language English
format Others
sources NDLTD
topic Aeronautics and Astronautics.
spellingShingle Aeronautics and Astronautics.
Ma, Fangchang
On maximum-reward motion in stochastic environments
description Thesis: S.M., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, 2015. === Cataloged from PDF version of thesis. === Includes bibliographical references (pages 75-77). === In this thesis, we consider the problem of an autonomous mobile robot operating in a stochastic reward field to maximize total rewards collected in an online setting. This is a generalization of the problem where an unmanned aerial vehicle (UAV) collects data from randomly deployed unattended ground sensors (UGS). Specifically, the rewards are assumed to be generated by a Poisson point process. The robot has a limited perception range, and thus it discovers the reward field on the fly. The robot is assumed to be a dynamical system with substantial drift in one direction, e.g., a high-speed airplane, so it cannot traverse the entire field. The task of the robot is to maximize the total rewards collected during the course of the mission, given above constraints. Under such assumptions, we analyze the performance of a simple receding-horizon planning algorithm with respect to the perception range, robot agility and computational resources available. Firstly, we show that, with highly limited perception range, the robot is able to collect as many rewards as if it could see the entire reward field, if and only if the reward distribution is light-tailed. The second result attained shows that the expected rewards collected scale proportionally to the square root of the robot agility. Finally, we are able to prove that the overall computational workload increases linearly with the mission length, i.e., the distance of travel. We verify our results in simulation examples. At the end, we present one interesting application of our theoretical study to the ground sensor selection problem. For an inference/estimation task, we prove that sensors with randomized quality outperform those with homogeneous precisions, since random sensors yield a higher confidence level of estimation (lower variance), under certain technical assumptions. This finding might have practical implications on the design of UAV-UGS systems. === by Fangchang Ma. === S.M.
author2 Sertac Karaman.
author_facet Sertac Karaman.
Ma, Fangchang
author Ma, Fangchang
author_sort Ma, Fangchang
title On maximum-reward motion in stochastic environments
title_short On maximum-reward motion in stochastic environments
title_full On maximum-reward motion in stochastic environments
title_fullStr On maximum-reward motion in stochastic environments
title_full_unstemmed On maximum-reward motion in stochastic environments
title_sort on maximum-reward motion in stochastic environments
publisher Massachusetts Institute of Technology
publishDate 2015
url http://hdl.handle.net/1721.1/98696
work_keys_str_mv AT mafangchang onmaximumrewardmotioninstochasticenvironments
_version_ 1719040140864651264