On maximum-reward motion in stochastic environments
Thesis: S.M., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, 2015. === Cataloged from PDF version of thesis. === Includes bibliographical references (pages 75-77). === In this thesis, we consider the problem of an autonomous mobile robot operating in a stochastic...
Main Author: | |
---|---|
Other Authors: | |
Format: | Others |
Language: | English |
Published: |
Massachusetts Institute of Technology
2015
|
Subjects: | |
Online Access: | http://hdl.handle.net/1721.1/98696 |
id |
ndltd-MIT-oai-dspace.mit.edu-1721.1-98696 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-MIT-oai-dspace.mit.edu-1721.1-986962019-05-02T16:23:35Z On maximum-reward motion in stochastic environments Ma, Fangchang Sertac Karaman. Massachusetts Institute of Technology. Department of Aeronautics and Astronautics. Massachusetts Institute of Technology. Department of Aeronautics and Astronautics. Aeronautics and Astronautics. Thesis: S.M., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, 2015. Cataloged from PDF version of thesis. Includes bibliographical references (pages 75-77). In this thesis, we consider the problem of an autonomous mobile robot operating in a stochastic reward field to maximize total rewards collected in an online setting. This is a generalization of the problem where an unmanned aerial vehicle (UAV) collects data from randomly deployed unattended ground sensors (UGS). Specifically, the rewards are assumed to be generated by a Poisson point process. The robot has a limited perception range, and thus it discovers the reward field on the fly. The robot is assumed to be a dynamical system with substantial drift in one direction, e.g., a high-speed airplane, so it cannot traverse the entire field. The task of the robot is to maximize the total rewards collected during the course of the mission, given above constraints. Under such assumptions, we analyze the performance of a simple receding-horizon planning algorithm with respect to the perception range, robot agility and computational resources available. Firstly, we show that, with highly limited perception range, the robot is able to collect as many rewards as if it could see the entire reward field, if and only if the reward distribution is light-tailed. The second result attained shows that the expected rewards collected scale proportionally to the square root of the robot agility. Finally, we are able to prove that the overall computational workload increases linearly with the mission length, i.e., the distance of travel. We verify our results in simulation examples. At the end, we present one interesting application of our theoretical study to the ground sensor selection problem. For an inference/estimation task, we prove that sensors with randomized quality outperform those with homogeneous precisions, since random sensors yield a higher confidence level of estimation (lower variance), under certain technical assumptions. This finding might have practical implications on the design of UAV-UGS systems. by Fangchang Ma. S.M. 2015-09-17T19:05:31Z 2015-09-17T19:05:31Z 2015 2015 Thesis http://hdl.handle.net/1721.1/98696 920688365 eng M.I.T. theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. See provided URL for inquiries about permission. http://dspace.mit.edu/handle/1721.1/7582 77 pages application/pdf Massachusetts Institute of Technology |
collection |
NDLTD |
language |
English |
format |
Others
|
sources |
NDLTD |
topic |
Aeronautics and Astronautics. |
spellingShingle |
Aeronautics and Astronautics. Ma, Fangchang On maximum-reward motion in stochastic environments |
description |
Thesis: S.M., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, 2015. === Cataloged from PDF version of thesis. === Includes bibliographical references (pages 75-77). === In this thesis, we consider the problem of an autonomous mobile robot operating in a stochastic reward field to maximize total rewards collected in an online setting. This is a generalization of the problem where an unmanned aerial vehicle (UAV) collects data from randomly deployed unattended ground sensors (UGS). Specifically, the rewards are assumed to be generated by a Poisson point process. The robot has a limited perception range, and thus it discovers the reward field on the fly. The robot is assumed to be a dynamical system with substantial drift in one direction, e.g., a high-speed airplane, so it cannot traverse the entire field. The task of the robot is to maximize the total rewards collected during the course of the mission, given above constraints. Under such assumptions, we analyze the performance of a simple receding-horizon planning algorithm with respect to the perception range, robot agility and computational resources available. Firstly, we show that, with highly limited perception range, the robot is able to collect as many rewards as if it could see the entire reward field, if and only if the reward distribution is light-tailed. The second result attained shows that the expected rewards collected scale proportionally to the square root of the robot agility. Finally, we are able to prove that the overall computational workload increases linearly with the mission length, i.e., the distance of travel. We verify our results in simulation examples. At the end, we present one interesting application of our theoretical study to the ground sensor selection problem. For an inference/estimation task, we prove that sensors with randomized quality outperform those with homogeneous precisions, since random sensors yield a higher confidence level of estimation (lower variance), under certain technical assumptions. This finding might have practical implications on the design of UAV-UGS systems. === by Fangchang Ma. === S.M. |
author2 |
Sertac Karaman. |
author_facet |
Sertac Karaman. Ma, Fangchang |
author |
Ma, Fangchang |
author_sort |
Ma, Fangchang |
title |
On maximum-reward motion in stochastic environments |
title_short |
On maximum-reward motion in stochastic environments |
title_full |
On maximum-reward motion in stochastic environments |
title_fullStr |
On maximum-reward motion in stochastic environments |
title_full_unstemmed |
On maximum-reward motion in stochastic environments |
title_sort |
on maximum-reward motion in stochastic environments |
publisher |
Massachusetts Institute of Technology |
publishDate |
2015 |
url |
http://hdl.handle.net/1721.1/98696 |
work_keys_str_mv |
AT mafangchang onmaximumrewardmotioninstochasticenvironments |
_version_ |
1719040140864651264 |