From first contact to close encounters : a developmentally deep perceptual system for a humanoid robot
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2003. === Includes bibliographical references (p. 139-152). === This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and...
Main Author: | |
---|---|
Other Authors: | |
Format: | Others |
Language: | English |
Published: |
Massachusetts Institute of Technology
2005
|
Subjects: | |
Online Access: | http://hdl.handle.net/1721.1/16951 |
id |
ndltd-MIT-oai-dspace.mit.edu-1721.1-16951 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-MIT-oai-dspace.mit.edu-1721.1-169512019-05-02T16:14:26Z From first contact to close encounters : a developmentally deep perceptual system for a humanoid robot Developmentally deep perceptual system for a humanoid robot Fitzpatrick, Paul Michael, 1974- Rodney A. Brooks. Massachusetts Institute of Technology. Dept. of Electrical Engineering and Computer Science. Massachusetts Institute of Technology. Dept. of Electrical Engineering and Computer Science. Electrical Engineering and Computer Science. Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2003. Includes bibliographical references (p. 139-152). This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections. This thesis presents a perceptual system for a humanoid robot that integrates abilities such as object localization and recognition with the deeper developmental machinery required to forge those competences out of raw physical experiences. It shows that a robotic platform can build up and maintain a system for object localization, segmentation, and recognition, starting from very little. What the robot starts with is a direct solution to achieving figure/ground separation: it simply 'pokes around' in a region of visual ambiguity and watches what happens. If the arm passes through an area, that area is recognized as free space. If the arm collides with an object, causing it to move, the robot can use that motion to segment the object from the background. Once the robot can acquire reliable segmented views of objects, it learns from them, and from then on recognizes and segments those objects without further contact. Both low-level and high-level visual features can also be learned in this way, and examples are presented for both: orientation detection and affordance recognition, respectively. The motivation for this work is simple. Training on large corpora of annotated real-world data has proven crucial for creating robust solutions to perceptual problems such as speech recognition and face detection. But the powerful tools used during training of such systems are typically stripped away at deployment. Ideally they should remain, particularly for unstable tasks such as object detection, where the set of objects needed in a task tomorrow might be different from the set of objects needed today. The key limiting factor is access to training data, but as this thesis shows, that need not be a problem on a robotic platform that can actively probe its environment, and carry out experiments to resolve ambiguity. (cont.) This work is an instance of a general approach to learning a new perceptual judgment: find special situations in which the perceptual judgment is easy and study these situations to find correlated features that can be observed more generally. by Paul Michael Fitzpatrick. Ph.D. 2005-05-19T15:24:57Z 2005-05-19T15:24:57Z 2003 2003 Thesis http://hdl.handle.net/1721.1/16951 53278821 eng M.I.T. theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. See provided URL for inquiries about permission. http://dspace.mit.edu/handle/1721.1/7582 152 p. 7101026 bytes 7100374 bytes application/pdf application/pdf application/pdf Massachusetts Institute of Technology |
collection |
NDLTD |
language |
English |
format |
Others
|
sources |
NDLTD |
topic |
Electrical Engineering and Computer Science. |
spellingShingle |
Electrical Engineering and Computer Science. Fitzpatrick, Paul Michael, 1974- From first contact to close encounters : a developmentally deep perceptual system for a humanoid robot |
description |
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2003. === Includes bibliographical references (p. 139-152). === This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections. === This thesis presents a perceptual system for a humanoid robot that integrates abilities such as object localization and recognition with the deeper developmental machinery required to forge those competences out of raw physical experiences. It shows that a robotic platform can build up and maintain a system for object localization, segmentation, and recognition, starting from very little. What the robot starts with is a direct solution to achieving figure/ground separation: it simply 'pokes around' in a region of visual ambiguity and watches what happens. If the arm passes through an area, that area is recognized as free space. If the arm collides with an object, causing it to move, the robot can use that motion to segment the object from the background. Once the robot can acquire reliable segmented views of objects, it learns from them, and from then on recognizes and segments those objects without further contact. Both low-level and high-level visual features can also be learned in this way, and examples are presented for both: orientation detection and affordance recognition, respectively. The motivation for this work is simple. Training on large corpora of annotated real-world data has proven crucial for creating robust solutions to perceptual problems such as speech recognition and face detection. But the powerful tools used during training of such systems are typically stripped away at deployment. Ideally they should remain, particularly for unstable tasks such as object detection, where the set of objects needed in a task tomorrow might be different from the set of objects needed today. The key limiting factor is access to training data, but as this thesis shows, that need not be a problem on a robotic platform that can actively probe its environment, and carry out experiments to resolve ambiguity. === (cont.) This work is an instance of a general approach to learning a new perceptual judgment: find special situations in which the perceptual judgment is easy and study these situations to find correlated features that can be observed more generally. === by Paul Michael Fitzpatrick. === Ph.D. |
author2 |
Rodney A. Brooks. |
author_facet |
Rodney A. Brooks. Fitzpatrick, Paul Michael, 1974- |
author |
Fitzpatrick, Paul Michael, 1974- |
author_sort |
Fitzpatrick, Paul Michael, 1974- |
title |
From first contact to close encounters : a developmentally deep perceptual system for a humanoid robot |
title_short |
From first contact to close encounters : a developmentally deep perceptual system for a humanoid robot |
title_full |
From first contact to close encounters : a developmentally deep perceptual system for a humanoid robot |
title_fullStr |
From first contact to close encounters : a developmentally deep perceptual system for a humanoid robot |
title_full_unstemmed |
From first contact to close encounters : a developmentally deep perceptual system for a humanoid robot |
title_sort |
from first contact to close encounters : a developmentally deep perceptual system for a humanoid robot |
publisher |
Massachusetts Institute of Technology |
publishDate |
2005 |
url |
http://hdl.handle.net/1721.1/16951 |
work_keys_str_mv |
AT fitzpatrickpaulmichael1974 fromfirstcontacttocloseencountersadevelopmentallydeepperceptualsystemforahumanoidrobot AT fitzpatrickpaulmichael1974 developmentallydeepperceptualsystemforahumanoidrobot |
_version_ |
1719036786683936768 |