Neural Network-Based Learning from Demonstration of an Autonomous Ground Robot

This paper presents and experimentally validates a concept of end-to-end imitation learning for autonomous systems by using a composite architecture of convolutional neural network (ConvNet) and Long Short Term Memory (LSTM) neural network. In particular, a spatio-temporal deep neural network is dev...

Full description

Bibliographic Details
Main Authors: Yiwei Fu, Devesh K. Jha, Zeyu Zhang, Zhenyuan Yuan, Asok Ray
Format: Article
Language:English
Published: MDPI AG 2019-04-01
Series:Machines
Subjects:
Online Access:https://www.mdpi.com/2075-1702/7/2/24
id doaj-f5a8756b133148598ed098c90c3df0a2
record_format Article
spelling doaj-f5a8756b133148598ed098c90c3df0a22020-11-25T00:29:52ZengMDPI AGMachines2075-17022019-04-01722410.3390/machines7020024machines7020024Neural Network-Based Learning from Demonstration of an Autonomous Ground RobotYiwei Fu0Devesh K. Jha1Zeyu Zhang2Zhenyuan Yuan3Asok Ray4Department of Mechanical Engineering, Pennsylvania State University, University Park, PA 16802, USAMitsubishi Electric Research Laboratories, Cambridge, MA 02139, USADepartment of Mechanical Engineering, Pennsylvania State University, University Park, PA 16802, USADepartment of Electrical Engineering, Pennsylvania State University, University Park, PA 16802, USADepartment of Mechanical Engineering, Pennsylvania State University, University Park, PA 16802, USAThis paper presents and experimentally validates a concept of end-to-end imitation learning for autonomous systems by using a composite architecture of convolutional neural network (ConvNet) and Long Short Term Memory (LSTM) neural network. In particular, a spatio-temporal deep neural network is developed, which learns to imitate the policy used by a human supervisor to drive a car-like robot in a maze environment. The spatial and temporal components of the imitation model are learned by using deep convolutional network and recurrent neural network architectures, respectively. The imitation model learns the policy of a human supervisor as a function of laser light detection and ranging (LIDAR) data, which is then used in real time to drive a robot in an autonomous fashion in a laboratory setting. The performance of the proposed model for imitation learning is compared with that of several other state-of-the-art methods, reported in the machine learning literature, for spatial and temporal modeling. The learned policy is implemented on a robot using a Nvidia Jetson TX2 board which, in turn, is validated on test tracks. The proposed spatio-temporal model outperforms several other off-the-shelf machine learning techniques to learn the policy.https://www.mdpi.com/2075-1702/7/2/24autonomous robotsneural networksimitation learning
collection DOAJ
language English
format Article
sources DOAJ
author Yiwei Fu
Devesh K. Jha
Zeyu Zhang
Zhenyuan Yuan
Asok Ray
spellingShingle Yiwei Fu
Devesh K. Jha
Zeyu Zhang
Zhenyuan Yuan
Asok Ray
Neural Network-Based Learning from Demonstration of an Autonomous Ground Robot
Machines
autonomous robots
neural networks
imitation learning
author_facet Yiwei Fu
Devesh K. Jha
Zeyu Zhang
Zhenyuan Yuan
Asok Ray
author_sort Yiwei Fu
title Neural Network-Based Learning from Demonstration of an Autonomous Ground Robot
title_short Neural Network-Based Learning from Demonstration of an Autonomous Ground Robot
title_full Neural Network-Based Learning from Demonstration of an Autonomous Ground Robot
title_fullStr Neural Network-Based Learning from Demonstration of an Autonomous Ground Robot
title_full_unstemmed Neural Network-Based Learning from Demonstration of an Autonomous Ground Robot
title_sort neural network-based learning from demonstration of an autonomous ground robot
publisher MDPI AG
series Machines
issn 2075-1702
publishDate 2019-04-01
description This paper presents and experimentally validates a concept of end-to-end imitation learning for autonomous systems by using a composite architecture of convolutional neural network (ConvNet) and Long Short Term Memory (LSTM) neural network. In particular, a spatio-temporal deep neural network is developed, which learns to imitate the policy used by a human supervisor to drive a car-like robot in a maze environment. The spatial and temporal components of the imitation model are learned by using deep convolutional network and recurrent neural network architectures, respectively. The imitation model learns the policy of a human supervisor as a function of laser light detection and ranging (LIDAR) data, which is then used in real time to drive a robot in an autonomous fashion in a laboratory setting. The performance of the proposed model for imitation learning is compared with that of several other state-of-the-art methods, reported in the machine learning literature, for spatial and temporal modeling. The learned policy is implemented on a robot using a Nvidia Jetson TX2 board which, in turn, is validated on test tracks. The proposed spatio-temporal model outperforms several other off-the-shelf machine learning techniques to learn the policy.
topic autonomous robots
neural networks
imitation learning
url https://www.mdpi.com/2075-1702/7/2/24
work_keys_str_mv AT yiweifu neuralnetworkbasedlearningfromdemonstrationofanautonomousgroundrobot
AT deveshkjha neuralnetworkbasedlearningfromdemonstrationofanautonomousgroundrobot
AT zeyuzhang neuralnetworkbasedlearningfromdemonstrationofanautonomousgroundrobot
AT zhenyuanyuan neuralnetworkbasedlearningfromdemonstrationofanautonomousgroundrobot
AT asokray neuralnetworkbasedlearningfromdemonstrationofanautonomousgroundrobot
_version_ 1725329367729438720