Terrain Adaptive Walking of Biped Neuromuscular Virtual Human Using Deep Reinforcement Learning

There have been some biomechanics-based control systems that have achieved better realistic virtual human motion. Yet their abilities to adapt the changing environments are weaker than the traditional control systems with characters driven by proportional derivative actuators directly. In our method...

Full description

Bibliographic Details
Main Authors: Jianpeng Wang, Wenhu Qin, Libo Sun
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8758114/
id doaj-54704c9f65474e13885b370f0171f624
record_format Article
spelling doaj-54704c9f65474e13885b370f0171f6242021-03-29T23:28:16ZengIEEEIEEE Access2169-35362019-01-017924659247510.1109/ACCESS.2019.29276068758114Terrain Adaptive Walking of Biped Neuromuscular Virtual Human Using Deep Reinforcement LearningJianpeng Wang0https://orcid.org/0000-0002-9716-1809Wenhu Qin1Libo Sun2https://orcid.org/0000-0002-7838-9410School of Instrument Science and Engineering, Southeast University, Nanjing, ChinaSchool of Instrument Science and Engineering, Southeast University, Nanjing, ChinaSchool of Instrument Science and Engineering, Southeast University, Nanjing, ChinaThere have been some biomechanics-based control systems that have achieved better realistic virtual human motion. Yet their abilities to adapt the changing environments are weaker than the traditional control systems with characters driven by proportional derivative actuators directly. In our method, we build a hierarchical neuromuscular virtual human (NMVH) motion control system that consists of a low-level spine reflex layer and a high-level policy control layer. The spine reflex layer uses a feedback net to map sensory information to excitations, which stimulate muscles to generate joint torques. The policy control layer includes a deep neural network, which provides a learned action policy to spine reflex layer for achieving terrain-adaptive motion skills. The particle swarm optimization algorithm is used to optimize the gain factors of the feedback net for finding out a basic policy to make the virtual human walk on the flat terrain autonomously. The proximal policy optimization algorithm is employed to train the deep neural network in policy control layer for learning how to modulate the actions to adapt to the changing terrain. The simulation results in Matlab show that virtual human can walk smoothly and better adapt to the given terrain changes. It demonstrates that our control system improves the terrain-adaptive walking skill of the neuromuscular virtual human.https://ieeexplore.ieee.org/document/8758114/Biped walkinghierarchical controlneuromuscular virtual humanproximal policy optimization
collection DOAJ
language English
format Article
sources DOAJ
author Jianpeng Wang
Wenhu Qin
Libo Sun
spellingShingle Jianpeng Wang
Wenhu Qin
Libo Sun
Terrain Adaptive Walking of Biped Neuromuscular Virtual Human Using Deep Reinforcement Learning
IEEE Access
Biped walking
hierarchical control
neuromuscular virtual human
proximal policy optimization
author_facet Jianpeng Wang
Wenhu Qin
Libo Sun
author_sort Jianpeng Wang
title Terrain Adaptive Walking of Biped Neuromuscular Virtual Human Using Deep Reinforcement Learning
title_short Terrain Adaptive Walking of Biped Neuromuscular Virtual Human Using Deep Reinforcement Learning
title_full Terrain Adaptive Walking of Biped Neuromuscular Virtual Human Using Deep Reinforcement Learning
title_fullStr Terrain Adaptive Walking of Biped Neuromuscular Virtual Human Using Deep Reinforcement Learning
title_full_unstemmed Terrain Adaptive Walking of Biped Neuromuscular Virtual Human Using Deep Reinforcement Learning
title_sort terrain adaptive walking of biped neuromuscular virtual human using deep reinforcement learning
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2019-01-01
description There have been some biomechanics-based control systems that have achieved better realistic virtual human motion. Yet their abilities to adapt the changing environments are weaker than the traditional control systems with characters driven by proportional derivative actuators directly. In our method, we build a hierarchical neuromuscular virtual human (NMVH) motion control system that consists of a low-level spine reflex layer and a high-level policy control layer. The spine reflex layer uses a feedback net to map sensory information to excitations, which stimulate muscles to generate joint torques. The policy control layer includes a deep neural network, which provides a learned action policy to spine reflex layer for achieving terrain-adaptive motion skills. The particle swarm optimization algorithm is used to optimize the gain factors of the feedback net for finding out a basic policy to make the virtual human walk on the flat terrain autonomously. The proximal policy optimization algorithm is employed to train the deep neural network in policy control layer for learning how to modulate the actions to adapt to the changing terrain. The simulation results in Matlab show that virtual human can walk smoothly and better adapt to the given terrain changes. It demonstrates that our control system improves the terrain-adaptive walking skill of the neuromuscular virtual human.
topic Biped walking
hierarchical control
neuromuscular virtual human
proximal policy optimization
url https://ieeexplore.ieee.org/document/8758114/
work_keys_str_mv AT jianpengwang terrainadaptivewalkingofbipedneuromuscularvirtualhumanusingdeepreinforcementlearning
AT wenhuqin terrainadaptivewalkingofbipedneuromuscularvirtualhumanusingdeepreinforcementlearning
AT libosun terrainadaptivewalkingofbipedneuromuscularvirtualhumanusingdeepreinforcementlearning
_version_ 1724189421504299008