Dynamical Motor Control Learned with Deep Deterministic Policy Gradient

Conventional models of motor control exploit the spatial representation of the controlled system to generate control commands. Typically, the control command is gained with the feedback state of a specific instant in time, which behaves like an optimal regulator or spatial filter to the feedback sta...

Full description

Bibliographic Details
Main Authors: Haibo Shi, Yaoru Sun, Jie Li
Format: Article
Language:English
Published: Hindawi Limited 2018-01-01
Series:Computational Intelligence and Neuroscience
Online Access:http://dx.doi.org/10.1155/2018/8535429
id doaj-469e629f5a9844cab76bc836a117623f
record_format Article
spelling doaj-469e629f5a9844cab76bc836a117623f2020-11-24T23:51:16ZengHindawi LimitedComputational Intelligence and Neuroscience1687-52651687-52732018-01-01201810.1155/2018/85354298535429Dynamical Motor Control Learned with Deep Deterministic Policy GradientHaibo Shi0Yaoru Sun1Jie Li2Laboratory of Cognition & Intelligent Computing, Department of Computer Science, Tongji University, Shanghai, ChinaLaboratory of Cognition & Intelligent Computing, Department of Computer Science, Tongji University, Shanghai, ChinaLaboratory of Cognition & Intelligent Computing, Department of Computer Science, Tongji University, Shanghai, ChinaConventional models of motor control exploit the spatial representation of the controlled system to generate control commands. Typically, the control command is gained with the feedback state of a specific instant in time, which behaves like an optimal regulator or spatial filter to the feedback state. Yet, recent neuroscience studies found that the motor network may constitute an autonomous dynamical system and the temporal patterns of the control command can be contained in the dynamics of the motor network, that is, the dynamical system hypothesis (DSH). Inspired by these findings, here we propose a computational model that incorporates this neural mechanism, in which the control command could be unfolded from a dynamical controller whose initial state is specified with the task parameters. The model is trained in a trial-and-error manner in the framework of deep deterministic policy gradient (DDPG). The experimental results show that the dynamical controller successfully learns the control policy for arm reaching movements, while the analysis of the internal activities of the dynamical controller provides the computational evidence to the DSH of the neural coding in motor cortices.http://dx.doi.org/10.1155/2018/8535429
collection DOAJ
language English
format Article
sources DOAJ
author Haibo Shi
Yaoru Sun
Jie Li
spellingShingle Haibo Shi
Yaoru Sun
Jie Li
Dynamical Motor Control Learned with Deep Deterministic Policy Gradient
Computational Intelligence and Neuroscience
author_facet Haibo Shi
Yaoru Sun
Jie Li
author_sort Haibo Shi
title Dynamical Motor Control Learned with Deep Deterministic Policy Gradient
title_short Dynamical Motor Control Learned with Deep Deterministic Policy Gradient
title_full Dynamical Motor Control Learned with Deep Deterministic Policy Gradient
title_fullStr Dynamical Motor Control Learned with Deep Deterministic Policy Gradient
title_full_unstemmed Dynamical Motor Control Learned with Deep Deterministic Policy Gradient
title_sort dynamical motor control learned with deep deterministic policy gradient
publisher Hindawi Limited
series Computational Intelligence and Neuroscience
issn 1687-5265
1687-5273
publishDate 2018-01-01
description Conventional models of motor control exploit the spatial representation of the controlled system to generate control commands. Typically, the control command is gained with the feedback state of a specific instant in time, which behaves like an optimal regulator or spatial filter to the feedback state. Yet, recent neuroscience studies found that the motor network may constitute an autonomous dynamical system and the temporal patterns of the control command can be contained in the dynamics of the motor network, that is, the dynamical system hypothesis (DSH). Inspired by these findings, here we propose a computational model that incorporates this neural mechanism, in which the control command could be unfolded from a dynamical controller whose initial state is specified with the task parameters. The model is trained in a trial-and-error manner in the framework of deep deterministic policy gradient (DDPG). The experimental results show that the dynamical controller successfully learns the control policy for arm reaching movements, while the analysis of the internal activities of the dynamical controller provides the computational evidence to the DSH of the neural coding in motor cortices.
url http://dx.doi.org/10.1155/2018/8535429
work_keys_str_mv AT haiboshi dynamicalmotorcontrollearnedwithdeepdeterministicpolicygradient
AT yaorusun dynamicalmotorcontrollearnedwithdeepdeterministicpolicygradient
AT jieli dynamicalmotorcontrollearnedwithdeepdeterministicpolicygradient
_version_ 1725476685011222528