Deep Deterministic Policy Gradient Based on Double Network Prioritized Experience Replay
The traditional deep deterministic policy gradient (DDPG) algorithm has the disadvantages of slow convergence velocity and ease of falling into the local optimum. From these two perspectives, a DDPG algorithm based on the double network prioritized experience replay mechanism (DNPER-DDPG) is propose...
Main Authors: | Chaohai Kang, Chuiting Rong, Weijian Ren, Fengcai Huo, Pengyun Liu |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9409070/ |
Similar Items
-
Motion Planning of Robot Manipulators for a Smoother Path Using a Twin Delayed Deep Deterministic Policy Gradient with Hindsight Experience Replay
by: MyeongSeop Kim, et al.
Published: (2020-01-01) -
THE FUTURE, THE CRISIS, AND THE FUTURE OF REPLAY STORY
by: Eleonora Teresa Imbierowicz
Published: (2021-06-01) -
Self-Adaptive Priority Correction for Prioritized Experience Replay
by: Hongjie Zhang, et al.
Published: (2020-10-01) -
Trajectory Based Prioritized Double Experience Buffer for Sample-Efficient Policy Optimization
by: Shengxiang Li, et al.
Published: (2021-01-01) -
On Event Reproduction Ratio in Stateless and Stateful Replay of Real-World Traffic
by: Ying-Dar Lin, et al.
Published: (2013-09-01)