A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems
碩士 === 國立中正大學 === 光機電整合工程研究所 === 100 === Traditional reinforcement learning algorithm, such as Q-learning, is based on one agent and one step learning without a model. In recent years, many have proposed the concepts of multi-agents and using a model for retraining to increase learning efficiency, s...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | en_US |
Published: |
2012
|
Online Access: | http://ndltd.ncl.edu.tw/handle/67686939437674667921 |
id |
ndltd-TW-101CCU00651001 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-101CCU006510012015-10-13T21:07:18Z http://ndltd.ncl.edu.tw/handle/67686939437674667921 A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems 具混合規劃架構之並行Dyna-Q學習演算法 Hung, Tsan-Shun 洪贊順 碩士 國立中正大學 光機電整合工程研究所 100 Traditional reinforcement learning algorithm, such as Q-learning, is based on one agent and one step learning without a model. In recent years, many have proposed the concepts of multi-agents and using a model for retraining to increase learning efficiency, such as Dyna-Q and multi-agent system. In this thesis, we integrated several algorithms of different domains, applied concepts from different domains in reinforcement learning, and made extensions in compliance with existing concepts such as Dyna-Q and multi-agent system. We added UCB algorithm to reinforce exploration efficiency of agents and shorten the time for virtual environment model establishment. For the virtual environment model of Dyna-Q, we added the concept of image processing to sharpened model. We also proposed a planning algorithm for environmental space paralleling, which can perform parallel computing and accelerate Dyna-Q learning. The concept of prioritized sweeping was integrated to further increase planning efficiency and resource management. After improving and integrating the above algorithms, the concept of GPGPU (General Purpose Computing on Graphics Processing Units) was used for simulation on CUDA (Compute Unified Device Architecture). The simulation was applied for verifying the impact of the above method on learning speed of Dyna-Q. Hwang, Kao-Shing 黃國勝 2012 學位論文 ; thesis 63 en_US |
collection |
NDLTD |
language |
en_US |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立中正大學 === 光機電整合工程研究所 === 100 === Traditional reinforcement learning algorithm, such as Q-learning, is based on one agent and one step learning without a model. In recent years, many have proposed the concepts of multi-agents and using a model for retraining to increase learning
efficiency, such as Dyna-Q and multi-agent system.
In this thesis, we integrated several algorithms of different domains, applied concepts from different domains in reinforcement learning, and made extensions in compliance with existing concepts such as Dyna-Q and multi-agent system.
We added UCB algorithm to reinforce exploration efficiency of agents and shorten the time for virtual environment model establishment. For the virtual environment model of Dyna-Q, we added the concept of image processing to sharpened model.
We also proposed a planning algorithm for environmental space paralleling, which can perform parallel computing and accelerate Dyna-Q learning. The concept of prioritized sweeping was integrated to further increase planning efficiency and resource management. After improving and integrating the above algorithms, the concept of GPGPU (General Purpose Computing on Graphics Processing Units) was
used for simulation on CUDA (Compute Unified Device Architecture). The simulation was applied for verifying the impact of the above method on learning speed of Dyna-Q.
|
author2 |
Hwang, Kao-Shing |
author_facet |
Hwang, Kao-Shing Hung, Tsan-Shun 洪贊順 |
author |
Hung, Tsan-Shun 洪贊順 |
spellingShingle |
Hung, Tsan-Shun 洪贊順 A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems |
author_sort |
Hung, Tsan-Shun |
title |
A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems |
title_short |
A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems |
title_full |
A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems |
title_fullStr |
A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems |
title_full_unstemmed |
A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems |
title_sort |
hybrid planning in concurrent dyna-q learning for multi-agent systems |
publishDate |
2012 |
url |
http://ndltd.ncl.edu.tw/handle/67686939437674667921 |
work_keys_str_mv |
AT hungtsanshun ahybridplanninginconcurrentdynaqlearningformultiagentsystems AT hóngzànshùn ahybridplanninginconcurrentdynaqlearningformultiagentsystems AT hungtsanshun jùhùnhéguīhuàjiàgòuzhībìngxíngdynaqxuéxíyǎnsuànfǎ AT hóngzànshùn jùhùnhéguīhuàjiàgòuzhībìngxíngdynaqxuéxíyǎnsuànfǎ AT hungtsanshun hybridplanninginconcurrentdynaqlearningformultiagentsystems AT hóngzànshùn hybridplanninginconcurrentdynaqlearningformultiagentsystems |
_version_ |
1718056548174921728 |