A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems

碩士 === 國立中正大學 === 光機電整合工程研究所 === 100 === Traditional reinforcement learning algorithm, such as Q-learning, is based on one agent and one step learning without a model. In recent years, many have proposed the concepts of multi-agents and using a model for retraining to increase learning efficiency, s...

Full description

Bibliographic Details
Main Authors: Hung, Tsan-Shun, 洪贊順
Other Authors: Hwang, Kao-Shing
Format: Others
Language:en_US
Published: 2012
Online Access:http://ndltd.ncl.edu.tw/handle/67686939437674667921
id ndltd-TW-101CCU00651001
record_format oai_dc
spelling ndltd-TW-101CCU006510012015-10-13T21:07:18Z http://ndltd.ncl.edu.tw/handle/67686939437674667921 A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems 具混合規劃架構之並行Dyna-Q學習演算法 Hung, Tsan-Shun 洪贊順 碩士 國立中正大學 光機電整合工程研究所 100 Traditional reinforcement learning algorithm, such as Q-learning, is based on one agent and one step learning without a model. In recent years, many have proposed the concepts of multi-agents and using a model for retraining to increase learning efficiency, such as Dyna-Q and multi-agent system. In this thesis, we integrated several algorithms of different domains, applied concepts from different domains in reinforcement learning, and made extensions in compliance with existing concepts such as Dyna-Q and multi-agent system. We added UCB algorithm to reinforce exploration efficiency of agents and shorten the time for virtual environment model establishment. For the virtual environment model of Dyna-Q, we added the concept of image processing to sharpened model. We also proposed a planning algorithm for environmental space paralleling, which can perform parallel computing and accelerate Dyna-Q learning. The concept of prioritized sweeping was integrated to further increase planning efficiency and resource management. After improving and integrating the above algorithms, the concept of GPGPU (General Purpose Computing on Graphics Processing Units) was used for simulation on CUDA (Compute Unified Device Architecture). The simulation was applied for verifying the impact of the above method on learning speed of Dyna-Q. Hwang, Kao-Shing 黃國勝 2012 學位論文 ; thesis 63 en_US
collection NDLTD
language en_US
format Others
sources NDLTD
description 碩士 === 國立中正大學 === 光機電整合工程研究所 === 100 === Traditional reinforcement learning algorithm, such as Q-learning, is based on one agent and one step learning without a model. In recent years, many have proposed the concepts of multi-agents and using a model for retraining to increase learning efficiency, such as Dyna-Q and multi-agent system. In this thesis, we integrated several algorithms of different domains, applied concepts from different domains in reinforcement learning, and made extensions in compliance with existing concepts such as Dyna-Q and multi-agent system. We added UCB algorithm to reinforce exploration efficiency of agents and shorten the time for virtual environment model establishment. For the virtual environment model of Dyna-Q, we added the concept of image processing to sharpened model. We also proposed a planning algorithm for environmental space paralleling, which can perform parallel computing and accelerate Dyna-Q learning. The concept of prioritized sweeping was integrated to further increase planning efficiency and resource management. After improving and integrating the above algorithms, the concept of GPGPU (General Purpose Computing on Graphics Processing Units) was used for simulation on CUDA (Compute Unified Device Architecture). The simulation was applied for verifying the impact of the above method on learning speed of Dyna-Q.
author2 Hwang, Kao-Shing
author_facet Hwang, Kao-Shing
Hung, Tsan-Shun
洪贊順
author Hung, Tsan-Shun
洪贊順
spellingShingle Hung, Tsan-Shun
洪贊順
A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems
author_sort Hung, Tsan-Shun
title A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems
title_short A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems
title_full A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems
title_fullStr A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems
title_full_unstemmed A Hybrid Planning in Concurrent Dyna-Q Learning for Multi-agent Systems
title_sort hybrid planning in concurrent dyna-q learning for multi-agent systems
publishDate 2012
url http://ndltd.ncl.edu.tw/handle/67686939437674667921
work_keys_str_mv AT hungtsanshun ahybridplanninginconcurrentdynaqlearningformultiagentsystems
AT hóngzànshùn ahybridplanninginconcurrentdynaqlearningformultiagentsystems
AT hungtsanshun jùhùnhéguīhuàjiàgòuzhībìngxíngdynaqxuéxíyǎnsuànfǎ
AT hóngzànshùn jùhùnhéguīhuàjiàgòuzhībìngxíngdynaqxuéxíyǎnsuànfǎ
AT hungtsanshun hybridplanninginconcurrentdynaqlearningformultiagentsystems
AT hóngzànshùn hybridplanninginconcurrentdynaqlearningformultiagentsystems
_version_ 1718056548174921728