A Decentralized Partially Observable Markov Decision Model with Action Duration for Goal Recognition in Real Time Strategy Games
Multiagent goal recognition is a tough yet important problem in many real time strategy games or simulation systems. Traditional modeling methods either are in great demand of detailed agents’ domain knowledge and training dataset for policy estimation or lack clear definition of action duration. To...
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Hindawi Limited
2017-01-01
|
Series: | Discrete Dynamics in Nature and Society |
Online Access: | http://dx.doi.org/10.1155/2017/4580206 |
id |
doaj-bc67c9fced804383984269a5c1d7199b |
---|---|
record_format |
Article |
spelling |
doaj-bc67c9fced804383984269a5c1d7199b2020-11-24T22:57:44ZengHindawi LimitedDiscrete Dynamics in Nature and Society1026-02261607-887X2017-01-01201710.1155/2017/45802064580206A Decentralized Partially Observable Markov Decision Model with Action Duration for Goal Recognition in Real Time Strategy GamesPeng Jiao0Kai Xu1Shiguang Yue2Xiangyu Wei3Lin Sun4College of Information System and Management, National University of Defense Technology, Changsha 410073, ChinaCollege of Information System and Management, National University of Defense Technology, Changsha 410073, ChinaCollege of Information System and Management, National University of Defense Technology, Changsha 410073, ChinaCollege of Information System and Management, National University of Defense Technology, Changsha 410073, ChinaCollege of Information System and Management, National University of Defense Technology, Changsha 410073, ChinaMultiagent goal recognition is a tough yet important problem in many real time strategy games or simulation systems. Traditional modeling methods either are in great demand of detailed agents’ domain knowledge and training dataset for policy estimation or lack clear definition of action duration. To solve the above problems, we propose a novel Dec-POMDM-T model, combining the classic Dec-POMDP, an observation model for recognizer, joint goal with its termination indicator, and time duration variables for actions with action termination variables. In this paper, a model-free algorithm named cooperative colearning based on Sarsa is used. Considering that Dec-POMDM-T usually encounters multiagent goal recognition problems with different sorts of noises, partially missing data, and unknown action durations, the paper exploits the SIS PF with resampling for inference under the dynamic Bayesian network structure of Dec-POMDM-T. In experiments, a modified predator-prey scenario is adopted to study multiagent joint goal recognition problem, which is the recognition of the joint target shared among cooperative predators. Experiment results show that (a) Dec-POMDM-T works effectively in multiagent goal recognition and adapts well to dynamic changing goals within agent group; (b) Dec-POMDM-T outperforms traditional Dec-MDP-based methods in terms of precision, recall, and F-measure.http://dx.doi.org/10.1155/2017/4580206 |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Peng Jiao Kai Xu Shiguang Yue Xiangyu Wei Lin Sun |
spellingShingle |
Peng Jiao Kai Xu Shiguang Yue Xiangyu Wei Lin Sun A Decentralized Partially Observable Markov Decision Model with Action Duration for Goal Recognition in Real Time Strategy Games Discrete Dynamics in Nature and Society |
author_facet |
Peng Jiao Kai Xu Shiguang Yue Xiangyu Wei Lin Sun |
author_sort |
Peng Jiao |
title |
A Decentralized Partially Observable Markov Decision Model with Action Duration for Goal Recognition in Real Time Strategy Games |
title_short |
A Decentralized Partially Observable Markov Decision Model with Action Duration for Goal Recognition in Real Time Strategy Games |
title_full |
A Decentralized Partially Observable Markov Decision Model with Action Duration for Goal Recognition in Real Time Strategy Games |
title_fullStr |
A Decentralized Partially Observable Markov Decision Model with Action Duration for Goal Recognition in Real Time Strategy Games |
title_full_unstemmed |
A Decentralized Partially Observable Markov Decision Model with Action Duration for Goal Recognition in Real Time Strategy Games |
title_sort |
decentralized partially observable markov decision model with action duration for goal recognition in real time strategy games |
publisher |
Hindawi Limited |
series |
Discrete Dynamics in Nature and Society |
issn |
1026-0226 1607-887X |
publishDate |
2017-01-01 |
description |
Multiagent goal recognition is a tough yet important problem in many real time strategy games or simulation systems. Traditional modeling methods either are in great demand of detailed agents’ domain knowledge and training dataset for policy estimation or lack clear definition of action duration. To solve the above problems, we propose a novel Dec-POMDM-T model, combining the classic Dec-POMDP, an observation model for recognizer, joint goal with its termination indicator, and time duration variables for actions with action termination variables. In this paper, a model-free algorithm named cooperative colearning based on Sarsa is used. Considering that Dec-POMDM-T usually encounters multiagent goal recognition problems with different sorts of noises, partially missing data, and unknown action durations, the paper exploits the SIS PF with resampling for inference under the dynamic Bayesian network structure of Dec-POMDM-T. In experiments, a modified predator-prey scenario is adopted to study multiagent joint goal recognition problem, which is the recognition of the joint target shared among cooperative predators. Experiment results show that (a) Dec-POMDM-T works effectively in multiagent goal recognition and adapts well to dynamic changing goals within agent group; (b) Dec-POMDM-T outperforms traditional Dec-MDP-based methods in terms of precision, recall, and F-measure. |
url |
http://dx.doi.org/10.1155/2017/4580206 |
work_keys_str_mv |
AT pengjiao adecentralizedpartiallyobservablemarkovdecisionmodelwithactiondurationforgoalrecognitioninrealtimestrategygames AT kaixu adecentralizedpartiallyobservablemarkovdecisionmodelwithactiondurationforgoalrecognitioninrealtimestrategygames AT shiguangyue adecentralizedpartiallyobservablemarkovdecisionmodelwithactiondurationforgoalrecognitioninrealtimestrategygames AT xiangyuwei adecentralizedpartiallyobservablemarkovdecisionmodelwithactiondurationforgoalrecognitioninrealtimestrategygames AT linsun adecentralizedpartiallyobservablemarkovdecisionmodelwithactiondurationforgoalrecognitioninrealtimestrategygames AT pengjiao decentralizedpartiallyobservablemarkovdecisionmodelwithactiondurationforgoalrecognitioninrealtimestrategygames AT kaixu decentralizedpartiallyobservablemarkovdecisionmodelwithactiondurationforgoalrecognitioninrealtimestrategygames AT shiguangyue decentralizedpartiallyobservablemarkovdecisionmodelwithactiondurationforgoalrecognitioninrealtimestrategygames AT xiangyuwei decentralizedpartiallyobservablemarkovdecisionmodelwithactiondurationforgoalrecognitioninrealtimestrategygames AT linsun decentralizedpartiallyobservablemarkovdecisionmodelwithactiondurationforgoalrecognitioninrealtimestrategygames |
_version_ |
1725649384569307136 |