Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning.
Multi-Agent Reinforcement Learning (MARL) algorithms face two main difficulties: the curse of dimensionality, and environment non-stationarity due to the independent learning processes carried out by the agents concurrently. In this paper we formalize and prove the convergence of a Distributed Round...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Public Library of Science (PLoS)
2015-01-01
|
Series: | PLoS ONE |
Online Access: | http://europepmc.org/articles/PMC4497621?pdf=render |
id |
doaj-336f1c736d20458f9ee8715b19d51a03 |
---|---|
record_format |
Article |
spelling |
doaj-336f1c736d20458f9ee8715b19d51a032020-11-24T21:58:38ZengPublic Library of Science (PLoS)PLoS ONE1932-62032015-01-01107e012712910.1371/journal.pone.0127129Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning.Borja Fernandez-GaunaIsmael Etxeberria-AgirianoManuel GrañaMulti-Agent Reinforcement Learning (MARL) algorithms face two main difficulties: the curse of dimensionality, and environment non-stationarity due to the independent learning processes carried out by the agents concurrently. In this paper we formalize and prove the convergence of a Distributed Round Robin Q-learning (D-RR-QL) algorithm for cooperative systems. The computational complexity of this algorithm increases linearly with the number of agents. Moreover, it eliminates environment non sta tionarity by carrying a round-robin scheduling of the action selection and execution. That this learning scheme allows the implementation of Modular State-Action Vetoes (MSAV) in cooperative multi-agent systems, which speeds up learning convergence in over-constrained systems by vetoing state-action pairs which lead to undesired termination states (UTS) in the relevant state-action subspace. Each agent's local state-action value function learning is an independent process, including the MSAV policies. Coordination of locally optimal policies to obtain the global optimal joint policy is achieved by a greedy selection procedure using message passing. We show that D-RR-QL improves over state-of-the-art approaches, such as Distributed Q-Learning, Team Q-Learning and Coordinated Reinforcement Learning in a paradigmatic Linked Multi-Component Robotic System (L-MCRS) control problem: the hose transportation task. L-MCRS are over-constrained systems with many UTS induced by the interaction of the passive linking element and the active mobile robots.http://europepmc.org/articles/PMC4497621?pdf=render |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Borja Fernandez-Gauna Ismael Etxeberria-Agiriano Manuel Graña |
spellingShingle |
Borja Fernandez-Gauna Ismael Etxeberria-Agiriano Manuel Graña Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning. PLoS ONE |
author_facet |
Borja Fernandez-Gauna Ismael Etxeberria-Agiriano Manuel Graña |
author_sort |
Borja Fernandez-Gauna |
title |
Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning. |
title_short |
Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning. |
title_full |
Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning. |
title_fullStr |
Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning. |
title_full_unstemmed |
Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning. |
title_sort |
learning multirobot hose transportation and deployment by distributed round-robin q-learning. |
publisher |
Public Library of Science (PLoS) |
series |
PLoS ONE |
issn |
1932-6203 |
publishDate |
2015-01-01 |
description |
Multi-Agent Reinforcement Learning (MARL) algorithms face two main difficulties: the curse of dimensionality, and environment non-stationarity due to the independent learning processes carried out by the agents concurrently. In this paper we formalize and prove the convergence of a Distributed Round Robin Q-learning (D-RR-QL) algorithm for cooperative systems. The computational complexity of this algorithm increases linearly with the number of agents. Moreover, it eliminates environment non sta tionarity by carrying a round-robin scheduling of the action selection and execution. That this learning scheme allows the implementation of Modular State-Action Vetoes (MSAV) in cooperative multi-agent systems, which speeds up learning convergence in over-constrained systems by vetoing state-action pairs which lead to undesired termination states (UTS) in the relevant state-action subspace. Each agent's local state-action value function learning is an independent process, including the MSAV policies. Coordination of locally optimal policies to obtain the global optimal joint policy is achieved by a greedy selection procedure using message passing. We show that D-RR-QL improves over state-of-the-art approaches, such as Distributed Q-Learning, Team Q-Learning and Coordinated Reinforcement Learning in a paradigmatic Linked Multi-Component Robotic System (L-MCRS) control problem: the hose transportation task. L-MCRS are over-constrained systems with many UTS induced by the interaction of the passive linking element and the active mobile robots. |
url |
http://europepmc.org/articles/PMC4497621?pdf=render |
work_keys_str_mv |
AT borjafernandezgauna learningmultirobothosetransportationanddeploymentbydistributedroundrobinqlearning AT ismaeletxeberriaagiriano learningmultirobothosetransportationanddeploymentbydistributedroundrobinqlearning AT manuelgrana learningmultirobothosetransportationanddeploymentbydistributedroundrobinqlearning |
_version_ |
1725851001451184128 |