Multiagent Learning with Bargaining - A Game Theoretic Approach

Learning in the real world occurs when an agent, which perceives its current state and takes actions, interacts with the environment, which in return provides a positive or negative feedback. The field of reinforcement learning studies such processes and attempts to find policies that map states of...

Full description

Bibliographic Details
Main Author: Qiao, Haiyan
Other Authors: Rozenblit, Jerzy
Language:EN
Published: The University of Arizona. 2007
Online Access:http://hdl.handle.net/10150/194382
id ndltd-arizona.edu-oai-arizona.openrepository.com-10150-194382
record_format oai_dc
spelling ndltd-arizona.edu-oai-arizona.openrepository.com-10150-1943822015-10-23T04:41:05Z Multiagent Learning with Bargaining - A Game Theoretic Approach Qiao, Haiyan Rozenblit, Jerzy Szidarovszky, Ferenc Rozenblit, Jerzy Szidarovszky, Ferenc Hariri, Salim Barnard, Kobus Learning in the real world occurs when an agent, which perceives its current state and takes actions, interacts with the environment, which in return provides a positive or negative feedback. The field of reinforcement learning studies such processes and attempts to find policies that map states of the world to the actions of agents in order to maximize cumulative reward over the long run. In multi-agent systems, agent learning becomes more challenging, since the optimal action of each agent generally depends on the actions of other agents. Most studies in multiagent learning research employ non-cooperative equilibrium as a learning objective. However, in many situations, the equilibrium gives worse payoffs to both players than their payoffs would be in the case of cooperation. Therefore the agents have strong desire to choose a cooperative solution instead of the non-cooperative equilibrium. In this work, we apply the Nash Bargaining Solution (NBS) to multi-agent systems with unknown parameters and design a multiagent learning algorithm based on bargaining, in which the agents can reach the NBS by learning through experience. We show that the solution is unique and is Pareto-optimal. We also prove theoretically that the algorithm converges. In addition, we extend the work to multi-agent systems with asymmetric agents having different powers in decision making and design a multiagent learning algorithm with asymmetric bargaining. To evaluate these learning algorithms and compare with the existing learning algorithms, the benchmark, grid world games, are adopted as the simulation test-bed. The simulation results demonstrate that our learning algorithms converge to the unique Pareto-optimal solution and the convergence is faster in comparison to the existing multiagent learning algorithms. Finally, we discuss an application of multiagent learning algorithms to a classic economic model, which is known as oligopoly. 2007 text Electronic Dissertation http://hdl.handle.net/10150/194382 659748098 2259 EN Copyright © is held by the author. Digital access to this material is made possible by the University Libraries, University of Arizona. Further transmission, reproduction or presentation (such as public display or performance) of protected items is prohibited except with permission of the author. The University of Arizona.
collection NDLTD
language EN
sources NDLTD
description Learning in the real world occurs when an agent, which perceives its current state and takes actions, interacts with the environment, which in return provides a positive or negative feedback. The field of reinforcement learning studies such processes and attempts to find policies that map states of the world to the actions of agents in order to maximize cumulative reward over the long run. In multi-agent systems, agent learning becomes more challenging, since the optimal action of each agent generally depends on the actions of other agents. Most studies in multiagent learning research employ non-cooperative equilibrium as a learning objective. However, in many situations, the equilibrium gives worse payoffs to both players than their payoffs would be in the case of cooperation. Therefore the agents have strong desire to choose a cooperative solution instead of the non-cooperative equilibrium. In this work, we apply the Nash Bargaining Solution (NBS) to multi-agent systems with unknown parameters and design a multiagent learning algorithm based on bargaining, in which the agents can reach the NBS by learning through experience. We show that the solution is unique and is Pareto-optimal. We also prove theoretically that the algorithm converges. In addition, we extend the work to multi-agent systems with asymmetric agents having different powers in decision making and design a multiagent learning algorithm with asymmetric bargaining. To evaluate these learning algorithms and compare with the existing learning algorithms, the benchmark, grid world games, are adopted as the simulation test-bed. The simulation results demonstrate that our learning algorithms converge to the unique Pareto-optimal solution and the convergence is faster in comparison to the existing multiagent learning algorithms. Finally, we discuss an application of multiagent learning algorithms to a classic economic model, which is known as oligopoly.
author2 Rozenblit, Jerzy
author_facet Rozenblit, Jerzy
Qiao, Haiyan
author Qiao, Haiyan
spellingShingle Qiao, Haiyan
Multiagent Learning with Bargaining - A Game Theoretic Approach
author_sort Qiao, Haiyan
title Multiagent Learning with Bargaining - A Game Theoretic Approach
title_short Multiagent Learning with Bargaining - A Game Theoretic Approach
title_full Multiagent Learning with Bargaining - A Game Theoretic Approach
title_fullStr Multiagent Learning with Bargaining - A Game Theoretic Approach
title_full_unstemmed Multiagent Learning with Bargaining - A Game Theoretic Approach
title_sort multiagent learning with bargaining - a game theoretic approach
publisher The University of Arizona.
publishDate 2007
url http://hdl.handle.net/10150/194382
work_keys_str_mv AT qiaohaiyan multiagentlearningwithbargainingagametheoreticapproach
_version_ 1718099262018945024