Empirically evaluating multiagent reinforcement learning algorithms in repeated games
This dissertation presents a platform for running experiments on multiagent reinforcement learning algorithms and an empirical evaluation that was conducted on the platform. The setting under consideration is game theoretic in which a single normal form game is repeatedly played. There has been a...
Main Author: | |
---|---|
Language: | English |
Published: |
2009
|
Online Access: | http://hdl.handle.net/2429/16633 |
id |
ndltd-UBC-oai-circle.library.ubc.ca-2429-16633 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-UBC-oai-circle.library.ubc.ca-2429-166332018-01-05T17:38:29Z Empirically evaluating multiagent reinforcement learning algorithms in repeated games Lipson, Asher G. This dissertation presents a platform for running experiments on multiagent reinforcement learning algorithms and an empirical evaluation that was conducted on the platform. The setting under consideration is game theoretic in which a single normal form game is repeatedly played. There has been a large body of work focusing on introducing new algorithms to achieve certain goals such as guaranteeing values in a game, converging to a Nash equilibrium or minimizing total regret. Currently, we have an understanding of how some of these algorithms work in limited settings, but lack a broader understanding of which algorithms perform well against each other and how they perform on a larger variety of games. We describe our development of a platform that allows large scale tests to be run, where multiple algorithms are played against one another on a variety of games. The platform has a set of builtin metrics that can be used to measure the performance of an algorithm, including convergence to a Nash equilibrium, regret, reward and number of wins. Visualising the results of the test can be automatically achieved through the platform, with all interaction taking place through graphical user interfaces. We also present the results of an empirical test that to our knowledge includes the largest combination of game instances and algorithms used in the multiagent learning literature. To demonstrate the usefulness of the platform, we provide evidence for a number of claims and hypotheses. This includes claims related to convergence to a Nash equilibrium, reward, regret and best response metrics and claims dealing with estimating an opponent's strategy. Some of our claims include that (1) no algorithm does best across all metrics and over all opponents, (2) algorithms do not often converge to an exact Nash equilibrium, but (3) do often reach a small window around a Nash equilibrium, (4) there is no apparent link between converging to a Nash equilibrium and obtaining high reward and (5) there is no linear trend between reward and the size of the game for any agent. The two major contributions of this work are a software platform for running large experimental tests and empirical results that provide insight into the performance of various algorithms. Science, Faculty of Computer Science, Department of Graduate 2009-12-11T23:51:30Z 2009-12-11T23:51:30Z 2005 2005-11 Text Thesis/Dissertation http://hdl.handle.net/2429/16633 eng For non-commercial purposes only, such as research, private study and education. Additional conditions apply, see Terms of Use https://open.library.ubc.ca/terms_of_use. |
collection |
NDLTD |
language |
English |
sources |
NDLTD |
description |
This dissertation presents a platform for running experiments on multiagent reinforcement learning
algorithms and an empirical evaluation that was conducted on the platform. The setting under
consideration is game theoretic in which a single normal form game is repeatedly played.
There has been a large body of work focusing on introducing new algorithms to achieve certain
goals such as guaranteeing values in a game, converging to a Nash equilibrium or minimizing total
regret. Currently, we have an understanding of how some of these algorithms work in limited
settings, but lack a broader understanding of which algorithms perform well against each other and
how they perform on a larger variety of games.
We describe our development of a platform that allows large scale tests to be run, where multiple
algorithms are played against one another on a variety of games. The platform has a set of builtin
metrics that can be used to measure the performance of an algorithm, including convergence to
a Nash equilibrium, regret, reward and number of wins. Visualising the results of the test can be
automatically achieved through the platform, with all interaction taking place through graphical user
interfaces.
We also present the results of an empirical test that to our knowledge includes the largest combination
of game instances and algorithms used in the multiagent learning literature. To demonstrate
the usefulness of the platform, we provide evidence for a number of claims and hypotheses. This includes
claims related to convergence to a Nash equilibrium, reward, regret and best response metrics
and claims dealing with estimating an opponent's strategy. Some of our claims include that (1) no
algorithm does best across all metrics and over all opponents, (2) algorithms do not often converge
to an exact Nash equilibrium, but (3) do often reach a small window around a Nash equilibrium, (4)
there is no apparent link between converging to a Nash equilibrium and obtaining high reward and
(5) there is no linear trend between reward and the size of the game for any agent.
The two major contributions of this work are a software platform for running large experimental
tests and empirical results that provide insight into the performance of various algorithms. === Science, Faculty of === Computer Science, Department of === Graduate |
author |
Lipson, Asher G. |
spellingShingle |
Lipson, Asher G. Empirically evaluating multiagent reinforcement learning algorithms in repeated games |
author_facet |
Lipson, Asher G. |
author_sort |
Lipson, Asher G. |
title |
Empirically evaluating multiagent reinforcement learning algorithms in repeated games |
title_short |
Empirically evaluating multiagent reinforcement learning algorithms in repeated games |
title_full |
Empirically evaluating multiagent reinforcement learning algorithms in repeated games |
title_fullStr |
Empirically evaluating multiagent reinforcement learning algorithms in repeated games |
title_full_unstemmed |
Empirically evaluating multiagent reinforcement learning algorithms in repeated games |
title_sort |
empirically evaluating multiagent reinforcement learning algorithms in repeated games |
publishDate |
2009 |
url |
http://hdl.handle.net/2429/16633 |
work_keys_str_mv |
AT lipsonasherg empiricallyevaluatingmultiagentreinforcementlearningalgorithmsinrepeatedgames |
_version_ |
1718590285497237504 |