A general framework for reducing variance in agent evaluation
In this work, we present a unified, general approach to variance reduction in agent evaluation using machine learning to minimize variance. Evaluating an agent's performance in a stochastic setting is necessary for agent development, scientific evaluation, and competitions. Traditionally, evalu...
Main Author: | |
---|---|
Other Authors: | |
Format: | Others |
Language: | en |
Published: |
2010
|
Subjects: | |
Online Access: | http://hdl.handle.net/10048/890 |
id |
ndltd-LACETR-oai-collectionscanada.gc.ca-AEU.10048-890 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-LACETR-oai-collectionscanada.gc.ca-AEU.10048-8902012-03-21T22:50:08ZBowling, Michael (Computing Science)Schuurmans, Dale (Computing Science)White, Martha2010-01-06T18:18:02Z2010-01-06T18:18:02Z2010-01-06T18:18:02Zhttp://hdl.handle.net/10048/890In this work, we present a unified, general approach to variance reduction in agent evaluation using machine learning to minimize variance. Evaluating an agent's performance in a stochastic setting is necessary for agent development, scientific evaluation, and competitions. Traditionally, evaluation is done using Monte Carlo estimation (sample averages); the magnitude of the stochasticity in the domain or the high cost of sampling, however, can often prevent the approach from resulting in statistically significant conclusions. Recently, an advantage sum technique based on control variates has been proposed for constructing unbiased, low variance estimates of agent performance. The technique requires an expert to define a value function over states of the system, essentially a guess of the state's unknown value. In this work, we propose learning this value function from past interactions between agents in some target population. Our learned value functions have two key advantages: they can be applied in domains where no expert value function is available and they can result in tuned evaluation for a specific population of agents (e.g., novice versus advanced agents). This work has three main contributions. First, we consolidate previous work in using control variates for variance reduction into one unified, general framework and summarize the connections between this previous work. Second, our framework makes variance reduction practically possible in any sequential decision making task where designing the expert value function is time-consuming, difficult or essentially impossible. We prove the optimality of our approach and extend the theoretical understanding of advantage sum estimators. In addition, we significantly extend the applicability of advantage sum estimators and discuss practical methods for using our framework in real-world scenarios. Finally, we provide low-variance estimators for three poker domains previously without variance reduction and improve strategy selection in the expert-level University of Alberta poker bot.392540 bytesapplication/pdfenMartha White and Michael Bowling. Learning a value analysis tool for agent evaluation. In Proceedings of the Twenty-First International Joint Conference on Artificial Intelligence, pages 1976-1981, 2009variance reductionmachine learningagent evaluationA general framework for reducing variance in agent evaluationThesisMaster of ScienceMaster'sDepartment of Computing ScienceUniversity of Alberta2010-06Szafron, Duane (Computing Science)Hooper, Peter (Mathematical and Statistical Sciences) |
collection |
NDLTD |
language |
en |
format |
Others
|
sources |
NDLTD |
topic |
variance reduction machine learning agent evaluation |
spellingShingle |
variance reduction machine learning agent evaluation White, Martha A general framework for reducing variance in agent evaluation |
description |
In this work, we present a unified, general approach to variance reduction in agent evaluation using machine learning to minimize variance. Evaluating an agent's performance in a stochastic setting is necessary for agent development, scientific evaluation, and competitions. Traditionally, evaluation is done using Monte Carlo estimation (sample averages); the magnitude of the stochasticity in the domain or the high cost of sampling, however, can often prevent the approach from resulting in statistically significant conclusions. Recently, an advantage sum technique based on control variates has been proposed for constructing unbiased, low variance estimates of agent performance. The technique requires an expert to define a value function over states of the system, essentially a guess of the state's unknown value. In this work, we propose learning this value function from past interactions between agents in some target population. Our learned value functions have two key advantages: they can be applied in domains where no expert value function is available and they can result in tuned evaluation for a specific population of agents (e.g., novice versus advanced agents). This work has three main contributions. First, we consolidate previous work in using control variates for variance reduction into one unified, general framework and summarize the connections between this previous work. Second, our framework makes variance reduction practically possible in any sequential decision making task where designing the expert value function is time-consuming, difficult or essentially impossible. We prove the optimality of our approach and extend the theoretical understanding of advantage sum estimators. In addition, we significantly extend the applicability of advantage sum estimators and discuss practical methods for using our framework in real-world scenarios. Finally, we provide low-variance estimators for three poker domains previously without variance reduction and improve strategy selection in the expert-level University of Alberta poker bot. |
author2 |
Bowling, Michael (Computing Science) |
author_facet |
Bowling, Michael (Computing Science) White, Martha |
author |
White, Martha |
author_sort |
White, Martha |
title |
A general framework for reducing variance in agent evaluation |
title_short |
A general framework for reducing variance in agent evaluation |
title_full |
A general framework for reducing variance in agent evaluation |
title_fullStr |
A general framework for reducing variance in agent evaluation |
title_full_unstemmed |
A general framework for reducing variance in agent evaluation |
title_sort |
general framework for reducing variance in agent evaluation |
publishDate |
2010 |
url |
http://hdl.handle.net/10048/890 |
work_keys_str_mv |
AT whitemartha ageneralframeworkforreducingvarianceinagentevaluation AT whitemartha generalframeworkforreducingvarianceinagentevaluation |
_version_ |
1716391444618412032 |