Meta-learning within Projective Simulation
Learning models of artificial intelligence can nowadays perform very well on a large variety of tasks. However, in practice, different task environments are best handled by different learning models, rather than a single universal approach. Most non-trivial models thus require the adjustment of seve...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2016-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/7458793/ |
id |
doaj-d8d49bcaf38d4f4da017ddc56aba13b9 |
---|---|
record_format |
Article |
spelling |
doaj-d8d49bcaf38d4f4da017ddc56aba13b92021-03-29T19:40:53ZengIEEEIEEE Access2169-35362016-01-0142110212210.1109/ACCESS.2016.25565797458793Meta-learning within Projective SimulationAdi Makmal0Alexey A. Melnikov1https://orcid.org/0000-0002-5033-4063Vedran Dunjko2Hans J. Briegel3Institute for Theoretical Physics, University of Innsbruck, Innsbruck, AustriaInstitute for Theoretical Physics, University of Innsbruck, Innsbruck, AustriaInstitute for Theoretical Physics, University of Innsbruck, Innsbruck, AustriaInstitute for Theoretical Physics, University of Innsbruck, Innsbruck, AustriaLearning models of artificial intelligence can nowadays perform very well on a large variety of tasks. However, in practice, different task environments are best handled by different learning models, rather than a single universal approach. Most non-trivial models thus require the adjustment of several to many learning parameters, which is often done on a case-by-case basis by an external party. Meta-learning refers to the ability of an agent to autonomously and dynamically adjust its own learning parameters or meta-parameters. In this paper, we show how projective simulation, a recently developed model of artificial intelligence, can naturally be extended to account for meta-learning in reinforcement learning settings. The projective simulation approach is based on a random walk process over a network of clips. The suggested meta-learning scheme builds upon the same design and employs clip networks to monitor the agent's performance and to adjust its meta-parameters on the fly. We distinguish between reflex-type adaptation and adaptation through learning, and show the utility of both approaches. In addition, a trade-off between flexibility and learning-time is addressed. The extended model is examined on three different kinds of reinforcement learning tasks, in which the agent has different optimal values of the meta-parameters, and is shown to perform well, reaching near-optimal to optimal success rates in all of them, without ever needing to manually adjust any meta-parameter.https://ieeexplore.ieee.org/document/7458793/Machine learningreinforcement learningadaptive algorithmmeta-learningrandom processesquantum mechanics |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Adi Makmal Alexey A. Melnikov Vedran Dunjko Hans J. Briegel |
spellingShingle |
Adi Makmal Alexey A. Melnikov Vedran Dunjko Hans J. Briegel Meta-learning within Projective Simulation IEEE Access Machine learning reinforcement learning adaptive algorithm meta-learning random processes quantum mechanics |
author_facet |
Adi Makmal Alexey A. Melnikov Vedran Dunjko Hans J. Briegel |
author_sort |
Adi Makmal |
title |
Meta-learning within Projective Simulation |
title_short |
Meta-learning within Projective Simulation |
title_full |
Meta-learning within Projective Simulation |
title_fullStr |
Meta-learning within Projective Simulation |
title_full_unstemmed |
Meta-learning within Projective Simulation |
title_sort |
meta-learning within projective simulation |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2016-01-01 |
description |
Learning models of artificial intelligence can nowadays perform very well on a large variety of tasks. However, in practice, different task environments are best handled by different learning models, rather than a single universal approach. Most non-trivial models thus require the adjustment of several to many learning parameters, which is often done on a case-by-case basis by an external party. Meta-learning refers to the ability of an agent to autonomously and dynamically adjust its own learning parameters or meta-parameters. In this paper, we show how projective simulation, a recently developed model of artificial intelligence, can naturally be extended to account for meta-learning in reinforcement learning settings. The projective simulation approach is based on a random walk process over a network of clips. The suggested meta-learning scheme builds upon the same design and employs clip networks to monitor the agent's performance and to adjust its meta-parameters on the fly. We distinguish between reflex-type adaptation and adaptation through learning, and show the utility of both approaches. In addition, a trade-off between flexibility and learning-time is addressed. The extended model is examined on three different kinds of reinforcement learning tasks, in which the agent has different optimal values of the meta-parameters, and is shown to perform well, reaching near-optimal to optimal success rates in all of them, without ever needing to manually adjust any meta-parameter. |
topic |
Machine learning reinforcement learning adaptive algorithm meta-learning random processes quantum mechanics |
url |
https://ieeexplore.ieee.org/document/7458793/ |
work_keys_str_mv |
AT adimakmal metalearningwithinprojectivesimulation AT alexeyamelnikov metalearningwithinprojectivesimulation AT vedrandunjko metalearningwithinprojectivesimulation AT hansjbriegel metalearningwithinprojectivesimulation |
_version_ |
1724195883262672896 |