Meta-learning within Projective Simulation

Learning models of artificial intelligence can nowadays perform very well on a large variety of tasks. However, in practice, different task environments are best handled by different learning models, rather than a single universal approach. Most non-trivial models thus require the adjustment of seve...

Full description

Bibliographic Details
Main Authors: Adi Makmal, Alexey A. Melnikov, Vedran Dunjko, Hans J. Briegel
Format: Article
Language:English
Published: IEEE 2016-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/7458793/
id doaj-d8d49bcaf38d4f4da017ddc56aba13b9
record_format Article
spelling doaj-d8d49bcaf38d4f4da017ddc56aba13b92021-03-29T19:40:53ZengIEEEIEEE Access2169-35362016-01-0142110212210.1109/ACCESS.2016.25565797458793Meta-learning within Projective SimulationAdi Makmal0Alexey A. Melnikov1https://orcid.org/0000-0002-5033-4063Vedran Dunjko2Hans J. Briegel3Institute for Theoretical Physics, University of Innsbruck, Innsbruck, AustriaInstitute for Theoretical Physics, University of Innsbruck, Innsbruck, AustriaInstitute for Theoretical Physics, University of Innsbruck, Innsbruck, AustriaInstitute for Theoretical Physics, University of Innsbruck, Innsbruck, AustriaLearning models of artificial intelligence can nowadays perform very well on a large variety of tasks. However, in practice, different task environments are best handled by different learning models, rather than a single universal approach. Most non-trivial models thus require the adjustment of several to many learning parameters, which is often done on a case-by-case basis by an external party. Meta-learning refers to the ability of an agent to autonomously and dynamically adjust its own learning parameters or meta-parameters. In this paper, we show how projective simulation, a recently developed model of artificial intelligence, can naturally be extended to account for meta-learning in reinforcement learning settings. The projective simulation approach is based on a random walk process over a network of clips. The suggested meta-learning scheme builds upon the same design and employs clip networks to monitor the agent's performance and to adjust its meta-parameters on the fly. We distinguish between reflex-type adaptation and adaptation through learning, and show the utility of both approaches. In addition, a trade-off between flexibility and learning-time is addressed. The extended model is examined on three different kinds of reinforcement learning tasks, in which the agent has different optimal values of the meta-parameters, and is shown to perform well, reaching near-optimal to optimal success rates in all of them, without ever needing to manually adjust any meta-parameter.https://ieeexplore.ieee.org/document/7458793/Machine learningreinforcement learningadaptive algorithmmeta-learningrandom processesquantum mechanics
collection DOAJ
language English
format Article
sources DOAJ
author Adi Makmal
Alexey A. Melnikov
Vedran Dunjko
Hans J. Briegel
spellingShingle Adi Makmal
Alexey A. Melnikov
Vedran Dunjko
Hans J. Briegel
Meta-learning within Projective Simulation
IEEE Access
Machine learning
reinforcement learning
adaptive algorithm
meta-learning
random processes
quantum mechanics
author_facet Adi Makmal
Alexey A. Melnikov
Vedran Dunjko
Hans J. Briegel
author_sort Adi Makmal
title Meta-learning within Projective Simulation
title_short Meta-learning within Projective Simulation
title_full Meta-learning within Projective Simulation
title_fullStr Meta-learning within Projective Simulation
title_full_unstemmed Meta-learning within Projective Simulation
title_sort meta-learning within projective simulation
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2016-01-01
description Learning models of artificial intelligence can nowadays perform very well on a large variety of tasks. However, in practice, different task environments are best handled by different learning models, rather than a single universal approach. Most non-trivial models thus require the adjustment of several to many learning parameters, which is often done on a case-by-case basis by an external party. Meta-learning refers to the ability of an agent to autonomously and dynamically adjust its own learning parameters or meta-parameters. In this paper, we show how projective simulation, a recently developed model of artificial intelligence, can naturally be extended to account for meta-learning in reinforcement learning settings. The projective simulation approach is based on a random walk process over a network of clips. The suggested meta-learning scheme builds upon the same design and employs clip networks to monitor the agent's performance and to adjust its meta-parameters on the fly. We distinguish between reflex-type adaptation and adaptation through learning, and show the utility of both approaches. In addition, a trade-off between flexibility and learning-time is addressed. The extended model is examined on three different kinds of reinforcement learning tasks, in which the agent has different optimal values of the meta-parameters, and is shown to perform well, reaching near-optimal to optimal success rates in all of them, without ever needing to manually adjust any meta-parameter.
topic Machine learning
reinforcement learning
adaptive algorithm
meta-learning
random processes
quantum mechanics
url https://ieeexplore.ieee.org/document/7458793/
work_keys_str_mv AT adimakmal metalearningwithinprojectivesimulation
AT alexeyamelnikov metalearningwithinprojectivesimulation
AT vedrandunjko metalearningwithinprojectivesimulation
AT hansjbriegel metalearningwithinprojectivesimulation
_version_ 1724195883262672896