Off-policy reinforcement learning with Gaussian processes
An off-policy Bayesian nonparameteric approximate reinforcement learning framework, termed as GPQ, that employs a Gaussian processes (GP) model of the value (Q) function is presented in both the batch and online settings. Sufficient conditions on GP hyperparameter selection are established to guaran...
Main Authors: | Chowdhary, Girish (Author), Liu, Miao (Author), Grande, Robert (Contributor), Walsh, Thomas (Contributor), How, Jonathan P. (Contributor), Carin, Lawrence (Author) |
---|---|
Other Authors: | Massachusetts Institute of Technology. Aerospace Controls Laboratory (Contributor), Massachusetts Institute of Technology. Department of Aeronautics and Astronautics (Contributor) |
Format: | Article |
Language: | English |
Published: |
Institute of Electrical and Electronics Engineers (IEEE),
2015-05-11T19:13:37Z.
|
Subjects: | |
Online Access: | Get fulltext |
Similar Items
-
Bayesian Nonparametric Adaptive Control Using Gaussian Processes
by: Chowdhary, Girish, et al.
Published: (2015) -
Stick-breaking policy learning in Dec-POMDPs
by: Amato, Christopher, et al.
Published: (2016) -
Efficient Reinforcement Learning using Gaussian Processes
Published: (2010) -
Off-Policy Reinforcement Learning for Robotics
by: Tosatto, Samuele
Published: (2021) -
Gaussian processes non‐linear inverse reinforcement learning
by: Qifeng Qiao, et al.
Published: (2021-06-01)