A leader-follower partially observed Markov game
The intent of this dissertation is to generate a set of non-dominated finite-memory policies from which one of two agents (the leader) can select a most preferred policy to control a dynamic system that is also affected by the control decisions of the other agent (the follower). The problem is descr...
Main Author: | Chang, Yanling |
---|---|
Other Authors: | Erera, Alan L. |
Format: | Others |
Published: |
Georgia Institute of Technology
2016
|
Subjects: | |
Online Access: | http://hdl.handle.net/1853/54407 |
Similar Items
-
Finite Memory Policies for Partially Observable Markov Decision Proesses
by: Lusena, Christopher
Published: (2001) -
Recent Advances in Deep Reinforcement Learning Applications for Solving Partially Observable Markov Decision Processes (POMDP) Problems: Part 1—Fundamentals and Applications in Games, Robotics and Natural Language Processing
by: Xuanchen Xiang, et al.
Published: (2021-07-01) -
Controlled Semi-Markov Processes With Partial Observation
by: Goswami, Anindya
Published: (2011) -
On the Practical Art of State Definitions for Markov Decision Process Construction
by: William T. Scherer, et al.
Published: (2018-01-01) -
Learning in Partially Observable Markov Decision Processes
by: Sachan, Mohit
Published: (2013)