Virtual State Feedback Reference Tuning and Value Iteration Reinforcement Learning for Unknown Observable Systems Control

In this paper, a novel Virtual State-feedback Reference Feedback Tuning (VSFRT) and Approximate Iterative Value Iteration Reinforcement Learning (AI-VIRL) are applied for learning linear reference model output (LRMO) tracking control of observable systems with unknown dynamics. For the observable sy...

Full description

Bibliographic Details
Main Authors: Mircea-Bogdan Radac, Anamaria-Ioana Borlea
Format: Article
Language:English
Published: MDPI AG 2021-02-01
Series:Energies
Subjects:
Online Access:https://www.mdpi.com/1996-1073/14/4/1006
id doaj-ee8b3a11f0b8430ba552855bc3fed662
record_format Article
spelling doaj-ee8b3a11f0b8430ba552855bc3fed6622021-02-16T00:00:16ZengMDPI AGEnergies1996-10732021-02-01141006100610.3390/en14041006Virtual State Feedback Reference Tuning and Value Iteration Reinforcement Learning for Unknown Observable Systems ControlMircea-Bogdan Radac0Anamaria-Ioana Borlea1Department of Automation and Applied Informatics, Politehnica University of Timisoara, 300223 Timisoara, RomaniaDepartment of Automation and Applied Informatics, Politehnica University of Timisoara, 300223 Timisoara, RomaniaIn this paper, a novel Virtual State-feedback Reference Feedback Tuning (VSFRT) and Approximate Iterative Value Iteration Reinforcement Learning (AI-VIRL) are applied for learning linear reference model output (LRMO) tracking control of observable systems with unknown dynamics. For the observable system, a new state representation in terms of input/output (IO) data is derived. Consequently, the Virtual State Feedback Tuning (VRFT)-based solution is redefined to accommodate virtual state feedback control, leading to an original stability-certified Virtual State-Feedback Reference Tuning (VSFRT) concept. Both VSFRT and AI-VIRL use neural networks controllers. We find that AI-VIRL is significantly more computationally demanding and more sensitive to the exploration settings, while leading to inferior LRMO tracking performance when compared to VSFRT. It is not helped either by transfer learning the VSFRT control as initialization for AI-VIRL. State dimensionality reduction using machine learning techniques such as principal component analysis and autoencoders does not improve on the best learned tracking performance however it trades off the learning complexity. Surprisingly, unlike AI-VIRL, the VSFRT control is one-shot (non-iterative) and learns stabilizing controllers even in poorly, open-loop explored environments, proving to be superior in learning LRMO tracking control. Validation on two nonlinear coupled multivariable complex systems serves as a comprehensive case study.https://www.mdpi.com/1996-1073/14/4/1006learning controlreference model output trackingneural networksstate-feedbackreinforcement learningobservability
collection DOAJ
language English
format Article
sources DOAJ
author Mircea-Bogdan Radac
Anamaria-Ioana Borlea
spellingShingle Mircea-Bogdan Radac
Anamaria-Ioana Borlea
Virtual State Feedback Reference Tuning and Value Iteration Reinforcement Learning for Unknown Observable Systems Control
Energies
learning control
reference model output tracking
neural networks
state-feedback
reinforcement learning
observability
author_facet Mircea-Bogdan Radac
Anamaria-Ioana Borlea
author_sort Mircea-Bogdan Radac
title Virtual State Feedback Reference Tuning and Value Iteration Reinforcement Learning for Unknown Observable Systems Control
title_short Virtual State Feedback Reference Tuning and Value Iteration Reinforcement Learning for Unknown Observable Systems Control
title_full Virtual State Feedback Reference Tuning and Value Iteration Reinforcement Learning for Unknown Observable Systems Control
title_fullStr Virtual State Feedback Reference Tuning and Value Iteration Reinforcement Learning for Unknown Observable Systems Control
title_full_unstemmed Virtual State Feedback Reference Tuning and Value Iteration Reinforcement Learning for Unknown Observable Systems Control
title_sort virtual state feedback reference tuning and value iteration reinforcement learning for unknown observable systems control
publisher MDPI AG
series Energies
issn 1996-1073
publishDate 2021-02-01
description In this paper, a novel Virtual State-feedback Reference Feedback Tuning (VSFRT) and Approximate Iterative Value Iteration Reinforcement Learning (AI-VIRL) are applied for learning linear reference model output (LRMO) tracking control of observable systems with unknown dynamics. For the observable system, a new state representation in terms of input/output (IO) data is derived. Consequently, the Virtual State Feedback Tuning (VRFT)-based solution is redefined to accommodate virtual state feedback control, leading to an original stability-certified Virtual State-Feedback Reference Tuning (VSFRT) concept. Both VSFRT and AI-VIRL use neural networks controllers. We find that AI-VIRL is significantly more computationally demanding and more sensitive to the exploration settings, while leading to inferior LRMO tracking performance when compared to VSFRT. It is not helped either by transfer learning the VSFRT control as initialization for AI-VIRL. State dimensionality reduction using machine learning techniques such as principal component analysis and autoencoders does not improve on the best learned tracking performance however it trades off the learning complexity. Surprisingly, unlike AI-VIRL, the VSFRT control is one-shot (non-iterative) and learns stabilizing controllers even in poorly, open-loop explored environments, proving to be superior in learning LRMO tracking control. Validation on two nonlinear coupled multivariable complex systems serves as a comprehensive case study.
topic learning control
reference model output tracking
neural networks
state-feedback
reinforcement learning
observability
url https://www.mdpi.com/1996-1073/14/4/1006
work_keys_str_mv AT mirceabogdanradac virtualstatefeedbackreferencetuningandvalueiterationreinforcementlearningforunknownobservablesystemscontrol
AT anamariaioanaborlea virtualstatefeedbackreferencetuningandvalueiterationreinforcementlearningforunknownobservablesystemscontrol
_version_ 1724268794871808000