Robust Online Learning Method Based on Dynamical Linear Quadratic Regulator

In this paper, a novel algorithm is proposed for inferring online learning tasks efficiently. By a carefully designed scheme, the online learning problem is first formulated as a state feedback control problem for a series of finite-dimensional systems. Then, the online linear quadratic regulator (O...

Full description

Bibliographic Details
Main Authors: Hanwen Ning, Jiaming Zhang, Xingjian Jing, Tianhai Tian
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8807177/
Description
Summary:In this paper, a novel algorithm is proposed for inferring online learning tasks efficiently. By a carefully designed scheme, the online learning problem is first formulated as a state feedback control problem for a series of finite-dimensional systems. Then, the online linear quadratic regulator (OLQR) learning algorithm is developed to obtain the optimal parameter updating. Solid mathematical analysis on the convergence and rationality of our method is also provided. Compared with the conventional learning methods, our learning framework represents a completely different approach with optimal control techniques, but does not introduce any assumption on the characteristics of noise or learning rate. The proposed method not only guarantees the fast and robust convergence but also achieves better performance in learning efficiency and accuracy, especially for the data streams with complex noise disturbances. In addition, under the proposed framework, new robust algorithms can be potentially developed for various machine learning tasks by using the powerful optimal control techniques. Numerical results on benchmark datasets and practical applications confirm the advantages of our new method.
ISSN:2169-3536