Back to Search Start Over

Data-Driven Nonzero-Sum Game for Discrete-Time Systems Using Off-Policy Reinforcement Learning

Authors :
Yongliang Yang
Sen Zhang
Jie Dong
Yixin Yin
Source :
IEEE Access, Vol 8, Pp 14074-14088 (2020)
Publication Year :
2020
Publisher :
IEEE, 2020.

Abstract

In this paper, we develop a data-driven algorithm to learn the Nash equilibrium solution for a two-player non-zero-sum (NZS) game with completely unknown linear discrete-time dynamics based on off-policy reinforcement learning (RL). This algorithm solves the coupled algebraic Riccati equations (CARE) forward in time in a model-free manner by using the online measured data. We first derive the CARE for solving the two-player NZS game. Then, model-free off-policy RL is developed to obviate the requirement of complete knowledge of system dynamics. Besides, on- and off-policy RL algorithms are compared in terms of the robustness against the probing noise. Finally, a simulation example is presented to show the efficacy of the presented approach.

Details

Language :
English
ISSN :
21693536
Volume :
8
Database :
Directory of Open Access Journals
Journal :
IEEE Access
Publication Type :
Academic Journal
Accession number :
edsdoj.80ea185393bd4d70ab7e3d29fb112364
Document Type :
article
Full Text :
https://doi.org/10.1109/ACCESS.2019.2960064