Back to Search Start Over

Byzantine-Resilient Decentralized Policy Evaluation With Linear Function Approximation.

Authors :
Wu, Zhaoxian
Shen, Han
Chen, Tianyi
Ling, Qing
Source :
IEEE Transactions on Signal Processing; 11/15/2021, p3839-3853, 15p
Publication Year :
2021

Abstract

In this paper, we consider the policy evaluation problem in reinforcement learning with agents on a decentralized and directed network. In order to evaluate the quality of a fixed policy in this decentralized setting, one option is for agents to run decentralized temporal-difference (TD) collaboratively. To account for the practical scenarios where the state and action spaces are large and malicious attacks emerge, we focus on the decentralized TD learning with linear function approximation in the presence of malicious agents (often termed as Byzantine agents). We propose a trimmed mean-based Byzantine-resilient decentralized TD algorithm to perform policy evaluation in this setting. We establish the finite-time convergence rate, as well as the asymptotic learning error that depends on the number of Byzantine agents. Numerical experiments corroborate the robustness of the proposed algorithm. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
1053587X
Database :
Complementary Index
Journal :
IEEE Transactions on Signal Processing
Publication Type :
Academic Journal
Accession number :
153880572
Full Text :
https://doi.org/10.1109/TSP.2021.3090952