Back to Search Start Over

Attentional Factorized Q-Learning for Many-Agent Learning

Authors :
Xiaoqiang Wang
Liangjun Ke
Qiang Fu
Source :
IEEE Access, Vol 10, Pp 108775-108784 (2022)
Publication Year :
2022
Publisher :
IEEE, 2022.

Abstract

The difficulty of Multi-Agent Reinforcement Learning (MARL) increases with the growing number of agents in system. The value function decomposition is an effective way to alleviate the curse of dimension. However, the existing methods usually either can only provide the low-order approximate decomposition of no more than the second-order, or need to spend a lot of effort manually designing the high-order interaction among agents according to experience. Therefore, the existing methods either tend to bear large decomposition error or are not convenient to use. In this paper, a high-order approximate value function decomposition method is proposed, which has the following prominent characteristics: the low-rank vector is exploited to represent value function, the low-order and high-order components share the same input (i.e., the embedding vector), the attention mechanism is used to select the agents participating in the high-order interaction, and all agents share the model parameters if the agents are homogeneous. To our knowledge, this is the first MARL method modeling low- and high-order interaction simultaneously among agents that can be trained end-to-end. Extensive experiments on two different multi-agent problems demonstrate the performance gain of our proposed approach in comparison with strong baselines, particularly when there are a large number of agents.

Details

Language :
English
ISSN :
21693536
Volume :
10
Database :
Directory of Open Access Journals
Journal :
IEEE Access
Publication Type :
Academic Journal
Accession number :
edsdoj.505f8c00ba44847a23c174e646312f7
Document Type :
article
Full Text :
https://doi.org/10.1109/ACCESS.2022.3214481