Back to Search
Start Over
MADiff: Offline Multi-agent Learning with Diffusion Models
- Publication Year :
- 2023
- Publisher :
- arXiv, 2023.
-
Abstract
- Diffusion model (DM), as a powerful generative model, recently achieved huge success in various scenarios including offline reinforcement learning, where the policy learns to conduct planning by generating trajectory in the online evaluation. However, despite the effectiveness shown for single-agent learning, it remains unclear how DMs can operate in multi-agent problems, where agents can hardly complete teamwork without good coordination by independently modeling each agent's trajectories. In this paper, we propose MADiff, a novel generative multi-agent learning framework to tackle this problem. MADiff is realized with an attention-based diffusion model to model the complex coordination among behaviors of multiple diffusion agents. To the best of our knowledge, MADiff is the first diffusion-based multi-agent offline RL framework, which behaves as both a decentralized policy and a centralized controller, which includes opponent modeling and can be used for multi-agent trajectory prediction. MADiff takes advantage of the powerful generative ability of diffusion while well-suited in modeling complex multi-agent interactions. Our experiments show the superior performance of MADiff compared to baseline algorithms in a range of multi-agent learning tasks.<br />Comment: 17 pages, 7 figures, 4 tables
Details
- Database :
- OpenAIRE
- Accession number :
- edsair.doi.dedup.....5ec0f1d11ab4405249bc1c7387da2044
- Full Text :
- https://doi.org/10.48550/arxiv.2305.17330