Back to Search
Start Over
ROMA-iQSS: An Objective Alignment Approach via State-Based Value Learning and ROund-Robin Multi-Agent Scheduling
- Source :
- Proceedings of the 2024 American Control Conference (ACC), 2024
- Publication Year :
- 2024
-
Abstract
- Effective multi-agent collaboration is imperative for solving complex, distributed problems. In this context, two key challenges must be addressed: first, autonomously identifying optimal objectives for collective outcomes; second, aligning these objectives among agents. Traditional frameworks, often reliant on centralized learning, struggle with scalability and efficiency in large multi-agent systems. To overcome these issues, we introduce a decentralized state-based value learning algorithm that enables agents to independently discover optimal states. Furthermore, we introduce a novel mechanism for multi-agent interaction, wherein less proficient agents follow and adopt policies from more experienced ones, thereby indirectly guiding their learning process. Our theoretical analysis shows that our approach leads decentralized agents to an optimal collective policy. Empirical experiments further demonstrate that our method outperforms existing decentralized state-based and action-based value learning strategies by effectively identifying and aligning optimal objectives.<br />Comment: 10 pages, 3 figures, extended version of our 2024 American Control Conference publication
Details
- Database :
- arXiv
- Journal :
- Proceedings of the 2024 American Control Conference (ACC), 2024
- Publication Type :
- Report
- Accession number :
- edsarx.2404.03984
- Document Type :
- Working Paper