Back to Search Start Over

ROMA-iQSS: An Objective Alignment Approach via State-Based Value Learning and ROund-Robin Multi-Agent Scheduling

Authors :
Lin, Chi-Hui
Koh, Joewie J.
Roncone, Alessandro
Chen, Lijun
Source :
Proceedings of the 2024 American Control Conference (ACC), 2024
Publication Year :
2024

Abstract

Effective multi-agent collaboration is imperative for solving complex, distributed problems. In this context, two key challenges must be addressed: first, autonomously identifying optimal objectives for collective outcomes; second, aligning these objectives among agents. Traditional frameworks, often reliant on centralized learning, struggle with scalability and efficiency in large multi-agent systems. To overcome these issues, we introduce a decentralized state-based value learning algorithm that enables agents to independently discover optimal states. Furthermore, we introduce a novel mechanism for multi-agent interaction, wherein less proficient agents follow and adopt policies from more experienced ones, thereby indirectly guiding their learning process. Our theoretical analysis shows that our approach leads decentralized agents to an optimal collective policy. Empirical experiments further demonstrate that our method outperforms existing decentralized state-based and action-based value learning strategies by effectively identifying and aligning optimal objectives.<br />Comment: 10 pages, 3 figures, extended version of our 2024 American Control Conference publication

Details

Database :
arXiv
Journal :
Proceedings of the 2024 American Control Conference (ACC), 2024
Publication Type :
Report
Accession number :
edsarx.2404.03984
Document Type :
Working Paper