Towards Cost-Efficient Federated Multi-agent RL with Learnable Aggregation
Paper
Paper/Presentation Title | Towards Cost-Efficient Federated Multi-agent RL with Learnable Aggregation |
---|---|
Presentation Type | Paper |
Authors | Zhang, Yi, Wang, Sen, Chen, Zhi, Xu, Xuwei, Funiak, Stano and Liu, Jiajun |
Journal or Proceedings Title | Proceedings of the 28th Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD 2024) |
Journal Citation | 14646, pp. 171-183 |
Number of Pages | 13 |
Year | 2024 |
Publisher | Springer |
ISBN | 9789819722525 |
9789819722532 | |
Digital Object Identifier (DOI) | https://doi.org/10.1007/978-981-97-2253-2_14 |
Web Address (URL) of Paper | https://link.springer.com/chapter/10.1007/978-981-97-2253-2_14 |
Web Address (URL) of Conference Proceedings | https://link.springer.com/book/10.1007/978-981-97-2253-2 |
Conference/Event | 28th Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD 2024) |
Event Details | 28th Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD 2024) Parent Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD) Delivery In person Event Date 07 to end of 10 May 2024 Event Location Taipei, Taiwan |
Abstract | Multi-agent reinforcement learning (MARL) often adopts centralized training with a decentralized execution (CTDE) framework to facilitate cooperation among agents. When it comes to deploying MARL algorithms in real-world scenarios, CTDE requires gradient transmission and parameter synchronization for each training step, which can incur disastrous communication overhead. To enhance communication efficiency, federated MARL is proposed to average the gradients periodically during communication. However, such straightforward averaging leads to poor coordination and slow convergence arising from the non-i.i.d. problem which is evidenced by our theoretical analysis. To address the two challenges, we propose a federated MARL framework, termed cost-efficient federated multi-agent reinforcement learning with learnable aggregation (FMRL-LA). Specifically, we use asynchronous critics to optimize communication efficiency by filtering out redundant local updates based on the estimation of agent utilities. A centralized aggregator rectifies these estimations conditioned on global information to improve cooperation and reduce non-i.i.d. impact by maximizing the composite system objectives. For a comprehensive evaluation, we extend a challenging multi-agent autonomous driving environment to the federated learning paradigm, comparing our method to competitive MARL baselines. Our findings indicate that FMRL-LA can adeptly balance performance and efficiency. Code and appendix can be found in https://github.com/ArronDZhang/FMRL_LA. |
Keywords | Multi-agent reinforcement learning; Federated |
Contains Sensitive Content | Does not contain sensitive content |
ANZSRC Field of Research 2020 | 4602. Artificial intelligence |
Public Notes | Files associated with this item cannot be displayed due to copyright restrictions. |
Series | Lecture Notes in Computer Science |
Byline Affiliations | University of Queensland |
Commonwealth Scientific and Industrial Research Organisation (CSIRO), Australia |
https://research.usq.edu.au/item/zyx49/towards-cost-efficient-federated-multi-agent-rl-with-learnable-aggregation
9
total views0
total downloads5
views this month0
downloads this month