MARPO: A Reflective Policy Optimization for Multi-Agent Reinforcement Learning

Authors

  • Cuiling Wu School of Computer Science and Technology, Beijing Institute of Technology QiYuan Lab
  • Yaozhong Gan QiYuan Lab
  • Junliang Xing QiYuan Lab
  • Ying Fu School of Computer Science and Technology, Beijing Institute of Technology

DOI:

https://doi.org/10.1609/aaai.v40i35.40219

Abstract

We propose Multi-Agent Reflective Policy Optimization MARPO to alleviate the issue of sample inefficiency in multi-agent reinforcement learning. MARPO consists of two key components: a reflection mechanism that leverages subsequent trajectories to enhance sample efficiency, and an asymmetric clipping mechanism that is derived from the KL divergence and dynamically adjusts the clipping range to improve training stability. We evaluate MARPO in classic multi-agent environments, where it consistently outperforms other methods.

Downloads

Published

2026-03-14

How to Cite

Wu, C., Gan, Y., Xing, J., & Fu, Y. (2026). MARPO: A Reflective Policy Optimization for Multi-Agent Reinforcement Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 40(35), 29740-29748. https://doi.org/10.1609/aaai.v40i35.40219

Issue

Section

AAAI Technical Track on Multiagent Systems