Learning to Play General-Sum Games against Multiple Boundedly Rational Agents

Authors

  • Eric Zhao University of California, Berkeley Salesforce Research
  • Alexander R. Trott Mosaic ML
  • Caiming Xiong Salesforce Research
  • Stephan Zheng Salesforce Research

DOI:

https://doi.org/10.1609/aaai.v37i10.26391

Keywords:

MAS: Multiagent Learning, MAS: Mechanism Design

Abstract

We study the problem of training a principal in a multi-agent general-sum game using reinforcement learning (RL). Learning a robust principal policy requires anticipating the worst possible strategic responses of other agents, which is generally NP-hard. However, we show that no-regret dynamics can identify these worst-case responses in poly-time in smooth games. We propose a framework that uses this policy evaluation method for efficiently learning a robust principal policy using RL. This framework can be extended to provide robustness to boundedly rational agents too. Our motivating application is automated mechanism design: we empirically demonstrate our framework learns robust mechanisms in both matrix games and complex spatiotemporal games. In particular, we learn a dynamic tax policy that improves the welfare of a simulated trade-and-barter economy by 15%, even when facing previously unseen boundedly rational RL taxpayers.

Downloads

Published

2023-06-26

How to Cite

Zhao, E., Trott, A. R., Xiong, C., & Zheng, S. (2023). Learning to Play General-Sum Games against Multiple Boundedly Rational Agents. Proceedings of the AAAI Conference on Artificial Intelligence, 37(10), 11781-11789. https://doi.org/10.1609/aaai.v37i10.26391

Issue

Section

AAAI Technical Track on Multiagent Systems