Multi-Expert Distillation for Few-Shot Coordination (Student Abstract)
DOI:
https://doi.org/10.1609/aaai.v38i21.30539Keywords:
Multiagent Learning, Multiagent Systems, Reinforcement LearningAbstract
Ad hoc teamwork is a crucial challenge that aims to design an agent capable of effective collaboration with teammates employing diverse strategies without prior coordination. However, current Population-Based Training (PBT) approaches train the ad hoc agent through interaction with diverse teammates from scratch, which suffer from low efficiency. We introduce Multi-Expert Distillation (MED), a novel approach that directly distills diverse strategies through modeling across-episodic sequences. Experiments show that our algorithm achieves more efficient and stable training and has the ability to improve its behavior using historical contexts. Our code is available at https://github.com/LAMDA-RL/MED.Downloads
Published
2024-03-24
How to Cite
Zhu, Y., Ding, H., & Zhang, Z. (2024). Multi-Expert Distillation for Few-Shot Coordination (Student Abstract). Proceedings of the AAAI Conference on Artificial Intelligence, 38(21), 23717–23719. https://doi.org/10.1609/aaai.v38i21.30539
Issue
Section
AAAI Student Abstract and Poster Program