MetaCARD: Meta-Reinforcement Learning with Task Uncertainty Feedback via Decoupled Context-Aware Reward and Dynamics Components

Authors

  • Min Wang Beijing Institute of Technology
  • Xin Li Beijing Institute of Technology
  • Leiji Zhang Beijing Institute of Technology
  • Mingzhong Wang The University of the Sunshine Coast

DOI:

https://doi.org/10.1609/aaai.v38i14.29482

Keywords:

ML: Reinforcement Learning, ML: Representation Learning, ML: Transfer, Domain Adaptation, Multi-Task Learning

Abstract

Meta-Reinforcement Learning (Meta-RL) aims to reveal shared characteristics in dynamics and reward functions across diverse training tasks. This objective is achieved by meta-learning a policy that is conditioned on task representations with encoded trajectory data or context, thus allowing rapid adaptation to new tasks from a known task distribution. However, since the trajectory data generated by the policy may be biased, the task inference module tends to form spurious correlations between trajectory data and specific tasks, thereby leading to poor adaptation to new tasks. To address this issue, we propose the Meta-RL with task unCertAinty feedback through decoupled context-aware Reward and Dynamics components (MetaCARD). MetaCARD distinctly decouples the dynamics and rewards when inferring tasks and integrates task uncertainty feedback from policy evaluation into the task inference module. This design effectively reduces uncertainty in tasks with changes in dynamics or/and reward functions, thereby enabling accurate task identification and adaptation. The experiment results on both Meta-World and classical MuJoCo benchmarks show that MetaCARD significantly outperforms prevailing Meta-RL baselines, demonstrating its remarkable adaptation ability in sophisticated environments that involve changes in both reward functions and dynamics.

Published

2024-03-24

How to Cite

Wang, M., Li, X., Zhang, L., & Wang, M. (2024). MetaCARD: Meta-Reinforcement Learning with Task Uncertainty Feedback via Decoupled Context-Aware Reward and Dynamics Components. Proceedings of the AAAI Conference on Artificial Intelligence, 38(14), 15555-15562. https://doi.org/10.1609/aaai.v38i14.29482

Issue

Section

AAAI Technical Track on Machine Learning V