Contextual Pre-planning on Reward Machine Abstractions for Enhanced Transfer in Deep Reinforcement Learning

Authors

  • Guy Azran Technion - Israel Institute of Technology
  • Mohamad H. Danesh McGill University
  • Stefano V. Albrecht University of Edinburgh
  • Sarah Keren Technion - Israel Institute of Technology

DOI:

https://doi.org/10.1609/aaai.v38i10.28970

Keywords:

ML: Transfer, Domain Adaptation, Multi-Task Learning, ML: Reinforcement Learning, PRS: Deterministic Planning, PRS: Model-Based Reasoning

Abstract

Recent studies show that deep reinforcement learning (DRL) agents tend to overfit to the task on which they were trained and fail to adapt to minor environment changes. To expedite learning when transferring to unseen tasks, we propose a novel approach to representing the current task using reward machines (RMs), state machine abstractions that induce subtasks based on the current task’s rewards and dynamics. Our method provides agents with symbolic representations of optimal transitions from their current abstract state and rewards them for achieving these transitions. These representations are shared across tasks, allowing agents to exploit knowledge of previously encountered symbols and transitions, thus enhancing transfer. Empirical results show that our representations improve sample efficiency and few-shot transfer in a variety of domains.

Published

2024-03-24

How to Cite

Azran, G., Danesh, M. H., Albrecht, S. V., & Keren, S. (2024). Contextual Pre-planning on Reward Machine Abstractions for Enhanced Transfer in Deep Reinforcement Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 38(10), 10953-10961. https://doi.org/10.1609/aaai.v38i10.28970

Issue

Section

AAAI Technical Track on Machine Learning I