Incremental Reinforcement Learning with Dual-Adaptive ε-Greedy Exploration

Authors

  • Wei Ding National Taiwan University
  • Siyang Jiang National Taiwan University
  • Hsi-Wen Chen National Taiwan University
  • Ming-Syan Chen National Taiwan University

DOI:

https://doi.org/10.1609/aaai.v37i6.25899

Keywords:

ML: Reinforcement Learning Algorithms, ML: Lifelong and Continual Learning

Abstract

Reinforcement learning (RL) has achieved impressive performance in various domains. However, most RL frameworks oversimplify the problem by assuming a fixed-yet-known environment and often have difficulty being generalized to real-world scenarios. In this paper, we address a new challenge with a more realistic setting, Incremental Reinforcement Learning, where the search space of the Markov Decision Process continually expands. While previous methods usually suffer from the lack of efficiency in exploring the unseen transitions, especially with increasing search space, we present a new exploration framework named Dual-Adaptive ϵ-greedy Exploration (DAE) to address the challenge of Incremental RL. Specifically, DAE employs a Meta Policy and an Explorer to avoid redundant computation on those sufficiently learned samples. Furthermore, we release a testbed based on a synthetic environment and the Atari benchmark to validate the effectiveness of any exploration algorithms under Incremental RL. Experimental results demonstrate that the proposed framework can efficiently learn the unseen transitions in new environments, leading to notable performance improvement, i.e., an average of more than 80%, over eight baselines examined.

Downloads

Published

2023-06-26

How to Cite

Ding, W., Jiang, S., Chen, H.-W., & Chen, M.-S. (2023). Incremental Reinforcement Learning with Dual-Adaptive ε-Greedy Exploration. Proceedings of the AAAI Conference on Artificial Intelligence, 37(6), 7387-7395. https://doi.org/10.1609/aaai.v37i6.25899

Issue

Section

AAAI Technical Track on Machine Learning I