Incremental Reinforcement Learning with Dual-Adaptive ε-Greedy Exploration
DOI:
https://doi.org/10.1609/aaai.v37i6.25899Keywords:
ML: Reinforcement Learning Algorithms, ML: Lifelong and Continual LearningAbstract
Reinforcement learning (RL) has achieved impressive performance in various domains. However, most RL frameworks oversimplify the problem by assuming a fixed-yet-known environment and often have difficulty being generalized to real-world scenarios. In this paper, we address a new challenge with a more realistic setting, Incremental Reinforcement Learning, where the search space of the Markov Decision Process continually expands. While previous methods usually suffer from the lack of efficiency in exploring the unseen transitions, especially with increasing search space, we present a new exploration framework named Dual-Adaptive ϵ-greedy Exploration (DAE) to address the challenge of Incremental RL. Specifically, DAE employs a Meta Policy and an Explorer to avoid redundant computation on those sufficiently learned samples. Furthermore, we release a testbed based on a synthetic environment and the Atari benchmark to validate the effectiveness of any exploration algorithms under Incremental RL. Experimental results demonstrate that the proposed framework can efficiently learn the unseen transitions in new environments, leading to notable performance improvement, i.e., an average of more than 80%, over eight baselines examined.Downloads
Published
2023-06-26
How to Cite
Ding, W., Jiang, S., Chen, H.-W., & Chen, M.-S. (2023). Incremental Reinforcement Learning with Dual-Adaptive ε-Greedy Exploration. Proceedings of the AAAI Conference on Artificial Intelligence, 37(6), 7387-7395. https://doi.org/10.1609/aaai.v37i6.25899
Issue
Section
AAAI Technical Track on Machine Learning I