Attentive Experience Replay

Authors

  • Peiquan Sun University of Science and Technology of China
  • Wengang Zhou University of Science and Technology of China
  • Houqiang Li University of Science and Technology of China

DOI:

https://doi.org/10.1609/aaai.v34i04.6049

Abstract

Experience replay (ER) has become an important component of deep reinforcement learning (RL) algorithms. ER enables RL algorithms to reuse past experiences for the update of current policy. By reusing a previous state for training, the RL agent would learn more accurate value estimation and better decision on that state. However, as the policy is continually updated, some states in past experiences become rarely visited, and optimization over these states might not improve the overall performance of current policy. To tackle this issue, we propose a new replay strategy to prioritize the transitions that contain states frequently visited by current policy. We introduce Attentive Experience Replay (AER), a novel experience replay algorithm that samples transitions according to the similarities between their states and the agent's state. We couple AER with different off-policy algorithms and demonstrate that AER makes consistent improvements on the suite of OpenAI gym tasks.

Downloads

Published

2020-04-03

How to Cite

Sun, P., Zhou, W., & Li, H. (2020). Attentive Experience Replay. Proceedings of the AAAI Conference on Artificial Intelligence, 34(04), 5900-5907. https://doi.org/10.1609/aaai.v34i04.6049

Issue

Section

AAAI Technical Track: Machine Learning