State-Conditioned Adversarial Subgoal Generation

Authors

  • Vivienne Huiling Wang Computing Sciences, Tampere University, Finland Department of Electrical Engineering and Automation, Aalto University, Finland
  • Joni Pajarinen Department of Electrical Engineering and Automation, Aalto University, Finland
  • Tinghuai Wang Huawei Helsinki Research Center, Finland
  • Joni-Kristian Kämäräinen Computing Sciences, Tampere University, Finland

DOI:

https://doi.org/10.1609/aaai.v37i8.26213

Keywords:

ML: Reinforcement Learning Theory, ML: Reinforcement Learning Algorithms

Abstract

Hierarchical reinforcement learning (HRL) proposes to solve difficult tasks by performing decision-making and control at successively higher levels of temporal abstraction. However, off-policy HRL often suffers from the problem of a non-stationary high-level policy since the low-level policy is constantly changing. In this paper, we propose a novel HRL approach for mitigating the non-stationarity by adversarially enforcing the high-level policy to generate subgoals compatible with the current instantiation of the low-level policy. In practice, the adversarial learning is implemented by training a simple state conditioned discriminator network concurrently with the high-level policy which determines the compatibility level of subgoals. Comparison to state-of-the-art algorithms shows that our approach improves both learning efficiency and performance in challenging continuous control tasks.

Downloads

Published

2023-06-26

How to Cite

Wang, V. H., Pajarinen, J., Wang, T., & Kämäräinen, J.-K. (2023). State-Conditioned Adversarial Subgoal Generation. Proceedings of the AAAI Conference on Artificial Intelligence, 37(8), 10184-10191. https://doi.org/10.1609/aaai.v37i8.26213

Issue

Section

AAAI Technical Track on Machine Learning III