Evolutionary Reinforcement Learning with Parameterized Action Primitives for Diverse Manipulation Tasks

Authors

  • Xianxu Qiu Shenzhen University
  • Haiming Huang Shenzhen University
  • Weiwei Chen Shenzhen University
  • Qiuzhen Lin Shenzhen University
  • Wei-Neng Chen South China University of Technology
  • Fuchun Sun Tsinghua University

DOI:

https://doi.org/10.1609/aaai.v39i14.33606

Abstract

Reinforcement learning (RL) has shown promising performance in tackling robotic manipulation tasks (RMTs), which require learning a prolonged sequence of manipulation actions to control robots efficiently. However, most RL algorithms often suffer from two problems when solving RMTs: inefficient exploration due to the extremely large action space and catastrophic forgetting due to the poor sampling efficiency. To alleviate these problems, this paper introduces an Evolutionary Reinforcement Learning algorithm with parameterized Action Primitives, called ERLAP, which combines the advantages of an evolutionary algorithm (EA) and hierarchical RL (HRL) to solve diverse RMTs. A library of heterogeneous action primitives is constructed in HRL to enhance the exploration efficiency of robots and dual populations with new evolutionary operators are run in EA to optimize these primitive sequences, which can diversify the distribution of replay buffer and avoid catastrophic forgetting. The experiments show that ERLAP outperforms four state-of-the-art RL algorithms in simulated RMTs with dense rewards and can effectively avoid catastrophic forgetting in a set of more challenging simulated RMTs with sparse rewards.

Downloads

Published

2025-04-11

How to Cite

Qiu, X., Huang, H., Chen, W., Lin, Q., Chen, W.-N., & Sun, F. (2025). Evolutionary Reinforcement Learning with Parameterized Action Primitives for Diverse Manipulation Tasks. Proceedings of the AAAI Conference on Artificial Intelligence, 39(14), 14655–14663. https://doi.org/10.1609/aaai.v39i14.33606

Issue

Section

AAAI Technical Track on Intelligent Robots