Learning Robust Policy against Disturbance in Transition Dynamics via State-Conservative Policy Optimization
Keywords:Machine Learning (ML)
AbstractDeep reinforcement learning algorithms can perform poorly in real-world tasks due to the discrepancy between source and target environments. This discrepancy is commonly viewed as the disturbance in transition dynamics. Many existing algorithms learn robust policies by modeling the disturbance and applying it to source environments during training, which usually requires prior knowledge about the disturbance and control of simulators. However, these algorithms can fail in scenarios where the disturbance from target environments is unknown or is intractable to model in simulators. To tackle this problem, we propose a novel model-free actor-critic algorithm---namely, state-conservative policy optimization (SCPO)---to learn robust policies without modeling the disturbance in advance. Specifically, SCPO reduces the disturbance in transition dynamics to that in state space and then approximates it by a simple gradient-based regularizer. The appealing features of SCPO include that it is simple to implement and does not require additional knowledge about the disturbance or specially designed simulators. Experiments in several robot control tasks demonstrate that SCPO learns robust policies against the disturbance in transition dynamics.
How to Cite
Kuang, Y., Lu, M., Wang, J., Zhou, Q., Li, B., & Li, H. (2022). Learning Robust Policy against Disturbance in Transition Dynamics via State-Conservative Policy Optimization. Proceedings of the AAAI Conference on Artificial Intelligence, 36(7), 7247-7254. https://doi.org/10.1609/aaai.v36i7.20686
AAAI Technical Track on Machine Learning II