Distributional Reinforcement Learning via Moment Matching
Keywords:Reinforcement Learning, Kernel Methods, (Deep) Neural Network Algorithms, Representation Learning
AbstractWe consider the problem of learning a set of probability distributions from the empirical Bellman dynamics in distributional reinforcement learning (RL), a class of state-of-the-art methods that estimate the distribution, as opposed to only the expectation, of the total return. We formulate a method that learns a finite set of statistics from each return distribution via neural networks, as in the distributional RL literature. Existing distributional RL methods however constrain the learned statistics to predefined functional forms of the return distribution which is both restrictive in representation and difficult in maintaining the predefined statistics. Instead, we learn unrestricted statistics, i.e., deterministic (pseudo-)samples, of the return distribution by leveraging a technique from hypothesis testing known as maximum mean discrepancy (MMD), which leads to a simpler objective amenable to backpropagation. Our method can be interpreted as implicitly matching all orders of moments between a return distribution and its Bellman target. We establish sufficient conditions for the contraction of the distributional Bellman operator and provide finite-sample analysis for the deterministic samples in distribution approximation. Experiments on the suite of Atari games show that our method outperforms the standard distributional RL baselines and sets a new record in the Atari games for non-distributed agents.
How to Cite
Nguyen-Tang, T., Gupta, S., & Venkatesh, S. (2021). Distributional Reinforcement Learning via Moment Matching. Proceedings of the AAAI Conference on Artificial Intelligence, 35(10), 9144-9152. Retrieved from https://ojs.aaai.org/index.php/AAAI/article/view/17104
AAAI Technical Track on Machine Learning III