TY - JOUR AU - Nguyen-Tang, Thanh AU - Gupta, Sunil AU - Venkatesh, Svetha PY - 2021/05/18 Y2 - 2024/03/29 TI - Distributional Reinforcement Learning via Moment Matching JF - Proceedings of the AAAI Conference on Artificial Intelligence JA - AAAI VL - 35 IS - 10 SE - AAAI Technical Track on Machine Learning III DO - 10.1609/aaai.v35i10.17104 UR - https://ojs.aaai.org/index.php/AAAI/article/view/17104 SP - 9144-9152 AB - We consider the problem of learning a set of probability distributions from the empirical Bellman dynamics in distributional reinforcement learning (RL), a class of state-of-the-art methods that estimate the distribution, as opposed to only the expectation, of the total return. We formulate a method that learns a finite set of statistics from each return distribution via neural networks, as in the distributional RL literature. Existing distributional RL methods however constrain the learned statistics to predefined functional forms of the return distribution which is both restrictive in representation and difficult in maintaining the predefined statistics. Instead, we learn unrestricted statistics, i.e., deterministic (pseudo-)samples, of the return distribution by leveraging a technique from hypothesis testing known as maximum mean discrepancy (MMD), which leads to a simpler objective amenable to backpropagation. Our method can be interpreted as implicitly matching all orders of moments between a return distribution and its Bellman target. We establish sufficient conditions for the contraction of the distributional Bellman operator and provide finite-sample analysis for the deterministic samples in distribution approximation. Experiments on the suite of Atari games show that our method outperforms the standard distributional RL baselines and sets a new record in the Atari games for non-distributed agents. ER -