Bilinear Exponential Family of MDPs: Frequentist Regret Bound with Tractable Exploration & Planning

Authors

  • Reda Ouhamma Univ. Lille, CNRS, Inria, Centrale Lille, UMR 9189 - CRIStAL, F-59000
  • Debabrota Basu Univ. Lille, CNRS, Inria, Centrale Lille, UMR 9189 - CRIStAL, F-59000
  • Odalric Maillard Univ. Lille, CNRS, Inria, Centrale Lille, UMR 9189 - CRIStAL, F-59000

DOI:

https://doi.org/10.1609/aaai.v37i8.26119

Keywords:

ML: Reinforcement Learning Theory, ML: Reinforcement Learning Algorithms

Abstract

We study the problem of episodic reinforcement learning in continuous state-action spaces with unknown rewards and transitions. Specifically, we consider the setting where the rewards and transitions are modeled using parametric bilinear exponential families. We propose an algorithm, that a) uses penalized maximum likelihood estimators to learn the unknown parameters, b) injects a calibrated Gaussian noise in the parameter of rewards to ensure exploration, and c) leverages linearity of the bilinear exponential family transitions with respect to an underlying RKHS to perform tractable planning. We provide a frequentist regret upper-bound for our algorithm which, in the case of tabular MDPs, is order-optimal with respect to H and K, where H is the episode length and K is the number of episodes. Our analysis improves the existing bounds for the bilinear exponential family of MDPs by square root of H and removes the handcrafted clipping deployed in existing RLSVI-type algorithms.

Downloads

Published

2023-06-26

How to Cite

Ouhamma, R., Basu, D., & Maillard, O. (2023). Bilinear Exponential Family of MDPs: Frequentist Regret Bound with Tractable Exploration & Planning. Proceedings of the AAAI Conference on Artificial Intelligence, 37(8), 9336-9344. https://doi.org/10.1609/aaai.v37i8.26119

Issue

Section

AAAI Technical Track on Machine Learning III