Robust Bayesian Inverse Reinforcement Learning with Sparse Behavior Noise


  • Jiangchuan Zheng Hong Kong University of Science and Technology
  • Siyuan Liu Carnegie Mellon University
  • Lionel M. Ni Hong Kong University of Science and Technology



Inverse Reinforcement Learning, Robust Model, Sparse Behavior Noise, Variational Inference


Inverse reinforcement learning (IRL) aims to recover the reward function underlying a Markov Decision Process from behaviors of experts in support of decision-making. Most recent work on IRL assumes the same level of trustworthiness of all expert behaviors, and frames IRL as a process of seeking reward function that makes those behaviors appear (near)-optimal. However, it is common in reality that noisy expert behaviors disobeying the optimal policy exist, which may degrade the IRL performance significantly. To address this issue, in this paper, we develop a robust IRL framework that can accurately estimate the reward function in the presence of behavior noise. In particular, we focus on a special type of behavior noise referred to as sparse noise due to its wide popularity in real-world behavior data. To model such noise, we introduce a novel latent variable characterizing the reliability of each expert action and use Laplace distribution as its prior. We then devise an EM algorithm with a novel variational inference procedure in the E-step, which can automatically identify and remove behavior noise in reward learning. Experiments on both synthetic data and real vehicle routing data with noticeable behavior noise show significant improvement of our method over previous approaches in learning accuracy, and also show its power in de-noising behavior data.




How to Cite

Zheng, J., Liu, S., & Ni, L. M. (2014). Robust Bayesian Inverse Reinforcement Learning with Sparse Behavior Noise. Proceedings of the AAAI Conference on Artificial Intelligence, 28(1).



Main Track: Novel Machine Learning Algorithms