Reward Shaping for Model-Based Bayesian Reinforcement Learning

Authors

  • Hyeoneun Kim KAIST
  • Woosang Lim KAIST
  • Kanghoon Lee KAIST
  • Yung-Kyun Noh KAIST
  • Kee-Eung Kim KAIST

DOI:

https://doi.org/10.1609/aaai.v29i1.9702

Abstract

Bayesian reinforcement learning (BRL) provides a formal framework for optimal exploration-exploitation tradeoff in reinforcement learning. Unfortunately, it is generally intractable to find the Bayes-optimal behavior except for restricted cases. As a consequence, many BRL algorithms, model-based approaches in particular, rely on approximated models or real-time search methods. In this paper, we present potential-based shaping for improving the learning performance in model-based BRL. We propose a number of potential functions that are particularly well suited for BRL, and are domain-independent in the sense that they do not require any prior knowledge about the actual environment. By incorporating the potential function into real-time heuristic search, we show that we can significantly improve the learning performance in standard benchmark domains.

Downloads

Published

2015-03-04

How to Cite

Kim, H., Lim, W., Lee, K., Noh, Y.-K., & Kim, K.-E. (2015). Reward Shaping for Model-Based Bayesian Reinforcement Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 29(1). https://doi.org/10.1609/aaai.v29i1.9702

Issue

Section

AAAI Technical Track: Reasoning under Uncertainty