Concurrent PAC RL

Authors

  • Zhaohan Guo Carnegie Mellon University
  • Emma Brunskill Carnegie Mellon University

DOI:

https://doi.org/10.1609/aaai.v29i1.9585

Keywords:

Reinforcement Learning

Abstract

In many real-world situations a decision maker may make decisions across many separate reinforcement learning tasks in parallel, yet there has been very little work on concurrent RL. Building on the efficient exploration RL literature, we introduce two new concurrent RL algorithms and bound their sample complexity. We show that under some mild conditions, both when the agent is known to be acting in many copies of the same MDP, and when they are not the same but are taken from a finite set, we can gain linear improvements in the sample complexity over not sharing information. This is quite exciting as a linear speedup is the most one might hope to gain. Our preliminary experiments confirm this result and show empirical benefits.

Downloads

Published

2015-02-21

How to Cite

Guo, Z., & Brunskill, E. (2015). Concurrent PAC RL. Proceedings of the AAAI Conference on Artificial Intelligence, 29(1). https://doi.org/10.1609/aaai.v29i1.9585

Issue

Section

Main Track: Novel Machine Learning Algorithms