Criticality-Based Advice in Reinforcement Learning (Student Abstract)
DOI:
https://doi.org/10.1609/aaai.v36i11.21665Keywords:
Human-in-the-loop Machine Learning, Advice-based Reinforcement Learning, Human-aided Reinforcement LearningAbstract
One of the ways to make reinforcement learning (RL) more efficient is by utilizing human advice. Because human advice is expensive, the central question in advice-based reinforcement learning is, how to decide in which states the agent should ask for advice. To approach this challenge, various advice strategies have been proposed. Although all of these strategies distribute advice more efficiently than naive strategies, they rely solely on the agent's estimate of the action-value function, and therefore, are rather inefficient when this estimate is not accurate, in particular, in the early stages of the learning process. To address this weakness, we present an approach to advice-based RL, in which the human’s role is not limited to giving advice in chosen states, but also includes hinting a-priori, before the learning procedure, in which sub-domains of the state space the agent might require more advice. For this purpose we use the concept of critical: states in which choosing the proper action is more important than in other states.Downloads
Published
2022-06-28
How to Cite
Spielberg, Y., & Azaria, A. (2022). Criticality-Based Advice in Reinforcement Learning (Student Abstract). Proceedings of the AAAI Conference on Artificial Intelligence, 36(11), 13057-13058. https://doi.org/10.1609/aaai.v36i11.21665
Issue
Section
AAAI Student Abstract and Poster Program