Active Grammatical Inference for Non-Markovian Planning
Keywords:Non-Markovian Planning, Regular Decision Processes, Grammatical Inference, Non-Markovian Reward, Markov Decision Processes
AbstractPlanning in finite stochastic environments is canonically posed as a Markov decision process where the transition and reward structures are explicitly known. Reinforcement learning (RL) lifts the explicitness assumption by working with sampling models instead. Further, with the advent of reward machines, we can relax the Markovian assumption on the reward. Angluin's active grammatical inference algorithm L* has found novel application in explicating reward machines for non-Markovian RL. We propose maintaining the assumption of explicit transition dynamics, but with an implicit non-Markovian reward signal, which must be inferred from experiments. We call this setting non-Markovian planning, as opposed to non-Markovian RL. The proposed approach leverages L* to explicate an automaton structure for the underlying planning objective. We exploit the environment model to learn an automaton faster and integrate it with value iteration to accelerate the planning. We compare against recent non-Markovian RL solutions which leverage grammatical inference, and establish complexity results that illustrate the difference in runtime between grammatical inference in planning and RL settings.
How to Cite
Topper, N., Atia, G., Trivedi, A., & Velasquez, A. (2022). Active Grammatical Inference for Non-Markovian Planning. Proceedings of the International Conference on Automated Planning and Scheduling, 32(1), 647-651. Retrieved from https://ojs.aaai.org/index.php/ICAPS/article/view/19853
Planning and Learning Track