Instance-Based Online Learning of Deterministic Relational Action Models

Authors

  • Joseph Xu University of Michigan
  • John Laird University of Michigan

DOI:

https://doi.org/10.1609/aaai.v24i1.7569

Keywords:

action modeling, soar, analogy, episodic memory, model-based reinforcement learning

Abstract

We present an instance-based, online method for learning action models in unanticipated, relational domains. Our algorithm memorizes pre- and post-states of transitions an agent encounters while experiencing the environment, and makes predictions by using analogy to map the recorded transitions to novel situations. Our algorithm is implemented in the Soar cognitive architecture, integrating its task-independent episodic memory module and analogical reasoning implemented in procedural memory. We evaluate this algorithm’s prediction performance in a modified version of the blocks world domain and the taxi domain. We also present a reinforcement learning agent that uses our model learning algorithm to significantly speed up its convergence to an optimal policy in the modified blocks world domain.

Downloads

Published

2010-07-05

How to Cite

Xu, J., & Laird, J. (2010). Instance-Based Online Learning of Deterministic Relational Action Models. Proceedings of the AAAI Conference on Artificial Intelligence, 24(1), 1574-1579. https://doi.org/10.1609/aaai.v24i1.7569