Learning from Demonstration for Goal-Driven Autonomy

Authors

  • Ben Weber University of California, Santa Cruz
  • Michael Mateas University of California, Santa Cruz
  • Arnav Jhala University of California, Santa Cruz

DOI:

https://doi.org/10.1609/aaai.v26i1.8311

Keywords:

Game AI, Agent Architecture, Learning from Demonstration, Reactive Planning, Real-Time Strategy, Goal-Driven Autonomy

Abstract

Goal-driven autonomy (GDA) is a conceptual model for creating an autonomous agent that monitors a set of expectations during plan execution, detects when discrepancies occur, builds explanations for the cause of failures, and formulates new goals to pursue when planning failures arise. While this framework enables the development of agents that can operate in complex and dynamic environments, implementing the logic for each of the subtasks in the model requires substantial domain engineering. We present a method using case-based reasoning and intent recognition in order to build GDA agents that learn from demonstrations. Our approach reduces the amount of domain engineering necessary to implement GDA agents and learns expectations, explanations, and goals from expert demonstrations. We have applied this approach to build an agent for the real-time strategy game StarCraft. Our results show that integrating the GDA conceptual model into the agent greatly improves its win rate.

Downloads

Published

2021-09-20

How to Cite

Weber, B., Mateas, M., & Jhala, A. (2021). Learning from Demonstration for Goal-Driven Autonomy. Proceedings of the AAAI Conference on Artificial Intelligence, 26(1), 1176-1182. https://doi.org/10.1609/aaai.v26i1.8311

Issue

Section

AAAI Technical Track: Machine Learning