Unsupervised Object Interaction Learning with Counterfactual Dynamics Models

Authors

  • Jongwook Choi University of Michigan
  • Sungtae Lee Individual Researcher
  • Xinyu Wang University of Michigan
  • Sungryull Sohn LG AI Research
  • Honglak Lee University of Michigan LG AI Research

DOI:

https://doi.org/10.1609/aaai.v38i10.29039

Keywords:

ML: Reinforcement Learning, ML: Unsupervised & Self-Supervised Learning

Abstract

We present COIL (Counterfactual Object Interaction Learning), a novel way of learning skills of object interactions on entity-centric environments. The goal is to learn primitive behaviors that can induce interactions without external reward or any supervision. Existing skill discovery methods are limited to locomotion, simple navigation tasks, or single-object manipulation tasks, mostly not inducing interaction between objects. Unlike a monolithic representation usually used in prior skill learning methods, we propose to use a structured goal representation that can query and scope which objects to interact with, which can serve as a basis for solving more complex downstream tasks. We design a novel counterfactual intrinsic reward through the use of either a forward model or successor features that can learn an interaction skill between a pair of objects given as a goal. Through experiments on continuous control environments such as Magnetic Block and 2.5-D Stacking Box, we demonstrate that an agent can learn object interaction behaviors (e.g., attaching or stacking one block to another) without any external rewards or domain-specific knowledge.

Published

2024-03-24

How to Cite

Choi, J., Lee, S., Wang, X., Sohn, S., & Lee, H. (2024). Unsupervised Object Interaction Learning with Counterfactual Dynamics Models. Proceedings of the AAAI Conference on Artificial Intelligence, 38(10), 11570-11578. https://doi.org/10.1609/aaai.v38i10.29039

Issue

Section

AAAI Technical Track on Machine Learning I