Reinforcement Learning of Causal Variables Using Mediation Analysis

Authors

  • Tue Herlau Technical University of Denmark
  • Rasmus Larsen Alexandra Institute

DOI:

https://doi.org/10.1609/aaai.v36i6.20648

Keywords:

Machine Learning (ML), Reasoning Under Uncertainty (RU), Knowledge Representation And Reasoning (KRR)

Abstract

We consider the problem of acquiring causal representations and concepts in a reinforcement learning setting. Our approach defines a causal variable as being both manipulable by a policy, and able to predict the outcome. We thereby obtain a parsimonious causal graph in which interventions occur at the level of policies. The approach avoids defining a generative model of the data, prior pre-processing, or learning the transition kernel of the Markov decision process. Instead, causal variables and policies are determined by maximizing a new optimization target inspired by mediation analysis, which differs from the expected return. The maximization is accomplished using a generalization of Bellman's equation which is shown to converge, and the method finds meaningful causal representations in a simulated environment.

Downloads

Published

2022-06-28

How to Cite

Herlau, T., & Larsen, R. (2022). Reinforcement Learning of Causal Variables Using Mediation Analysis. Proceedings of the AAAI Conference on Artificial Intelligence, 36(6), 6910-6917. https://doi.org/10.1609/aaai.v36i6.20648

Issue

Section

AAAI Technical Track on Machine Learning I