CATAL: Causally Disentangled Task Representation Learning for Offline Meta-Reinforcement Learning

Authors

  • Shan Cong Sun Yat-sen University Pengcheng Laboratory
  • Chao Yu Sun Yat-sen University Pengcheng Laboratory
  • Xiangyuan Lan Sun Yat-sen University

DOI:

https://doi.org/10.1609/aaai.v40i25.39200

Abstract

Context-based Offline Meta Reinforcement Learning (COMRL) has shown promising results in improving the cross-task generalization ability of meta-policies. However, current methods often lead to entangled task representations, in which each latent dimension is influenced by multiple causal factors that govern variations in environment dynamics and reward mechanisms. This entanglement can degrade generalization performance, particularly when multiple causal factors vary simultaneously across tasks. To address this limitation, we propose CAusally disentangled TAsk representation Learning (CATAL) method for COMRL that aims to improve the generalization ability of the meta-policy, where each latent dimension in the task representations aligns to a single causal factor.Theoretically, we show that under mild conditions, the task representations learned by CATAL are causally disentangled. Empirically, extensive results on multi-task MuJoCo benchmarks show that CATAL consistently outperforms existing COMRL baselines in both in-distribution and out-of-distribution generalization.

Published

2026-03-14

How to Cite

Cong, S., Yu, C., & Lan, X. (2026). CATAL: Causally Disentangled Task Representation Learning for Offline Meta-Reinforcement Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 40(25), 20634–20641. https://doi.org/10.1609/aaai.v40i25.39200

Issue

Section

AAAI Technical Track on Machine Learning II