CATAL: Causally Disentangled Task Representation Learning for Offline Meta-Reinforcement Learning
DOI:
https://doi.org/10.1609/aaai.v40i25.39200Abstract
Context-based Offline Meta Reinforcement Learning (COMRL) has shown promising results in improving the cross-task generalization ability of meta-policies. However, current methods often lead to entangled task representations, in which each latent dimension is influenced by multiple causal factors that govern variations in environment dynamics and reward mechanisms. This entanglement can degrade generalization performance, particularly when multiple causal factors vary simultaneously across tasks. To address this limitation, we propose CAusally disentangled TAsk representation Learning (CATAL) method for COMRL that aims to improve the generalization ability of the meta-policy, where each latent dimension in the task representations aligns to a single causal factor.Theoretically, we show that under mild conditions, the task representations learned by CATAL are causally disentangled. Empirically, extensive results on multi-task MuJoCo benchmarks show that CATAL consistently outperforms existing COMRL baselines in both in-distribution and out-of-distribution generalization.Downloads
Published
2026-03-14
How to Cite
Cong, S., Yu, C., & Lan, X. (2026). CATAL: Causally Disentangled Task Representation Learning for Offline Meta-Reinforcement Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 40(25), 20634–20641. https://doi.org/10.1609/aaai.v40i25.39200
Issue
Section
AAAI Technical Track on Machine Learning II