SkillTree: Explainable Skill-Based Deep Reinforcement Learning for Long-Horizon Control Tasks
DOI:
https://doi.org/10.1609/aaai.v39i20.35451Abstract
Deep reinforcement learning (DRL) has achieved remarkable success in various domains, yet its reliance on neural networks results in a lack of transparency, which limits its practical applications in safety-critical and human-agent interaction domains. Decision trees, known for their notable explainability, have emerged as a promising alternative to neural networks. However, decision trees often struggle in long-horizon continuous control tasks with high-dimensional observation space due to their limited expressiveness. To address this challenge, we propose SkillTree, a novel hierarchical framework that reduces the complex continuous action space of challenging control tasks into discrete skill space. By integrating the differentiable decision tree within the high-level policy, SkillTree generates discrete skill embeddings that guide low-level policy execution. Furthermore, through distillation, we obtain a simplified decision tree model that improves performance while further reducing complexity. Experiment results validate SkillTree’s effectiveness across various robotic manipulation tasks, providing clear skill-level insights into the decision-making process. The proposed approach not only achieves performance comparable to neural network based methods in complex long-horizon control tasks but also significantly enhances the transparency and explainability of the decision-making process.Downloads
Published
2025-04-11
How to Cite
Wen, Y., Li, S., Zuo, R., Yuan, L., Mao, H., & Liu, P. (2025). SkillTree: Explainable Skill-Based Deep Reinforcement Learning for Long-Horizon Control Tasks. Proceedings of the AAAI Conference on Artificial Intelligence, 39(20), 21491–21500. https://doi.org/10.1609/aaai.v39i20.35451
Issue
Section
AAAI Technical Track on Machine Learning VI