Good-for-MDP State Reduction for Stochastic LTL Planning
DOI:
https://doi.org/10.1609/aaai.v40i43.40967Abstract
We study stochastic planning problems in Markov Decision Processes (MDPs) with goals specified in Linear Temporal Logic (LTL). The state-of-the-art approach transforms LTL formulas into good-for-MDP (GFM) automata, which feature a restricted form of nondeterminism. These automata are then composed with the MDP, allowing the agent to resolve the nondeterminism during policy synthesis. A major factor affecting the scalability of this approach is the size of the generated automata. In this paper, we propose a novel GFM state-space reduction technique that significantly reduces the number of automata states. Our method employs a sophisticated chain of transformations, leveraging recent advances in good-for-games minimisation developed for adversarial settings. In addition to our theoretical contributions, we present empirical results demonstrating the practical effectiveness of our state-reduction technique. Furthermore, we introduce a direct construction method for formulas of the form GFφ, where φ is a co-safety formula. This construction is provably single-exponential in the worst case, in contrast to the general doubly-exponential complexity. Our experiments confirm the scalability advantages of this specialised construction.Downloads
Published
2026-03-14
How to Cite
Weinhuber, C., De Giacomo, G., Li, Y., Schewe, S., & Tang, Q. (2026). Good-for-MDP State Reduction for Stochastic LTL Planning. Proceedings of the AAAI Conference on Artificial Intelligence, 40(43), 36457–36465. https://doi.org/10.1609/aaai.v40i43.40967
Issue
Section
AAAI Technical Track on Planning, Routing, and Scheduling