Good-for-MDP State Reduction for Stochastic LTL Planning

Authors

  • Christoph Weinhuber University of Oxford
  • Giuseppe De Giacomo University of Oxford
  • Yong Li Institute of Software, Chinese Academy of Sciences
  • Sven Schewe University of Liverpool
  • Qiyi Tang University of Liverpool

DOI:

https://doi.org/10.1609/aaai.v40i43.40967

Abstract

We study stochastic planning problems in Markov Decision Processes (MDPs) with goals specified in Linear Temporal Logic (LTL). The state-of-the-art approach transforms LTL formulas into good-for-MDP (GFM) automata, which feature a restricted form of nondeterminism. These automata are then composed with the MDP, allowing the agent to resolve the nondeterminism during policy synthesis. A major factor affecting the scalability of this approach is the size of the generated automata. In this paper, we propose a novel GFM state-space reduction technique that significantly reduces the number of automata states. Our method employs a sophisticated chain of transformations, leveraging recent advances in good-for-games minimisation developed for adversarial settings. In addition to our theoretical contributions, we present empirical results demonstrating the practical effectiveness of our state-reduction technique. Furthermore, we introduce a direct construction method for formulas of the form GFφ, where φ is a co-safety formula. This construction is provably single-exponential in the worst case, in contrast to the general doubly-exponential complexity. Our experiments confirm the scalability advantages of this specialised construction.

Published

2026-03-14

How to Cite

Weinhuber, C., De Giacomo, G., Li, Y., Schewe, S., & Tang, Q. (2026). Good-for-MDP State Reduction for Stochastic LTL Planning. Proceedings of the AAAI Conference on Artificial Intelligence, 40(43), 36457–36465. https://doi.org/10.1609/aaai.v40i43.40967

Issue

Section

AAAI Technical Track on Planning, Routing, and Scheduling