Optimizing Local Satisfaction of Long-Run Average Objectives in Markov Decision Processes

Authors

  • David Klaška Masaryk University
  • Antonín Kučera Masaryk University
  • Vojtěch Kůr Masaryk University
  • Vít Musil Masaryk University
  • Vojtěch Řehák Masaryk University

DOI:

https://doi.org/10.1609/aaai.v38i18.29993

Keywords:

PRS: Planning with Markov Models (MDPs, POMDPs), ROB: Motion and Path Planning, MAS: Multiagent Planning

Abstract

Long-run average optimization problems for Markov decision processes (MDPs) require constructing policies with optimal steady-state behavior, i.e., optimal limit frequency of visits to the states. However, such policies may suffer from local instability in the sense that the frequency of states visited in a bounded time horizon along a run differs significantly from the limit frequency. In this work, we propose an efficient algorithmic solution to this problem.

Published

2024-03-24

How to Cite

Klaška, D., Kučera, A., Kůr, V., Musil, V., & Řehák, V. (2024). Optimizing Local Satisfaction of Long-Run Average Objectives in Markov Decision Processes. Proceedings of the AAAI Conference on Artificial Intelligence, 38(18), 20143-20150. https://doi.org/10.1609/aaai.v38i18.29993

Issue

Section

AAAI Technical Track on Planning, Routing, and Scheduling