Howard's Policy Iteration is Subexponential for Deterministic Markov Decision Problems with Rewards of Fixed Bit-size and Arbitrary Discount Factor

Authors

  • Dibyangshu Mukherjee Indian Institute of Technology, Bombay
  • Shivaram Kalyanakrishnan Indian Institute of Technology, Bombay

DOI:

https://doi.org/10.1609/icaps.v35i1.36104

Abstract

Howard's Policy Iteration (HPI) is a classic algorithm for solving Markov Decision Problems (MDPs). HPI uses a "greedy" switching rule to update from any non-optimal policy to a dominating one, iterating until an optimal policy is found. Despite its introduction over 60 years ago, the best-known upper bounds on HPI's running time remain exponential in the number of states---indeed even on the restricted class of MDPs with only deterministic transitions (DMDPs). Meanwhile, the tightest lower bound for HPI for MDPs with a constant number of actions per state is only linear. In this paper, we report a significant improvement: a subexponential upper bound for HPI on DMDPs, which is parameterised by the bit-size of the rewards, while independent of the discount factor. The same upper bound also applies to DMDPs with only two possible rewards (which may be of arbitrary size).

Downloads

Published

2025-09-16

How to Cite

Mukherjee, D., & Kalyanakrishnan, S. (2025). Howard’s Policy Iteration is Subexponential for Deterministic Markov Decision Problems with Rewards of Fixed Bit-size and Arbitrary Discount Factor. Proceedings of the International Conference on Automated Planning and Scheduling, 35(1), 84-92. https://doi.org/10.1609/icaps.v35i1.36104