Deep Neural Network Approximated Dynamic Programming for Combinatorial Optimization

Authors

  • Shenghe Xu New York University
  • Shivendra S. Panwar New York University
  • Murali Kodialam Nokia Bell Labs
  • T.V. Lakshman Nokia Bell Labs

DOI:

https://doi.org/10.1609/aaai.v34i02.5531

Abstract

In this paper, we propose a general framework for combining deep neural networks (DNNs) with dynamic programming to solve combinatorial optimization problems. For problems that can be broken into smaller subproblems and solved by dynamic programming, we train a set of neural networks to replace value or policy functions at each decision step. Two variants of the neural network approximated dynamic programming (NDP) methods are proposed; in the value-based NDP method, the networks learn to estimate the value of each choice at the corresponding step, while in the policy-based NDP method the DNNs only estimate the best decision at each step. The training procedure of the NDP starts from the smallest problem size and a new DNN for the next size is trained to cooperate with previous DNNs. After all the DNNs are trained, the networks are fine-tuned together to further improve overall performance. We test NDP on the linear sum assignment problem, the traveling salesman problem and the talent scheduling problem. Experimental results show that NDP can achieve considerable computation time reduction on hard problems with reasonable performance loss. In general, NDP can be applied to reducible combinatorial optimization problems for the purpose of computation time reduction.

Downloads

Published

2020-04-03

How to Cite

Xu, S., Panwar, S. S., Kodialam, M., & Lakshman, T. (2020). Deep Neural Network Approximated Dynamic Programming for Combinatorial Optimization. Proceedings of the AAAI Conference on Artificial Intelligence, 34(02), 1684-1691. https://doi.org/10.1609/aaai.v34i02.5531

Issue

Section

AAAI Technical Track: Constraint Satisfaction and Optimization