Distilling Autoregressive Models to Obtain High-Performance Non-autoregressive Solvers for Vehicle Routing Problems with Faster Inference Speed

Authors

  • Yubin Xiao Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, China
  • Di Wang Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly, Nanyang Technological University, Singapore WeBank-NTU Joint Research Institute on Fintech, Nanyang Technological University, Singapore
  • Boyang Li School of Computer Science and Engineering, Nanyang Technological University, Singapore
  • Mingzhao Wang Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, China
  • Xuan Wu Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, China
  • Changliang Zhou School of System Design and Intelligent Manufacturing, Southern University of Science and Technology, China
  • You Zhou Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, China

DOI:

https://doi.org/10.1609/aaai.v38i18.30008

Keywords:

PRS: Learning for Planning and Scheduling

Abstract

Neural construction models have shown promising performance for Vehicle Routing Problems (VRPs) by adopting either the Autoregressive (AR) or Non-Autoregressive (NAR) learning approach. While AR models produce high-quality solutions, they generally have a high inference latency due to their sequential generation nature. Conversely, NAR models generate solutions in parallel with a low inference latency but generally exhibit inferior performance. In this paper, we propose a generic Guided Non-Autoregressive Knowledge Distillation (GNARKD) method to obtain high-performance NAR models having a low inference latency. GNARKD removes the constraint of sequential generation in AR models while preserving the learned pivotal components in the network architecture to obtain the corresponding NAR models through knowledge distillation. We evaluate GNARKD by applying it to three widely adopted AR models to obtain NAR VRP solvers for both synthesized and real-world instances. The experimental results demonstrate that GNARKD significantly reduces the inference time (4-5 times faster) with acceptable performance drop (2-3%). To the best of our knowledge, this study is first-of-its-kind to obtain NAR VRP solvers from AR ones through knowledge distillation.

Published

2024-03-24

How to Cite

Xiao, Y., Wang, D., Li, B., Wang, M., Wu, X., Zhou, C., & Zhou, Y. (2024). Distilling Autoregressive Models to Obtain High-Performance Non-autoregressive Solvers for Vehicle Routing Problems with Faster Inference Speed. Proceedings of the AAAI Conference on Artificial Intelligence, 38(18), 20274-20283. https://doi.org/10.1609/aaai.v38i18.30008

Issue

Section

AAAI Technical Track on Planning, Routing, and Scheduling