End-to-End Learning for Optimization via Constraint-Enforcing Approximators

Authors

  • Rares Cristian Massachusetts Institute of Technology
  • Pavithra Harsha IBM Research
  • Georgia Perakis Massachusetts Institute of Technology
  • Brian L Quanz IBM Research
  • Ioannis Spantidakis Massachusetts Institute of Technology

DOI:

https://doi.org/10.1609/aaai.v37i6.25884

Keywords:

ML: Optimization, RU: Stochastic Optimization

Abstract

In many real-world applications, predictive methods are used to provide inputs for downstream optimization problems. It has been shown that using the downstream task-based objective to learn the intermediate predictive model is often better than using only intermediate task objectives, such as prediction error. The learning task in the former approach is referred to as end-to-end learning. The difficulty in end-to-end learning lies in differentiating through the optimization problem. Therefore, we propose a neural network architecture that can learn to approximately solve these optimization problems, particularly ensuring its output satisfies the feasibility constraints via alternate projections. We show these projections converge at a geometric rate to the exact projection. Our approach is more computationally efficient than existing methods as we do not need to solve the original optimization problem at each iteration. Furthermore, our approach can be applied to a wider range of optimization problems. We apply this to a shortest path problem for which the first stage forecasting problem is a computer vision task of predicting edge costs from terrain maps, a capacitated multi-product newsvendor problem, and a maximum matching problem. We show that this method out-performs existing approaches in terms of final task-based loss and training time.

Downloads

Published

2023-06-26

How to Cite

Cristian, R., Harsha, P., Perakis, G., Quanz, B. L., & Spantidakis, I. (2023). End-to-End Learning for Optimization via Constraint-Enforcing Approximators. Proceedings of the AAAI Conference on Artificial Intelligence, 37(6), 7253-7260. https://doi.org/10.1609/aaai.v37i6.25884

Issue

Section

AAAI Technical Track on Machine Learning I