Taming Binarized Neural Networks and Mixed-Integer Programs

Authors

  • Johannes Aspman Czech Technical University
  • Georgios Korpas HSBC Czech Technical University
  • Jakub Marecek Czech Technical University

DOI:

https://doi.org/10.1609/aaai.v38i10.28968

Keywords:

ML: Deep Learning Theory, ML: Optimization

Abstract

There has been a great deal of recent interest in binarized neural networks, especially because of their explainability. At the same time, automatic differentiation algorithms such as backpropagation fail for binarized neural networks, which limits their applicability. We show that binarized neural networks admit a tame representation by reformulating the problem of training binarized neural networks as a subadditive dual of a mixed-integer program, which we show to have nice properties. This makes it possible to use the framework of Bolte et al. for implicit differentiation, which offers the possibility for practical implementation of backpropagation in the context of binarized neural networks. This approach could also be used for a broader class of mixed-integer programs, beyond the training of binarized neural networks, as encountered in symbolic approaches to AI and beyond.

Published

2024-03-24

How to Cite

Aspman, J., Korpas, G., & Marecek, J. (2024). Taming Binarized Neural Networks and Mixed-Integer Programs. Proceedings of the AAAI Conference on Artificial Intelligence, 38(10), 10935-10943. https://doi.org/10.1609/aaai.v38i10.28968

Issue

Section

AAAI Technical Track on Machine Learning I