Implicit Bilevel Optimization: Differentiating through Bilevel Optimization Programming
DOI:
https://doi.org/10.1609/aaai.v37i12.26716Keywords:
GeneralAbstract
Bilevel Optimization Programming is used to model complex and conflicting interactions between agents, for example in Robust AI or Privacy preserving AI. Integrating bilevel mathematical programming within deep learning is thus an essential objective for the Machine Learning community. Previously proposed approaches only consider single-level programming. In this paper, we extend existing single-level optimization programming approaches and thus propose Differentiating through Bilevel Optimization Programming (BiGrad) for end-to-end learning of models that use Bilevel Programming as a layer. BiGrad has wide applicability and can be used in modern machine learning frameworks. BiGrad is applicable to both continuous and combinatorial Bilevel optimization problems. We describe a class of gradient estimators for the combinatorial case which reduces the requirements in terms of computation complexity; for the case of the continuous variable, the gradient computation takes advantage of the push-back approach (i.e. vector-jacobian product) for an efficient implementation. Experiments show that the BiGrad successfully extends existing single-level approaches to Bilevel Programming.Downloads
Published
2023-06-26
How to Cite
Alesiani, F. (2023). Implicit Bilevel Optimization: Differentiating through Bilevel Optimization Programming. Proceedings of the AAAI Conference on Artificial Intelligence, 37(12), 14683-14691. https://doi.org/10.1609/aaai.v37i12.26716
Issue
Section
AAAI Special Track on Safe and Robust AI