Efficient Gradient Approximation Method for Constrained Bilevel Optimization
DOI:
https://doi.org/10.1609/aaai.v37i10.26473Keywords:
SO: Evaluation and Analysis, SO: ApplicationsAbstract
Bilevel optimization has been developed for many machine learning tasks with large-scale and high-dimensional data. This paper considers a constrained bilevel optimization problem, where the lower-level optimization problem is convex with equality and inequality constraints and the upper-level optimization problem is non-convex. The overall objective function is non-convex and non-differentiable. To solve the problem, we develop a gradient-based approach, called gradient approximation method, which determines the descent direction by computing several representative gradients of the objective function inside a neighborhood of the current estimate. We show that the algorithm asymptotically converges to the set of Clarke stationary points, and demonstrate the efficacy of the algorithm by the experiments on hyperparameter optimization and meta-learning.Downloads
Published
2023-06-26
How to Cite
Xu, S., & Zhu, M. (2023). Efficient Gradient Approximation Method for Constrained Bilevel Optimization. Proceedings of the AAAI Conference on Artificial Intelligence, 37(10), 12509-12517. https://doi.org/10.1609/aaai.v37i10.26473
Issue
Section
AAAI Technical Track on Search and Optimization