Efficient Gradient Approximation Method for Constrained Bilevel Optimization

Authors

  • Siyuan Xu The Pennsylvania State University
  • Minghui Zhu The Pennsylvania State University

DOI:

https://doi.org/10.1609/aaai.v37i10.26473

Keywords:

SO: Evaluation and Analysis, SO: Applications

Abstract

Bilevel optimization has been developed for many machine learning tasks with large-scale and high-dimensional data. This paper considers a constrained bilevel optimization problem, where the lower-level optimization problem is convex with equality and inequality constraints and the upper-level optimization problem is non-convex. The overall objective function is non-convex and non-differentiable. To solve the problem, we develop a gradient-based approach, called gradient approximation method, which determines the descent direction by computing several representative gradients of the objective function inside a neighborhood of the current estimate. We show that the algorithm asymptotically converges to the set of Clarke stationary points, and demonstrate the efficacy of the algorithm by the experiments on hyperparameter optimization and meta-learning.

Downloads

Published

2023-06-26

How to Cite

Xu, S., & Zhu, M. (2023). Efficient Gradient Approximation Method for Constrained Bilevel Optimization. Proceedings of the AAAI Conference on Artificial Intelligence, 37(10), 12509-12517. https://doi.org/10.1609/aaai.v37i10.26473

Issue

Section

AAAI Technical Track on Search and Optimization