Provably Convergent Federated Trilevel Learning

Authors

  • Yang Jiao Department of Computer Science and Technology, Tongji University
  • Kai Yang Department of Computer Science and Technology, Tongji University Key Laboratory of Embedded System and Service Computing Ministry of Education at Tongji University Shanghai Research Institute for Intelligent Autonomous Systems
  • Tiancheng Wu Department of Computer Science and Technology, Tongji University
  • Chengtao Jian Department of Computer Science and Technology, Tongji University
  • Jianwei Huang School of Science and Engineering, The Chinese University of Hong Kong, Shenzhen Shenzhen Institute of Artificial Intelligence and Robotics for Society

DOI:

https://doi.org/10.1609/aaai.v38i11.29190

Keywords:

ML: Distributed Machine Learning & Federated Learning, ML: Optimization

Abstract

Trilevel learning, also called trilevel optimization (TLO), has been recognized as a powerful modelling tool for hierarchical decision process and widely applied in many machine learning applications, such as robust neural architecture search, hyperparameter optimization, and domain adaptation. Tackling TLO problems has presented a great challenge due to their nested decision-making structure. In addition, existing works on TLO face the following key challenges: 1) they all focus on the non-distributed setting, which may lead to privacy breach; 2) they do not offer any non-asymptotic convergence analysis which characterizes how fast an algorithm converges. To address the aforementioned challenges, this paper proposes an asynchronous federated trilevel optimization method to solve TLO problems. The proposed method utilizes u-cuts to construct a hyper-polyhedral approximation for the TLO problem and solve it in an asynchronous manner. We demonstrate that the proposed u-cuts are applicable to not only convex functions but also a wide range of non-convex functions that meet the u-weakly convex assumption. Furthermore, we theoretically analyze the non-asymptotic convergence rate for the proposed method by showing its iteration complexity to obtain ϵ-stationary point is upper bounded by O(1/ϵ²). Extensive experiments on real-world datasets have been conducted to elucidate the superiority of the proposed method, e.g., it has a faster convergence rate with a maximum acceleration of approximately 80%.

Downloads

Published

2024-03-24

How to Cite

Jiao, Y., Yang, K., Wu, T., Jian, C., & Huang, J. (2024). Provably Convergent Federated Trilevel Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 38(11), 12928-12937. https://doi.org/10.1609/aaai.v38i11.29190

Issue

Section

AAAI Technical Track on Machine Learning II