TurboSVM-FL: Boosting Federated Learning through SVM Aggregation for Lazy Clients

Authors

  • Mengdi Wang Technical University of Munich, Munich, Bavaria, Germany
  • Anna Bodonhelyi Technical University of Munich, Munich, Bavaria, Germany
  • Efe Bozkir Technical University of Munich, Munich, Bavaria, Germany
  • Enkelejda Kasneci Technical University of Munich, Munich, Bavaria, Germany

DOI:

https://doi.org/10.1609/aaai.v38i14.29481

Keywords:

ML: Distributed Machine Learning & Federated Learning, ML: Deep Learning Algorithms, ML: Optimization

Abstract

Federated learning is a distributed collaborative machine learning paradigm that has gained strong momentum in recent years. In federated learning, a central server periodically coordinates models with clients and aggregates the models trained locally by clients without necessitating access to local data. Despite its potential, the implementation of federated learning continues to encounter several challenges, predominantly the slow convergence that is largely due to data heterogeneity. The slow convergence becomes particularly problematic in cross-device federated learning scenarios where clients may be strongly limited by computing power and storage space, and hence counteracting methods that induce additional computation or memory cost on the client side such as auxiliary objective terms and larger training iterations can be impractical. In this paper, we propose a novel federated aggregation strategy, TurboSVM-FL, that poses no additional computation burden on the client side and can significantly accelerate convergence for federated classification task, especially when clients are "lazy" and train their models solely for few epochs for next global aggregation. TurboSVM-FL extensively utilizes support vector machine to conduct selective aggregation and max-margin spread-out regularization on class embeddings. We evaluate TurboSVM-FL on multiple datasets including FEMNIST, CelebA, and Shakespeare using user-independent validation with non-iid data distribution. Our results show that TurboSVM-FL can significantly outperform existing popular algorithms on convergence rate and reduce communication rounds while delivering better test metrics including accuracy, F1 score, and MCC.

Published

2024-03-24

How to Cite

Wang, M., Bodonhelyi, A., Bozkir, E., & Kasneci, E. (2024). TurboSVM-FL: Boosting Federated Learning through SVM Aggregation for Lazy Clients. Proceedings of the AAAI Conference on Artificial Intelligence, 38(14), 15546-15554. https://doi.org/10.1609/aaai.v38i14.29481

Issue

Section

AAAI Technical Track on Machine Learning V