Sparse Transfer Learning Accelerates and Enhances Certified Robustness: A Comprehensive Study

Authors

  • Zhangheng Li University of Texas at Austin
  • Tianlong Chen University of Texas at Austin
  • Linyi Li University of Illinois, Urbana Champaign
  • Bo Li University of Illinois, Urbana Champaign University of Chicago
  • Zhangyang Wang University of Texas at Austin

DOI:

https://doi.org/10.1609/aaai.v39i5.32539

Abstract

Certified robustness is a critical measure for assessing the reliability of machine learning systems. Traditionally, the computational burden associated with certifying the robustness of machine learning models has posed a substantial challenge, particularly with the continuous expansion of model sizes. In this paper, we introduce an innovative approach to expedite the verification process for L2-norm certified robustness through sparse transfer learning. Our approach is both efficient and effective. It leverages verification results obtained from pre-training tasks and applies sparse updates to these results. To enhance performance, we incorporate dynamic sparse mask selection and introduce a novel stability-based regularizer called DiffStab. Empirical results demonstrate that our method accelerates the verification process for downstream tasks by as much as 70-80%, with only slight reductions in certified accuracy compared to dense parameter updates. We further validate that this performance improvement is even more pronounced in the few-shot transfer learning scenario.

Downloads

Published

2025-04-11

How to Cite

Li, Z., Chen, T., Li, L., Li, B., & Wang, Z. (2025). Sparse Transfer Learning Accelerates and Enhances Certified Robustness: A Comprehensive Study. Proceedings of the AAAI Conference on Artificial Intelligence, 39(5), 5084–5091. https://doi.org/10.1609/aaai.v39i5.32539

Issue

Section

AAAI Technical Track on Computer Vision IV