Privacy-Preserving Low-Rank Adaptation Against Membership Inference Attacks for Latent Diffusion Models

Authors

  • Zihao Luo University of Auckland
  • Xilie Xu National University of Singapore
  • Feng Liu University of Melbourne
  • Yun Sing Koh University of Auckland
  • Di Wang King Abdullah University of Science and Technology
  • Jingfeng Zhang University of Auckland, King Abdullah University of Science and Technology

DOI:

https://doi.org/10.1609/aaai.v39i6.32628

Abstract

Low-rank adaptation (LoRA) is an efficient strategy for adapting latent diffusion models (LDMs) on a private dataset to generate specific images by minimizing the adaptation loss. However, the LoRA-adapted LDMs are vulnerable to membership inference (MI) attacks that can judge whether a particular data point belongs to the private dataset, thus leading to the privacy leakage. To defend against MI attacks, we first propose a straightforward solution: Membership-Privacy-preserving LoRA (MP-LoRA). MP-LoRA is formulated as a min-max optimization problem where a proxy attack model is trained by maximizing its MI gain while the LDM is adapted by minimizing the sum of the adaptation loss and the MI gain of the proxy attack model. However, we empirically find that MP-LoRA has the issue of unstable optimization, and theoretically analyze that the potential reason is the unconstrained local smoothness, which impedes the privacy-preserving adaptation. To mitigate this issue, we further propose a Stable Membership-Privacy-preserving LoRA (SMP-LoRA) that adapts the LDM by minimizing the ratio of the adaptation loss to the MI gain. Besides, we theoretically prove that the local smoothness of SMP-LoRA can be constrained by the gradient norm, leading to improved convergence. Our experimental results corroborate that SMP-LoRA can indeed defend against MI attacks and generate high-quality images.

Downloads

Published

2025-04-11

How to Cite

Luo, Z., Xu, X., Liu, F., Koh, Y. S., Wang, D., & Zhang, J. (2025). Privacy-Preserving Low-Rank Adaptation Against Membership Inference Attacks for Latent Diffusion Models. Proceedings of the AAAI Conference on Artificial Intelligence, 39(6), 5883–5891. https://doi.org/10.1609/aaai.v39i6.32628

Issue

Section

AAAI Technical Track on Computer Vision V