SDGAN: Disentangling Semantic Manipulation for Facial Attribute Editing

Authors

  • Wenmin Huang Sun Yat-sen University
  • Weiqi Luo Sun Yat-sen University
  • Jiwu Huang Shenzhen University
  • Xiaochun Cao Sun Yat-sen University

DOI:

https://doi.org/10.1609/aaai.v38i3.28012

Keywords:

CV: Computational Photography, Image & Video Synthesis

Abstract

Facial attribute editing has garnered significant attention, yet prevailing methods struggle with achieving precise attribute manipulation while preserving irrelevant details and controlling attribute styles. This challenge primarily arises from the strong correlations between different attributes and the interplay between attributes and identity. In this paper, we propose Semantic Disentangled GAN (SDGAN), a novel method addressing this challenge. SDGAN introduces two key concepts: a semantic disentanglement generator that assigns facial representations to distinct attribute-specific editing modules, enabling the decoupling of the facial attribute editing process, and a semantic mask alignment strategy that confines attribute editing to appropriate regions, thereby avoiding undesired modifications. Leveraging these concepts, SDGAN demonstrates accurate attribute editing and achieves high-quality attribute style manipulation through both latent-guided and reference-guided manners. We extensively evaluate our method on the CelebA-HQ database, providing both qualitative and quantitative analyses. Our results establish that SDGAN significantly outperforms state-of-the-art techniques, showcasing the effectiveness of our approach. To foster reproducibility and further research, we will provide the code for our method.

Published

2024-03-24

How to Cite

Huang, W., Luo, W., Huang, J., & Cao, X. (2024). SDGAN: Disentangling Semantic Manipulation for Facial Attribute Editing. Proceedings of the AAAI Conference on Artificial Intelligence, 38(3), 2374-2381. https://doi.org/10.1609/aaai.v38i3.28012

Issue

Section

AAAI Technical Track on Computer Vision II