SDGAN: Disentangling Semantic Manipulation for Facial Attribute Editing
DOI:
https://doi.org/10.1609/aaai.v38i3.28012Keywords:
CV: Computational Photography, Image & Video SynthesisAbstract
Facial attribute editing has garnered significant attention, yet prevailing methods struggle with achieving precise attribute manipulation while preserving irrelevant details and controlling attribute styles. This challenge primarily arises from the strong correlations between different attributes and the interplay between attributes and identity. In this paper, we propose Semantic Disentangled GAN (SDGAN), a novel method addressing this challenge. SDGAN introduces two key concepts: a semantic disentanglement generator that assigns facial representations to distinct attribute-specific editing modules, enabling the decoupling of the facial attribute editing process, and a semantic mask alignment strategy that confines attribute editing to appropriate regions, thereby avoiding undesired modifications. Leveraging these concepts, SDGAN demonstrates accurate attribute editing and achieves high-quality attribute style manipulation through both latent-guided and reference-guided manners. We extensively evaluate our method on the CelebA-HQ database, providing both qualitative and quantitative analyses. Our results establish that SDGAN significantly outperforms state-of-the-art techniques, showcasing the effectiveness of our approach. To foster reproducibility and further research, we will provide the code for our method.Downloads
Published
2024-03-24
How to Cite
Huang, W., Luo, W., Huang, J., & Cao, X. (2024). SDGAN: Disentangling Semantic Manipulation for Facial Attribute Editing. Proceedings of the AAAI Conference on Artificial Intelligence, 38(3), 2374-2381. https://doi.org/10.1609/aaai.v38i3.28012
Issue
Section
AAAI Technical Track on Computer Vision II