CA-Edit: Causality-Aware Condition Adapter for High-Fidelity Local Facial Attribute Editing
DOI:
https://doi.org/10.1609/aaai.v39i8.32928Abstract
For efficient and high-fidelity local facial attribute editing, most existing editing methods either require additional fine-tuning for different editing effects or tend to affect beyond the editing regions. Alternatively, inpainting methods can edit the target image region while preserving external areas. However, current inpainting methods still suffer from the generation misalignment with facial attributes description and the loss of facial skin details. To address these challenges, (i) a novel data utilization strategy is introduced to construct datasets consisting of attribute-text-image triples from a data-driven perspective, (ii) a Causality-Aware Condition Adapter is proposed to enhance the contextual causality modeling of specific details, which encodes the skin details from the original image while preventing conflicts between these cues and textual conditions. In addition, a Skin Transition Frequency Guidance technique is introduced for the local modeling of contextual causality via sampling guidance driven by low-frequency alignment. Extensive quantitative and qualitative experiments demonstrate the effectiveness of our method in boosting both fidelity and editability for localized attribute editing. Our codes will be made publicly available.Downloads
Published
2025-04-11
How to Cite
Xian, X., He, X., Niu, Z., Zhang, J., Xie, W., Song, S., … Shen, L. (2025). CA-Edit: Causality-Aware Condition Adapter for High-Fidelity Local Facial Attribute Editing. Proceedings of the AAAI Conference on Artificial Intelligence, 39(8), 8593–8601. https://doi.org/10.1609/aaai.v39i8.32928
Issue
Section
AAAI Technical Track on Computer Vision VII