Style-Guided and Disentangled Representation for Robust Image-to-Image Translation

Authors

  • Jaewoong Choi Inha University, Republic of Korea
  • Daeha Kim Inha University, Republic of Korea
  • Byung Cheol Song Inha University, Republic of Korea

DOI:

https://doi.org/10.1609/aaai.v36i1.19924

Keywords:

Computer Vision (CV), Machine Learning (ML), Humans And AI (HAI)

Abstract

Recently, various image-to-image translation (I2I) methods have improved mode diversity and visual quality in terms of neural networks or regularization terms. However, conventional I2I methods relies on a static decision boundary and the encoded representations in those methods are entangled with each other, so they often face with ‘mode collapse’ phenomenon. To mitigate mode collapse, 1) we design a so-called style-guided discriminator that guides an input image to the target image style based on the strategy of flexible decision boundary. 2) Also, we make the encoded representations include independent domain attributes. Based on two ideas, this paper proposes Style-Guided and Disentangled Representation for Robust Image-to-Image Translation (SRIT). SRIT showed outstanding FID by 8%, 22.8%, and 10.1% for CelebA-HQ, AFHQ, and Yosemite datasets, respectively. The translated images of SRIT reflect the styles of target domain successfully. This indicates that SRIT shows better mode diversity than previous works.

Downloads

Published

2022-06-28

How to Cite

Choi, J., Kim, D., & Song, B. C. (2022). Style-Guided and Disentangled Representation for Robust Image-to-Image Translation. Proceedings of the AAAI Conference on Artificial Intelligence, 36(1), 463-471. https://doi.org/10.1609/aaai.v36i1.19924

Issue

Section

AAAI Technical Track on Computer Vision I