Patch-Wise Graph Contrastive Learning for Image Translation

Authors

  • Chanyong Jung Department of Brain and Bio Engineering, KAIST, Daejeon, Republic of Korea
  • Gihyun Kwon Department of Brain and Bio Engineering, KAIST, Daejeon, Republic of Korea
  • Jong Chul Ye Department of Brain and Bio Engineering, KAIST, Daejeon, Republic of Korea Kim Jaechul Graduate School of AI, KAIST, Daejeon, Republic of Korea

DOI:

https://doi.org/10.1609/aaai.v38i12.29199

Keywords:

ML: Deep Generative Models & Autoencoders, CV: Representation Learning for Vision, ML: Representation Learning, ML: Graph-based Machine Learning

Abstract

Recently, patch-wise contrastive learning is drawing attention for the image translation by exploring the semantic correspondence between the input image and the output image. To further explore the patch-wise topology for high-level semantic understanding, here we exploit the graph neural network to capture the topology-aware features. Specifically, we construct the graph based on the patch-wise similarity from a pretrained encoder, whose adjacency matrix is shared to enhance the consistency of patch-wise relation between the input and the output. Then, we obtain the node feature from the graph neural network, and enhance the correspondence between the nodes by increasing mutual information using the contrastive loss. In order to capture the hierarchical semantic structure, we further propose the graph pooling. Experimental results demonstrate the state-of-art results for the image translation thanks to the semantic encoding by the constructed graphs.

Published

2024-03-24

How to Cite

Jung, C., Kwon, G., & Ye, J. C. (2024). Patch-Wise Graph Contrastive Learning for Image Translation. Proceedings of the AAAI Conference on Artificial Intelligence, 38(12), 13013-13021. https://doi.org/10.1609/aaai.v38i12.29199

Issue

Section

AAAI Technical Track on Machine Learning III