Latent Space Editing in Transformer-Based Flow Matching

Authors

  • Vincent Tao Hu University of Amsterdam
  • Wei Zhang University of Amsterdam
  • Meng Tang University of California Merced
  • Pascal Mettes University of Amsterdam
  • Deli Zhao Alibaba Group
  • Cees Snoek University of Amsterdam

DOI:

https://doi.org/10.1609/aaai.v38i3.27998

Keywords:

CV: Computational Photography, Image & Video Synthesis, CV: Representation Learning for Vision, CV: Language and Vision, CV: Applications

Abstract

This paper strives for image editing via generative models. Flow Matching is an emerging generative modeling technique that offers the advantage of simple and efficient training. Simultaneously, a new transformer-based U-ViT has recently been proposed to replace the commonly used UNet for better scalability and performance in generative modeling. Hence, Flow Matching with a transformer backbone offers the potential for scalable and high-quality generative modeling, but their latent structure and editing ability are as of yet unknown. Hence, we adopt this setting and explore how to edit images through latent space manipulation. We introduce an editing space, which we call u-space, that can be manipulated in a controllable, accumulative, and composable manner. Additionally, we propose a tailored sampling solution to enable sampling with the more efficient adaptive step-size ODE solvers. Lastly, we put forth a straightforward yet powerful method for achieving fine-grained and nuanced editing using text prompts. Our framework is simple and efficient, all while being highly effective at editing images while preserving the essence of the original content. Our code will be publicly available at https://taohu.me/lfm/

Downloads

Published

2024-03-24

How to Cite

Hu, V. T., Zhang, W., Tang, M., Mettes, P., Zhao, D., & Snoek, C. . (2024). Latent Space Editing in Transformer-Based Flow Matching. Proceedings of the AAAI Conference on Artificial Intelligence, 38(3), 2247-2255. https://doi.org/10.1609/aaai.v38i3.27998

Issue

Section

AAAI Technical Track on Computer Vision II