VersaGen: Unleashing Versatile Visual Control for Text-to-Image Synthesis

Authors

  • Zhipeng Chen School of Artificial Intelligence, Beijing University of Posts and Telecommunications, China
  • Lan Yang School of Artificial Intelligence, Beijing University of Posts and Telecommunications, China SketchX, CVSSP, University of Surrey, United Kingdom
  • Yonggang Qi School of Artificial Intelligence, Beijing University of Posts and Telecommunications, China SketchX, CVSSP, University of Surrey, United Kingdom
  • Honggang Zhang School of Artificial Intelligence, Beijing University of Posts and Telecommunications, China
  • Kaiyue Pang SketchX, CVSSP, University of Surrey, United Kingdom
  • Ke Li School of Artificial Intelligence, Beijing University of Posts and Telecommunications, China SketchX, CVSSP, University of Surrey, United Kingdom
  • Yi-Zhe Song SketchX, CVSSP, University of Surrey, United Kingdom

DOI:

https://doi.org/10.1609/aaai.v39i3.32240

Abstract

Despite the rapid advancements in text-to-image (T2I) synthesis, enabling precise visual control remains a significant challenge. Existing works attempted to incorporate multi-facet controls (text and sketch), aiming to enhance the creative control over generated images. However, our pilot study reveals that the expressive power of humans far surpasses the capabilities of current methods. Users desire a more versatile approach that can accommodate their diverse creative intents, ranging from controlling individual subjects to manipulating the entire scene composition. We present VersaGen, a generative AI agent that enables versatile visual control in T2I synthesis. VersaGen admits four types of visual controls: i) single visual subject; ii) multiple visual subjects; iii) scene background; iv) any combination of the three above or merely no control at all. We train an adaptor upon a frozen T2I model to accommodate the visual information into the text-dominated diffusion process. We introduce three optimization strategies during the inference phase of VersaGen to improve generation results and enhance user experience. Comprehensive experiments on COCO and Sketchy validate the effectiveness and flexibility of VersaGen, as evidenced by both qualitative and quantitative results.

Published

2025-04-11

How to Cite

Chen, Z., Yang, L., Qi, Y., Zhang, H., Pang, K., Li, K., & Song, Y.-Z. (2025). VersaGen: Unleashing Versatile Visual Control for Text-to-Image Synthesis. Proceedings of the AAAI Conference on Artificial Intelligence, 39(3), 2394–2402. https://doi.org/10.1609/aaai.v39i3.32240

Issue

Section

AAAI Technical Track on Computer Vision II