ShiftDDPMs: Exploring Conditional Diffusion Models by Shifting Diffusion Trajectories

Authors

  • Zijian Zhang Zhejiang University
  • Zhou Zhao Zhejiang University
  • Jun Yu Hangzhou Dianzi University
  • Qi Tian Huawei Cloud & AI

DOI:

https://doi.org/10.1609/aaai.v37i3.25465

Keywords:

CV: Computational Photography, Image & Video Synthesis, CV: Representation Learning for Vision

Abstract

Diffusion models have recently exhibited remarkable abilities to synthesize striking image samples since the introduction of denoising diffusion probabilistic models (DDPMs). Their key idea is to disrupt images into noise through a fixed forward process and learn its reverse process to generate samples from noise in a denoising way. For conditional DDPMs, most existing practices relate conditions only to the reverse process and fit it to the reversal of unconditional forward process. We find this will limit the condition modeling and generation in a small time window. In this paper, we propose a novel and flexible conditional diffusion model by introducing conditions into the forward process. We utilize extra latent space to allocate an exclusive diffusion trajectory for each condition based on some shifting rules, which will disperse condition modeling to all timesteps and improve the learning capacity of model. We formulate our method, which we call ShiftDDPMs, and provide a unified point of view on existing related methods. Extensive qualitative and quantitative experiments on image synthesis demonstrate the feasibility and effectiveness of ShiftDDPMs.

Downloads

Published

2023-06-26

How to Cite

Zhang, Z., Zhao, Z., Yu, J., & Tian, Q. (2023). ShiftDDPMs: Exploring Conditional Diffusion Models by Shifting Diffusion Trajectories. Proceedings of the AAAI Conference on Artificial Intelligence, 37(3), 3552-3560. https://doi.org/10.1609/aaai.v37i3.25465

Issue

Section

AAAI Technical Track on Computer Vision III