Text Diffusion with Reinforced Conditioning

Authors

  • Yuxuan Liu Peking University
  • Tianchi Yang Microsoft Corporation
  • Shaohan Huang Microsoft Corporation
  • Zihan Zhang Microsoft Corporation
  • Haizhen Huang Microsoft Corporation
  • Furu Wei Microsoft Corporation
  • Weiwei Deng Microsoft Corporation
  • Feng Sun Microsoft Corporation
  • Qi Zhang Microsoft Corporation

DOI:

https://doi.org/10.1609/aaai.v38i12.29316

Keywords:

ML: Deep Generative Models & Autoencoders, NLP: Generation, NLP: (Large) Language Models, ML: Reinforcement Learning

Abstract

Diffusion models have demonstrated exceptional capability in generating high-quality images, videos, and audio. Due to their adaptiveness in iterative refinement, they provide a strong potential for achieving better non-autoregressive sequence generation. However, existing text diffusion models still fall short in their performance due to a challenge in handling the discreteness of language. This paper thoroughly analyzes text diffusion models and uncovers two significant limitations: degradation of self-conditioning during training and misalignment between training and sampling. Motivated by our findings, we propose a novel Text Diffusion model called TReC, which mitigates the degradation with Reinforced Conditioning and the misalignment by Time-Aware Variance Scaling. Our extensive experiments demonstrate the competitiveness of TReC against autoregressive, non-autoregressive, and diffusion baselines. Moreover, qualitative analysis shows its advanced ability to fully utilize the diffusion process in refining samples.

Published

2024-03-24

How to Cite

Liu, Y., Yang, T., Huang, S., Zhang, Z., Huang, H., Wei, F., Deng, W., Sun, F., & Zhang, Q. (2024). Text Diffusion with Reinforced Conditioning. Proceedings of the AAAI Conference on Artificial Intelligence, 38(12), 14069-14077. https://doi.org/10.1609/aaai.v38i12.29316

Issue

Section

AAAI Technical Track on Machine Learning III