HuTuMotion: Human-Tuned Navigation of Latent Motion Diffusion Models with Minimal Feedback

Authors

  • Gaoge Han Northwest A&F University
  • Shaoli Huang Tencent AI-Lab
  • Mingming Gong The University of Melbourne Mohamed bin Zayed University of Artificial Intelligence
  • Jinglei Tang Northwest A&F University

DOI:

https://doi.org/10.1609/aaai.v38i3.27974

Keywords:

CV: Multi-modal Vision, CV: Representation Learning for Vision

Abstract

We introduce HuTuMotion, an innovative approach for generating natural human motions that navigates latent motion diffusion models by leveraging few-shot human feedback. Unlike existing approaches that sample latent variables from a standard normal prior distribution, our method adapts the prior distribution to better suit the characteristics of the data, as indicated by human feedback, thus enhancing the quality of motion generation. Furthermore, our findings reveal that utilizing few-shot feedback can yield performance levels on par with those attained through extensive human feedback. This discovery emphasizes the potential and efficiency of incorporating few-shot human-guided optimization within latent diffusion models for personalized and style-aware human motion generation applications. The experimental results show the significantly superior performance of our method over existing state-of-the-art approaches.

Published

2024-03-24

How to Cite

Han, G., Huang, S., Gong, M., & Tang, J. (2024). HuTuMotion: Human-Tuned Navigation of Latent Motion Diffusion Models with Minimal Feedback. Proceedings of the AAAI Conference on Artificial Intelligence, 38(3), 2031–2039. https://doi.org/10.1609/aaai.v38i3.27974

Issue

Section

AAAI Technical Track on Computer Vision II