MotionGPT: Finetuned LLMs Are General-Purpose Motion Generators

Authors

  • Yaqi Zhang University of Science and Technology of China CAS Key Laboratory of Electromagnetic Space Information
  • Di Huang The University of Sydney
  • Bin Liu University of Science and Technology of China CAS Key Laboratory of Electromagnetic Space Information
  • Shixiang Tang The University of Sydney
  • Yan Lu The University of Sydney
  • Lu Chen Shanghai AI Laboratory
  • Lei Bai Shanghai AI Laboratory
  • Qi Chu University of Science and Technology of China CAS Key Laboratory of Electromagnetic Space Information
  • Nenghai Yu University of Science and Technology of China CAS Key Laboratory of Electromagnetic Space Information
  • Wanli Ouyang Shanghai AI Laboratory

DOI:

https://doi.org/10.1609/aaai.v38i7.28567

Keywords:

CV: Motion & Tracking, CV: 3D Computer Vision, CV: Biometrics, Face, Gesture & Pose, CV: Multi-modal Vision

Abstract

Generating realistic human motion from given action descriptions has experienced significant advancements because of the emerging requirement of digital humans. While recent works have achieved impressive results in generating motion directly from textual action descriptions, they often support only a single modality of the control signal, which limits their application in the real digital human industry. This paper presents a Motion General-Purpose generaTor (MotionGPT) that can use multimodal control signals, e.g., text and single-frame poses, for generating consecutive human motions by treating multimodal signals as special input tokens in large language models (LLMs). Specifically, we first quantize multimodal control signals into discrete codes and then formulate them in a unified prompt instruction to ask the LLMs to generate the motion answer. Our MotionGPT demonstrates a unified human motion generation model with multimodal control signals by tuning a mere 0.4% of LLM parameters. To the best of our knowledge, MotionGPT is the first method to generate human motion by multimodal control signals, which we hope can shed light on this new direction. Visit our webpage at https://qiqiapink.github.io/MotionGPT/.

Published

2024-03-24

How to Cite

Zhang, Y., Huang, D., Liu, B., Tang, S., Lu, Y., Chen, L., Bai, L., Chu, Q., Yu, N., & Ouyang, W. (2024). MotionGPT: Finetuned LLMs Are General-Purpose Motion Generators. Proceedings of the AAAI Conference on Artificial Intelligence, 38(7), 7368-7376. https://doi.org/10.1609/aaai.v38i7.28567

Issue

Section

AAAI Technical Track on Computer Vision VI