Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/180277
Title: Large motion model for unified multi-modal motion generation
Authors: Zhang, Mingyuan
Jin, Daisheng
Gu, Chenyang
Hong, Fangzhou
Cai, Zhongang
Huang, Jingfang
Zhang, Chongzhi
Guo, Xinying
Yang, Lei
He, Ying
Liu, Ziwei
Keywords: Computer and Information Science
Issue Date: 2024
Source: Zhang, M., Jin, D., Gu, C., Hong, F., Cai, Z., Huang, J., Zhang, C., Guo, X., Yang, L., He, Y. & Liu, Z. (2024). Large motion model for unified multi-modal motion generation. 2024 European Conference on Computer Vision (ECCV). https://dx.doi.org/10.48550/arXiv.2404.01284
Project: MOET2EP20221-0012
NTU NAP
IAF-ICP
Conference: 2024 European Conference on Computer Vision (ECCV)
Abstract: Human motion generation, a cornerstone technique in animation and video production, has widespread applications in various tasks like text-to-motion and music-to-dance. Previous works focus on developing specialist models tailored for each task without scalability. In this work, we present Large Motion Model (LMM), a motion-centric, multi-modal framework that unifies mainstream motion generation tasks into a generalist model. A unified motion model is appealing since it can leverage a wide range of motion data to achieve broad generalization beyond a single task. However, it is also challenging due to the heterogeneous nature of substantially different motion data and tasks. LMM tackles these challenges from three principled aspects: 1) Data: We consolidate datasets with different modalities, formats and tasks into a comprehensive yet unified motion generation dataset, MotionVerse, comprising 10 tasks, 16 datasets, a total of 320k sequences, and 100 million frames. 2) Architecture: We design an articulated attention mechanism ArtAttention that incorporates body part-aware modeling into Diffusion Transformer backbone. 3) Pre-Training: We propose a novel pre-training strategy for LMM, which employs variable frame rates and masking forms, to better exploit knowledge from diverse training data. Extensive experiments demonstrate that our generalist LMM achieves competitive performance across various standard motion generation tasks over state-of-the-art specialist models. Notably, LMM exhibits strong generalization capabilities and emerging properties across many unseen tasks. Additionally, our ablation studies reveal valuable insights about training and scaling up large motion models for future research.
URI: https://hdl.handle.net/10356/180277
URL: http://arxiv.org/abs/2404.01284v1
DOI: 10.48550/arXiv.2404.01284
Schools: College of Computing and Data Science 
Research Centres: S-Lab
Rights: © 2024 ECCV. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder.
Fulltext Permission: open
Fulltext Availability: With Fulltext
Appears in Collections:CCDS Conference Papers

Files in This Item:
File Description SizeFormat 
Supplementary Material.pdf3.3 MBAdobe PDFView/Open
02125.pdfPreprint13.24 MBAdobe PDFView/Open

Page view(s)

48
Updated on Dec 4, 2024

Download(s)

16
Updated on Dec 4, 2024

Google ScholarTM

Check

Altmetric


Plumx

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.