Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/165971
Title: Animate your avatar: learning conditional human motion prior
Authors: Singh, Ananya
Keywords: Engineering::Computer science and engineering
Issue Date: 2023
Publisher: Nanyang Technological University
Source: Singh, A. (2023). Animate your avatar: learning conditional human motion prior. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/165971
Project: SCSE22-0197 
Abstract: This research project focuses on the concept of virtual humans and aims to enable natural language control of 3D avatars, allowing them to perform human-like movements that are coherent with their surrounding environment. To achieve this goal, the project proposes to learn a "conditional" human motion prior that takes into account scene information and/or language descriptions. This approach is intended to generate more coherent and meaningful human-like motions that are better suited to specific scenarios and objectives. The potential applications of this research include gaming, virtual avatars, and the Metaverse. The project employs a two-stage approach that combines language-conditioned human motion generation and physics-based character control techniques to generate diverse and physically plausible human motions in a physical world from language descriptions. The results of the project demonstrate the effectiveness of this approach in terms of motion diversity, faithfulness to the language description, and physical plausibility.
URI: https://hdl.handle.net/10356/165971
Schools: School of Computer Science and Engineering 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:SCSE Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
FYP Final Report_Singh Ananya.pdf
  Restricted Access
1.61 MBAdobe PDFView/Open

Page view(s)

232
Updated on May 7, 2025

Download(s)

20
Updated on May 7, 2025

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.