Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/184643
Title: Music generation using deep learning
Authors: Harish Vasanth
Keywords: Computer and Information Science
Issue Date: 2025
Publisher: Nanyang Technological University
Source: Harish Vasanth (2025). Music generation using deep learning. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/184643
Abstract: This research studies the generation and the results of a Text-to-Midi Transformer model, focusing on converting textual descriptions to functional and melodious sounds in MIDI format. A novel dataset on combinational music is used to enable the model to learn richer representations of music. This study uses an advanced music information framework to label music files, to improve accuracy on music captioning. An advanced text LLM is employed to process music captions, to improve understanding of their nuanced meanings. A user study is used to evaluate the hypothesis that a Text-to-Midi deep learning model can be designed to generate short, emotion-driven thematic music from textual prompts. The results verify that the model can generate music that is melodious, and able to capture the nuances of the textual prompt. A model that can produce melodious sounds better than previous implementations of other students while being able to capture emotional context from textual prompts is established.
URI: https://hdl.handle.net/10356/184643
Schools: College of Computing and Data Science 
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:CCDS Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
FYP_REPORT_HARISH_VASANTH.pdf
  Restricted Access
2.35 MBAdobe PDFView/Open

Page view(s)

8
Updated on May 7, 2025

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.