Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/184643
Title: | Music generation using deep learning | Authors: | Harish Vasanth | Keywords: | Computer and Information Science | Issue Date: | 2025 | Publisher: | Nanyang Technological University | Source: | Harish Vasanth (2025). Music generation using deep learning. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/184643 | Abstract: | This research studies the generation and the results of a Text-to-Midi Transformer model, focusing on converting textual descriptions to functional and melodious sounds in MIDI format. A novel dataset on combinational music is used to enable the model to learn richer representations of music. This study uses an advanced music information framework to label music files, to improve accuracy on music captioning. An advanced text LLM is employed to process music captions, to improve understanding of their nuanced meanings. A user study is used to evaluate the hypothesis that a Text-to-Midi deep learning model can be designed to generate short, emotion-driven thematic music from textual prompts. The results verify that the model can generate music that is melodious, and able to capture the nuances of the textual prompt. A model that can produce melodious sounds better than previous implementations of other students while being able to capture emotional context from textual prompts is established. | URI: | https://hdl.handle.net/10356/184643 | Schools: | College of Computing and Data Science | Fulltext Permission: | restricted | Fulltext Availability: | With Fulltext |
Appears in Collections: | CCDS Student Reports (FYP/IA/PA/PI) |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
FYP_REPORT_HARISH_VASANTH.pdf Restricted Access | 2.35 MB | Adobe PDF | View/Open |
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.