Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Music Understanding LLaMA: Advancing Text-to-Music Generation with Question Answering and Captioning

About

Text-to-music generation (T2M-Gen) faces a major obstacle due to the scarcity of large-scale publicly available music datasets with natural language captions. To address this, we propose the Music Understanding LLaMA (MU-LLaMA), capable of answering music-related questions and generating captions for music files. Our model utilizes audio representations from a pretrained MERT model to extract music features. However, obtaining a suitable dataset for training the MU-LLaMA model remains challenging, as existing publicly accessible audio question answering datasets lack the necessary depth for open-ended music question answering. To fill this gap, we present a methodology for generating question-answer pairs from existing audio captioning datasets and introduce the MusicQA Dataset designed for answering open-ended music-related questions. The experiments demonstrate that the proposed MU-LLaMA model, trained on our designed MusicQA dataset, achieves outstanding performance in both music question answering and music caption generation across various metrics, outperforming current state-of-the-art (SOTA) models in both fields and offering a promising advancement in the T2M-Gen research field.

Shansong Liu, Atin Sakkeer Hussain, Chenshuo Sun, Ying Shan• 2023

Related benchmarks

TaskDatasetResultRank
Audio UnderstandingMMAU v05.15.25 (test)
Sound Score31
53
Multimodal Audio UnderstandingMMAU mini v05.15.25 (test)
Sound Accuracy33
25
Multimodal Audio ReasoningMMAR
Mean Score13.9
22
Music CaptioningMTG-eval-Cap 1.0 (test)
BLEU Score0.278
5
Multi-turn audio dialogueAF-Dialogue-MusicCaps (test)
CIDEr0.585
4
Music Question AnsweringMTG-eval-QA (test)
B-U0.306
3
Showing 6 of 6 rows

Other info

Follow for update