Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

VidMuse: A Simple Video-to-Music Generation Framework with Long-Short-Term Modeling

About

In this work, we systematically study music generation conditioned solely on the video. First, we present a large-scale dataset comprising 360K video-music pairs, including various genres such as movie trailers, advertisements, and documentaries. Furthermore, we propose VidMuse, a simple framework for generating music aligned with video inputs. VidMuse stands out by producing high-fidelity music that is both acoustically and semantically aligned with the video. By incorporating local and global visual cues, VidMuse enables the creation of musically coherent audio tracks that consistently match the video content through Long-Short-Term modeling. Through extensive experiments, VidMuse outperforms existing models in terms of audio quality, diversity, and audio-visual alignment. The code and datasets are available at https://vidmuse.github.io/.

Zeyue Tian, Zhaoyang Liu, Ruibin Yuan, Jiahao Pan, Qifeng Liu, Xu Tan, Qifeng Chen, Wei Xue, Yike Guo• 2024

Related benchmarks

TaskDatasetResultRank
Video-to-Music GenerationV2M-bench (test)
Fréchet Audio Distance (FAD)2.459
12
Video-to-Music GenerationAIST++
BCS0.9997
10
Video-to-Music GenerationMovieGenBench Music
FAD2.98
7
Video-to-Music GenerationOES-Pub
FAD*10.4
7
Video-to-Music GenerationVideo-to-Music Generation Evaluation Dataset (test)
FAD2.459
6
Video-to-Music GenerationShort- and Mid-length Videos
EDC3.09
5
Video-to-Music GenerationLong-form Videos Subjective Study
EDC2.29
5
Video-to-Music GenerationLORIS
BCS96.3
4
Video-to-MusicV2M Evaluation Set
OVL73.5
3
Video-to-Music GenerationTikTok
BCS79.8
3
Showing 10 of 10 rows

Other info

Code

Follow for update