It's Time for Artistic Correspondence in Music and Video
About
We present an approach for recommending a music track for a given video, and vice versa, based on both their temporal alignment and their correspondence at an artistic level. We propose a self-supervised approach that learns this correspondence directly from data, without any need of human annotations. In order to capture the high-level concepts that are required to solve the task, we propose modeling the long-term temporal context of both the video and the music signals, using Transformer networks for each modality. Experiments show that this approach strongly outperforms alternatives that do not exploit the temporal context. The combination of our contributions improve retrieval accuracy up to 10x over prior state of the art. This strong improvement allows us to introduce a wide range of analyses and applications. For instance, we can condition music retrieval based on visually defined attributes.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Segment-level Music-to-Video Retrieval | MusicVid-YT8M (test) | Median Rank6 | 10 | |
| Segment-level Video-to-Music Retrieval | MusicVid-YT8M (test) | Median Rank7 | 10 | |
| Music-to-Video Retrieval | MusicVid-YT8M track-level (test) | Median Rank13 | 7 | |
| Video-to-Music Retrieval | MusicVid-YT8M track-level (test) | Median Rank13 | 7 | |
| Music Retrieval | YouTube8M MusicVideo (test) | Median Rank5 | 6 | |
| Segment-level Music-to-Video Retrieval | MovieClips (test) | Median Rank21 | 5 | |
| Segment-level Video-to-Music Retrieval | MovieClips (test) | Median Rank21 | 5 |