Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Auto-ACD: A Large-scale Dataset for Audio-Language Representation Learning

About

Recently, the AI community has made significant strides in developing powerful foundation models, driven by large-scale multimodal datasets. However, for audio representation learning, existing datasets suffer from limitations in the following aspects: insufficient volume, simplistic content, and arduous collection procedures. To establish an audio dataset with high-quality captions, we propose an innovative, automatic approach leveraging multimodal inputs, such as video frames, audio streams. Specifically, we construct a large-scale, high-quality, audio-language dataset, named as Auto-ACD, comprising over 1.5M audio-text pairs. We exploit a series of pre-trained models or APIs, to determine audio-visual synchronisation, generate image captions, object detection, or audio tags for specific videos. Subsequently, we employ LLM to paraphrase a congruent caption for each audio, guided by the extracted multi-modality clues. To demonstrate the effectiveness of the proposed dataset, we train widely used models on our dataset and show performance improvement on various downstream tasks, for example, audio-language retrieval, audio captioning, zero-shot classification. In addition, we establish a novel benchmark with environmental information and provide a benchmark for audio-text tasks.

Luoyi Sun, Xuenan Xu, Mengyue Wu, Weidi Xie• 2023

Related benchmarks

TaskDatasetResultRank
Audio CaptioningAudioSet
LA-CLAP0.396
4
Audio CaptioningVGGSound
LA-CLAP0.409
3
Audio CaptioningAudioCaps
MWR-S (MLLM)0.39
3
Showing 3 of 3 rows

Other info

Follow for update