Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Scaling up masked audio encoder learning for general audio classification

About

Despite progress in audio classification, a generalization gap remains between speech and other sound domains, such as environmental sounds and music. Models trained for speech tasks often fail to perform well on environmental or musical audio tasks, and vice versa. While self-supervised (SSL) audio representations offer an alternative, there has been limited exploration of scaling both model and dataset sizes for SSL-based general audio classification. We introduce Dasheng, a simple SSL audio encoder, based on the efficient masked autoencoder framework. Trained with 1.2 billion parameters on 272,356 hours of diverse audio, Dasheng obtains significant performance gains on the HEAR benchmark. It outperforms previous works on CREMA-D, LibriCount, Speech Commands, VoxLingua, and competes well in music and environment classification. Dasheng features inherently contain rich speech, music, and environmental information, as shown in nearest-neighbor classification experiments. Code is available https://github.com/richermans/dasheng/.

Heinrich Dinkel, Zhiyong Yan, Yongqing Wang, Junbo Zhang, Yujun Wang, Bin Wang• 2024

Related benchmarks

TaskDatasetResultRank
Fault DiagnosisRMIS Fault Diagnosis Suite (IICA, IIEE, WTPG, MaFaulDa, SDUST, UMGED, PU)
Overall Mean Score53.66
28
Vocal Sound ClassificationVocalSound
Accuracy92.5
21
Bioacoustic AnalysisBeans
wtkn77.3
20
Music Genre ClassificationGTZAN
Accuracy88.6
19
Speech Emotion RecognitionRAVDESS--
19
Fake DetectionASVspoof5 (dev)
EER1.625
16
Fault DiagnosisWTPG
Area under Multi-Split Curve86.04
14
Fault DiagnosisUMGED
Vib Score6.59
14
Fault DiagnosisPU
Vibration Score72.63
14
Fault DiagnosisIICA
Area under Multi-Split Curve70.68
14
Showing 10 of 29 rows

Other info

Follow for update