Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Listenable Maps for Zero-Shot Audio Classifiers

About

Interpreting the decisions of deep learning models, including audio classifiers, is crucial for ensuring the transparency and trustworthiness of this technology. In this paper, we introduce LMAC-ZS (Listenable Maps for Audio Classifiers in the Zero-Shot context), which, to the best of our knowledge, is the first decoder-based post-hoc interpretation method for explaining the decisions of zero-shot audio classifiers. The proposed method utilizes a novel loss function that maximizes the faithfulness to the original similarity between a given text-and-audio pair. We provide an extensive evaluation using the Contrastive Language-Audio Pretraining (CLAP) model to showcase that our interpreter remains faithful to the decisions in a zero-shot classification context. Moreover, we qualitatively show that our method produces meaningful explanations that correlate well with different text prompts.

Francesco Paissan, Luca Della Libera, Mirco Ravanelli, Cem Subakan• 2024

Related benchmarks

TaskDatasetResultRank
Zero-shot Audio Classification ExplanationESC50 contamination
AI39.4
24
Audio ClassificationESC50 (In-Domain)
AI43.35
12
Audio ClassificationUrbanSound8K In-Domain
AI40.85
12
Interpretation for Audio ClassificationUrbanSound8K Mel-Masking US8K contamination (test)
AI42.7
12
Zero-shot Audio Classification ExplanationESC50 White Noise contamination
AI Score32.97
12
Interpretation for Audio ClassificationUrbanSound8K STFT-Masking, ESC50 contamination (test)
AI39.42
6
Interpretation for Audio ClassificationUrbanSound8K STFT-Masking White Noise contamination (test)
Attribution Importance Score46.51
6
Interpretation for Audio ClassificationUrbanSound8K Mel-Masking LJ-Speech contamination (test)
AI36.24
6
Interpretation for Audio ClassificationUrbanSound8K STFT-Masking, LJ-Speech contamination (test)
Attribution Importance (AI)35.96
6
Showing 9 of 9 rows

Other info

Code

Follow for update