MALEFA: Multi-grAnularity Learning and Effective False Alarm Suppression for Zero-shot Keyword Spotting
About
User-defined keyword spotting (KWS) without resorting to domain-specific pre-labeled training data is of fundamental importance in building adaptable and personalized voice interfaces. However, such systems are still faced with arduous challenges, including constrained computational resources and limited annotated training data. Existing methods also struggle to distinguish acoustically similar keywords, often leading to a pesky false alarm rate (FAR) in real-world deployments. To mitigate these limitations, we put forward MALEFA, a novel lightweight zero-shot KWS framework that jointly learns utterance- and phoneme-level alignments via cross-attention and a multi-granularity contrastive learning objective. Evaluations on four public benchmark datasets show that MALEFA achieves a high accuracy of 90%, significantly reducing FAR to 0.007% on the AMI dataset. Beyond its strong performance, MALEFA demonstrates high computational efficiency and can readily support real-time deployment on resource-constrained devices.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Zero-shot Keyword Spotting | LibriPhrase Easy (LPE) Low phonetic confusion other-500 (train) | AUC99.98 | 9 | |
| Zero-shot Keyword Spotting | LibriPhrase Hard High phonetic confusion (train-other-500) | AUC93.58 | 9 | |
| Zero-shot Keyword Spotting | Google Speech Commands G V2 | AUC99.41 | 6 | |
| Zero-shot Keyword Spotting | Qualcomm Keyword Speech Q (evaluation) | AUC99.91 | 6 | |
| Keyword Spotting | AMI | FAR0.007 | 5 | |
| Keyword Spotting | Google Speech Commands G V2 | False Alarm Rate (FAR)0.002 | 5 | |
| Keyword Spotting | Qualcomm Keyword Speech (Q) | FAR0.00e+0 | 5 |