Attention-Free Keyword Spotting
About
Till now, attention-based models have been used with great success in the keyword spotting problem domain. However, in light of recent advances in deep learning, the question arises whether self-attention is truly irreplaceable for recognizing speech keywords. We thus explore the usage of gated MLPs --previously shown to be alternatives to transformers in vision tasks-- for the keyword spotting task. We provide a family of highly efficient MLP-based models for keyword spotting, with less than 0.5 million parameters. We show that our approach achieves competitive performance on Google Speech Commands V2-12 and V2-35 benchmarks with much fewer parameters than self-attention-based methods.
Mashrur M. Morshed, Ahmad Omar Ahsan• 2021
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Keyword Spotting | Google Speech Commands V2-35 | Accuracy97.56 | 42 | |
| Keyword Spotting | Google Speech Commands V2-12 2018 | Accuracy97.63 | 16 |
Showing 2 of 2 rows