End-to-end anti-spoofing with RawNet2
About
Spoofing countermeasures aim to protect automatic speaker verification systems from attempts to manipulate their reliability with the use of spoofed speech signals. While results from the most recent ASVspoof 2019 evaluation show great potential to detect most forms of attack, some continue to evade detection. This paper reports the first application of RawNet2 to anti-spoofing. RawNet2 ingests raw audio and has potential to learn cues that are not detectable using more traditional countermeasure solutions. We describe modifications made to the original RawNet2 architecture so that it can be applied to anti-spoofing. For A17 attacks, our RawNet2 systems results are the second-best reported, while the fusion of RawNet2 and baseline countermeasures gives the second-best results reported for the full ASVspoof 2019 logical access condition. Our results are reproducible with open source software.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Spoof Speech Detection | ASVspoof LA 2021 (eval) | min-tDCF0.4257 | 36 | |
| Audio Deepfake Detection | ASVspoof 2021 | EER2.85 | 27 | |
| Synthetic Speech Detection | ASVspoof DF 2021 (eval) | EER (%)22.38 | 19 | |
| Audio Deepfake Detection | ADD-C 0.5s duration (test) | C0 Score23.4 | 12 | |
| Audio Deepfake Detection | ADD-C 2.0s duration (test) | Class 0 Score2.05 | 12 | |
| Audio Deepfake Detection | ADD-C 1.0s duration (test) | C0 Score4.8 | 12 | |
| Audio Deepfake Detection | ADD-C 1.5s duration (test) | C0 Score2.35 | 12 | |
| Automatic Speaker Verification Anti-spoofing | ASVspoof pooled 2019 (evaluation) | min t-DCF0.033 | 11 | |
| Automatic Speaker Verification Anti-spoofing | ASVspoof A17 2019 (evaluation) | Min t-DCF0.0808 | 11 | |
| Authenticity Classification | Mixed In-Domain | Accuracy91.24 | 11 |