Multi-Task Transformer for Explainable Speech Deepfake Detection via Formant Modeling
About
In this work, we introduce a multi-task transformer for speech deepfake detection, capable of predicting formant trajectories and voicing patterns over time, ultimately classifying speech as real or fake, and highlighting whether its decisions rely more on voiced or unvoiced regions. Building on a prior speaker-formant transformer architecture, we streamline the model with an improved input segmentation strategy, redesign the decoding process, and integrate built-in explainability. Compared to the baseline, our model requires fewer parameters, trains faster, and provides better interpretability, without sacrificing prediction performance.
Viola Negroni, Luca Cuccovillo, Paolo Bestagini, Patrick Aichroth, Stefano Tubaro• 2026
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Audio Deepfake Detection | in the wild | EER17.29 | 58 | |
| Audio Deepfake Detection | ASVspoof 5 | EER4.41 | 9 | |
| Speech Deepfake Detection | FakeOrReal | EER20.33 | 9 | |
| Speech Deepfake Detection | TIMIT TTS | EER20.93 | 2 |
Showing 4 of 4 rows