Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

SNAP: Speaker Nulling for Artifact Projection in Speech Deepfake Detection

About

Recent advancements in text-to-speech technologies enable generating high-fidelity synthetic speech nearly indistinguishable from real human voices. While recent studies show the efficacy of self-supervised learning-based speech encoders for deepfake detection, these models struggle to generalize across unseen speakers. Our quantitative analysis suggests these encoder representations are substantially influenced by speaker information, causing detectors to exploit speaker-specific correlations rather than artifact-related cues. We call this phenomenon speaker entanglement. To mitigate this reliance, we introduce SNAP, a speaker-nulling framework. We estimate a speaker subspace and apply orthogonal projection to suppress speaker-dependent components, isolating synthesis artifacts within the residual features. By reducing speaker entanglement, SNAP encourages detectors to focus on artifact-related patterns, leading to state-of-the-art performance.

Kyudan Jung, Jihwan Kim, Minwoo Lee, Soyoon Kim, Jeonghoon Kim, Jaegul Choo, Cheonbok Park• 2026

Related benchmarks

TaskDatasetResultRank
Audio Deepfake DetectionASVspoof LA 2021
EER9.46
41
Synthetic Speech DetectionASVspoof DF 2021 (eval)
EER (%)5.42
25
Synthetic Speech DetectionIn-The-Wild (ITW)
EER15.39
6
Synthetic Speech DetectionASVspoof LA 2019
EER0.35
6
Showing 4 of 4 rows

Other info

Follow for update