Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Beyond Flicker: Detecting Kinematic Inconsistencies for Generalizable Deepfake Video Detection

About

Generalizing deepfake detection to unseen manipulations remains a key challenge. A recent approach to tackle this issue is to train a network with pristine face images that have been manipulated with hand-crafted artifacts to extract more generalizable clues. While effective for static images, extending this to the video domain is an open issue. Existing methods model temporal artifacts as frame-to-frame instabilities, overlooking a key vulnerability: the violation of natural motion dependencies between different facial regions. In this paper, we propose a synthetic video generation method that creates training data with subtle kinematic inconsistencies. We train an autoencoder to decompose facial landmark configurations into motion bases. By manipulating these bases, we selectively break the natural correlations in facial movements and introduce these artifacts into pristine videos via face morphing. A network trained on our data learns to spot these sophisticated biomechanical flaws, achieving state-of-the-art generalization results on several popular benchmarks.

Alejandro Cobo, Roberto Valle, Jos\'e Miguel Buenaposada, Luis Baumela• 2025

Related benchmarks

TaskDatasetResultRank
Deepfake DetectionDFD
Video AUC0.9704
23
Video Deepfake DetectionCeleb-DF (CDF)
Video-level AUC94.74
21
Deepfake DetectionDFDCP
Video-level AUC0.9389
20
Deepfake DetectionWildDeepfake (WDF)
Video-level AUC0.8352
17
Deepfake DetectionDeeperForensics (DFo)
Video-level AUC98.47
16
Deepfake DetectionDF40
BlendFace Score98.06
14
Showing 6 of 6 rows

Other info

Follow for update