Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MARE: Multimodal Alignment and Reinforcement for Explainable Deepfake Detection via Vision-Language Models

About

Deepfake detection is a widely researched topic that is crucial for combating the spread of malicious content, with existing methods mainly modeling the problem as classification or spatial localization. The rapid advancements in generative models impose new demands on Deepfake detection. In this paper, we propose multimodal alignment and reinforcement for explainable Deepfake detection via vision-language models, termed MARE, which aims to enhance the accuracy and reliability of Vision-Language Models (VLMs) in Deepfake detection and reasoning. Specifically, MARE designs comprehensive reward functions, incorporating reinforcement learning from human feedback (RLHF), to incentivize the generation of text-spatially aligned reasoning content that adheres to human preferences. Besides, MARE introduces a forgery disentanglement module to capture intrinsic forgery traces from high-level facial semantics, thereby improving its authenticity detection capability. We conduct thorough evaluations on the reasoning content generated by MARE. Both quantitative and qualitative experimental results demonstrate that MARE achieves state-of-the-art performance in terms of accuracy and reliability.

Wenbo Xu, Wei Lu, Xiangyang Luo, Jiantao Zhou• 2026

Related benchmarks

TaskDatasetResultRank
Deepfake DetectionDFDC (test)
AUC99.77
87
Deepfake DetectionFF++ (test)
AUC99.28
39
Deepfake DetectionCeleb-DF (test)
Accuracy100
24
Deepfake DetectionWildDeepfake (test)
AUC0.937
19
Deepfake DetectionFaceForensics++ (test)
OA96.55
13
Deepfake DetectionDMA dataset (test)
Accuracy0.9809
7
Deepfake DetectionWDF (test)
Accuracy87.72
5
Deepfake DetectionDFD (test)
Accuracy98.25
5
Showing 8 of 8 rows

Other info

Follow for update