Learning Counterfactually Decoupled Attention for Open-World Model Attribution
About
In this paper, we propose a Counterfactually Decoupled Attention Learning (CDAL) method for open-world model attribution. Existing methods rely on handcrafted design of region partitioning or feature space, which could be confounded by the spurious statistical correlations and struggle with novel attacks in open-world scenarios. To address this, CDAL explicitly models the causal relationships between the attentional visual traces and source model attribution, and counterfactually decouples the discriminative model-specific artifacts from confounding source biases for comparison. In this way, the resulting causal effect provides a quantification on the quality of learned attention maps, thus encouraging the network to capture essential generation patterns that generalize to unseen source models by maximizing the effect. Extensive experiments on existing open-world model attribution benchmarks show that with minimal computational overhead, our method consistently improves state-of-the-art models by large margins, particularly for unseen novel attacks. Source code: https://github.com/yzheng97/CDAL.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Attribution | Celeb-DF | Accuracy33.79 | 14 | |
| Deepfake Attribution | Unseen Advanced Generators VAE, HART, FLUX | VAE Accuracy37.89 | 14 | |
| Deepfake Attribution | DF40 and FFHQ unseen generators | FFHQ Accuracy8.65 | 14 | |
| Attribution | DFDC | Accuracy40.66 | 14 | |
| Attribution | WildDeepfake | Accuracy43.8 | 14 | |
| Attribution | Unseen Datasets Average | Accuracy39.42 | 14 | |
| Deepfake Detection | DF40 and FFHQ unseen generators | Average Accuracy (ACC)80.2 | 14 | |
| Detection | Unseen Datasets Average | Accuracy72.04 | 14 | |
| Deepfake Detection | Unseen Advanced Generators VAE, HART, FLUX | Average Accuracy35.2 | 14 | |
| Deepfake Attribution | OW-DFA-40 Protocol-2 | All Accuracy80.2 | 10 |