Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

AutoMIA: Improved Baselines for Membership Inference Attack via Agentic Self-Exploration

About

Membership Inference Attacks (MIAs) serve as a fundamental auditing tool for evaluating training data leakage in machine learning models. However, existing methodologies predominantly rely on static, handcrafted heuristics that lack adaptability, often leading to suboptimal performance when transferred across different large models. In this work, we propose AutoMIA, an agentic framework that reformulates membership inference as an automated process of self-exploration and strategy evolution. Given high-level scenario specifications, AutoMIA self-explores the attack space by generating executable logits-level strategies and progressively refining them through closed-loop evaluation feedback. By decoupling abstract strategy reasoning from low-level execution, our framework enables a systematic, model-agnostic traversal of the attack search space. Extensive experiments demonstrate that AutoMIA consistently matches or outperforms state-of-the-art baselines while eliminating the need for manual feature engineering.

Ruhao Liu, Weiqi Huang, Qi Li, Xinchao Wang• 2026

Related benchmarks

TaskDatasetResultRank
Membership Inference AttackFlickr
TPR @ 5% FPR21
142
Membership Inference AttackFlickr
Accuracy68.7
71
Membership Inference AttackDALL·E (test)
TPR @ 5% FPR29.4
54
Membership Inference Attack (TEXTLEN=32)VL-MIA TEXT
Accuracy78.2
33
Membership Inference AttackDALL-E
Accuracy72.3
26
Membership Inference AttackVL-MIA Text LLaVA TEXTLEN=32
AUC81
21
Membership Inference AttackVL-MIA Text LLaVA TEXTLEN=64
AUC0.994
21
Membership Inference AttackVL-MIA Text MiniGPT-4 TEXTLEN=32
AUC82.4
21
Membership Inference AttackVL-MIA Text MiniGPT-4 TEXTLEN=64
AUC89.1
21
Membership Inference AttackVL-MIA Text LLaMA Adapter TEXTLEN=32
AUC82.8
21
Showing 10 of 14 rows

Other info

GitHub

Follow for update