Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Membership Inference on LLMs in the Wild

About

Membership Inference Attacks (MIAs) act as a crucial auditing tool for the opaque training data of Large Language Models (LLMs). However, existing techniques predominantly rely on inaccessible model internals (e.g., logits) or suffer from poor generalization across domains in strict black-box settings where only generated text is available. In this work, we propose SimMIA, a robust MIA framework tailored for this text-only regime by leveraging an advanced sampling strategy and scoring mechanism. Furthermore, we present WikiMIA-25, a new benchmark curated to evaluate MIA performance on modern proprietary LLMs. Experiments demonstrate that SimMIA achieves state-of-the-art results in the black-box setting, rivaling baselines that exploit internal model information.

Jiatong Yi, Yanyang Li• 2026

Related benchmarks

TaskDatasetResultRank
Membership Inference AttackWikiMIA length 64
AUC0.875
84
Membership Inference AttackWikiMIA length 32
TPR @ 5% FPR0.334
54
Membership Inference AttackMIMIR Github
AUC86.1
52
Membership Inference AttackWikipedia
AUC0.65
52
Membership Inference AttackMIMIR Average
AUC70.7
48
Membership Inference AttackWikiMIA-25
AUC0.906
33
Membership Inference AttackWikiMIA length 128
AUC0.794
28
Showing 7 of 7 rows

Other info

Follow for update