Have you forgotten? A method to assess if machine learning models have forgotten data
About
In the era of deep learning, aggregation of data from several sources is a common approach to ensuring data diversity. Let us consider a scenario where several providers contribute data to a consortium for the joint development of a classification model (hereafter the target model), but, now one of the providers decides to leave. This provider requests that their data (hereafter the query dataset) be removed from the databases but also that the model `forgets' their data. In this paper, for the first time, we want to address the challenging question of whether data have been forgotten by a model. We assume knowledge of the query dataset and the distribution of a model's output. We establish statistical methods that compare the target's outputs with outputs of models trained with different datasets. We evaluate our approach on several benchmark datasets (MNIST, CIFAR-10 and SVHN) and on a cardiac pathology diagnosis task using data from the Automated Cardiac Diagnosis Challenge (ACDC). We hope to encourage studies on what information a model retains and inspire extensions in more complex settings.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Auditing Data Removal | MNIST Memorized (train folds M1-M5) | pks1.07 | 36 | |
| Auditing Data Removal | Chest X-ray C1 | PKS Score1.32 | 10 | |
| Auditing Data Removal | Chest X-ray C2 | PKS Score1.32 | 10 | |
| Auditing Data Removal | Chest X-ray C3 | PKS Score1.31 | 10 | |
| Auditing Data Removal | Chest X-ray C4 | pks Score1.26 | 10 | |
| Auditing Data Removal | Chest X-ray C5 | PKS Score1.31 | 10 | |
| Auditing Data Removal | Chest X-ray C6 | PKS Score109 | 10 | |
| Auditing Data Removal | Chest X-ray R | PKS Score106 | 10 | |
| Auditing Data Removal | MNIST M6 Disjoint Set | PKS94 | 7 |