Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Statistical Roughness-Informed Machine Unlearning

About

Machine unlearning aims to remove the influence of a designated forget set from a trained model while preserving utility on the retained data. In modern deep networks, approximate unlearning frequently fails under large or adversarial deletions due to pronounced layer-wise heterogeneity: some layers exhibit stable, well-regularized representations while others are brittle, undertrained, or overfit, so naive update allocation can trigger catastrophic forgetting or unstable dynamics. We propose Statistical-Roughness Adaptive Gradient Unlearning (SRAGU), a mechanism-first unlearning algorithm that reallocates unlearning updates using layer-wise statistical roughness operationalized via heavy-tailed spectral diagnostics of layer weight matrices. Starting from an Adaptive Gradient Unlearning (AGU) sensitivity signal computed on the forget set, SRAGU estimates a WeightWatcher-style heavy-tailed exponent for each layer, maps it to a bounded spectral stability weight, and uses this stability signal to spectrally reweight the AGU sensitivities before applying the same minibatch update form. This concentrates unlearning motion in spectrally stable layers while damping updates in unstable or overfit layers, improving stability under hard deletions. We evaluate unlearning via behavioral alignment to a gold retrained reference model trained from scratch on the retained data, using empirical prediction-divergence and KL-to-gold proxies on a forget-focused query set; we additionally report membership inference auditing as a complementary leakage signal, treating forget-set points as should-be-forgotten members during evaluation.

Mohammad Partohaghighi, Roummel Marcia, Bruce J. West, YangQuan Chen• 2026

Related benchmarks

TaskDatasetResultRank
Machine UnlearningMNIST
Model Accuracy89.33
44
Machine UnlearningCIFAR-10
Accuracy63.31
24
Machine UnlearningUCI Adult
Accuracy85.25
24
Machine UnlearningCIFAR-10 Low-margin deletion (test)
Accuracy64.89
8
Machine UnlearningCIFAR-10 High-gradient-norm deletion (test)
Accuracy65.88
8
Machine UnlearningCIFAR-10 Influence-inspired deletion (test)
Accuracy64.86
8
Machine UnlearningImageNet100 30% influence-style deletion (retain)
Retention Accuracy42.99
8
Machine UnlearningCIFAR-10 High-loss deletion (test)
Acc64.32
8
Membership Inference AttackCIFAR-10 10% Random deletion (test)
AUC0.522
8
Membership Inference AttackCIFAR-10 10% Class-specific deletion (test)
AUC0.536
8
Showing 10 of 11 rows

Other info

Follow for update