Deep Contrastive Unlearning for Language Models
About
The past a few years have witnessed the great success of large language models, demonstrating powerful capabilities in comprehending textual data and generating human-like languages. Large language models achieve success by being trained on vast amounts of textual data, including online sources with copyrighted content and user-generated knowledge. However, this comes at a cost: the potential risk of exposing users' privacy and violating copyright protections. Thus, to safeguard individuals' "right to be forgotten", there has been increasing interests in machine unlearning -- the process of removing information carried by particular training samples from a model while not deteriorating its predictive quality. This is a challenging task due to the black-box nature of language models. Most existing studies focus on mitigating the impact of those forgot samples upon a model's outputs, and do not explicitly consider the geometric distributions of samples in the latent space of a model. To address this issue, we propose a machine unlearning framework, named Deep Contrastive Unlearning for fine-Tuning (DeepCUT) language models. Our proposed model achieves machine unlearning by directly optimizing the latent space of a model. Comprehensive experiments on real-world datasets demonstrate the effectiveness and efficiency of DeepCUT with consistent and significant improvement over baseline methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| LLM Unlearning | RWKU | USR80.1 | 16 | |
| Machine Unlearning | MUSE | -- | 16 | |
| Machine Unlearning | WaterDrum | USR76.3 | 8 | |
| Relearning Attack | RWKU | RAP24.9 | 8 | |
| Relearning Attack | WMDP | RAP26.8 | 8 | |
| Relearning Attack | MUSE | RAP29.2 | 8 | |
| Relearning Attack | WaterDrum | RAP25.7 | 8 | |
| Machine Unlearning | WMDP | MIA0.068 | 8 |