9
0

BLUR: A Benchmark for LLM Unlearning Robust to Forget-Retain Overlap

Main:9 Pages
2 Figures
Bibliography:2 Pages
12 Tables
Appendix:3 Pages
Abstract

Machine unlearning has the potential to improve the safety of large language models (LLMs) by removing sensitive or harmful information post hoc. A key challenge in unlearning involves balancing between forget quality (effectively unlearning undesirable information) and retain quality (maintaining good performance on other, general tasks). Unfortunately, as we show, current LLM unlearning benchmarks contain highly disparate forget and retain sets -- painting a false picture of the effectiveness of LLM unlearning methods. This can be particularly problematic because it opens the door for benign perturbations, such as relearning attacks, to easily reveal supposedly unlearned knowledge once models are deployed. To address this, we present BLUR\texttt{BLUR}: a benchmark for LLM unlearning that provides more realistic scenarios of forget-retain overlap. BLUR\texttt{BLUR} significantly expands on existing unlearning benchmarks by providing extended evaluation tasks, combined forget/retain queries, and relearning datasets of varying degrees of difficulty. Despite the benign nature of the queries considered, we find that the performance of existing methods drops significantly when evaluated on BLUR\texttt{BLUR}, with simple approaches performing better on average than more recent methods. These results highlight the importance of robust evaluation and suggest several important directions of future study. Our benchmark is publicly available at:this https URL

View on arXiv
@article{hu2025_2506.15699,
  title={ BLUR: A Benchmark for LLM Unlearning Robust to Forget-Retain Overlap },
  author={ Shengyuan Hu and Neil Kale and Pratiksha Thaker and Yiwei Fu and Steven Wu and Virginia Smith },
  journal={arXiv preprint arXiv:2506.15699},
  year={ 2025 }
}
Comments on this paper