ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.06658
19
0

A Neuro-inspired Interpretation of Unlearning in Large Language Models through Sample-level Unlearning Difficulty

9 April 2025
Xiaohua Feng
Yuyuan Li
C. Wang
Junlin Liu
L. Zhang
Chaochao Chen
    MU
ArXivPDFHTML
Abstract

Driven by privacy protection laws and regulations, unlearning in Large Language Models (LLMs) is gaining increasing attention. However, current research often neglects the interpretability of the unlearning process, particularly concerning sample-level unlearning difficulty. Existing studies typically assume a uniform unlearning difficulty across samples. This simplification risks attributing the performance of unlearning algorithms to sample selection rather than the algorithm's design, potentially steering the development of LLM unlearning in the wrong direction. Thus, we investigate the relationship between LLM unlearning and sample characteristics, with a focus on unlearning difficulty. Drawing inspiration from neuroscience, we propose a Memory Removal Difficulty (MRD\mathrm{MRD}MRD) metric to quantify sample-level unlearning difficulty. Using MRD\mathrm{MRD}MRD, we analyze the characteristics of hard-to-unlearn versus easy-to-unlearn samples. Furthermore, we propose an MRD\mathrm{MRD}MRD-based weighted sampling method to optimize existing unlearning algorithms, which prioritizes easily forgettable samples, thereby improving unlearning efficiency and effectiveness. We validate the proposed metric and method using public benchmarks and datasets, with results confirming its effectiveness.

View on arXiv
@article{feng2025_2504.06658,
  title={ A Neuro-inspired Interpretation of Unlearning in Large Language Models through Sample-level Unlearning Difficulty },
  author={ Xiaohua Feng and Yuyuan Li and Chengye Wang and Junlin Liu and Li Zhang and Chaochao Chen },
  journal={arXiv preprint arXiv:2504.06658},
  year={ 2025 }
}
Comments on this paper