ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.12681
57
0

GRAIL: Gradient-Based Adaptive Unlearning for Privacy and Copyright in LLMs

17 April 2025
Kun-Woo Kim
Ji-Hoon Park
Ju-Min Han
Seong-Whan Lee
    MU
    PILM
ArXivPDFHTML
Abstract

Large Language Models (LLMs) trained on extensive datasets often learn sensitive information, which raises significant social and legal concerns under principles such as the "Right to be forgotten." Retraining entire models from scratch to remove undesired information is both costly and impractical. Furthermore, existing single-domain unlearning methods fail to address multi-domain scenarios, where knowledge is interwoven across domains such as privacy and copyright, creating overlapping representations that lead to excessive knowledge removal or degraded performance. To tackle these issues, we propose GRAIL (GRadient-based AdaptIve unLearning), a novel multi-domain unlearning framework. GRAIL leverages gradient information from multiple domains to precisely distinguish the unlearning scope from the retention scope, and applies an adaptive parameter-wise localization strategy to selectively remove targeted knowledge while preserving critical parameters for each domain. Experimental results on unlearning benchmarks show that GRAIL achieves unlearning success on par with the existing approaches, while also demonstrating up to 17% stronger knowledge retention success compared to the previous state-of-art method. Our findings establish a new paradigm for effectively managing and regulating sensitive information in large-scale pre-trained language models.

View on arXiv
@article{kim2025_2504.12681,
  title={ GRAIL: Gradient-Based Adaptive Unlearning for Privacy and Copyright in LLMs },
  author={ Kun-Woo Kim and Ji-Hoon Park and Ju-Min Han and Seong-Whan Lee },
  journal={arXiv preprint arXiv:2504.12681},
  year={ 2025 }
}
Comments on this paper