Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 93% Match Research Paper ML Researchers,AI Ethicists,LLM Developers,Legal/Compliance Officers 2 weeks ago

BLUR: A Bi-Level Optimization Approach for LLM Unlearning

large-language-models › training-methods
📄 Abstract

Abstract: Enabling large language models (LLMs) to unlearn knowledge and capabilities acquired during training has proven vital for ensuring compliance with data regulations and promoting ethical practices in generative AI. Although there are growing interests in developing various unlearning algorithms, it remains unclear how to best formulate the unlearning problem. The most popular formulation uses a weighted sum of forget and retain loss, but it often leads to performance degradation due to the inherent trade-off between forget and retain losses. In this work, we argue that it is important to model the hierarchical structure of the unlearning problem, where the forget problem (which \textit{unlearns} certain knowledge and/or capabilities) takes priority over the retain problem (which preserves model utility). This hierarchical structure naturally leads to a bi-level optimization formulation where the lower-level objective focuses on minimizing the forget loss, while the upper-level objective aims to maintain the model's utility. Based on this new formulation, we propose a novel algorithm, termed Bi-Level UnleaRning (\texttt{BLUR}), which not only possesses strong theoretical guarantees but more importantly, delivers superior performance. In particular, our extensive experiments demonstrate that \texttt{BLUR} consistently outperforms all the state-of-the-art algorithms across various unlearning tasks, models, and metrics. Codes are available at https://github.com/OptimAI-Lab/BLURLLMUnlearning.
Authors (9)
Hadi Reisizadeh
Jinghan Jia
Zhiqi Bu
Bhanukiran Vinzamuri
Anil Ramakrishna
Kai-Wei Chang
+3 more
Submitted
June 9, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

Introduces a bi-level optimization formulation for LLM unlearning, arguing for a hierarchical structure where forgetting takes priority over retaining utility. This approach aims to better manage the trade-off between forgetting specific knowledge and preserving overall model performance.

Business Value

Enables organizations to comply with data privacy regulations (like GDPR's 'right to be forgotten') by selectively removing data or knowledge from LLMs, reducing legal and reputational risks.