14
3

SWEA: Updating Factual Knowledge in Large Language Models via Subject Word Embedding Altering

Abstract

The general capabilities of large language models (LLMs) make them the infrastructure for various AI applications, but updating their inner knowledge requires significant resources. Recent model editing is a promising technique for efficiently updating a small amount of knowledge of LLMs and has attracted much attention. In particular, local editing methods, which directly update model parameters, are proven suitable for updating small amounts of knowledge. Local editing methods update weights by computing least squares closed-form solutions and identify edited knowledge by vector-level matching in inference, which achieve promising results. However, these methods still require a lot of time and resources to complete the computation. Moreover, vector-level matching lacks reliability, and such updates disrupt the original organization of the model's parameters. To address these issues, we propose a detachable and expandable Subject Word Embedding Altering (SWEA) framework, which finds the editing embeddings through token-level matching and adds them to the subject word embeddings in Transformer input. To get these editing embeddings, we propose optimizing then suppressing fusion method, which first optimizes learnable embedding vectors for the editing target and then suppresses the Knowledge Embedding Dimensions (KEDs) to obtain final editing embeddings. We thus propose SWEA\oplusOS method for editing factual knowledge in LLMs. We demonstrate the overall state-of-the-art (SOTA) performance of SWEA\oplusOS on the CounterFact and zsRE datasets. To further validate the reasoning ability of SWEA\oplusOS in editing knowledge, we evaluate it on the more complex RippleEdits benchmark. The results demonstrate that SWEA\oplusOS possesses SOTA reasoning ability.

View on arXiv
@article{li2025_2401.17809,
  title={ SWEA: Updating Factual Knowledge in Large Language Models via Subject Word Embedding Altering },
  author={ Xiaopeng Li and Shasha Li and Shezheng Song and Huijun Liu and Bin Ji and Xi Wang and Jun Ma and Jie Yu and Xiaodong Liu and Jing Wang and Weimin Zhang },
  journal={arXiv preprint arXiv:2401.17809},
  year={ 2025 }
}
Comments on this paper