ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2402.12233
19
5

Empirical Study on Updating Key-Value Memories in Transformer Feed-forward Layers

19 February 2024
Zihan Qiu
Zeyu Huang
Youcheng Huang
Jie Fu
    KELM
ArXivPDFHTML
Abstract

The feed-forward networks (FFNs) in transformers are recognized as a group of key-value neural memories to restore abstract high-level knowledge. In this work, we conduct an empirical ablation study on updating keys (the 1st layer in the FFNs layer) or values (the 2nd layer in the FFNs layer). We compare those two methods in various knowledge editing and fine-tuning tasks of large language models to draw insights to understand FFNs further. Code is available at \href\href{https://github.com/qiuzh20/Tuning-keys-v.s.-values}{this\,repo}\href.

View on arXiv
Comments on this paper