ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.06257
62
0

K-ON: Stacking Knowledge On the Head Layer of Large Language Model

10 February 2025
Lingbing Guo
Y. Zhang
Zhongpu Bo
Zhuo Chen
Mengshu Sun
Zhiqiang Zhang
Wen Zhang
H. Chen
ArXivPDFHTML
Abstract

Recent advancements in large language models (LLMs) have significantly improved various natural language processing (NLP) tasks. Typically, LLMs are trained to predict the next token, aligning well with many NLP tasks. However, in knowledge graph (KG) scenarios, entities are the fundamental units and identifying an entity requires at least several tokens. This leads to a granularity mismatch between KGs and natural languages. To address this issue, we propose K-ON, which integrates KG knowledge into the LLM by employing multiple head layers for next k-step prediction. K-ON can not only generate entity-level results in one step, but also enables contrastive loss against entities, which is the most powerful tool in KG representation learning. Experimental results show that K-ON outperforms state-of-the-art methods that incorporate text and even the other modalities.

View on arXiv
@article{guo2025_2502.06257,
  title={ K-ON: Stacking Knowledge On the Head Layer of Large Language Model },
  author={ Lingbing Guo and Yichi Zhang and Zhongpu Bo and Zhuo Chen and Mengshu Sun and Zhiqiang Zhang and Wen Zhang and Huajun Chen },
  journal={arXiv preprint arXiv:2502.06257},
  year={ 2025 }
}
Comments on this paper