316

LambdaKG: A Library for Pre-trained Language Model-Based Knowledge Graph Embeddings

International Joint Conference on Natural Language Processing (IJCNLP), 2022
Ningyu Zhang
Jintian Zhang
Siyuan Cheng
Bozhong Tian
Shumin Deng
Feiyu Xiong
Huajun Chen
Abstract

Knowledge Graphs (KGs) often have two characteristics: heterogeneous graph structure and text-rich entity/relation information. Text-based KG embeddings can represent entities by encoding descriptions with pre-trained language models, but no open-sourced library is specifically designed for KGs with PLMs at present. In this paper, we present LambdaKG, a library for KGE that equips with many pre-trained language models (e.g., BERT, BART, T5, GPT-3), and supports various tasks (e.g., knowledge graph completion, question answering, recommendation, and knowledge probing). LambdaKG is publicly open-sourced at https://github.com/zjunlp/PromptKG/tree/main/lambdaKG, with a demo video at http://deepke.zjukg.cn/lambdakg.mp4 and long-term maintenance.

View on arXiv
Comments on this paper