ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2305.11038
222
27
v1v2v3 (latest)

Learning In-context Learning for Named Entity Recognition

Annual Meeting of the Association for Computational Linguistics (ACL), 2023
18 May 2023
Jiawei Chen
Yaojie Lu
Hongyu Lin
Jie Lou
Wei Jia
Dai Dai
Hua Wu
Boxi Cao
Xianpei Han
Le Sun
    NAI
ArXiv (abs)PDFHTML
Abstract

Named entity recognition in real-world applications suffers from the diversity of entity types, the emergence of new entity types, and the lack of high-quality annotations. To address the above problems, this paper proposes an in-context learning-based NER approach, which can effectively inject in-context NER ability into PLMs and recognize entities of novel types on-the-fly using only a few demonstrative instances. Specifically, we model PLMs as a meta-function λinstruction, demonstrations, text.M\mathcal{ \lambda_ {\text{instruction, demonstrations, text}}. M}λinstruction, demonstrations, text​.M, and a new entity extractor can be implicitly constructed by applying new instruction and demonstrations to PLMs, i.e., (λ.M)\mathcal{ (\lambda . M) }(λ.M)(instruction, demonstrations) →\to→ F\mathcal{F}F where F\mathcal{F}F will be a new entity extractor, i.e., F\mathcal{F}F: text →\to→ entities. To inject the above in-context NER ability into PLMs, we propose a meta-function pre-training algorithm, which pre-trains PLMs by comparing the (instruction, demonstration)-initialized extractor with a surrogate golden extractor. Experimental results on 4 few-shot NER datasets show that our method can effectively inject in-context NER ability into PLMs and significantly outperforms the PLMs+fine-tuning counterparts.

View on arXiv
Comments on this paper