Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2209.06453
Cited By
Prompt Combines Paraphrase: Teaching Pre-trained Models to Understand Rare Biomedical Words
14 September 2022
Hao Wang
Chi-Liang Liu
Nuwa Xi
Sendong Zhao
Meizhi Ju
Shiwei Zhang
Ziheng Zhang
Yefeng Zheng
Bing Qin
Ting Liu
VLM
AAML
LM&MA
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Prompt Combines Paraphrase: Teaching Pre-trained Models to Understand Rare Biomedical Words"
5 / 5 papers shown
Title
P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks
Xiao Liu
Kaixuan Ji
Yicheng Fu
Weng Lam Tam
Zhengxiao Du
Zhilin Yang
Jie Tang
VLM
228
780
0
14 Oct 2021
Dict-BERT: Enhancing Language Model Pre-training with Dictionary
W. Yu
Chenguang Zhu
Yuwei Fang
Donghan Yu
Shuohang Wang
Yichong Xu
Michael Zeng
Meng-Long Jiang
45
58
0
13 Oct 2021
Making Pre-trained Language Models Better Few-shot Learners
Tianyu Gao
Adam Fisch
Danqi Chen
238
1,898
0
31 Dec 2020
Pre-trained Models for Natural Language Processing: A Survey
Xipeng Qiu
Tianxiang Sun
Yige Xu
Yunfan Shao
Ning Dai
Xuanjing Huang
LM&MA
VLM
224
1,281
0
18 Mar 2020
Exploiting Cloze Questions for Few Shot Text Classification and Natural Language Inference
Timo Schick
Hinrich Schütze
248
1,382
0
21 Jan 2020
1