ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2006.03654
  4. Cited By
DeBERTa: Decoding-enhanced BERT with Disentangled Attention

DeBERTa: Decoding-enhanced BERT with Disentangled Attention

5 June 2020
Pengcheng He
Xiaodong Liu
Jianfeng Gao
Weizhu Chen
    AAML
ArXivPDFHTML

Papers citing "DeBERTa: Decoding-enhanced BERT with Disentangled Attention"

37 / 1,037 papers shown
Title
Improving Controllability of Educational Question Generation by Keyword
  Provision
Improving Controllability of Educational Question Generation by Keyword Provision
Ying-Hong Chan
Ho-Lam Chung
Yao-Chung Fan
11
3
0
02 Dec 2021
Interactive Model with Structural Loss for Language-based Abductive
  Reasoning
Interactive Model with Structural Loss for Language-based Abductive Reasoning
Linhao Li
Ming Xu
Yongfeng Dong
Xin Li
Ao Wang
12
2
0
01 Dec 2021
Backdoor Pre-trained Models Can Transfer to All
Backdoor Pre-trained Models Can Transfer to All
Lujia Shen
S. Ji
Xuhong Zhang
Jinfeng Li
Jing Chen
Jie Shi
Chengfang Fang
Jianwei Yin
Ting Wang
AAML
SILM
26
117
0
30 Oct 2021
Team Enigma at ArgMining-EMNLP 2021: Leveraging Pre-trained Language
  Models for Key Point Matching
Team Enigma at ArgMining-EMNLP 2021: Leveraging Pre-trained Language Models for Key Point Matching
Chao Fan
Yang Yang
Siba Smarak Panigrahi
Varun Madhavan
Abhilash Nandy
6
9
0
24 Oct 2021
Overview of the 2021 Key Point Analysis Shared Task
Overview of the 2021 Key Point Analysis Shared Task
Roni Friedman
Lena Dankin
Yufang Hou
R. Aharonov
Yoav Katz
Noam Slonim
16
22
0
20 Oct 2021
Sharpness-Aware Minimization Improves Language Model Generalization
Sharpness-Aware Minimization Improves Language Model Generalization
Dara Bahri
H. Mobahi
Yi Tay
119
97
0
16 Oct 2021
Clean or Annotate: How to Spend a Limited Data Collection Budget
Clean or Annotate: How to Spend a Limited Data Collection Budget
Derek Chen
Zhou Yu
Samuel R. Bowman
27
13
0
15 Oct 2021
P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally
  Across Scales and Tasks
P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks
Xiao Liu
Kaixuan Ji
Yicheng Fu
Weng Lam Tam
Zhengxiao Du
Zhilin Yang
Jie Tang
VLM
236
804
0
14 Oct 2021
Taming Sparsely Activated Transformer with Stochastic Experts
Taming Sparsely Activated Transformer with Stochastic Experts
Simiao Zuo
Xiaodong Liu
Jian Jiao
Young Jin Kim
Hany Hassan
Ruofei Zhang
T. Zhao
Jianfeng Gao
MoE
23
108
0
08 Oct 2021
ContractNLI: A Dataset for Document-level Natural Language Inference for
  Contracts
ContractNLI: A Dataset for Document-level Natural Language Inference for Contracts
Yuta Koreeda
Christopher D. Manning
AILaw
87
96
0
05 Oct 2021
Small-Bench NLP: Benchmark for small single GPU trained models in
  Natural Language Processing
Small-Bench NLP: Benchmark for small single GPU trained models in Natural Language Processing
K. Kanakarajan
Bhuvana Kundumani
Malaikannan Sankarasubbu
ALM
MoE
6
5
0
22 Sep 2021
Fine-Tuned Transformers Show Clusters of Similar Representations Across
  Layers
Fine-Tuned Transformers Show Clusters of Similar Representations Across Layers
Jason Phang
Haokun Liu
Samuel R. Bowman
14
25
0
17 Sep 2021
The Stem Cell Hypothesis: Dilemma behind Multi-Task Learning with
  Transformer Encoders
The Stem Cell Hypothesis: Dilemma behind Multi-Task Learning with Transformer Encoders
Han He
Jinho D. Choi
43
87
0
14 Sep 2021
The Impact of Positional Encodings on Multilingual Compression
The Impact of Positional Encodings on Multilingual Compression
Vinit Ravishankar
Anders Søgaard
9
5
0
11 Sep 2021
Tiered Reasoning for Intuitive Physics: Toward Verifiable Commonsense
  Language Understanding
Tiered Reasoning for Intuitive Physics: Toward Verifiable Commonsense Language Understanding
Shane Storks
Qiaozi Gao
Yichi Zhang
J. Chai
ReLM
LRM
34
22
0
10 Sep 2021
It's not Rocket Science : Interpreting Figurative Language in Narratives
It's not Rocket Science : Interpreting Figurative Language in Narratives
Tuhin Chakrabarty
Yejin Choi
Vered Shwartz
8
55
0
31 Aug 2021
How to Query Language Models?
How to Query Language Models?
Leonard Adolphs
S. Dhuliawala
Thomas Hofmann
KELM
16
15
0
04 Aug 2021
Domain-matched Pre-training Tasks for Dense Retrieval
Domain-matched Pre-training Tasks for Dense Retrieval
Barlas Oğuz
Kushal Lakhotia
Anchit Gupta
Patrick Lewis
Vladimir Karpukhin
...
Xilun Chen
Sebastian Riedel
Wen-tau Yih
Sonal Gupta
Yashar Mehdad
RALM
19
66
0
28 Jul 2021
Trusting RoBERTa over BERT: Insights from CheckListing the Natural
  Language Inference Task
Trusting RoBERTa over BERT: Insights from CheckListing the Natural Language Inference Task
Ishan Tarunesh
Somak Aditya
Monojit Choudhury
11
17
0
15 Jul 2021
Doing Good or Doing Right? Exploring the Weakness of Commonsense Causal
  Reasoning Models
Doing Good or Doing Right? Exploring the Weakness of Commonsense Causal Reasoning Models
Mingyue Han
Yinglin Wang
LRM
11
10
0
05 Jul 2021
Draw Me a Flower: Processing and Grounding Abstraction in Natural
  Language
Draw Me a Flower: Processing and Grounding Abstraction in Natural Language
R. Lachmy
Valentina Pyatkin
Avshalom Manevich
Reut Tsarfaty
21
18
0
27 Jun 2021
Can Transformer Language Models Predict Psychometric Properties?
Can Transformer Language Models Predict Psychometric Properties?
Antonio Laverghetta
Animesh Nighojkar
Jamshidbek Mirzakhalov
John Licato
LM&MA
22
14
0
12 Jun 2021
Generate, Annotate, and Learn: NLP with Synthetic Text
Generate, Annotate, and Learn: NLP with Synthetic Text
Xuanli He
Islam Nassar
J. Kiros
Gholamreza Haffari
Mohammad Norouzi
19
51
0
11 Jun 2021
Poolingformer: Long Document Modeling with Pooling Attention
Poolingformer: Long Document Modeling with Pooling Attention
Hang Zhang
Yeyun Gong
Yelong Shen
Weisheng Li
Jiancheng Lv
Nan Duan
Weizhu Chen
29
98
0
10 May 2021
REPT: Bridging Language Models and Machine Reading Comprehension via
  Retrieval-Based Pre-training
REPT: Bridging Language Models and Machine Reading Comprehension via Retrieval-Based Pre-training
Fangkai Jiao
Yangyang Guo
Yilin Niu
Feng Ji
Feng-Lin Li
Liqiang Nie
LRM
18
12
0
10 May 2021
Logic-Driven Context Extension and Data Augmentation for Logical
  Reasoning of Text
Logic-Driven Context Extension and Data Augmentation for Logical Reasoning of Text
Siyuan Wang
Wanjun Zhong
Duyu Tang
Zhongyu Wei
Zhihao Fan
Daxin Jiang
Ming Zhou
Nan Duan
NAI
24
70
0
08 May 2021
Evaluating Attribution in Dialogue Systems: The BEGIN Benchmark
Evaluating Attribution in Dialogue Systems: The BEGIN Benchmark
Nouha Dziri
Hannah Rashkin
Tal Linzen
David Reitter
ALM
185
79
0
30 Apr 2021
RoFormer: Enhanced Transformer with Rotary Position Embedding
RoFormer: Enhanced Transformer with Rotary Position Embedding
Jianlin Su
Yu Lu
Shengfeng Pan
Ahmed Murtadha
Bo Wen
Yunfeng Liu
33
2,142
0
20 Apr 2021
AMMU : A Survey of Transformer-based Biomedical Pretrained Language
  Models
AMMU : A Survey of Transformer-based Biomedical Pretrained Language Models
Katikapalli Subramanyam Kalyan
A. Rajasekharan
S. Sangeetha
LM&MA
MedIm
18
163
0
16 Apr 2021
CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review
CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review
Dan Hendrycks
Collin Burns
Anya Chen
Spencer Ball
ELM
AILaw
6
178
0
10 Mar 2021
Measuring Mathematical Problem Solving With the MATH Dataset
Measuring Mathematical Problem Solving With the MATH Dataset
Dan Hendrycks
Collin Burns
Saurav Kadavath
Akul Arora
Steven Basart
Eric Tang
D. Song
Jacob Steinhardt
ReLM
FaML
43
1,787
0
05 Mar 2021
ZJUKLAB at SemEval-2021 Task 4: Negative Augmentation with Language
  Model for Reading Comprehension of Abstract Meaning
ZJUKLAB at SemEval-2021 Task 4: Negative Augmentation with Language Model for Reading Comprehension of Abstract Meaning
Xin Xie
Xiangnan Chen
Xiang Chen
Yong Wang
Ningyu Zhang
Shumin Deng
Huajun Chen
32
2
0
25 Feb 2021
COCO-LM: Correcting and Contrasting Text Sequences for Language Model
  Pretraining
COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Yu Meng
Chenyan Xiong
Payal Bajaj
Saurabh Tiwary
Paul N. Bennett
Jiawei Han
Xia Song
119
202
0
16 Feb 2021
WARP: Word-level Adversarial ReProgramming
WARP: Word-level Adversarial ReProgramming
Karen Hambardzumyan
Hrant Khachatrian
Jonathan May
AAML
254
341
0
01 Jan 2021
ANLIzing the Adversarial Natural Language Inference Dataset
ANLIzing the Adversarial Natural Language Inference Dataset
Adina Williams
Tristan Thrush
Douwe Kiela
AAML
166
45
0
24 Oct 2020
Megatron-LM: Training Multi-Billion Parameter Language Models Using
  Model Parallelism
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
243
1,815
0
17 Sep 2019
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language
  Understanding
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
294
6,943
0
20 Apr 2018
Previous
123...192021