ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1804.07461
  4. Cited By
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language
  Understanding
v1v2v3 (latest)

GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding

20 April 2018
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
    ELM
ArXiv (abs)PDFHTML

Papers citing "GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding"

50 / 4,447 papers shown
Title
Guiding Attention for Self-Supervised Learning with Transformers
Guiding Attention for Self-Supervised Learning with Transformers
Ameet Deshpande
Karthik Narasimhan
69
21
0
06 Oct 2020
Mixup-Transformer: Dynamic Data Augmentation for NLP Tasks
Mixup-Transformer: Dynamic Data Augmentation for NLP Tasks
Lichao Sun
Congying Xia
Wenpeng Yin
Tingting Liang
Philip S. Yu
Lifang He
62
36
0
05 Oct 2020
Conversational Document Prediction to Assist Customer Care Agents
Conversational Document Prediction to Assist Customer Care Agents
Jatin Ganhotra
Haggai Roitman
Doron Cohen
Nathaniel Mills
Chulaka Gunasekara
Y. Mass
Sachindra Joshi
Luis A. Lastras
D. Konopnicki
50
4
0
05 Oct 2020
Learning to Generalize for Sequential Decision Making
Learning to Generalize for Sequential Decision Making
Xusen Yin
R. Weischedel
Jonathan May
OffRLLRM
70
7
0
05 Oct 2020
PMI-Masking: Principled masking of correlated spans
PMI-Masking: Principled masking of correlated spans
Yoav Levine
Barak Lenz
Opher Lieber
Omri Abend
Kevin Leyton-Brown
Moshe Tennenholtz
Y. Shoham
66
73
0
05 Oct 2020
Pruning Redundant Mappings in Transformer Models via Spectral-Normalized
  Identity Prior
Pruning Redundant Mappings in Transformer Models via Spectral-Normalized Identity Prior
Zi Lin
Jeremiah Zhe Liu
Ziao Yang
Nan Hua
Dan Roth
94
47
0
05 Oct 2020
How Effective is Task-Agnostic Data Augmentation for Pretrained
  Transformers?
How Effective is Task-Agnostic Data Augmentation for Pretrained Transformers?
Shayne Longpre
Yu Wang
Christopher DuBois
ViT
86
85
0
05 Oct 2020
Effective Unsupervised Domain Adaptation with Adversarially Trained
  Language Models
Effective Unsupervised Domain Adaptation with Adversarially Trained Language Models
Thuy-Trang Vu
Dinh Q. Phung
Gholamreza Haffari
86
25
0
05 Oct 2020
On Losses for Modern Language Models
On Losses for Modern Language Models
Stephane Aroca-Ouellette
Frank Rudzicz
81
32
0
04 Oct 2020
Tell Me How to Ask Again: Question Data Augmentation with Controllable
  Rewriting in Continuous Space
Tell Me How to Ask Again: Question Data Augmentation with Controllable Rewriting in Continuous Space
Dayiheng Liu
Yeyun Gong
Jie Fu
Yu Yan
Jiusheng Chen
Jiancheng Lv
Nan Duan
M. Zhou
51
37
0
04 Oct 2020
Code to Comment "Translation": Data, Metrics, Baselining & Evaluation
Code to Comment "Translation": Data, Metrics, Baselining & Evaluation
David Gros
Hariharan Sezhiyan
Prem Devanbu
Zhou Yu
100
72
0
03 Oct 2020
Mining Knowledge for Natural Language Inference from Wikipedia
  Categories
Mining Knowledge for Natural Language Inference from Wikipedia Categories
Mingda Chen
Zewei Chu
K. Stratos
Kevin Gimpel
57
8
0
03 Oct 2020
Cost-effective Selection of Pretraining Data: A Case Study of
  Pretraining BERT on Social Media
Cost-effective Selection of Pretraining Data: A Case Study of Pretraining BERT on Social Media
Xiang Dai
Sarvnaz Karimi
Ben Hachey
Cécile Paris
67
35
0
02 Oct 2020
Cross-Lingual Transfer Learning for Complex Word Identification
Cross-Lingual Transfer Learning for Complex Word Identification
George-Eduard Zaharia
Dumitru-Clementin Cercel
M. Dascalu
47
13
0
02 Oct 2020
Which *BERT? A Survey Organizing Contextualized Encoders
Which *BERT? A Survey Organizing Contextualized Encoders
Patrick Xia
Shijie Wu
Benjamin Van Durme
62
50
0
02 Oct 2020
CoLAKE: Contextualized Language and Knowledge Embedding
CoLAKE: Contextualized Language and Knowledge Embedding
Tianxiang Sun
Yunfan Shao
Xipeng Qiu
Qipeng Guo
Yaru Hu
Xuanjing Huang
Zheng Zhang
KELM
111
185
0
01 Oct 2020
An Empirical Investigation Towards Efficient Multi-Domain Language Model
  Pre-training
An Empirical Investigation Towards Efficient Multi-Domain Language Model Pre-training
Kristjan Arumae
Q. Sun
Parminder Bhatia
58
15
0
01 Oct 2020
CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked
  Language Models
CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models
Nikita Nangia
Clara Vania
Rasika Bhalerao
Samuel R. Bowman
161
690
0
30 Sep 2020
Pea-KD: Parameter-efficient and Accurate Knowledge Distillation on BERT
Pea-KD: Parameter-efficient and Accurate Knowledge Distillation on BERT
Ikhyun Cho
U. Kang
25
1
0
30 Sep 2020
TaxiNLI: Taking a Ride up the NLU Hill
TaxiNLI: Taking a Ride up the NLU Hill
Pratik M. Joshi
Somak Aditya
Aalok Sathe
Monojit Choudhury
66
36
0
30 Sep 2020
AUBER: Automated BERT Regularization
AUBER: Automated BERT Regularization
Hyun Dong Lee
Seongmin Lee
U. Kang
38
9
0
30 Sep 2020
Contrastive Distillation on Intermediate Representations for Language
  Model Compression
Contrastive Distillation on Intermediate Representations for Language Model Compression
S. Sun
Zhe Gan
Yu Cheng
Yuwei Fang
Shuohang Wang
Jingjing Liu
VLM
78
73
0
29 Sep 2020
Utility is in the Eye of the User: A Critique of NLP Leaderboards
Utility is in the Eye of the User: A Critique of NLP Leaderboards
Kawin Ethayarajh
Dan Jurafsky
ELM
90
52
0
29 Sep 2020
DialoGLUE: A Natural Language Understanding Benchmark for Task-Oriented
  Dialogue
DialoGLUE: A Natural Language Understanding Benchmark for Task-Oriented Dialogue
Shikib Mehri
Mihail Eric
Dilek Z. Hakkani-Tür
ELM
108
140
0
28 Sep 2020
Domain Adversarial Fine-Tuning as an Effective Regularizer
Domain Adversarial Fine-Tuning as an Effective Regularizer
Giorgos Vernikos
Katerina Margatina
Alexandra Chronopoulou
Ion Androutsopoulos
70
15
0
28 Sep 2020
TernaryBERT: Distillation-aware Ultra-low Bit BERT
TernaryBERT: Distillation-aware Ultra-low Bit BERT
Wei Zhang
Lu Hou
Yichun Yin
Lifeng Shang
Xiao Chen
Xin Jiang
Qun Liu
MQ
101
211
0
27 Sep 2020
BET: A Backtranslation Approach for Easy Data Augmentation in
  Transformer-based Paraphrase Identification Context
BET: A Backtranslation Approach for Easy Data Augmentation in Transformer-based Paraphrase Identification Context
Jean-Philippe Corbeil
Hadi Abdi Ghadivel
40
28
0
25 Sep 2020
Towards Debiasing NLU Models from Unknown Biases
Towards Debiasing NLU Models from Unknown Biases
Prasetya Ajie Utama
N. Moosavi
Iryna Gurevych
119
155
0
25 Sep 2020
RecoBERT: A Catalog Language Model for Text-Based Recommendations
RecoBERT: A Catalog Language Model for Text-Based Recommendations
Itzik Malkiel
Oren Barkan
Avi Caciularu
Noam Razin
Ori Katz
Noam Koenigstein
103
13
0
25 Sep 2020
MinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems
MinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems
Zhaojiang Lin
Andrea Madotto
Genta Indra Winata
Pascale Fung
81
173
0
25 Sep 2020
Hierarchical Pre-training for Sequence Labelling in Spoken Dialog
Hierarchical Pre-training for Sequence Labelling in Spoken Dialog
E. Chapuis
Pierre Colombo
Matteo Manica
Matthieu Labeau
Chloé Clavel
170
59
0
23 Sep 2020
Worst-Case-Aware Curriculum Learning for Zero and Few Shot Transfer
Worst-Case-Aware Curriculum Learning for Zero and Few Shot Transfer
Sheng Zhang
Xin Zhang
Weiming Zhang
Anders Søgaard
VLM
37
10
0
23 Sep 2020
Dataset Cartography: Mapping and Diagnosing Datasets with Training
  Dynamics
Dataset Cartography: Mapping and Diagnosing Datasets with Training Dynamics
Swabha Swayamdipta
Roy Schwartz
Nicholas Lourie
Yizhong Wang
Hannaneh Hajishirzi
Noah A. Smith
Yejin Choi
147
452
0
22 Sep 2020
VirtualFlow: Decoupling Deep Learning Models from the Underlying
  Hardware
VirtualFlow: Decoupling Deep Learning Models from the Underlying Hardware
Andrew Or
Haoyu Zhang
M. Freedman
73
10
0
20 Sep 2020
Repulsive Attention: Rethinking Multi-head Attention as Bayesian
  Inference
Repulsive Attention: Rethinking Multi-head Attention as Bayesian Inference
Bang An
Jie Lyu
Zhenyi Wang
Chunyuan Li
Changwei Hu
Fei Tan
Ruiyi Zhang
Yifan Hu
Changyou Chen
AAML
97
28
0
20 Sep 2020
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning
  in NLP Using Fewer Parameters & Less Data
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning in NLP Using Fewer Parameters & Less Data
Jonathan Pilault
Amine Elhattami
C. Pal
CLLMoE
90
91
0
19 Sep 2020
Self-Supervised Meta-Learning for Few-Shot Natural Language
  Classification Tasks
Self-Supervised Meta-Learning for Few-Shot Natural Language Classification Tasks
Trapit Bansal
Rishikesh Jha
Tsendsuren Munkhdalai
Andrew McCallum
SSLVLM
108
88
0
17 Sep 2020
Efficient Transformer-based Large Scale Language Representations using
  Hardware-friendly Block Structured Pruning
Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Bingbing Li
Zhenglun Kong
Tianyun Zhang
Ji Li
Zechao Li
Hang Liu
Caiwen Ding
VLM
192
65
0
17 Sep 2020
Contextualized Perturbation for Textual Adversarial Attack
Contextualized Perturbation for Textual Adversarial Attack
Dianqi Li
Yizhe Zhang
Hao Peng
Liqun Chen
Chris Brockett
Ming-Ting Sun
Bill Dolan
AAMLSILM
194
236
0
16 Sep 2020
Critical Thinking for Language Models
Critical Thinking for Language Models
Gregor Betz
Christian Voigt
Kyle Richardson
SyDaReLMLRMAI4CE
111
35
0
15 Sep 2020
Real-Time Execution of Large-scale Language Models on Mobile
Real-Time Execution of Large-scale Language Models on Mobile
Wei Niu
Zhenglun Kong
Geng Yuan
Weiwen Jiang
Jiexiong Guan
Caiwen Ding
Pu Zhao
Sijia Liu
Bin Ren
Yanzhi Wang
MQ
62
7
0
15 Sep 2020
Efficient Transformers: A Survey
Efficient Transformers: A Survey
Yi Tay
Mostafa Dehghani
Dara Bahri
Donald Metzler
VLM
230
1,133
0
14 Sep 2020
GeDi: Generative Discriminator Guided Sequence Generation
GeDi: Generative Discriminator Guided Sequence Generation
Ben Krause
Akhilesh Deepak Gotmare
Bryan McCann
N. Keskar
Shafiq Joty
R. Socher
Nazneen Rajani
169
408
0
14 Sep 2020
Margin-Based Regularization and Selective Sampling in Deep Neural
  Networks
Margin-Based Regularization and Selective Sampling in Deep Neural Networks
Berry Weinstein
Shai Fine
Y. Hel-Or
MQ
32
2
0
13 Sep 2020
BARS-CTR: Open Benchmarking for Click-Through Rate Prediction
BARS-CTR: Open Benchmarking for Click-Through Rate Prediction
Jieming Zhu
Jinyang Liu
Shuai Yang
Qi Zhang
Xiuqiang He
96
130
0
12 Sep 2020
Compressed Deep Networks: Goodbye SVD, Hello Robust Low-Rank
  Approximation
Compressed Deep Networks: Goodbye SVD, Hello Robust Low-Rank Approximation
M. Tukan
Alaa Maalouf
Matan Weksler
Dan Feldman
77
9
0
11 Sep 2020
UPB at SemEval-2020 Task 6: Pretrained Language Models for Definition
  Extraction
UPB at SemEval-2020 Task 6: Pretrained Language Models for Definition Extraction
Andrei-Marius Avram
Dumitru-Clementin Cercel
Costin-Gabriel Chiru
31
7
0
11 Sep 2020
A Comparison of LSTM and BERT for Small Corpus
A Comparison of LSTM and BERT for Small Corpus
Aysu Ezen-Can
60
113
0
11 Sep 2020
Multi-Task Learning with Deep Neural Networks: A Survey
Multi-Task Learning with Deep Neural Networks: A Survey
M. Crawshaw
CVBM
220
626
0
10 Sep 2020
Modern Methods for Text Generation
Modern Methods for Text Generation
Dimas Muñoz-Montesinos
24
5
0
10 Sep 2020
Previous
123...777879...878889
Next