Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
1805.12471
Cited By
v1
v2
v3 (latest)
Neural Network Acceptability Judgments
31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Neural Network Acceptability Judgments"
50 / 950 papers shown
Towards Debiasing Sentence Representations
Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Paul Pu Liang
Irene Li
Emily Zheng
Y. Lim
Ruslan Salakhutdinov
Louis-Philippe Morency
220
271
0
16 Jul 2020
Can neural networks acquire a structural bias from raw linguistic data?
Annual Meeting of the Cognitive Science Society (CogSci), 2020
Alex Warstadt
Samuel R. Bowman
AI4CE
200
56
0
14 Jul 2020
HyperGrid: Efficient Multi-Task Transformers with Grid-wise Decomposable Hyper Projections
Yi Tay
Zhe Zhao
Dara Bahri
Donald Metzler
Da-Cheng Juan
179
9
0
12 Jul 2020
Unsupervised Paraphrasing via Deep Reinforcement Learning
A.B. Siddique
Samet Oymak
Vagelis Hristidis
201
60
0
05 Jul 2020
Building Interpretable Interaction Trees for Deep NLP Models
Die Zhang
Huilin Zhou
Hao Zhang
Xiaoyi Bao
Da Huo
Ruizhao Chen
Feng He
Mengyue Wu
Quanshi Zhang
FAtt
131
3
0
29 Jun 2020
MaxVA: Fast Adaptation of Step Sizes by Maximizing Observed Variance of Gradients
Chenfei Zhu
Yu Cheng
Zhe Gan
Furong Huang
Jingjing Liu
Tom Goldstein
ODL
208
2
0
21 Jun 2020
SqueezeBERT: What can computer vision teach NLP about efficient neural networks?
F. Iandola
Albert Eaton Shaw
Ravi Krishna
Kurt Keutzer
VLM
251
136
0
19 Jun 2020
Revisiting Few-sample BERT Fine-tuning
International Conference on Learning Representations (ICLR), 2020
Tianyi Zhang
Felix Wu
Arzoo Katiyar
Kilian Q. Weinberger
Yoav Artzi
407
488
0
10 Jun 2020
On the Stability of Fine-tuning BERT: Misconceptions, Explanations, and Strong Baselines
Marius Mosbach
Maksym Andriushchenko
Dietrich Klakow
452
417
0
08 Jun 2020
BERT Loses Patience: Fast and Robust Inference with Early Exit
Wangchunshu Zhou
Canwen Xu
Tao Ge
Julian McAuley
Ke Xu
Furu Wei
376
399
0
07 Jun 2020
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
Pengcheng He
Xiaodong Liu
Jianfeng Gao
Weizhu Chen
AAML
616
3,407
0
05 Jun 2020
Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing
Zihang Dai
Guokun Lai
Yiming Yang
Quoc V. Le
260
256
0
05 Jun 2020
Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Transactions of the Association for Computational Linguistics (TACL), 2020
A. Kuncoro
Lingpeng Kong
Daniel Fried
Dani Yogatama
Laura Rimell
Chris Dyer
Phil Blunsom
203
36
0
27 May 2020
Common Sense or World Knowledge? Investigating Adapter-Based Knowledge Injection into Pretrained Transformers
Anne Lauscher
Olga Majewska
Leonardo F. R. Ribeiro
Iryna Gurevych
Nikolai Rozanov
Goran Glavaš
KELM
168
87
0
24 May 2020
CERT: Contrastive Self-supervised Learning for Language Understanding
Hongchao Fang
Sicheng Wang
Meng Zhou
Jiayuan Ding
P. Xie
ELM
SSL
213
368
0
16 May 2020
A Systematic Assessment of Syntactic Generalization in Neural Language Models
Jennifer Hu
Jon Gauthier
Peng Qian
Ethan Gotlieb Wilcox
R. Levy
ELM
303
250
0
07 May 2020
Spying on your neighbors: Fine-grained probing of contextual embeddings for information about surrounding words
Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Josef Klafka
Allyson Ettinger
184
47
0
04 May 2020
Intermediate-Task Transfer Learning with Pretrained Models for Natural Language Understanding: When and Why Does It Work?
Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Yada Pruksachatkun
Jason Phang
Haokun Liu
Phu Mon Htut
Xiaoyi Zhang
Richard Yuanzhe Pang
Clara Vania
Katharina Kann
Samuel R. Bowman
CLL
LRM
226
204
0
01 May 2020
When BERT Plays the Lottery, All Tickets Are Winning
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Sai Prasanna
Anna Rogers
Anna Rumshisky
MILM
298
200
0
01 May 2020
Cross-Linguistic Syntactic Evaluation of Word Prediction Models
Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Aaron Mueller
Garrett Nicolai
Panayiota Petrou-Zeniou
N. Talmina
Tal Linzen
238
63
0
01 May 2020
Segatron: Segment-Aware Transformer for Language Modeling and Understanding
Richard He Bai
Peng Shi
Jimmy J. Lin
Yuqing Xie
Luchen Tan
Kun Xiong
Wen Gao
Ming Li
154
8
0
30 Apr 2020
Investigating Transferability in Pretrained Language Models
Findings (Findings), 2020
Alex Tamkin
Trisha Singh
D. Giovanardi
Noah D. Goodman
MILM
140
48
0
30 Apr 2020
TAVAT: Token-Aware Virtual Adversarial Training for Language Understanding
Linyang Li
Xipeng Qiu
161
17
0
30 Apr 2020
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Sanyuan Chen
Yutai Hou
Yiming Cui
Wanxiang Che
Ting Liu
Xiangzhan Yu
KELM
CLL
297
256
0
27 Apr 2020
Masking as an Efficient Alternative to Finetuning for Pretrained Language Models
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Mengjie Zhao
Tao Lin
Fei Mi
Martin Jaggi
Hinrich Schütze
224
126
0
26 Apr 2020
Reevaluating Adversarial Examples in Natural Language
Findings (Findings), 2020
John X. Morris
Eli Lifland
Jack Lanchantin
Yangfeng Ji
Yanjun Qi
SILM
AAML
341
122
0
25 Apr 2020
How fine can fine-tuning be? Learning efficient language models
International Conference on Artificial Intelligence and Statistics (AISTATS), 2020
Evani Radiya-Dixit
Xin Wang
147
73
0
24 Apr 2020
Considering Likelihood in NLP Classification Explanations with Occlusion and Language Modeling
David Harbecke
Christoph Alt
130
14
0
21 Apr 2020
MPNet: Masked and Permuted Pre-training for Language Understanding
Neural Information Processing Systems (NeurIPS), 2020
Kaitao Song
Xu Tan
Tao Qin
Jianfeng Lu
Tie-Yan Liu
240
1,459
0
20 Apr 2020
Adversarial Training for Large Neural Language Models
Xiaodong Liu
Hao Cheng
Pengcheng He
Weizhu Chen
Yu Wang
Hoifung Poon
Jianfeng Gao
AAML
264
204
0
20 Apr 2020
Coreferential Reasoning Learning for Language Representation
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Deming Ye
Yankai Lin
Jiaju Du
Zhenghao Liu
Peng Li
Maosong Sun
Zhiyuan Liu
236
185
0
15 Apr 2020
VGCN-BERT: Augmenting BERT with Graph Embedding for Text Classification
European Conference on Information Retrieval (ECIR), 2020
Zhibin Lu
Pan Du
J. Nie
267
144
0
12 Apr 2020
Frequency, Acceptability, and Selection: A case study of clause-embedding
Glossa (Glossa), 2020
Aaron Steven White
Kyle Rawlins
113
15
0
08 Apr 2020
Improving BERT with Self-Supervised Attention
IEEE Access (IEEE Access), 2020
Yiren Chen
Xiaoyu Kou
Jiangang Bai
Yunhai Tong
123
11
0
08 Apr 2020
CALM: Continuous Adaptive Learning for Language Modeling
Kristjan Arumae
Parminder Bhatia
CLL
KELM
67
7
0
08 Apr 2020
MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices
Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Zhiqing Sun
Hongkun Yu
Xiaodan Song
Renjie Liu
Yiming Yang
Denny Zhou
MQ
402
930
0
06 Apr 2020
How Furiously Can Colourless Green Ideas Sleep? Sentence Acceptability in Context
Transactions of the Association for Computational Linguistics (TACL), 2020
Jey Han Lau
C. S. Armendariz
Shalom Lappin
Matthew Purver
Chang Shu
174
43
0
02 Apr 2020
UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training
International Conference on Machine Learning (ICML), 2020
Hangbo Bao
Li Dong
Furu Wei
Wenhui Wang
Nan Yang
...
Yu Wang
Songhao Piao
Jianfeng Gao
Ming Zhou
H. Hon
AI4CE
182
417
0
28 Feb 2020
MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers
Neural Information Processing Systems (NeurIPS), 2020
Wenhui Wang
Furu Wei
Li Dong
Hangbo Bao
Nan Yang
Ming Zhou
VLM
1.2K
1,742
0
25 Feb 2020
Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation
Journal of Computational Science and Technology (JCST), 2020
Yige Xu
Xipeng Qiu
L. Zhou
Xuanjing Huang
144
73
0
24 Feb 2020
The Microsoft Toolkit of Multi-Task Deep Neural Networks for Natural Language Understanding
Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Xiaodong Liu
Yu Wang
Jianshu Ji
Hao Cheng
Xueyun Zhu
...
Pengcheng He
Weizhu Chen
Hoifung Poon
Guihong Cao
Jianfeng Gao
AI4CE
182
62
0
19 Feb 2020
Fine-Tuning Pretrained Language Models: Weight Initializations, Data Orders, and Early Stopping
Jesse Dodge
Gabriel Ilharco
Roy Schwartz
Ali Farhadi
Hannaneh Hajishirzi
Noah A. Smith
278
674
0
15 Feb 2020
BERT-of-Theseus: Compressing BERT by Progressive Module Replacing
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Canwen Xu
Wangchunshu Zhou
Tao Ge
Furu Wei
Ming Zhou
668
219
0
07 Feb 2020
BLiMP: The Benchmark of Linguistic Minimal Pairs for English
Transactions of the Association for Computational Linguistics (TACL), 2019
Alex Warstadt
Alicia Parrish
Haokun Liu
Anhad Mohananey
Wei Peng
Sheng-Fu Wang
Samuel R. Bowman
465
616
0
02 Dec 2019
Neural language modeling of free word order argument structure
Charlotte Rochereau
Benoît Sagot
Emmanuel Dupoux
134
0
0
30 Nov 2019
Do Attention Heads in BERT Track Syntactic Dependencies?
Phu Mon Htut
Jason Phang
Shikha Bordia
Samuel R. Bowman
232
144
0
27 Nov 2019
Learning to Few-Shot Learn Across Diverse Natural Language Classification Tasks
International Conference on Computational Linguistics (COLING), 2019
Trapit Bansal
Rishikesh Jha
Andrew McCallum
SSL
270
126
0
10 Nov 2019
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
Annual Meeting of the Association for Computational Linguistics (ACL), 2019
Haoming Jiang
Pengcheng He
Weizhu Chen
Xiaodong Liu
Jianfeng Gao
T. Zhao
627
590
0
08 Nov 2019
What Would Elsa Do? Freezing Layers During Transformer Fine-Tuning
Jaejun Lee
Raphael Tang
Jimmy J. Lin
263
138
0
08 Nov 2019
MML: Maximal Multiverse Learning for Robust Fine-Tuning of Language Models
Itzik Malkiel
Lior Wolf
67
2
0
05 Nov 2019
Previous
1
2
3
...
17
18
19
Next