Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1805.12471
Cited By
Neural Network Acceptability Judgments
31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Neural Network Acceptability Judgments"
50 / 877 papers shown
Title
Exploring the Role of BERT Token Representations to Explain Sentence Probing Results
Hosein Mohebbi
Ali Modarressi
Mohammad Taher Pilehvar
MILM
19
23
0
03 Apr 2021
Evaluating the Morphosyntactic Well-formedness of Generated Texts
Adithya Pratapa
Antonios Anastasopoulos
Shruti Rijhwani
Aditi Chaudhary
David R. Mortensen
Graham Neubig
Yulia Tsvetkov
25
8
0
30 Mar 2021
Retraining DistilBERT for a Voice Shopping Assistant by Using Universal Dependencies
P. Jayarao
Arpit Sharma
16
2
0
29 Mar 2021
A Practical Survey on Faster and Lighter Transformers
Quentin Fournier
G. Caron
Daniel Aloise
14
93
0
26 Mar 2021
Approximating Instance-Dependent Noise via Instance-Confidence Embedding
Yivan Zhang
Masashi Sugiyama
31
8
0
25 Mar 2021
UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New Multitask Benchmark
Nicholas Lourie
Ronan Le Bras
Chandra Bhagavatula
Yejin Choi
LRM
22
137
0
24 Mar 2021
The NLP Cookbook: Modern Recipes for Transformer based Deep Learning Architectures
Sushant Singh
A. Mahmood
AI4TS
60
92
0
23 Mar 2021
Unsupervised Contextual Paraphrase Generation using Lexical Control and Reinforcement Learning
Sonal Garg
Sumanth Prabhu
Hemant Misra
G. Srinivasaraghavan
9
14
0
23 Mar 2021
TAG: Gradient Attack on Transformer-based Language Models
Jieren Deng
Yijue Wang
Ji Li
Chao Shang
Hang Liu
Sanguthevar Rajasekaran
Caiwen Ding
FedML
PILM
6
73
0
11 Mar 2021
FairFil: Contrastive Neural Debiasing Method for Pretrained Text Encoders
Pengyu Cheng
Weituo Hao
Siyang Yuan
Shijing Si
Lawrence Carin
25
100
0
11 Mar 2021
Split Computing and Early Exiting for Deep Learning Applications: Survey and Research Challenges
Yoshitomo Matsubara
Marco Levorato
Francesco Restuccia
22
199
0
08 Mar 2021
Rissanen Data Analysis: Examining Dataset Characteristics via Description Length
Ethan Perez
Douwe Kiela
Kyunghyun Cho
22
24
0
05 Mar 2021
Token-Modification Adversarial Attacks for Natural Language Processing: A Survey
Tom Roth
Yansong Gao
A. Abuadbba
Surya Nepal
Wei Liu
AAML
23
12
0
01 Mar 2021
SparseBERT: Rethinking the Importance Analysis in Self-attention
Han Shi
Jiahui Gao
Xiaozhe Ren
Hang Xu
Xiaodan Liang
Zhenguo Li
James T. Kwok
23
54
0
25 Feb 2021
Analyzing Curriculum Learning for Sentiment Analysis along Task Difficulty, Pacing and Visualization Axes
Anvesh Rao Vijjini
Kaveri Anuranjana
R. Mamidi
30
2
0
19 Feb 2021
COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Yu Meng
Chenyan Xiong
Payal Bajaj
Saurabh Tiwary
Paul N. Bennett
Jiawei Han
Xia Song
125
202
0
16 Feb 2021
AutoFreeze: Automatically Freezing Model Blocks to Accelerate Fine-tuning
Yuhan Liu
Saurabh Agarwal
Shivaram Venkataraman
OffRL
11
53
0
02 Feb 2021
Explaining Natural Language Processing Classifiers with Occlusion and Language Modeling
David Harbecke
AAML
27
2
0
28 Jan 2021
CLiMP: A Benchmark for Chinese Language Model Evaluation
Beilei Xiang
Changbing Yang
Yu Li
Alex Warstadt
Katharina Kann
ALM
17
38
0
26 Jan 2021
Muppet: Massive Multi-task Representations with Pre-Finetuning
Armen Aghajanyan
Anchit Gupta
Akshat Shrivastava
Xilun Chen
Luke Zettlemoyer
Sonal Gupta
22
266
0
26 Jan 2021
WER-BERT: Automatic WER Estimation with BERT in a Balanced Ordinal Classification Paradigm
Akshay Krishna Sheshadri
Anvesh Rao Vijjini
S. Kharbanda
11
8
0
14 Jan 2021
I-BERT: Integer-only BERT Quantization
Sehoon Kim
A. Gholami
Z. Yao
Michael W. Mahoney
Kurt Keutzer
MQ
96
341
0
05 Jan 2021
WARP: Word-level Adversarial ReProgramming
Karen Hambardzumyan
Hrant Khachatrian
Jonathan May
AAML
254
342
0
01 Jan 2021
MiniLMv2: Multi-Head Self-Attention Relation Distillation for Compressing Pretrained Transformers
Wenhui Wang
Hangbo Bao
Shaohan Huang
Li Dong
Furu Wei
MQ
19
257
0
31 Dec 2020
Making Pre-trained Language Models Better Few-shot Learners
Tianyu Gao
Adam Fisch
Danqi Chen
241
1,919
0
31 Dec 2020
Out of Order: How Important Is The Sequential Order of Words in a Sentence in Natural Language Understanding Tasks?
Thang M. Pham
Trung Bui
Long Mai
Anh Totti Nguyen
214
122
0
30 Dec 2020
Accurate Word Representations with Universal Visual Guidance
Zhuosheng Zhang
Haojie Yu
Hai Zhao
Rui-cang Wang
Masao Utiyama
19
0
0
30 Dec 2020
BURT: BERT-inspired Universal Representation from Learning Meaningful Segment
Yian Li
Hai Zhao
SSL
11
0
0
28 Dec 2020
ALP-KD: Attention-Based Layer Projection for Knowledge Distillation
Peyman Passban
Yimeng Wu
Mehdi Rezagholizadeh
Qun Liu
11
122
0
27 Dec 2020
Pre-Training Transformers as Energy-Based Cloze Models
Kevin Clark
Minh-Thang Luong
Quoc V. Le
Christopher D. Manning
15
78
0
15 Dec 2020
LRC-BERT: Latent-representation Contrastive Knowledge Distillation for Natural Language Understanding
Hao Fu
Shaojun Zhou
Qihong Yang
Junjie Tang
Guiquan Liu
Kaikui Liu
Xiaolong Li
37
57
0
14 Dec 2020
Infusing Finetuning with Semantic Dependencies
Zhaofeng Wu
Hao Peng
Noah A. Smith
19
36
0
10 Dec 2020
Mapping the Space of Chemical Reactions Using Attention-Based Neural Networks
P. Schwaller
Daniel Probst
Alain C. Vaucher
Vishnu H. Nair
D. Kreutter
Teodoro Laino
J. Reymond
144
224
0
09 Dec 2020
Towards Coinductive Models for Natural Language Understanding. Bringing together Deep Learning and Deep Semantics
Wlodek Zadrozny
AI4CE
17
1
0
09 Dec 2020
An Investigation of Language Model Interpretability via Sentence Editing
Samuel Stevens
Yu-Chuan Su
LRM
11
8
0
28 Nov 2020
CoRe: An Efficient Coarse-refined Training Framework for BERT
Cheng Yang
Shengnan Wang
Yuechuan Li
Chao Yang
Ming Yan
Jingqiao Zhang
Fangquan Lin
12
0
0
27 Nov 2020
AGenT Zero: Zero-shot Automatic Multiple-Choice Question Generation for Skill Assessments
Eric Li
Jingyi Su
Hao Sheng
Lawrence Wai
14
2
0
25 Nov 2020
Data-Informed Global Sparseness in Attention Mechanisms for Deep Neural Networks
Ileana Rugina
Rumen Dangovski
L. Jing
Preslav Nakov
Marin Soljacic
20
0
0
20 Nov 2020
Mixing ADAM and SGD: a Combined Optimization Method
Nicola Landro
I. Gallo
Riccardo La Grassa
ODL
6
23
0
16 Nov 2020
Know What You Don't Need: Single-Shot Meta-Pruning for Attention Heads
Zhengyan Zhang
Fanchao Qi
Zhiyuan Liu
Qun Liu
Maosong Sun
VLM
36
30
0
07 Nov 2020
CharBERT: Character-aware Pre-trained Language Model
Wentao Ma
Yiming Cui
Chenglei Si
Ting Liu
Shijin Wang
Guoping Hu
20
104
0
03 Nov 2020
Word Frequency Does Not Predict Grammatical Knowledge in Language Models
Charles Yu
Ryan Sie
Nicolas Tedeschi
Leon Bergen
9
3
0
26 Oct 2020
ANLIzing the Adversarial Natural Language Inference Dataset
Adina Williams
Tristan Thrush
Douwe Kiela
AAML
168
46
0
24 Oct 2020
Improving Classification through Weak Supervision in Context-specific Conversational Agent Development for Teacher Education
Debajyoti Datta
Maria Phillips
Jennifer L. Chiu
G. Watson
J. Bywater
Laura E. Barnes
Donald E. Brown
15
7
0
23 Oct 2020
ERNIE-Gram: Pre-Training with Explicitly N-Gram Masked Language Modeling for Natural Language Understanding
Dongling Xiao
Yukun Li
Han Zhang
Yu Sun
Hao Tian
Hua-Hong Wu
Haifeng Wang
19
38
0
23 Oct 2020
Detecting and Exorcising Statistical Demons from Language Models with Anti-Models of Negative Data
Michael L. Wick
Kate Silverstein
Jean-Baptiste Tristan
Adam Craig Pocock
Mark Johnson
17
3
0
22 Oct 2020
Towards Fully Bilingual Deep Language Modeling
Li-Hsin Chang
S. Pyysalo
Jenna Kanerva
Filip Ginter
26
3
0
22 Oct 2020
Bayesian Attention Modules
Xinjie Fan
Shujian Zhang
Bo Chen
Mingyuan Zhou
111
59
0
20 Oct 2020
Optimal Subarchitecture Extraction For BERT
Adrian de Wynter
Daniel J. Perry
MQ
43
18
0
20 Oct 2020
TweetBERT: A Pretrained Language Representation Model for Twitter Text Analysis
Mohiuddin Md Abdul Qudar
Vijay K. Mago
SSeg
15
34
0
17 Oct 2020
Previous
1
2
3
...
14
15
16
17
18
Next