ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.12471
  4. Cited By
Neural Network Acceptability Judgments

Neural Network Acceptability Judgments

31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
ArXivPDFHTML

Papers citing "Neural Network Acceptability Judgments"

50 / 877 papers shown
Title
Exploring the Role of BERT Token Representations to Explain Sentence
  Probing Results
Exploring the Role of BERT Token Representations to Explain Sentence Probing Results
Hosein Mohebbi
Ali Modarressi
Mohammad Taher Pilehvar
MILM
19
23
0
03 Apr 2021
Evaluating the Morphosyntactic Well-formedness of Generated Texts
Evaluating the Morphosyntactic Well-formedness of Generated Texts
Adithya Pratapa
Antonios Anastasopoulos
Shruti Rijhwani
Aditi Chaudhary
David R. Mortensen
Graham Neubig
Yulia Tsvetkov
25
8
0
30 Mar 2021
Retraining DistilBERT for a Voice Shopping Assistant by Using Universal
  Dependencies
Retraining DistilBERT for a Voice Shopping Assistant by Using Universal Dependencies
P. Jayarao
Arpit Sharma
16
2
0
29 Mar 2021
A Practical Survey on Faster and Lighter Transformers
A Practical Survey on Faster and Lighter Transformers
Quentin Fournier
G. Caron
Daniel Aloise
14
93
0
26 Mar 2021
Approximating Instance-Dependent Noise via Instance-Confidence Embedding
Approximating Instance-Dependent Noise via Instance-Confidence Embedding
Yivan Zhang
Masashi Sugiyama
31
8
0
25 Mar 2021
UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New
  Multitask Benchmark
UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New Multitask Benchmark
Nicholas Lourie
Ronan Le Bras
Chandra Bhagavatula
Yejin Choi
LRM
22
137
0
24 Mar 2021
The NLP Cookbook: Modern Recipes for Transformer based Deep Learning
  Architectures
The NLP Cookbook: Modern Recipes for Transformer based Deep Learning Architectures
Sushant Singh
A. Mahmood
AI4TS
60
92
0
23 Mar 2021
Unsupervised Contextual Paraphrase Generation using Lexical Control and
  Reinforcement Learning
Unsupervised Contextual Paraphrase Generation using Lexical Control and Reinforcement Learning
Sonal Garg
Sumanth Prabhu
Hemant Misra
G. Srinivasaraghavan
9
14
0
23 Mar 2021
TAG: Gradient Attack on Transformer-based Language Models
TAG: Gradient Attack on Transformer-based Language Models
Jieren Deng
Yijue Wang
Ji Li
Chao Shang
Hang Liu
Sanguthevar Rajasekaran
Caiwen Ding
FedML
PILM
6
73
0
11 Mar 2021
FairFil: Contrastive Neural Debiasing Method for Pretrained Text
  Encoders
FairFil: Contrastive Neural Debiasing Method for Pretrained Text Encoders
Pengyu Cheng
Weituo Hao
Siyang Yuan
Shijing Si
Lawrence Carin
25
100
0
11 Mar 2021
Split Computing and Early Exiting for Deep Learning Applications: Survey
  and Research Challenges
Split Computing and Early Exiting for Deep Learning Applications: Survey and Research Challenges
Yoshitomo Matsubara
Marco Levorato
Francesco Restuccia
22
199
0
08 Mar 2021
Rissanen Data Analysis: Examining Dataset Characteristics via
  Description Length
Rissanen Data Analysis: Examining Dataset Characteristics via Description Length
Ethan Perez
Douwe Kiela
Kyunghyun Cho
22
24
0
05 Mar 2021
Token-Modification Adversarial Attacks for Natural Language Processing:
  A Survey
Token-Modification Adversarial Attacks for Natural Language Processing: A Survey
Tom Roth
Yansong Gao
A. Abuadbba
Surya Nepal
Wei Liu
AAML
23
12
0
01 Mar 2021
SparseBERT: Rethinking the Importance Analysis in Self-attention
SparseBERT: Rethinking the Importance Analysis in Self-attention
Han Shi
Jiahui Gao
Xiaozhe Ren
Hang Xu
Xiaodan Liang
Zhenguo Li
James T. Kwok
23
54
0
25 Feb 2021
Analyzing Curriculum Learning for Sentiment Analysis along Task
  Difficulty, Pacing and Visualization Axes
Analyzing Curriculum Learning for Sentiment Analysis along Task Difficulty, Pacing and Visualization Axes
Anvesh Rao Vijjini
Kaveri Anuranjana
R. Mamidi
30
2
0
19 Feb 2021
COCO-LM: Correcting and Contrasting Text Sequences for Language Model
  Pretraining
COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Yu Meng
Chenyan Xiong
Payal Bajaj
Saurabh Tiwary
Paul N. Bennett
Jiawei Han
Xia Song
125
202
0
16 Feb 2021
AutoFreeze: Automatically Freezing Model Blocks to Accelerate
  Fine-tuning
AutoFreeze: Automatically Freezing Model Blocks to Accelerate Fine-tuning
Yuhan Liu
Saurabh Agarwal
Shivaram Venkataraman
OffRL
11
53
0
02 Feb 2021
Explaining Natural Language Processing Classifiers with Occlusion and
  Language Modeling
Explaining Natural Language Processing Classifiers with Occlusion and Language Modeling
David Harbecke
AAML
27
2
0
28 Jan 2021
CLiMP: A Benchmark for Chinese Language Model Evaluation
CLiMP: A Benchmark for Chinese Language Model Evaluation
Beilei Xiang
Changbing Yang
Yu Li
Alex Warstadt
Katharina Kann
ALM
17
38
0
26 Jan 2021
Muppet: Massive Multi-task Representations with Pre-Finetuning
Muppet: Massive Multi-task Representations with Pre-Finetuning
Armen Aghajanyan
Anchit Gupta
Akshat Shrivastava
Xilun Chen
Luke Zettlemoyer
Sonal Gupta
22
266
0
26 Jan 2021
WER-BERT: Automatic WER Estimation with BERT in a Balanced Ordinal
  Classification Paradigm
WER-BERT: Automatic WER Estimation with BERT in a Balanced Ordinal Classification Paradigm
Akshay Krishna Sheshadri
Anvesh Rao Vijjini
S. Kharbanda
11
8
0
14 Jan 2021
I-BERT: Integer-only BERT Quantization
I-BERT: Integer-only BERT Quantization
Sehoon Kim
A. Gholami
Z. Yao
Michael W. Mahoney
Kurt Keutzer
MQ
96
341
0
05 Jan 2021
WARP: Word-level Adversarial ReProgramming
WARP: Word-level Adversarial ReProgramming
Karen Hambardzumyan
Hrant Khachatrian
Jonathan May
AAML
254
342
0
01 Jan 2021
MiniLMv2: Multi-Head Self-Attention Relation Distillation for
  Compressing Pretrained Transformers
MiniLMv2: Multi-Head Self-Attention Relation Distillation for Compressing Pretrained Transformers
Wenhui Wang
Hangbo Bao
Shaohan Huang
Li Dong
Furu Wei
MQ
19
257
0
31 Dec 2020
Making Pre-trained Language Models Better Few-shot Learners
Making Pre-trained Language Models Better Few-shot Learners
Tianyu Gao
Adam Fisch
Danqi Chen
241
1,919
0
31 Dec 2020
Out of Order: How Important Is The Sequential Order of Words in a
  Sentence in Natural Language Understanding Tasks?
Out of Order: How Important Is The Sequential Order of Words in a Sentence in Natural Language Understanding Tasks?
Thang M. Pham
Trung Bui
Long Mai
Anh Totti Nguyen
214
122
0
30 Dec 2020
Accurate Word Representations with Universal Visual Guidance
Accurate Word Representations with Universal Visual Guidance
Zhuosheng Zhang
Haojie Yu
Hai Zhao
Rui-cang Wang
Masao Utiyama
19
0
0
30 Dec 2020
BURT: BERT-inspired Universal Representation from Learning Meaningful
  Segment
BURT: BERT-inspired Universal Representation from Learning Meaningful Segment
Yian Li
Hai Zhao
SSL
11
0
0
28 Dec 2020
ALP-KD: Attention-Based Layer Projection for Knowledge Distillation
ALP-KD: Attention-Based Layer Projection for Knowledge Distillation
Peyman Passban
Yimeng Wu
Mehdi Rezagholizadeh
Qun Liu
11
122
0
27 Dec 2020
Pre-Training Transformers as Energy-Based Cloze Models
Pre-Training Transformers as Energy-Based Cloze Models
Kevin Clark
Minh-Thang Luong
Quoc V. Le
Christopher D. Manning
15
78
0
15 Dec 2020
LRC-BERT: Latent-representation Contrastive Knowledge Distillation for
  Natural Language Understanding
LRC-BERT: Latent-representation Contrastive Knowledge Distillation for Natural Language Understanding
Hao Fu
Shaojun Zhou
Qihong Yang
Junjie Tang
Guiquan Liu
Kaikui Liu
Xiaolong Li
37
57
0
14 Dec 2020
Infusing Finetuning with Semantic Dependencies
Infusing Finetuning with Semantic Dependencies
Zhaofeng Wu
Hao Peng
Noah A. Smith
19
36
0
10 Dec 2020
Mapping the Space of Chemical Reactions Using Attention-Based Neural
  Networks
Mapping the Space of Chemical Reactions Using Attention-Based Neural Networks
P. Schwaller
Daniel Probst
Alain C. Vaucher
Vishnu H. Nair
D. Kreutter
Teodoro Laino
J. Reymond
144
224
0
09 Dec 2020
Towards Coinductive Models for Natural Language Understanding. Bringing
  together Deep Learning and Deep Semantics
Towards Coinductive Models for Natural Language Understanding. Bringing together Deep Learning and Deep Semantics
Wlodek Zadrozny
AI4CE
17
1
0
09 Dec 2020
An Investigation of Language Model Interpretability via Sentence Editing
An Investigation of Language Model Interpretability via Sentence Editing
Samuel Stevens
Yu-Chuan Su
LRM
11
8
0
28 Nov 2020
CoRe: An Efficient Coarse-refined Training Framework for BERT
CoRe: An Efficient Coarse-refined Training Framework for BERT
Cheng Yang
Shengnan Wang
Yuechuan Li
Chao Yang
Ming Yan
Jingqiao Zhang
Fangquan Lin
12
0
0
27 Nov 2020
AGenT Zero: Zero-shot Automatic Multiple-Choice Question Generation for
  Skill Assessments
AGenT Zero: Zero-shot Automatic Multiple-Choice Question Generation for Skill Assessments
Eric Li
Jingyi Su
Hao Sheng
Lawrence Wai
14
2
0
25 Nov 2020
Data-Informed Global Sparseness in Attention Mechanisms for Deep Neural
  Networks
Data-Informed Global Sparseness in Attention Mechanisms for Deep Neural Networks
Ileana Rugina
Rumen Dangovski
L. Jing
Preslav Nakov
Marin Soljacic
20
0
0
20 Nov 2020
Mixing ADAM and SGD: a Combined Optimization Method
Mixing ADAM and SGD: a Combined Optimization Method
Nicola Landro
I. Gallo
Riccardo La Grassa
ODL
6
23
0
16 Nov 2020
Know What You Don't Need: Single-Shot Meta-Pruning for Attention Heads
Know What You Don't Need: Single-Shot Meta-Pruning for Attention Heads
Zhengyan Zhang
Fanchao Qi
Zhiyuan Liu
Qun Liu
Maosong Sun
VLM
36
30
0
07 Nov 2020
CharBERT: Character-aware Pre-trained Language Model
CharBERT: Character-aware Pre-trained Language Model
Wentao Ma
Yiming Cui
Chenglei Si
Ting Liu
Shijin Wang
Guoping Hu
20
104
0
03 Nov 2020
Word Frequency Does Not Predict Grammatical Knowledge in Language Models
Word Frequency Does Not Predict Grammatical Knowledge in Language Models
Charles Yu
Ryan Sie
Nicolas Tedeschi
Leon Bergen
9
3
0
26 Oct 2020
ANLIzing the Adversarial Natural Language Inference Dataset
ANLIzing the Adversarial Natural Language Inference Dataset
Adina Williams
Tristan Thrush
Douwe Kiela
AAML
168
46
0
24 Oct 2020
Improving Classification through Weak Supervision in Context-specific
  Conversational Agent Development for Teacher Education
Improving Classification through Weak Supervision in Context-specific Conversational Agent Development for Teacher Education
Debajyoti Datta
Maria Phillips
Jennifer L. Chiu
G. Watson
J. Bywater
Laura E. Barnes
Donald E. Brown
15
7
0
23 Oct 2020
ERNIE-Gram: Pre-Training with Explicitly N-Gram Masked Language Modeling
  for Natural Language Understanding
ERNIE-Gram: Pre-Training with Explicitly N-Gram Masked Language Modeling for Natural Language Understanding
Dongling Xiao
Yukun Li
Han Zhang
Yu Sun
Hao Tian
Hua-Hong Wu
Haifeng Wang
19
38
0
23 Oct 2020
Detecting and Exorcising Statistical Demons from Language Models with
  Anti-Models of Negative Data
Detecting and Exorcising Statistical Demons from Language Models with Anti-Models of Negative Data
Michael L. Wick
Kate Silverstein
Jean-Baptiste Tristan
Adam Craig Pocock
Mark Johnson
17
3
0
22 Oct 2020
Towards Fully Bilingual Deep Language Modeling
Towards Fully Bilingual Deep Language Modeling
Li-Hsin Chang
S. Pyysalo
Jenna Kanerva
Filip Ginter
26
3
0
22 Oct 2020
Bayesian Attention Modules
Bayesian Attention Modules
Xinjie Fan
Shujian Zhang
Bo Chen
Mingyuan Zhou
111
59
0
20 Oct 2020
Optimal Subarchitecture Extraction For BERT
Optimal Subarchitecture Extraction For BERT
Adrian de Wynter
Daniel J. Perry
MQ
43
18
0
20 Oct 2020
TweetBERT: A Pretrained Language Representation Model for Twitter Text
  Analysis
TweetBERT: A Pretrained Language Representation Model for Twitter Text Analysis
Mohiuddin Md Abdul Qudar
Vijay K. Mago
SSeg
15
34
0
17 Oct 2020
Previous
123...1415161718
Next