ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.12471
  4. Cited By
Neural Network Acceptability Judgments
v1v2v3 (latest)

Neural Network Acceptability Judgments

31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
ArXiv (abs)PDFHTML

Papers citing "Neural Network Acceptability Judgments"

50 / 950 papers shown
Towards Debiasing Sentence Representations
Towards Debiasing Sentence RepresentationsAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Paul Pu Liang
Irene Li
Emily Zheng
Y. Lim
Ruslan Salakhutdinov
Louis-Philippe Morency
220
271
0
16 Jul 2020
Can neural networks acquire a structural bias from raw linguistic data?
Can neural networks acquire a structural bias from raw linguistic data?Annual Meeting of the Cognitive Science Society (CogSci), 2020
Alex Warstadt
Samuel R. Bowman
AI4CE
200
56
0
14 Jul 2020
HyperGrid: Efficient Multi-Task Transformers with Grid-wise Decomposable
  Hyper Projections
HyperGrid: Efficient Multi-Task Transformers with Grid-wise Decomposable Hyper Projections
Yi Tay
Zhe Zhao
Dara Bahri
Donald Metzler
Da-Cheng Juan
179
9
0
12 Jul 2020
Unsupervised Paraphrasing via Deep Reinforcement Learning
Unsupervised Paraphrasing via Deep Reinforcement Learning
A.B. Siddique
Samet Oymak
Vagelis Hristidis
201
60
0
05 Jul 2020
Building Interpretable Interaction Trees for Deep NLP Models
Building Interpretable Interaction Trees for Deep NLP Models
Die Zhang
Huilin Zhou
Hao Zhang
Xiaoyi Bao
Da Huo
Ruizhao Chen
Feng He
Mengyue Wu
Quanshi Zhang
FAtt
131
3
0
29 Jun 2020
MaxVA: Fast Adaptation of Step Sizes by Maximizing Observed Variance of
  Gradients
MaxVA: Fast Adaptation of Step Sizes by Maximizing Observed Variance of Gradients
Chenfei Zhu
Yu Cheng
Zhe Gan
Furong Huang
Jingjing Liu
Tom Goldstein
ODL
208
2
0
21 Jun 2020
SqueezeBERT: What can computer vision teach NLP about efficient neural
  networks?
SqueezeBERT: What can computer vision teach NLP about efficient neural networks?
F. Iandola
Albert Eaton Shaw
Ravi Krishna
Kurt Keutzer
VLM
251
136
0
19 Jun 2020
Revisiting Few-sample BERT Fine-tuning
Revisiting Few-sample BERT Fine-tuningInternational Conference on Learning Representations (ICLR), 2020
Tianyi Zhang
Felix Wu
Arzoo Katiyar
Kilian Q. Weinberger
Yoav Artzi
407
488
0
10 Jun 2020
On the Stability of Fine-tuning BERT: Misconceptions, Explanations, and
  Strong Baselines
On the Stability of Fine-tuning BERT: Misconceptions, Explanations, and Strong Baselines
Marius Mosbach
Maksym Andriushchenko
Dietrich Klakow
452
417
0
08 Jun 2020
BERT Loses Patience: Fast and Robust Inference with Early Exit
BERT Loses Patience: Fast and Robust Inference with Early Exit
Wangchunshu Zhou
Canwen Xu
Tao Ge
Julian McAuley
Ke Xu
Furu Wei
376
399
0
07 Jun 2020
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
Pengcheng He
Xiaodong Liu
Jianfeng Gao
Weizhu Chen
AAML
616
3,407
0
05 Jun 2020
Funnel-Transformer: Filtering out Sequential Redundancy for Efficient
  Language Processing
Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing
Zihang Dai
Guokun Lai
Yiming Yang
Quoc V. Le
260
256
0
05 Jun 2020
Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Syntactic Structure Distillation Pretraining For Bidirectional EncodersTransactions of the Association for Computational Linguistics (TACL), 2020
A. Kuncoro
Lingpeng Kong
Daniel Fried
Dani Yogatama
Laura Rimell
Chris Dyer
Phil Blunsom
203
36
0
27 May 2020
Common Sense or World Knowledge? Investigating Adapter-Based Knowledge
  Injection into Pretrained Transformers
Common Sense or World Knowledge? Investigating Adapter-Based Knowledge Injection into Pretrained Transformers
Anne Lauscher
Olga Majewska
Leonardo F. R. Ribeiro
Iryna Gurevych
Nikolai Rozanov
Goran Glavaš
KELM
168
87
0
24 May 2020
CERT: Contrastive Self-supervised Learning for Language Understanding
CERT: Contrastive Self-supervised Learning for Language Understanding
Hongchao Fang
Sicheng Wang
Meng Zhou
Jiayuan Ding
P. Xie
ELMSSL
213
368
0
16 May 2020
A Systematic Assessment of Syntactic Generalization in Neural Language
  Models
A Systematic Assessment of Syntactic Generalization in Neural Language Models
Jennifer Hu
Jon Gauthier
Peng Qian
Ethan Gotlieb Wilcox
R. Levy
ELM
303
250
0
07 May 2020
Spying on your neighbors: Fine-grained probing of contextual embeddings
  for information about surrounding words
Spying on your neighbors: Fine-grained probing of contextual embeddings for information about surrounding wordsAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Josef Klafka
Allyson Ettinger
184
47
0
04 May 2020
Intermediate-Task Transfer Learning with Pretrained Models for Natural
  Language Understanding: When and Why Does It Work?
Intermediate-Task Transfer Learning with Pretrained Models for Natural Language Understanding: When and Why Does It Work?Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Yada Pruksachatkun
Jason Phang
Haokun Liu
Phu Mon Htut
Xiaoyi Zhang
Richard Yuanzhe Pang
Clara Vania
Katharina Kann
Samuel R. Bowman
CLLLRM
226
204
0
01 May 2020
When BERT Plays the Lottery, All Tickets Are Winning
When BERT Plays the Lottery, All Tickets Are WinningConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Sai Prasanna
Anna Rogers
Anna Rumshisky
MILM
298
200
0
01 May 2020
Cross-Linguistic Syntactic Evaluation of Word Prediction Models
Cross-Linguistic Syntactic Evaluation of Word Prediction ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Aaron Mueller
Garrett Nicolai
Panayiota Petrou-Zeniou
N. Talmina
Tal Linzen
238
63
0
01 May 2020
Segatron: Segment-Aware Transformer for Language Modeling and
  Understanding
Segatron: Segment-Aware Transformer for Language Modeling and Understanding
Richard He Bai
Peng Shi
Jimmy J. Lin
Yuqing Xie
Luchen Tan
Kun Xiong
Wen Gao
Ming Li
154
8
0
30 Apr 2020
Investigating Transferability in Pretrained Language Models
Investigating Transferability in Pretrained Language ModelsFindings (Findings), 2020
Alex Tamkin
Trisha Singh
D. Giovanardi
Noah D. Goodman
MILM
140
48
0
30 Apr 2020
TAVAT: Token-Aware Virtual Adversarial Training for Language
  Understanding
TAVAT: Token-Aware Virtual Adversarial Training for Language Understanding
Linyang Li
Xipeng Qiu
161
17
0
30 Apr 2020
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less
  Forgetting
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less ForgettingConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Sanyuan Chen
Yutai Hou
Yiming Cui
Wanxiang Che
Ting Liu
Xiangzhan Yu
KELMCLL
297
256
0
27 Apr 2020
Masking as an Efficient Alternative to Finetuning for Pretrained
  Language Models
Masking as an Efficient Alternative to Finetuning for Pretrained Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Mengjie Zhao
Tao Lin
Fei Mi
Martin Jaggi
Hinrich Schütze
224
126
0
26 Apr 2020
Reevaluating Adversarial Examples in Natural Language
Reevaluating Adversarial Examples in Natural LanguageFindings (Findings), 2020
John X. Morris
Eli Lifland
Jack Lanchantin
Yangfeng Ji
Yanjun Qi
SILMAAML
341
122
0
25 Apr 2020
How fine can fine-tuning be? Learning efficient language models
How fine can fine-tuning be? Learning efficient language modelsInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2020
Evani Radiya-Dixit
Xin Wang
147
73
0
24 Apr 2020
Considering Likelihood in NLP Classification Explanations with Occlusion
  and Language Modeling
Considering Likelihood in NLP Classification Explanations with Occlusion and Language Modeling
David Harbecke
Christoph Alt
130
14
0
21 Apr 2020
MPNet: Masked and Permuted Pre-training for Language Understanding
MPNet: Masked and Permuted Pre-training for Language UnderstandingNeural Information Processing Systems (NeurIPS), 2020
Kaitao Song
Xu Tan
Tao Qin
Jianfeng Lu
Tie-Yan Liu
240
1,459
0
20 Apr 2020
Adversarial Training for Large Neural Language Models
Adversarial Training for Large Neural Language Models
Xiaodong Liu
Hao Cheng
Pengcheng He
Weizhu Chen
Yu Wang
Hoifung Poon
Jianfeng Gao
AAML
264
204
0
20 Apr 2020
Coreferential Reasoning Learning for Language Representation
Coreferential Reasoning Learning for Language RepresentationConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Deming Ye
Yankai Lin
Jiaju Du
Zhenghao Liu
Peng Li
Maosong Sun
Zhiyuan Liu
236
185
0
15 Apr 2020
VGCN-BERT: Augmenting BERT with Graph Embedding for Text Classification
VGCN-BERT: Augmenting BERT with Graph Embedding for Text ClassificationEuropean Conference on Information Retrieval (ECIR), 2020
Zhibin Lu
Pan Du
J. Nie
267
144
0
12 Apr 2020
Frequency, Acceptability, and Selection: A case study of
  clause-embedding
Frequency, Acceptability, and Selection: A case study of clause-embeddingGlossa (Glossa), 2020
Aaron Steven White
Kyle Rawlins
113
15
0
08 Apr 2020
Improving BERT with Self-Supervised Attention
Improving BERT with Self-Supervised AttentionIEEE Access (IEEE Access), 2020
Yiren Chen
Xiaoyu Kou
Jiangang Bai
Yunhai Tong
123
11
0
08 Apr 2020
CALM: Continuous Adaptive Learning for Language Modeling
CALM: Continuous Adaptive Learning for Language Modeling
Kristjan Arumae
Parminder Bhatia
CLLKELM
67
7
0
08 Apr 2020
MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices
MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited DevicesAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Zhiqing Sun
Hongkun Yu
Xiaodan Song
Renjie Liu
Yiming Yang
Denny Zhou
MQ
402
930
0
06 Apr 2020
How Furiously Can Colourless Green Ideas Sleep? Sentence Acceptability
  in Context
How Furiously Can Colourless Green Ideas Sleep? Sentence Acceptability in ContextTransactions of the Association for Computational Linguistics (TACL), 2020
Jey Han Lau
C. S. Armendariz
Shalom Lappin
Matthew Purver
Chang Shu
174
43
0
02 Apr 2020
UniLMv2: Pseudo-Masked Language Models for Unified Language Model
  Pre-Training
UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-TrainingInternational Conference on Machine Learning (ICML), 2020
Hangbo Bao
Li Dong
Furu Wei
Wenhui Wang
Nan Yang
...
Yu Wang
Songhao Piao
Jianfeng Gao
Ming Zhou
H. Hon
AI4CE
182
417
0
28 Feb 2020
MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression
  of Pre-Trained Transformers
MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained TransformersNeural Information Processing Systems (NeurIPS), 2020
Wenhui Wang
Furu Wei
Li Dong
Hangbo Bao
Nan Yang
Ming Zhou
VLM
1.2K
1,742
0
25 Feb 2020
Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation
Improving BERT Fine-Tuning via Self-Ensemble and Self-DistillationJournal of Computational Science and Technology (JCST), 2020
Yige Xu
Xipeng Qiu
L. Zhou
Xuanjing Huang
144
73
0
24 Feb 2020
The Microsoft Toolkit of Multi-Task Deep Neural Networks for Natural
  Language Understanding
The Microsoft Toolkit of Multi-Task Deep Neural Networks for Natural Language UnderstandingAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Xiaodong Liu
Yu Wang
Jianshu Ji
Hao Cheng
Xueyun Zhu
...
Pengcheng He
Weizhu Chen
Hoifung Poon
Guihong Cao
Jianfeng Gao
AI4CE
182
62
0
19 Feb 2020
Fine-Tuning Pretrained Language Models: Weight Initializations, Data
  Orders, and Early Stopping
Fine-Tuning Pretrained Language Models: Weight Initializations, Data Orders, and Early Stopping
Jesse Dodge
Gabriel Ilharco
Roy Schwartz
Ali Farhadi
Hannaneh Hajishirzi
Noah A. Smith
278
674
0
15 Feb 2020
BERT-of-Theseus: Compressing BERT by Progressive Module Replacing
BERT-of-Theseus: Compressing BERT by Progressive Module ReplacingConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Canwen Xu
Wangchunshu Zhou
Tao Ge
Furu Wei
Ming Zhou
668
219
0
07 Feb 2020
BLiMP: The Benchmark of Linguistic Minimal Pairs for English
BLiMP: The Benchmark of Linguistic Minimal Pairs for EnglishTransactions of the Association for Computational Linguistics (TACL), 2019
Alex Warstadt
Alicia Parrish
Haokun Liu
Anhad Mohananey
Wei Peng
Sheng-Fu Wang
Samuel R. Bowman
465
616
0
02 Dec 2019
Neural language modeling of free word order argument structure
Neural language modeling of free word order argument structure
Charlotte Rochereau
Benoît Sagot
Emmanuel Dupoux
134
0
0
30 Nov 2019
Do Attention Heads in BERT Track Syntactic Dependencies?
Do Attention Heads in BERT Track Syntactic Dependencies?
Phu Mon Htut
Jason Phang
Shikha Bordia
Samuel R. Bowman
232
144
0
27 Nov 2019
Learning to Few-Shot Learn Across Diverse Natural Language
  Classification Tasks
Learning to Few-Shot Learn Across Diverse Natural Language Classification TasksInternational Conference on Computational Linguistics (COLING), 2019
Trapit Bansal
Rishikesh Jha
Andrew McCallum
SSL
270
126
0
10 Nov 2019
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language
  Models through Principled Regularized Optimization
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized OptimizationAnnual Meeting of the Association for Computational Linguistics (ACL), 2019
Haoming Jiang
Pengcheng He
Weizhu Chen
Xiaodong Liu
Jianfeng Gao
T. Zhao
627
590
0
08 Nov 2019
What Would Elsa Do? Freezing Layers During Transformer Fine-Tuning
What Would Elsa Do? Freezing Layers During Transformer Fine-Tuning
Jaejun Lee
Raphael Tang
Jimmy J. Lin
263
138
0
08 Nov 2019
MML: Maximal Multiverse Learning for Robust Fine-Tuning of Language
  Models
MML: Maximal Multiverse Learning for Robust Fine-Tuning of Language Models
Itzik Malkiel
Lior Wolf
67
2
0
05 Nov 2019
Previous
123...171819
Next