ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.12471
  4. Cited By
Neural Network Acceptability Judgments

Neural Network Acceptability Judgments

31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
ArXivPDFHTML

Papers citing "Neural Network Acceptability Judgments"

50 / 877 papers shown
Title
Masking as an Efficient Alternative to Finetuning for Pretrained
  Language Models
Masking as an Efficient Alternative to Finetuning for Pretrained Language Models
Mengjie Zhao
Tao R. Lin
Fei Mi
Martin Jaggi
Hinrich Schütze
22
119
0
26 Apr 2020
Reevaluating Adversarial Examples in Natural Language
Reevaluating Adversarial Examples in Natural Language
John X. Morris
Eli Lifland
Jack Lanchantin
Yangfeng Ji
Yanjun Qi
SILM
AAML
20
111
0
25 Apr 2020
How fine can fine-tuning be? Learning efficient language models
How fine can fine-tuning be? Learning efficient language models
Evani Radiya-Dixit
Xin Wang
6
63
0
24 Apr 2020
Considering Likelihood in NLP Classification Explanations with Occlusion
  and Language Modeling
Considering Likelihood in NLP Classification Explanations with Occlusion and Language Modeling
David Harbecke
Christoph Alt
12
10
0
21 Apr 2020
MPNet: Masked and Permuted Pre-training for Language Understanding
MPNet: Masked and Permuted Pre-training for Language Understanding
Kaitao Song
Xu Tan
Tao Qin
Jianfeng Lu
Tie-Yan Liu
28
1,073
0
20 Apr 2020
Adversarial Training for Large Neural Language Models
Adversarial Training for Large Neural Language Models
Xiaodong Liu
Hao Cheng
Pengcheng He
Weizhu Chen
Yu-Chiang Frank Wang
Hoifung Poon
Jianfeng Gao
AAML
26
183
0
20 Apr 2020
Coreferential Reasoning Learning for Language Representation
Coreferential Reasoning Learning for Language Representation
Deming Ye
Yankai Lin
Jiaju Du
Zhenghao Liu
Peng Li
Maosong Sun
Zhiyuan Liu
26
177
0
15 Apr 2020
VGCN-BERT: Augmenting BERT with Graph Embedding for Text Classification
VGCN-BERT: Augmenting BERT with Graph Embedding for Text Classification
Zhibin Lu
Pan Du
J. Nie
31
126
0
12 Apr 2020
Frequency, Acceptability, and Selection: A case study of
  clause-embedding
Frequency, Acceptability, and Selection: A case study of clause-embedding
Aaron Steven White
Kyle Rawlins
17
12
0
08 Apr 2020
Improving BERT with Self-Supervised Attention
Improving BERT with Self-Supervised Attention
Yiren Chen
Xiaoyu Kou
Jiangang Bai
Yunhai Tong
13
10
0
08 Apr 2020
CALM: Continuous Adaptive Learning for Language Modeling
CALM: Continuous Adaptive Learning for Language Modeling
Kristjan Arumae
Parminder Bhatia
CLL
KELM
16
6
0
08 Apr 2020
MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices
MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices
Zhiqing Sun
Hongkun Yu
Xiaodan Song
Renjie Liu
Yiming Yang
Denny Zhou
MQ
8
795
0
06 Apr 2020
How Furiously Can Colourless Green Ideas Sleep? Sentence Acceptability
  in Context
How Furiously Can Colourless Green Ideas Sleep? Sentence Acceptability in Context
Jey Han Lau
C. S. Armendariz
Shalom Lappin
Matthew Purver
Chang Shu
6
40
0
02 Apr 2020
UniLMv2: Pseudo-Masked Language Models for Unified Language Model
  Pre-Training
UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training
Hangbo Bao
Li Dong
Furu Wei
Wenhui Wang
Nan Yang
...
Yu-Chiang Frank Wang
Songhao Piao
Jianfeng Gao
Ming Zhou
H. Hon
AI4CE
22
391
0
28 Feb 2020
MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression
  of Pre-Trained Transformers
MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers
Wenhui Wang
Furu Wei
Li Dong
Hangbo Bao
Nan Yang
Ming Zhou
VLM
45
1,199
0
25 Feb 2020
Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation
Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation
Yige Xu
Xipeng Qiu
L. Zhou
Xuanjing Huang
17
65
0
24 Feb 2020
The Microsoft Toolkit of Multi-Task Deep Neural Networks for Natural
  Language Understanding
The Microsoft Toolkit of Multi-Task Deep Neural Networks for Natural Language Understanding
Xiaodong Liu
Yu-Chiang Frank Wang
Jianshu Ji
Hao Cheng
Xueyun Zhu
...
Pengcheng He
Weizhu Chen
Hoifung Poon
Guihong Cao
Jianfeng Gao
AI4CE
23
60
0
19 Feb 2020
Fine-Tuning Pretrained Language Models: Weight Initializations, Data
  Orders, and Early Stopping
Fine-Tuning Pretrained Language Models: Weight Initializations, Data Orders, and Early Stopping
Jesse Dodge
Gabriel Ilharco
Roy Schwartz
Ali Farhadi
Hannaneh Hajishirzi
Noah A. Smith
27
583
0
15 Feb 2020
BERT-of-Theseus: Compressing BERT by Progressive Module Replacing
BERT-of-Theseus: Compressing BERT by Progressive Module Replacing
Canwen Xu
Wangchunshu Zhou
Tao Ge
Furu Wei
Ming Zhou
221
197
0
07 Feb 2020
BLiMP: The Benchmark of Linguistic Minimal Pairs for English
BLiMP: The Benchmark of Linguistic Minimal Pairs for English
Alex Warstadt
Alicia Parrish
Haokun Liu
Anhad Mohananey
Wei Peng
Sheng-Fu Wang
Samuel R. Bowman
18
465
0
02 Dec 2019
Neural language modeling of free word order argument structure
Neural language modeling of free word order argument structure
Charlotte Rochereau
Benoît Sagot
Emmanuel Dupoux
14
0
0
30 Nov 2019
Do Attention Heads in BERT Track Syntactic Dependencies?
Do Attention Heads in BERT Track Syntactic Dependencies?
Phu Mon Htut
Jason Phang
Shikha Bordia
Samuel R. Bowman
19
136
0
27 Nov 2019
Learning to Few-Shot Learn Across Diverse Natural Language
  Classification Tasks
Learning to Few-Shot Learn Across Diverse Natural Language Classification Tasks
Trapit Bansal
Rishikesh Jha
Andrew McCallum
SSL
11
118
0
10 Nov 2019
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language
  Models through Principled Regularized Optimization
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
Haoming Jiang
Pengcheng He
Weizhu Chen
Xiaodong Liu
Jianfeng Gao
T. Zhao
22
557
0
08 Nov 2019
What Would Elsa Do? Freezing Layers During Transformer Fine-Tuning
What Would Elsa Do? Freezing Layers During Transformer Fine-Tuning
Jaejun Lee
Raphael Tang
Jimmy J. Lin
24
121
0
08 Nov 2019
MML: Maximal Multiverse Learning for Robust Fine-Tuning of Language
  Models
MML: Maximal Multiverse Learning for Robust Fine-Tuning of Language Models
Itzik Malkiel
Lior Wolf
11
2
0
05 Nov 2019
Deepening Hidden Representations from Pre-trained Language Models
Deepening Hidden Representations from Pre-trained Language Models
Junjie Yang
Hai Zhao
14
10
0
05 Nov 2019
Harnessing the linguistic signal to predict scalar inferences
Harnessing the linguistic signal to predict scalar inferences
Sebastian Schuster
Yuxing Chen
Judith Degen
8
33
0
31 Oct 2019
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language
  Generation, Translation, and Comprehension
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
M. Lewis
Yinhan Liu
Naman Goyal
Marjan Ghazvininejad
Abdel-rahman Mohamed
Omer Levy
Veselin Stoyanov
Luke Zettlemoyer
AIMat
VLM
41
10,583
0
29 Oct 2019
Exploring Multilingual Syntactic Sentence Representations
Exploring Multilingual Syntactic Sentence Representations
Chen Cecilia Liu
Anderson de Andrade
Muhammad Osama
15
4
0
25 Oct 2019
Exploring the Limits of Transfer Learning with a Unified Text-to-Text
  Transformer
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Colin Raffel
Noam M. Shazeer
Adam Roberts
Katherine Lee
Sharan Narang
Michael Matena
Yanqi Zhou
Wei Li
Peter J. Liu
AIMat
77
19,422
0
23 Oct 2019
Demon: Improved Neural Network Training with Momentum Decay
Demon: Improved Neural Network Training with Momentum Decay
John Chen
Cameron R. Wolfe
Zhaoqi Li
Anastasios Kyrillidis
ODL
19
15
0
11 Oct 2019
Multilingual Question Answering from Formatted Text applied to
  Conversational Agents
Multilingual Question Answering from Formatted Text applied to Conversational Agents
W. Siblini
Charlotte Pasqual
Axel Lavielle
Mohamed Challal
Cyril Cauchois
17
17
0
10 Oct 2019
Knowledge Distillation from Internal Representations
Knowledge Distillation from Internal Representations
Gustavo Aguilar
Yuan Ling
Yu Zhang
Benjamin Yao
Xing Fan
Edward Guo
25
178
0
08 Oct 2019
MinWikiSplit: A Sentence Splitting Corpus with Minimal Propositions
MinWikiSplit: A Sentence Splitting Corpus with Minimal Propositions
C. Niklaus
André Freitas
Siegfried Handschuh
10
15
0
26 Sep 2019
ALBERT: A Lite BERT for Self-supervised Learning of Language
  Representations
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
Zhenzhong Lan
Mingda Chen
Sebastian Goodman
Kevin Gimpel
Piyush Sharma
Radu Soricut
SSL
AIMat
64
6,370
0
26 Sep 2019
FreeLB: Enhanced Adversarial Training for Natural Language Understanding
FreeLB: Enhanced Adversarial Training for Natural Language Understanding
Chen Zhu
Yu Cheng
Zhe Gan
S. Sun
Tom Goldstein
Jingjing Liu
AAML
223
437
0
25 Sep 2019
CAT: Compression-Aware Training for bandwidth reduction
CAT: Compression-Aware Training for bandwidth reduction
Chaim Baskin
Brian Chmiel
Evgenii Zheltonozhskii
Ron Banner
A. Bronstein
A. Mendelson
MQ
9
10
0
25 Sep 2019
Mixout: Effective Regularization to Finetune Large-scale Pretrained
  Language Models
Mixout: Effective Regularization to Finetune Large-scale Pretrained Language Models
Cheolhyoung Lee
Kyunghyun Cho
Wanmo Kang
MoE
249
205
0
25 Sep 2019
TinyBERT: Distilling BERT for Natural Language Understanding
TinyBERT: Distilling BERT for Natural Language Understanding
Xiaoqi Jiao
Yichun Yin
Lifeng Shang
Xin Jiang
Xiao Chen
Linlin Li
F. Wang
Qun Liu
VLM
11
1,813
0
23 Sep 2019
Slice-based Learning: A Programming Model for Residual Learning in
  Critical Data Slices
Slice-based Learning: A Programming Model for Residual Learning in Critical Data Slices
V. Chen
Sen Wu
Zhenzhen Weng
Alexander Ratner
Christopher Ré
13
56
0
13 Sep 2019
Learning to Discriminate Perturbations for Blocking Adversarial Attacks
  in Text Classification
Learning to Discriminate Perturbations for Blocking Adversarial Attacks in Text Classification
Yichao Zhou
Jyun-Yu Jiang
Kai-Wei Chang
Wei Wang
AAML
11
117
0
06 Sep 2019
Investigating BERT's Knowledge of Language: Five Analysis Methods with
  NPIs
Investigating BERT's Knowledge of Language: Five Analysis Methods with NPIs
Alex Warstadt
Yuning Cao
Ioana Grosu
Wei Peng
Hagen Blix
...
Jason Phang
Anhad Mohananey
Phu Mon Htut
Paloma Jeretic
Samuel R. Bowman
13
122
0
05 Sep 2019
Specializing Unsupervised Pretraining Models for Word-Level Semantic
  Similarity
Specializing Unsupervised Pretraining Models for Word-Level Semantic Similarity
Anne Lauscher
Ivan Vulić
E. Ponti
Anna Korhonen
Goran Glavavs
SSL
23
57
0
05 Sep 2019
Semantics-aware BERT for Language Understanding
Semantics-aware BERT for Language Understanding
Zhuosheng Zhang
Yuwei Wu
Zhao Hai
Z. Li
Shuailiang Zhang
Xi Zhou
Xiang Zhou
19
363
0
05 Sep 2019
Transfer Fine-Tuning: A BERT Case Study
Transfer Fine-Tuning: A BERT Case Study
Yuki Arase
Junichi Tsujii
6
41
0
03 Sep 2019
Investigating Meta-Learning Algorithms for Low-Resource Natural Language
  Understanding Tasks
Investigating Meta-Learning Algorithms for Low-Resource Natural Language Understanding Tasks
Zi-Yi Dou
Keyi Yu
Antonios Anastasopoulos
8
126
0
27 Aug 2019
Empirical Evaluation of Multi-task Learning in Deep Neural Networks for
  Natural Language Processing
Empirical Evaluation of Multi-task Learning in Deep Neural Networks for Natural Language Processing
Jianquan Li
Xiaokang Liu
Wenpeng Yin
Min Yang
Liqun Ma
Yaohong Jin
AIMat
25
13
0
16 Aug 2019
StructBERT: Incorporating Language Structures into Pre-training for Deep
  Language Understanding
StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding
Wei Wang
Bin Bi
Ming Yan
Chen Henry Wu
Zuyi Bao
Jiangnan Xia
Liwei Peng
Luo Si
20
260
0
13 Aug 2019
On Identifiability in Transformers
On Identifiability in Transformers
Gino Brunner
Yang Liu
Damian Pascual
Oliver Richter
Massimiliano Ciaramita
Roger Wattenhofer
ViT
22
186
0
12 Aug 2019
Previous
123...161718
Next