ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.12471
  4. Cited By
Neural Network Acceptability Judgments

Neural Network Acceptability Judgments

31 May 2018
Alex Warstadt
Amanpreet Singh
Samuel R. Bowman
ArXivPDFHTML

Papers citing "Neural Network Acceptability Judgments"

50 / 877 papers shown
Title
TextHide: Tackling Data Privacy in Language Understanding Tasks
TextHide: Tackling Data Privacy in Language Understanding Tasks
Yangsibo Huang
Zhao-quan Song
Danqi Chen
K. Li
Sanjeev Arora
FedML
6
55
0
12 Oct 2020
EFSG: Evolutionary Fooling Sentences Generator
EFSG: Evolutionary Fooling Sentences Generator
Marco Di Giovanni
Marco Brambilla
AAML
27
2
0
12 Oct 2020
Reformulating Unsupervised Style Transfer as Paraphrase Generation
Reformulating Unsupervised Style Transfer as Paraphrase Generation
Kalpesh Krishna
John Wieting
Mohit Iyyer
19
237
0
12 Oct 2020
SMYRF: Efficient Attention using Asymmetric Clustering
SMYRF: Efficient Attention using Asymmetric Clustering
Giannis Daras
Nikita Kitaev
Augustus Odena
A. Dimakis
23
44
0
11 Oct 2020
Interpreting Multivariate Shapley Interactions in DNNs
Interpreting Multivariate Shapley Interactions in DNNs
Hao Zhang
Yichen Xie
Longjie Zheng
Die Zhang
Quanshi Zhang
TDI
FAtt
15
7
0
10 Oct 2020
Latent Tree Learning with Ordered Neurons: What Parses Does It Produce?
Latent Tree Learning with Ordered Neurons: What Parses Does It Produce?
Yian Zhang
6
1
0
10 Oct 2020
Large Product Key Memory for Pretrained Language Models
Large Product Key Memory for Pretrained Language Models
Gyuwan Kim
Tae-Hwan Jung
VLM
KELM
15
2
0
08 Oct 2020
Unsupervised Parsing via Constituency Tests
Unsupervised Parsing via Constituency Tests
Steven Cao
Nikita Kitaev
Dan Klein
16
25
0
07 Oct 2020
On the Interplay Between Fine-tuning and Sentence-level Probing for
  Linguistic Knowledge in Pre-trained Transformers
On the Interplay Between Fine-tuning and Sentence-level Probing for Linguistic Knowledge in Pre-trained Transformers
Marius Mosbach
A. Khokhlova
Michael A. Hedderich
Dietrich Klakow
15
44
0
06 Oct 2020
GRUEN for Evaluating Linguistic Quality of Generated Text
GRUEN for Evaluating Linguistic Quality of Generated Text
Wanzheng Zhu
S. Bhat
20
60
0
06 Oct 2020
Mixup-Transformer: Dynamic Data Augmentation for NLP Tasks
Mixup-Transformer: Dynamic Data Augmentation for NLP Tasks
Lichao Sun
Congying Xia
Wenpeng Yin
Tingting Liang
Philip S. Yu
Lifang He
6
36
0
05 Oct 2020
Investigating representations of verb bias in neural language models
Investigating representations of verb bias in neural language models
Robert D. Hawkins
Takateru Yamakoshi
Thomas L. Griffiths
A. Goldberg
12
28
0
05 Oct 2020
PMI-Masking: Principled masking of correlated spans
PMI-Masking: Principled masking of correlated spans
Yoav Levine
Barak Lenz
Opher Lieber
Omri Abend
Kevin Leyton-Brown
Moshe Tennenholtz
Y. Shoham
14
72
0
05 Oct 2020
An Empirical Investigation Towards Efficient Multi-Domain Language Model
  Pre-training
An Empirical Investigation Towards Efficient Multi-Domain Language Model Pre-training
Kristjan Arumae
Q. Sun
Parminder Bhatia
6
14
0
01 Oct 2020
AUBER: Automated BERT Regularization
AUBER: Automated BERT Regularization
Hyun Dong Lee
Seongmin Lee
U. Kang
6
7
0
30 Sep 2020
Contrastive Distillation on Intermediate Representations for Language
  Model Compression
Contrastive Distillation on Intermediate Representations for Language Model Compression
S. Sun
Zhe Gan
Yu Cheng
Yuwei Fang
Shuohang Wang
Jingjing Liu
VLM
17
68
0
29 Sep 2020
Domain Adversarial Fine-Tuning as an Effective Regularizer
Domain Adversarial Fine-Tuning as an Effective Regularizer
Giorgos Vernikos
Katerina Margatina
Alexandra Chronopoulou
Ion Androutsopoulos
14
15
0
28 Sep 2020
Repulsive Attention: Rethinking Multi-head Attention as Bayesian
  Inference
Repulsive Attention: Rethinking Multi-head Attention as Bayesian Inference
Bang An
Jie Lyu
Zhenyi Wang
Chunyuan Li
Changwei Hu
Fei Tan
Ruiyi Zhang
Yifan Hu
Changyou Chen
AAML
12
28
0
20 Sep 2020
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning
  in NLP Using Fewer Parameters & Less Data
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning in NLP Using Fewer Parameters & Less Data
Jonathan Pilault
Amine Elhattami
C. Pal
CLL
MoE
19
89
0
19 Sep 2020
Efficient Transformer-based Large Scale Language Representations using
  Hardware-friendly Block Structured Pruning
Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Bingbing Li
Zhenglun Kong
Tianyun Zhang
Ji Li
Z. Li
Hang Liu
Caiwen Ding
VLM
24
64
0
17 Sep 2020
Real-Time Execution of Large-scale Language Models on Mobile
Real-Time Execution of Large-scale Language Models on Mobile
Wei Niu
Zhenglun Kong
Geng Yuan
Weiwen Jiang
Jiexiong Guan
Caiwen Ding
Pu Zhao
Sijia Liu
Bin Ren
Yanzhi Wang
MQ
20
7
0
15 Sep 2020
BoostingBERT:Integrating Multi-Class Boosting into BERT for NLP Tasks
BoostingBERT:Integrating Multi-Class Boosting into BERT for NLP Tasks
Tongwen Huang
Qingyun She
Junlin Zhang
27
15
0
13 Sep 2020
AMBERT: A Pre-trained Language Model with Multi-Grained Tokenization
AMBERT: A Pre-trained Language Model with Multi-Grained Tokenization
Xinsong Zhang
Pengshuai Li
Hang Li
14
51
0
27 Aug 2020
ConvBERT: Improving BERT with Span-based Dynamic Convolution
ConvBERT: Improving BERT with Span-based Dynamic Convolution
Zihang Jiang
Weihao Yu
Daquan Zhou
Yunpeng Chen
Jiashi Feng
Shuicheng Yan
32
156
0
06 Aug 2020
Better Fine-Tuning by Reducing Representational Collapse
Better Fine-Tuning by Reducing Representational Collapse
Armen Aghajanyan
Akshat Shrivastava
Anchit Gupta
Naman Goyal
Luke Zettlemoyer
S. Gupta
AAML
29
208
0
06 Aug 2020
Learning Representations for Axis-Aligned Decision Forests through Input
  Perturbation
Learning Representations for Axis-Aligned Decision Forests through Input Perturbation
Sebastian Bruch
Jan Pfeifer
Mathieu Guillame-Bert
21
7
0
29 Jul 2020
Towards Debiasing Sentence Representations
Towards Debiasing Sentence Representations
Paul Pu Liang
Irene Z Li
Emily Zheng
Y. Lim
Ruslan Salakhutdinov
Louis-Philippe Morency
16
231
0
16 Jul 2020
Can neural networks acquire a structural bias from raw linguistic data?
Can neural networks acquire a structural bias from raw linguistic data?
Alex Warstadt
Samuel R. Bowman
AI4CE
20
53
0
14 Jul 2020
HyperGrid: Efficient Multi-Task Transformers with Grid-wise Decomposable
  Hyper Projections
HyperGrid: Efficient Multi-Task Transformers with Grid-wise Decomposable Hyper Projections
Yi Tay
Zhe Zhao
Dara Bahri
Donald Metzler
Da-Cheng Juan
40
9
0
12 Jul 2020
Unsupervised Paraphrasing via Deep Reinforcement Learning
Unsupervised Paraphrasing via Deep Reinforcement Learning
A. B. Siddique
Samet Oymak
Vagelis Hristidis
11
56
0
05 Jul 2020
Building Interpretable Interaction Trees for Deep NLP Models
Building Interpretable Interaction Trees for Deep NLP Models
Die Zhang
Huilin Zhou
Hao Zhang
Xiaoyi Bao
Da Huo
Ruizhao Chen
Xu Cheng
Mengyue Wu
Quanshi Zhang
FAtt
6
3
0
29 Jun 2020
MaxVA: Fast Adaptation of Step Sizes by Maximizing Observed Variance of
  Gradients
MaxVA: Fast Adaptation of Step Sizes by Maximizing Observed Variance of Gradients
Chenfei Zhu
Yu Cheng
Zhe Gan
Furong Huang
Jingjing Liu
Tom Goldstein
ODL
27
2
0
21 Jun 2020
SqueezeBERT: What can computer vision teach NLP about efficient neural
  networks?
SqueezeBERT: What can computer vision teach NLP about efficient neural networks?
F. Iandola
Albert Eaton Shaw
Ravi Krishna
Kurt Keutzer
VLM
20
127
0
19 Jun 2020
Revisiting Few-sample BERT Fine-tuning
Revisiting Few-sample BERT Fine-tuning
Tianyi Zhang
Felix Wu
Arzoo Katiyar
Kilian Q. Weinberger
Yoav Artzi
30
441
0
10 Jun 2020
On the Stability of Fine-tuning BERT: Misconceptions, Explanations, and
  Strong Baselines
On the Stability of Fine-tuning BERT: Misconceptions, Explanations, and Strong Baselines
Marius Mosbach
Maksym Andriushchenko
Dietrich Klakow
19
352
0
08 Jun 2020
BERT Loses Patience: Fast and Robust Inference with Early Exit
BERT Loses Patience: Fast and Robust Inference with Early Exit
Wangchunshu Zhou
Canwen Xu
Tao Ge
Julian McAuley
Ke Xu
Furu Wei
6
329
0
07 Jun 2020
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
Pengcheng He
Xiaodong Liu
Jianfeng Gao
Weizhu Chen
AAML
62
2,614
0
05 Jun 2020
Funnel-Transformer: Filtering out Sequential Redundancy for Efficient
  Language Processing
Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing
Zihang Dai
Guokun Lai
Yiming Yang
Quoc V. Le
31
229
0
05 Jun 2020
Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Syntactic Structure Distillation Pretraining For Bidirectional Encoders
A. Kuncoro
Lingpeng Kong
Daniel Fried
Dani Yogatama
Laura Rimell
Chris Dyer
Phil Blunsom
31
33
0
27 May 2020
Common Sense or World Knowledge? Investigating Adapter-Based Knowledge
  Injection into Pretrained Transformers
Common Sense or World Knowledge? Investigating Adapter-Based Knowledge Injection into Pretrained Transformers
Anne Lauscher
Olga Majewska
Leonardo F. R. Ribeiro
Iryna Gurevych
Nikolai Rozanov
Goran Glavavs
KELM
31
79
0
24 May 2020
CERT: Contrastive Self-supervised Learning for Language Understanding
CERT: Contrastive Self-supervised Learning for Language Understanding
Hongchao Fang
Sicheng Wang
Meng Zhou
Jiayuan Ding
P. Xie
ELM
SSL
30
337
0
16 May 2020
A Systematic Assessment of Syntactic Generalization in Neural Language
  Models
A Systematic Assessment of Syntactic Generalization in Neural Language Models
Jennifer Hu
Jon Gauthier
Peng Qian
Ethan Gotlieb Wilcox
R. Levy
ELM
10
212
0
07 May 2020
Spying on your neighbors: Fine-grained probing of contextual embeddings
  for information about surrounding words
Spying on your neighbors: Fine-grained probing of contextual embeddings for information about surrounding words
Josef Klafka
Allyson Ettinger
51
42
0
04 May 2020
Intermediate-Task Transfer Learning with Pretrained Models for Natural
  Language Understanding: When and Why Does It Work?
Intermediate-Task Transfer Learning with Pretrained Models for Natural Language Understanding: When and Why Does It Work?
Yada Pruksachatkun
Jason Phang
Haokun Liu
Phu Mon Htut
Xiaoyi Zhang
Richard Yuanzhe Pang
Clara Vania
Katharina Kann
Samuel R. Bowman
CLL
LRM
11
194
0
01 May 2020
When BERT Plays the Lottery, All Tickets Are Winning
When BERT Plays the Lottery, All Tickets Are Winning
Sai Prasanna
Anna Rogers
Anna Rumshisky
MILM
6
185
0
01 May 2020
Cross-Linguistic Syntactic Evaluation of Word Prediction Models
Cross-Linguistic Syntactic Evaluation of Word Prediction Models
Aaron Mueller
Garrett Nicolai
Panayiota Petrou-Zeniou
N. Talmina
Tal Linzen
14
54
0
01 May 2020
Segatron: Segment-Aware Transformer for Language Modeling and
  Understanding
Segatron: Segment-Aware Transformer for Language Modeling and Understanding
Richard He Bai
Peng Shi
Jimmy J. Lin
Yuqing Xie
Luchen Tan
Kun Xiong
Wen Gao
Ming Li
16
8
0
30 Apr 2020
Investigating Transferability in Pretrained Language Models
Investigating Transferability in Pretrained Language Models
Alex Tamkin
Trisha Singh
D. Giovanardi
Noah D. Goodman
MILM
38
48
0
30 Apr 2020
TAVAT: Token-Aware Virtual Adversarial Training for Language
  Understanding
TAVAT: Token-Aware Virtual Adversarial Training for Language Understanding
Linyang Li
Xipeng Qiu
6
17
0
30 Apr 2020
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less
  Forgetting
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting
Sanyuan Chen
Yutai Hou
Yiming Cui
Wanxiang Che
Ting Liu
Xiangzhan Yu
KELM
CLL
10
212
0
27 Apr 2020
Previous
123...15161718
Next