Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1905.06316
Cited By
What do you learn from context? Probing for sentence structure in contextualized word representations
15 May 2019
Ian Tenney
Patrick Xia
Berlin Chen
Alex Jinpeng Wang
Adam Poliak
R. Thomas McCoy
Najoung Kim
Benjamin Van Durme
Samuel R. Bowman
Dipanjan Das
Ellie Pavlick
Re-assign community
ArXiv
PDF
HTML
Papers citing
"What do you learn from context? Probing for sentence structure in contextualized word representations"
50 / 532 papers shown
Title
A Transformer-based Neural Language Model that Synthesizes Brain Activation Maps from Free-Form Text Queries
G. Ngo
Minh Le Nguyen
Nancy F. Chen
M. Sabuncu
MedIm
17
7
0
24 Jul 2022
Pretraining on Interactions for Learning Grounded Affordance Representations
Jack Merullo
Dylan Ebert
Carsten Eickhoff
Ellie Pavlick
8
4
0
05 Jul 2022
Probing via Prompting
Jiaoda Li
Ryan Cotterell
Mrinmaya Sachan
29
13
0
04 Jul 2022
Is neural language acquisition similar to natural? A chronological probing study
E. Voloshina
O. Serikov
Tatiana Shavrina
30
4
0
01 Jul 2022
A Unified Understanding of Deep NLP Models for Text Classification
Zhuguo Li
Xiting Wang
Weikai Yang
Jing Wu
Zhengyan Zhang
Zhiyuan Liu
Maosong Sun
Hui Zhang
Shixia Liu
VLM
12
30
0
19 Jun 2022
AnyMorph: Learning Transferable Polices By Inferring Agent Morphology
Brandon Trabucco
Mariano Phielipp
Glen Berseth
26
27
0
17 Jun 2022
Transition-based Abstract Meaning Representation Parsing with Contextual Embeddings
Yi Liang
47
0
0
13 Jun 2022
Sort by Structure: Language Model Ranking as Dependency Probing
Max Müller-Eberstein
Rob van der Goot
Barbara Plank
30
3
0
10 Jun 2022
Abstraction not Memory: BERT and the English Article System
Harish Tayyar Madabushi
Dagmar Divjak
P. Milin
7
4
0
08 Jun 2022
Latent Topology Induction for Understanding Contextualized Representations
Yao Fu
Mirella Lapata
BDL
67
6
0
03 Jun 2022
Garden-Path Traversal in GPT-2
William Jurayj
William Rudman
Carsten Eickhoff
11
4
0
24 May 2022
What company do words keep? Revisiting the distributional semantics of J.R. Firth & Zellig Harris
Mikael Brunila
J. LaViolette
32
20
0
16 May 2022
Discovering Latent Concepts Learned in BERT
Fahim Dalvi
A. Khan
Firoj Alam
Nadir Durrani
Jia Xu
Hassan Sajjad
SSL
11
56
0
15 May 2022
Improving Contextual Representation with Gloss Regularized Pre-training
Yu Lin
Zhecheng An
Peihao Wu
Zejun Ma
19
5
0
13 May 2022
ElitePLM: An Empirical Study on General Language Ability Evaluation of Pretrained Language Models
Junyi Li
Tianyi Tang
Zheng Gong
Lixin Yang
Zhuohao Yu
Z. Chen
Jingyuan Wang
Wayne Xin Zhao
Ji-Rong Wen
LM&MA
ELM
14
7
0
03 May 2022
Probing for the Usage of Grammatical Number
Karim Lasri
Tiago Pimentel
Alessandro Lenci
Thierry Poibeau
Ryan Cotterell
25
55
0
19 Apr 2022
Probing Script Knowledge from Pre-Trained Models
Zijian Jin
Xingyu Zhang
Mo Yu
Lifu Huang
10
4
0
16 Apr 2022
On the Role of Pre-trained Language Models in Word Ordering: A Case Study with BART
Zebin Ou
Meishan Zhang
Yue Zhang
19
2
0
15 Apr 2022
Curriculum: A Broad-Coverage Benchmark for Linguistic Phenomena in Natural Language Understanding
Zeming Chen
Qiyue Gao
ELM
25
4
0
13 Apr 2022
Probing for Constituency Structure in Neural Language Models
David Arps
Younes Samih
Laura Kallmeyer
Hassan Sajjad
19
12
0
13 Apr 2022
A Review on Language Models as Knowledge Bases
Badr AlKhamissi
Millicent Li
Asli Celikyilmaz
Mona T. Diab
Marjan Ghazvininejad
KELM
22
175
0
12 Apr 2022
What do Toothbrushes do in the Kitchen? How Transformers Think our World is Structured
Alexander Henlein
Alexander Mehler
25
6
0
12 Apr 2022
A Comparative Study of Pre-trained Encoders for Low-Resource Named Entity Recognition
Yuxuan Chen
Jonas Mikkelsen
Arne Binder
Christoph Alt
Leonhard Hennig
19
2
0
11 Apr 2022
Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Yu Meng
Chenyan Xiong
Payal Bajaj
Saurabh Tiwary
Paul N. Bennett
Jiawei Han
Xia Song
MoE
31
16
0
07 Apr 2022
An Exploratory Study on Code Attention in BERT
Rishab Sharma
Fuxiang Chen
Fatemeh H. Fard
David Lo
19
25
0
05 Apr 2022
An Analysis of Semantically-Aligned Speech-Text Embeddings
M. Huzaifah
Ivan Kukanov
25
7
0
04 Apr 2022
Effect and Analysis of Large-scale Language Model Rescoring on Competitive ASR Systems
Takuma Udagawa
Masayuki Suzuki
Gakuto Kurata
N. Itoh
G. Saon
34
23
0
01 Apr 2022
Interpretation of Black Box NLP Models: A Survey
Shivani Choudhary
N. Chatterjee
S. K. Saha
FAtt
32
10
0
31 Mar 2022
Metaphors in Pre-Trained Language Models: Probing and Generalization Across Datasets and Languages
Ehsan Aghazadeh
Mohsen Fayyaz
Yadollah Yaghoobzadeh
31
51
0
26 Mar 2022
How does the pre-training objective affect what large language models learn about linguistic properties?
Ahmed Alajrami
Nikolaos Aletras
21
20
0
20 Mar 2022
On the Importance of Data Size in Probing Fine-tuned Models
Houman Mehrafarin
S. Rajaee
Mohammad Taher Pilehvar
9
18
0
17 Mar 2022
Finding Structural Knowledge in Multimodal-BERT
Victor Milewski
Miryam de Lhoneux
Marie-Francine Moens
19
9
0
17 Mar 2022
A Simple but Effective Pluggable Entity Lookup Table for Pre-trained Language Models
Deming Ye
Yankai Lin
Peng Li
Maosong Sun
Zhiyuan Liu
KELM
6
11
0
27 Feb 2022
On the data requirements of probing
Zining Zhu
Jixuan Wang
Bai Li
Frank Rudzicz
19
5
0
25 Feb 2022
BERTVision -- A Parameter-Efficient Approach for Question Answering
Siduo Jiang
Cristopher Benge
Will King
11
1
0
24 Feb 2022
Evaluating the Construct Validity of Text Embeddings with Application to Survey Questions
Qixiang Fang
D. Nguyen
Daniel L. Oberski
19
12
0
18 Feb 2022
Probing Pretrained Models of Source Code
Sergey Troshin
Nadezhda Chirkova
ELM
25
38
0
16 Feb 2022
ZeroGen: Efficient Zero-shot Learning via Dataset Generation
Jiacheng Ye
Jiahui Gao
Qintong Li
Hang Xu
Jiangtao Feng
Zhiyong Wu
Tao Yu
Lingpeng Kong
SyDa
43
211
0
16 Feb 2022
Do Transformers Encode a Foundational Ontology? Probing Abstract Classes in Natural Language
Mael Jullien
Marco Valentino
André Freitas
24
9
0
25 Jan 2022
A Latent-Variable Model for Intrinsic Probing
Karolina Stañczak
Lucas Torroba Hennigen
Adina Williams
Ryan Cotterell
Isabelle Augenstein
21
4
0
20 Jan 2022
Zero-Shot and Few-Shot Classification of Biomedical Articles in Context of the COVID-19 Pandemic
Simon Lupart
Benoit Favre
Vassilina Nikoulina
Salah Ait-Mokhtar
10
3
0
09 Jan 2022
Does Entity Abstraction Help Generative Transformers Reason?
Nicolas Angelard-Gontier
Siva Reddy
C. Pal
19
5
0
05 Jan 2022
Discrete and continuous representations and processing in deep learning: Looking forward
Ruben Cartuyvels
Graham Spinks
Marie-Francine Moens
OCL
20
20
0
04 Jan 2022
Is "My Favorite New Movie" My Favorite Movie? Probing the Understanding of Recursive Noun Phrases
Qing Lyu
Hua Zheng
Daoxin Li
Li Zhang
Marianna Apidianaki
Chris Callison-Burch
18
4
0
15 Dec 2021
Linguistic Frameworks Go Toe-to-Toe at Neuro-Symbolic Language Modeling
Jakob Prange
Nathan Schneider
Lingpeng Kong
17
9
0
15 Dec 2021
LMTurk: Few-Shot Learners as Crowdsourcing Workers in a Language-Model-as-a-Service Framework
Mengjie Zhao
Fei Mi
Yasheng Wang
Minglei Li
Xin Jiang
Qun Liu
Hinrich Schütze
RALM
33
11
0
14 Dec 2021
Human Guided Exploitation of Interpretable Attention Patterns in Summarization and Topic Segmentation
Raymond Li
Wen Xiao
Linzi Xing
Lanjun Wang
Gabriel Murray
Giuseppe Carenini
ViT
25
7
0
10 Dec 2021
Open Vocabulary Electroencephalography-To-Text Decoding and Zero-shot Sentiment Classification
Zhenhailong Wang
Heng Ji
84
71
0
05 Dec 2021
LoNLI: An Extensible Framework for Testing Diverse Logical Reasoning Capabilities for NLI
Ishan Tarunesh
Somak Aditya
Monojit Choudhury
ELM
LRM
23
4
0
04 Dec 2021
Probing Linguistic Information For Logical Inference In Pre-trained Language Models
Zeming Chen
Qiyue Gao
20
7
0
03 Dec 2021
Previous
1
2
3
4
5
6
...
9
10
11
Next