ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.06316
  4. Cited By
What do you learn from context? Probing for sentence structure in
  contextualized word representations

What do you learn from context? Probing for sentence structure in contextualized word representations

International Conference on Learning Representations (ICLR), 2019
15 May 2019
Ian Tenney
Patrick Xia
Berlin Chen
Alex Jinpeng Wang
Adam Poliak
R. Thomas McCoy
Najoung Kim
Benjamin Van Durme
Samuel R. Bowman
Dipanjan Das
Ellie Pavlick
ArXiv (abs)PDFHTML

Papers citing "What do you learn from context? Probing for sentence structure in contextualized word representations"

50 / 555 papers shown
Mediators in Determining what Processing BERT Performs First
Mediators in Determining what Processing BERT Performs FirstNorth American Chapter of the Association for Computational Linguistics (NAACL), 2021
Aviv Slobodkin
Leshem Choshen
Omri Abend
MoE
190
16
0
13 Apr 2021
SpartQA: : A Textual Question Answering Benchmark for Spatial Reasoning
SpartQA: : A Textual Question Answering Benchmark for Spatial ReasoningNorth American Chapter of the Association for Computational Linguistics (NAACL), 2021
Roshanak Mirzaee
Hossein Rajaby Faghihi
Qiang Ning
Parisa Kordjmashidi
184
103
0
12 Apr 2021
Evaluating Saliency Methods for Neural Language Models
Evaluating Saliency Methods for Neural Language ModelsNorth American Chapter of the Association for Computational Linguistics (NAACL), 2021
Shuoyang Ding
Philipp Koehn
FAttXAI
125
61
0
12 Apr 2021
Does My Representation Capture X? Probe-Ably
Does My Representation Capture X? Probe-AblyAnnual Meeting of the Association for Computational Linguistics (ACL), 2021
Deborah Ferreira
Julia Rozanova
Mokanarangan Thayaparan
Marco Valentino
André Freitas
111
12
0
12 Apr 2021
On the Inductive Bias of Masked Language Modeling: From Statistical to
  Syntactic Dependencies
On the Inductive Bias of Masked Language Modeling: From Statistical to Syntactic DependenciesNorth American Chapter of the Association for Computational Linguistics (NAACL), 2021
Tianyi Zhang
Tatsunori Hashimoto
AI4CE
212
30
0
12 Apr 2021
Factual Probing Is [MASK]: Learning vs. Learning to Recall
Factual Probing Is [MASK]: Learning vs. Learning to RecallNorth American Chapter of the Association for Computational Linguistics (NAACL), 2021
Zexuan Zhong
Dan Friedman
Danqi Chen
334
441
0
12 Apr 2021
Connecting Attributions and QA Model Behavior on Realistic
  Counterfactuals
Connecting Attributions and QA Model Behavior on Realistic CounterfactualsConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
Xi Ye
Rohan Nair
Greg Durrett
244
28
0
09 Apr 2021
Transformers: "The End of History" for NLP?
Transformers: "The End of History" for NLP?
Anton Chernyavskiy
Dmitry Ilvovsky
Preslav Nakov
186
34
0
09 Apr 2021
Probing BERT in Hyperbolic Spaces
Probing BERT in Hyperbolic SpacesInternational Conference on Learning Representations (ICLR), 2021
Boli Chen
Yao Fu
Guangwei Xu
Pengjun Xie
Chuanqi Tan
Mosha Chen
L. Jing
133
64
0
08 Apr 2021
Better Neural Machine Translation by Extracting Linguistic Information
  from BERT
Better Neural Machine Translation by Extracting Linguistic Information from BERTConference of the European Chapter of the Association for Computational Linguistics (EACL), 2021
Hassan S. Shavarani
Anoop Sarkar
195
17
0
07 Apr 2021
Exploring the Role of BERT Token Representations to Explain Sentence
  Probing Results
Exploring the Role of BERT Token Representations to Explain Sentence Probing ResultsConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
Hosein Mohebbi
Ali Modarressi
Mohammad Taher Pilehvar
MILM
226
33
0
03 Apr 2021
Transformer visualization via dictionary learning: contextualized
  embedding as a linear superposition of transformer factors
Transformer visualization via dictionary learning: contextualized embedding as a linear superposition of transformer factorsWorkshop on Knowledge Extraction and Integration for Deep Learning Architectures; Deep Learning Inside Out (DEELIO), 2021
Zeyu Yun
Yubei Chen
Bruno A. Olshausen
Yann LeCun
288
109
0
29 Mar 2021
Synthesis of Compositional Animations from Textual Descriptions
Synthesis of Compositional Animations from Textual DescriptionsIEEE International Conference on Computer Vision (ICCV), 2021
Anindita Ghosh
N. Cheema
Cennet Oguz
Christian Theobalt
P. Slusallek
550
214
0
26 Mar 2021
Local Interpretations for Explainable Natural Language Processing: A
  Survey
Local Interpretations for Explainable Natural Language Processing: A SurveyACM Computing Surveys (CSUR), 2021
Siwen Luo
Michal Guerquin
S. Han
Josiah Poon
MILM
406
64
0
20 Mar 2021
The Interplay of Variant, Size, and Task Type in Arabic Pre-trained
  Language Models
The Interplay of Variant, Size, and Task Type in Arabic Pre-trained Language ModelsWorkshop on Arabic Natural Language Processing (WANLP), 2021
Go Inoue
Bashar Alhafni
Nurpeiis Baimukan
Houda Bouamor
Farah E. Shamout
356
299
0
11 Mar 2021
Large Pre-trained Language Models Contain Human-like Biases of What is
  Right and Wrong to Do
Large Pre-trained Language Models Contain Human-like Biases of What is Right and Wrong to DoNature Machine Intelligence (Nat. Mach. Intell.), 2021
P. Schramowski
Cigdem Turan
Nico Andersen
Constantin Rothkopf
Kristian Kersting
308
359
0
08 Mar 2021
Few-shot Learning for Slot Tagging with Attentive Relational Network
Few-shot Learning for Slot Tagging with Attentive Relational NetworkConference of the European Chapter of the Association for Computational Linguistics (EACL), 2021
Cennet Oguz
Ngoc Thang Vu
127
10
0
03 Mar 2021
The Rediscovery Hypothesis: Language Models Need to Meet Linguistics
The Rediscovery Hypothesis: Language Models Need to Meet LinguisticsJournal of Artificial Intelligence Research (JAIR), 2021
Vassilina Nikoulina
Maxat Tezekbayev
Nuradil Kozhakhmet
Madina Babazhanova
Matthias Gallé
Z. Assylbekov
209
8
0
02 Mar 2021
Vyākarana: A Colorless Green Benchmark for Syntactic Evaluation in
  Indic Languages
Vyākarana: A Colorless Green Benchmark for Syntactic Evaluation in Indic Languages
Rajaswa Patil
Jasleen Dhillon
Siddhant Mahurkar
Saumitra Kulkarni
M. Malhotra
V. Baths
173
2
0
01 Mar 2021
RuSentEval: Linguistic Source, Encoder Force!
RuSentEval: Linguistic Source, Encoder Force!Workshop on Balto-Slavic Natural Language Processing (BSNLP), 2021
Vladislav Mikhailov
Ekaterina Taktasheva
Elina Sigdel
Ekaterina Artemova
VLM
149
6
0
28 Feb 2021
Chess as a Testbed for Language Model State Tracking
Chess as a Testbed for Language Model State TrackingAAAI Conference on Artificial Intelligence (AAAI), 2021
Shubham Toshniwal
Sam Wiseman
Karen Livescu
Kevin Gimpel
225
66
0
26 Feb 2021
Probing Classifiers: Promises, Shortcomings, and Advances
Probing Classifiers: Promises, Shortcomings, and AdvancesInternational Conference on Computational Logic (ICCL), 2021
Yonatan Belinkov
753
593
0
24 Feb 2021
Using Prior Knowledge to Guide BERT's Attention in Semantic Textual
  Matching Tasks
Using Prior Knowledge to Guide BERT's Attention in Semantic Textual Matching TasksThe Web Conference (WWW), 2021
Tingyu Xia
Yue Wang
Yuan Tian
Yi-Ju Chang
147
55
0
22 Feb 2021
Evaluating Contextualized Language Models for Hungarian
Evaluating Contextualized Language Models for Hungarian
Judit Ács
Dániel Lévai
D. Nemeskey
András Kornai
92
2
0
22 Feb 2021
The Singleton Fallacy: Why Current Critiques of Language Models Miss the
  Point
The Singleton Fallacy: Why Current Critiques of Language Models Miss the PointFrontiers in Artificial Intelligence (Front. Artif. Intell.), 2021
Magnus Sahlgren
F. Carlsson
99
31
0
08 Feb 2021
Clinical Outcome Prediction from Admission Notes using Self-Supervised
  Knowledge Integration
Clinical Outcome Prediction from Admission Notes using Self-Supervised Knowledge IntegrationConference of the European Chapter of the Association for Computational Linguistics (EACL), 2021
Betty van Aken
Jens-Michalis Papaioannou
M. Mayrdorfer
Klemens Budde
Felix Alexander Gers
Alexander Loser
133
85
0
08 Feb 2021
On the Evolution of Syntactic Information Encoded by BERT's
  Contextualized Representations
On the Evolution of Syntactic Information Encoded by BERT's Contextualized RepresentationsConference of the European Chapter of the Association for Computational Linguistics (EACL), 2021
Laura Pérez-Mayos
Roberto Carlini
Miguel Ballesteros
Leo Wanner
196
9
0
27 Jan 2021
CLiMP: A Benchmark for Chinese Language Model Evaluation
CLiMP: A Benchmark for Chinese Language Model EvaluationConference of the European Chapter of the Association for Computational Linguistics (EACL), 2021
Beilei Xiang
Changbing Yang
Yu Li
Alex Warstadt
Katharina Kann
ALM
172
56
0
26 Jan 2021
The heads hypothesis: A unifying statistical approach towards
  understanding multi-headed attention in BERT
The heads hypothesis: A unifying statistical approach towards understanding multi-headed attention in BERTAAAI Conference on Artificial Intelligence (AAAI), 2021
Madhura Pande
Aakriti Budhraja
Preksha Nema
Pratyush Kumar
Mitesh M. Khapra
179
20
0
22 Jan 2021
Of Non-Linearity and Commutativity in BERT
Of Non-Linearity and Commutativity in BERTIEEE International Joint Conference on Neural Network (IJCNN), 2021
Sumu Zhao
Damian Pascual
Gino Brunner
Roger Wattenhofer
308
18
0
12 Jan 2021
Learning Better Sentence Representation with Syntax Information
Learning Better Sentence Representation with Syntax Information
Chen Yang
99
2
0
09 Jan 2021
FiD-Ex: Improving Sequence-to-Sequence Models for Extractive Rationale
  Generation
FiD-Ex: Improving Sequence-to-Sequence Models for Extractive Rationale GenerationConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Kushal Lakhotia
Bhargavi Paranjape
Asish Ghoshal
Anuj Kumar
Yashar Mehdad
Srini Iyer
130
32
0
31 Dec 2020
Inserting Information Bottlenecks for Attribution in Transformers
Inserting Information Bottlenecks for Attribution in TransformersFindings (Findings), 2020
Zhiying Jiang
Raphael Tang
Ji Xin
Jimmy J. Lin
224
6
0
27 Dec 2020
Pre-Training a Language Model Without Human Language
Pre-Training a Language Model Without Human Language
Cheng-Han Chiang
Hung-yi Lee
159
13
0
22 Dec 2020
Learning from Mistakes: Using Mis-predictions as Harm Alerts in Language
  Pre-Training
Learning from Mistakes: Using Mis-predictions as Harm Alerts in Language Pre-Training
Chen Xing
Wenhao Liu
Caiming Xiong
112
0
0
16 Dec 2020
Infusing Finetuning with Semantic Dependencies
Infusing Finetuning with Semantic DependenciesTransactions of the Association for Computational Linguistics (TACL), 2020
Zhaofeng Wu
Hao Peng
Noah A. Smith
227
38
0
10 Dec 2020
Circles are like Ellipses, or Ellipses are like Circles? Measuring the
  Degree of Asymmetry of Static and Contextual Embeddings and the Implications
  to Representation Learning
Circles are like Ellipses, or Ellipses are like Circles? Measuring the Degree of Asymmetry of Static and Contextual Embeddings and the Implications to Representation LearningAAAI Conference on Artificial Intelligence (AAAI), 2020
Wei Zhang
Murray Campbell
Yang Yu
Yara Rizk
90
0
0
03 Dec 2020
Picking BERT's Brain: Probing for Linguistic Dependencies in
  Contextualized Embeddings Using Representational Similarity Analysis
Picking BERT's Brain: Probing for Linguistic Dependencies in Contextualized Embeddings Using Representational Similarity AnalysisInternational Conference on Computational Linguistics (COLING), 2020
Michael A. Lepori
R. Thomas McCoy
131
26
0
24 Nov 2020
FLERT: Document-Level Features for Named Entity Recognition
FLERT: Document-Level Features for Named Entity Recognition
Stefan Schweter
Alan Akbik
213
122
0
13 Nov 2020
When Do You Need Billions of Words of Pretraining Data?
When Do You Need Billions of Words of Pretraining Data?
Yian Zhang
Alex Warstadt
Haau-Sing Li
Samuel R. Bowman
258
155
0
10 Nov 2020
Language Through a Prism: A Spectral Approach for Multiscale Language
  Representations
Language Through a Prism: A Spectral Approach for Multiscale Language Representations
Alex Tamkin
Dan Jurafsky
Noah D. Goodman
164
56
0
09 Nov 2020
Positional Artefacts Propagate Through Masked Language Model Embeddings
Positional Artefacts Propagate Through Masked Language Model Embeddings
Ziyang Luo
Artur Kulmizev
Xiaoxi Mao
305
41
0
09 Nov 2020
CxGBERT: BERT meets Construction Grammar
CxGBERT: BERT meets Construction Grammar
Harish Tayyar Madabushi
Laurence Romain
Dagmar Divjak
P. Milin
165
54
0
09 Nov 2020
A Closer Look at Linguistic Knowledge in Masked Language Models: The
  Case of Relative Clauses in American English
A Closer Look at Linguistic Knowledge in Masked Language Models: The Case of Relative Clauses in American EnglishInternational Conference on Computational Linguistics (COLING), 2020
Marius Mosbach
Stefania Degaetano-Ortlieb
Marie-Pauline Krielke
Badr M. Abdullah
Dietrich Klakow
160
7
0
02 Nov 2020
Influence Patterns for Explaining Information Flow in BERT
Influence Patterns for Explaining Information Flow in BERTNeural Information Processing Systems (NeurIPS), 2020
Kaiji Lu
Zifan Wang
Piotr (Peter) Mardziel
Anupam Datta
GNN
242
19
0
02 Nov 2020
Vec2Sent: Probing Sentence Embeddings with Natural Language Generation
Vec2Sent: Probing Sentence Embeddings with Natural Language GenerationInternational Conference on Computational Linguistics (COLING), 2020
M. Kerscher
Steffen Eger
172
1
0
01 Nov 2020
Image Representations Learned With Unsupervised Pre-Training Contain
  Human-like Biases
Image Representations Learned With Unsupervised Pre-Training Contain Human-like BiasesConference on Fairness, Accountability and Transparency (FAccT), 2020
Ryan Steed
Aylin Caliskan
SSL
333
173
0
28 Oct 2020
Deep Clustering of Text Representations for Supervision-free Probing of
  Syntax
Deep Clustering of Text Representations for Supervision-free Probing of SyntaxAAAI Conference on Artificial Intelligence (AAAI), 2020
Vikram Gupta
Haoyue Shi
Kevin Gimpel
Mrinmaya Sachan
314
9
0
24 Oct 2020
Applying Occam's Razor to Transformer-Based Dependency Parsing: What
  Works, What Doesn't, and What is Really Necessary
Applying Occam's Razor to Transformer-Based Dependency Parsing: What Works, What Doesn't, and What is Really NecessaryInternational Workshop/Conference on Parsing Technologies (IWPT), 2020
Stefan Grünewald
Annemarie Friedrich
Jonas Kuhn
271
12
0
23 Oct 2020
Dynamic Contextualized Word Embeddings
Dynamic Contextualized Word EmbeddingsAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Valentin Hofmann
J. Pierrehumbert
Hinrich Schütze
408
57
0
23 Oct 2020
Previous
123...101112789
Next
Page 8 of 12
Pageof 12