ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.06316
  4. Cited By
What do you learn from context? Probing for sentence structure in
  contextualized word representations

What do you learn from context? Probing for sentence structure in contextualized word representations

International Conference on Learning Representations (ICLR), 2019
15 May 2019
Ian Tenney
Patrick Xia
Berlin Chen
Alex Jinpeng Wang
Adam Poliak
R. Thomas McCoy
Najoung Kim
Benjamin Van Durme
Samuel R. Bowman
Dipanjan Das
Ellie Pavlick
ArXiv (abs)PDFHTML

Papers citing "What do you learn from context? Probing for sentence structure in contextualized word representations"

50 / 555 papers shown
Title
Picking the Underused Heads: A Network Pruning Perspective of Attention
  Head Selection for Fusing Dialogue Coreference Information
Picking the Underused Heads: A Network Pruning Perspective of Attention Head Selection for Fusing Dialogue Coreference InformationIEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2023
Zhengyuan Liu
Nancy F. Chen
203
1
0
15 Dec 2023
INSPECT: Intrinsic and Systematic Probing Evaluation for Code
  Transformers
INSPECT: Intrinsic and Systematic Probing Evaluation for Code TransformersIEEE Transactions on Software Engineering (TSE), 2023
Anjan Karmakar
Romain Robbes
209
5
0
08 Dec 2023
Is Bigger and Deeper Always Better? Probing LLaMA Across Scales and
  Layers
Is Bigger and Deeper Always Better? Probing LLaMA Across Scales and Layers
Polydoros Giannouris
Ning Wu
Shining Liang
Ming Gong
Linjun Shou
Dongmei Zhang
Jia Li
LRM
486
20
0
07 Dec 2023
The Case for Scalable, Data-Driven Theory: A Paradigm for Scientific
  Progress in NLP
The Case for Scalable, Data-Driven Theory: A Paradigm for Scientific Progress in NLP
Julian Michael
181
1
0
01 Dec 2023
Probabilistic Transformer: A Probabilistic Dependency Model for
  Contextual Word Representation
Probabilistic Transformer: A Probabilistic Dependency Model for Contextual Word RepresentationAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Haoyi Wu
Kewei Tu
806
4
0
26 Nov 2023
Physical Reasoning and Object Planning for Household Embodied Agents
Physical Reasoning and Object Planning for Household Embodied Agents
Ayush Agrawal
Raghav Prabhakar
Anirudh Goyal
Dianbo Liu
LM&RoLRM
102
3
0
22 Nov 2023
MELA: Multilingual Evaluation of Linguistic Acceptability
MELA: Multilingual Evaluation of Linguistic AcceptabilityAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Ziyin Zhang
Yikang Liu
Wei-Ping Huang
Junyu Mao
Rui Wang
Hai Hu
264
15
0
15 Nov 2023
Graph-Induced Syntactic-Semantic Spaces in Transformer-Based Variational
  AutoEncoders
Graph-Induced Syntactic-Semantic Spaces in Transformer-Based Variational AutoEncoders
Yingji Zhang
Marco Valentino
Danilo S. Carvalho
Ian Pratt-Hartmann
André Freitas
DRL
162
11
0
14 Nov 2023
How Abstract Is Linguistic Generalization in Large Language Models?
  Experiments with Argument Structure
How Abstract Is Linguistic Generalization in Large Language Models? Experiments with Argument Structure
Michael Wilson
Jackson Petty
Robert Frank
236
21
0
08 Nov 2023
Uncovering Intermediate Variables in Transformers using Circuit Probing
Uncovering Intermediate Variables in Transformers using Circuit Probing
Michael A. Lepori
Thomas Serre
Ellie Pavlick
360
11
0
07 Nov 2023
Perturbation-based Active Learning for Question Answering
Perturbation-based Active Learning for Question Answering
Fan Luo
Mihai Surdeanu
209
0
0
04 Nov 2023
Emergence of Abstract State Representations in Embodied Sequence
  Modeling
Emergence of Abstract State Representations in Embodied Sequence ModelingConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Tian Yun
Zilai Zeng
Kunal Handa
Ashish V. Thapliyal
Bo Pang
Ellie Pavlick
Chen Sun
LM&Ro
160
9
0
03 Nov 2023
Mean BERTs make erratic language teachers: the effectiveness of latent
  bootstrapping in low-resource settings
Mean BERTs make erratic language teachers: the effectiveness of latent bootstrapping in low-resource settings
David Samuel
155
4
0
30 Oct 2023
Probing LLMs for Joint Encoding of Linguistic Categories
Probing LLMs for Joint Encoding of Linguistic CategoriesConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Giulio Starace
Konstantinos Papakostas
Rochelle Choenni
Apostolos Panagiotopoulos
Matteo Rosati
Alina Leidinger
Ekaterina Shutova
229
12
0
28 Oct 2023
Understanding the Role of Input Token Characters in Language Models: How
  Does Information Loss Affect Performance?
Understanding the Role of Input Token Characters in Language Models: How Does Information Loss Affect Performance?Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Ahmed Alajrami
Katerina Margatina
Nikolaos Aletras
AAML
132
2
0
26 Oct 2023
Give Me the Facts! A Survey on Factual Knowledge Probing in Pre-trained
  Language Models
Give Me the Facts! A Survey on Factual Knowledge Probing in Pre-trained Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Paul Youssef
Osman Alperen Koracs
Meijie Li
Jorg Schlotterer
Christin Seifert
KELM
268
26
0
25 Oct 2023
Evaluating, Understanding, and Improving Constrained Text Generation for
  Large Language Models
Evaluating, Understanding, and Improving Constrained Text Generation for Large Language Models
Xiang Chen
Xiaojun Wan
171
2
0
25 Oct 2023
The Distributional Hypothesis Does Not Fully Explain the Benefits of
  Masked Language Model Pretraining
The Distributional Hypothesis Does Not Fully Explain the Benefits of Masked Language Model PretrainingConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Ting-Rui Chiang
Dani Yogatama
146
1
0
25 Oct 2023
Mixture-of-Linguistic-Experts Adapters for Improving and Interpreting
  Pre-trained Language Models
Mixture-of-Linguistic-Experts Adapters for Improving and Interpreting Pre-trained Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Raymond Li
Gabriel Murray
Giuseppe Carenini
MoE
145
4
0
24 Oct 2023
On the Convergence and Sample Complexity Analysis of Deep Q-Networks
  with $ε$-Greedy Exploration
On the Convergence and Sample Complexity Analysis of Deep Q-Networks with εεε-Greedy ExplorationNeural Information Processing Systems (NeurIPS), 2023
Shuai Zhang
Hongkang Li
Meng Wang
Miao Liu
Pin-Yu Chen
Songtao Lu
Sijia Liu
K. Murugesan
Subhajit Chaudhury
285
38
0
24 Oct 2023
From Heuristic to Analytic: Cognitively Motivated Strategies for
  Coherent Physical Commonsense Reasoning
From Heuristic to Analytic: Cognitively Motivated Strategies for Coherent Physical Commonsense ReasoningConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Zheyuan Zhang
Shane Storks
Fengyuan Hu
Sungryull Sohn
Moontae Lee
Honglak Lee
Joyce Chai
LRM
195
4
0
24 Oct 2023
Visually Grounded Continual Language Learning with Selective
  Specialization
Visually Grounded Continual Language Learning with Selective SpecializationConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Kyra Ahrens
Lennart Bengtson
Jae Hee Lee
Stefan Wermter
275
0
0
24 Oct 2023
EpiK-Eval: Evaluation for Language Models as Epistemic Models
EpiK-Eval: Evaluation for Language Models as Epistemic ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Gabriele Prato
Jerry Huang
Prasannna Parthasarathi
Shagun Sodhani
Sarath Chandar
ELM
217
6
0
23 Oct 2023
Probing Representations for Document-level Event Extraction
Probing Representations for Document-level Event ExtractionConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Barry Wang
Xinya Du
Claire Cardie
99
2
0
23 Oct 2023
Verb Conjugation in Transformers Is Determined by Linear Encodings of
  Subject Number
Verb Conjugation in Transformers Is Determined by Linear Encodings of Subject NumberConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Sophie Hao
Tal Linzen
138
6
0
23 Oct 2023
Large Language Models are biased to overestimate profoundness
Large Language Models are biased to overestimate profoundnessConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Eugenio Herrera-Berg
Tomás Vergara Browne
Pablo León-Villagrá
Marc-Lluís Vives
Cristian Buc Calderon
ELM
94
9
0
22 Oct 2023
Implications of Annotation Artifacts in Edge Probing Test Datasets
Implications of Annotation Artifacts in Edge Probing Test DatasetsConference on Computational Natural Language Learning (CoNLL), 2023
Sagnik Ray Choudhury
Jushaan Kalra
123
1
0
20 Oct 2023
Are Structural Concepts Universal in Transformer Language Models?
  Towards Interpretable Cross-Lingual Generalization
Are Structural Concepts Universal in Transformer Language Models? Towards Interpretable Cross-Lingual Generalization
Ningyu Xu
Tao Gui
Jingting Ye
Menghan Zhang
Xuanjing Huang
279
6
0
19 Oct 2023
Grounded and Well-rounded: A Methodological Approach to the Study of
  Cross-modal and Cross-lingual Grounding
Grounded and Well-rounded: A Methodological Approach to the Study of Cross-modal and Cross-lingual GroundingConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Timothee Mickus
Elaine Zosa
Denis Paperno
159
0
0
18 Oct 2023
Disentangling the Linguistic Competence of Privacy-Preserving BERT
Disentangling the Linguistic Competence of Privacy-Preserving BERTBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackboxNLP), 2023
Stefan Arnold
Nils Kemmerzell
Annika Schreiner
232
0
0
17 Oct 2023
Assessment of Pre-Trained Models Across Languages and Grammars
Assessment of Pre-Trained Models Across Languages and GrammarsInternational Joint Conference on Natural Language Processing (IJCNLP), 2023
Alberto Muñoz-Ortiz
David Vilares
Carlos Gómez-Rodríguez
179
4
0
20 Sep 2023
Leveraging Contextual Information for Effective Entity Salience
  Detection
Leveraging Contextual Information for Effective Entity Salience Detection
Rajarshi Bhowmik
Marco Ponza
Atharva Tendle
Anant Gupta
Rebecca Jiang
Xingyu Lu
Qian Zhao
Daniel Preoţiuc-Pietro
177
3
0
14 Sep 2023
Semantic reconstruction of continuous language from MEG signals
Semantic reconstruction of continuous language from MEG signalsIEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2023
Bo Wang
Xiran Xu
Longxiang Zhang
Boda Xiao
Xihong Wu
Jingshu Chen
94
5
0
14 Sep 2023
Are Emergent Abilities in Large Language Models just In-Context
  Learning?
Are Emergent Abilities in Large Language Models just In-Context Learning?Annual Meeting of the Association for Computational Linguistics (ACL), 2023
Sheng Lu
Irina Bigoulaeva
Rachneet Sachdeva
Harish Tayyar Madabushi
Iryna Gurevych
LRMELMReLM
415
131
0
04 Sep 2023
Explainability for Large Language Models: A Survey
Explainability for Large Language Models: A SurveyACM Transactions on Intelligent Systems and Technology (ACM TIST), 2023
Haiyan Zhao
Hanjie Chen
Fan Yang
Ninghao Liu
Huiqi Deng
Hengyi Cai
Shuaiqiang Wang
D. Yin
Jundong Li
LRM
399
690
0
02 Sep 2023
Construction Grammar and Language Models
Construction Grammar and Language Models
Harish Tayyar Madabushi
Laurence Romain
P. Milin
Dagmar Divjak
372
7
0
25 Aug 2023
Why Linguistics Will Thrive in the 21st Century: A Reply to Piantadosi
  (2023)
Why Linguistics Will Thrive in the 21st Century: A Reply to Piantadosi (2023)
Jordan Kodner
Sarah Payne
Jeffrey Heinz
LRM
175
16
0
06 Aug 2023
Agentività e telicità in GilBERTo: implicazioni cognitive
Agentività e telicità in GilBERTo: implicazioni cognitiveItalian Conference on Computational Linguistics (ICL), 2023
A. Lombardi
Alessandro Lenci
141
1
0
06 Jul 2023
Concept-Based Explanations to Test for False Causal Relationships
  Learned by Abusive Language Classifiers
Concept-Based Explanations to Test for False Causal Relationships Learned by Abusive Language Classifiers
I. Nejadgholi
S. Kiritchenko
Kathleen C. Fraser
Esma Balkir
198
1
0
04 Jul 2023
Comparing the Efficacy of Fine-Tuning and Meta-Learning for Few-Shot
  Policy Imitation
Comparing the Efficacy of Fine-Tuning and Meta-Learning for Few-Shot Policy Imitation
Massimiliano Patacchiola
Mingfei Sun
Katja Hofmann
Richard Turner
OffRL
188
1
0
23 Jun 2023
Towards Explainable Evaluation Metrics for Machine Translation
Towards Explainable Evaluation Metrics for Machine TranslationJournal of machine learning research (JMLR), 2023
Christoph Leiter
Piyawat Lertvittayakumjorn
M. Fomicheva
Wei Zhao
Yang Gao
Steffen Eger
ELM
296
23
0
22 Jun 2023
Limits for Learning with Language Models
Limits for Learning with Language Models
Nicholas M. Asher
Swarnadeep Bhar
Akshay Chaturvedi
Julie Hunter
Soumya Paul
166
29
0
21 Jun 2023
Towards Understanding What Code Language Models Learned
Towards Understanding What Code Language Models Learned
Toufique Ahmed
Dian Yu
Chen Huang
Cathy Wang
Prem Devanbu
Kenji Sagae
ELM
174
5
0
20 Jun 2023
Operationalising Representation in Natural Language Processing
Operationalising Representation in Natural Language ProcessingBritish Journal for the Philosophy of Science (BJPS), 2023
J. Harding
317
16
0
14 Jun 2023
Morphosyntactic probing of multilingual BERT models
Morphosyntactic probing of multilingual BERT modelsNatural Language Engineering (NLE), 2023
Judit Ács
Endre Hamerlik
Roy Schwartz
Noah A. Smith
András Kornai
171
16
0
09 Jun 2023
Table and Image Generation for Investigating Knowledge of Entities in
  Pre-trained Vision and Language Models
Table and Image Generation for Investigating Knowledge of Entities in Pre-trained Vision and Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Hidetaka Kamigaito
Katsuhiko Hayashi
Taro Watanabe
VLM
161
1
0
03 Jun 2023
Analyzing Text Representations by Measuring Task Alignment
Analyzing Text Representations by Measuring Task AlignmentAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
César González-Gutiérrez
Audi Primadhanty
Francesco Cazzaro
A. Quattoni
127
4
0
31 May 2023
Preserving Pre-trained Features Helps Calibrate Fine-tuned Language
  Models
Preserving Pre-trained Features Helps Calibrate Fine-tuned Language ModelsInternational Conference on Learning Representations (ICLR), 2023
Guande He
Jianfei Chen
Jun Zhu
236
25
0
30 May 2023
Empirical Sufficiency Lower Bounds for Language Modeling with
  Locally-Bootstrapped Semantic Structures
Empirical Sufficiency Lower Bounds for Language Modeling with Locally-Bootstrapped Semantic Structures
Jakob Prange
Emmanuele Chersoni
187
0
0
30 May 2023
Diagnosing Transformers: Illuminating Feature Spaces for Clinical
  Decision-Making
Diagnosing Transformers: Illuminating Feature Spaces for Clinical Decision-MakingInternational Conference on Learning Representations (ICLR), 2023
Aliyah R. Hsu
Yeshwanth Cherapanamjeri
Briton Park
Tristan Naumann
A. Odisho
Bin Yu
MedIm
204
1
0
27 May 2023
Previous
123456...101112
Next