ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1906.01698
  4. Cited By
Open Sesame: Getting Inside BERT's Linguistic Knowledge

Open Sesame: Getting Inside BERT's Linguistic Knowledge

4 June 2019
Yongjie Lin
Y. Tan
Robert Frank
ArXiv (abs)PDFHTML

Papers citing "Open Sesame: Getting Inside BERT's Linguistic Knowledge"

50 / 166 papers shown
Exploring the Role of BERT Token Representations to Explain Sentence
  Probing Results
Exploring the Role of BERT Token Representations to Explain Sentence Probing ResultsConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
Hosein Mohebbi
Ali Modarressi
Mohammad Taher Pilehvar
MILM
226
35
0
03 Apr 2021
Explaining the Road Not Taken
Explaining the Road Not Taken
Hua Shen
Ting-Hao 'Kenneth' Huang
FAttXAI
200
9
0
27 Mar 2021
Bertinho: Galician BERT Representations
Bertinho: Galician BERT Representations
David Vilares
Marcos Garcia
Carlos Gómez-Rodríguez
167
24
0
25 Mar 2021
Bridging the gap between supervised classification and unsupervised
  topic modelling for social-media assisted crisis management
Bridging the gap between supervised classification and unsupervised topic modelling for social-media assisted crisis management
Mikael Brunila
Rosie Zhao
Andrei Mircea
Sam Lumley
R. Sieber
110
0
0
22 Mar 2021
Local Interpretations for Explainable Natural Language Processing: A
  Survey
Local Interpretations for Explainable Natural Language Processing: A SurveyACM Computing Surveys (CSUR), 2021
Siwen Luo
Michal Guerquin
S. Han
Josiah Poon
MILM
414
64
0
20 Mar 2021
Large Pre-trained Language Models Contain Human-like Biases of What is
  Right and Wrong to Do
Large Pre-trained Language Models Contain Human-like Biases of What is Right and Wrong to DoNature Machine Intelligence (Nat. Mach. Intell.), 2021
P. Schramowski
Cigdem Turan
Nico Andersen
Constantin Rothkopf
Kristian Kersting
317
359
0
08 Mar 2021
Vyākarana: A Colorless Green Benchmark for Syntactic Evaluation in
  Indic Languages
Vyākarana: A Colorless Green Benchmark for Syntactic Evaluation in Indic Languages
Rajaswa Patil
Jasleen Dhillon
Siddhant Mahurkar
Saumitra Kulkarni
M. Malhotra
V. Baths
176
2
0
01 Mar 2021
On the Evolution of Syntactic Information Encoded by BERT's
  Contextualized Representations
On the Evolution of Syntactic Information Encoded by BERT's Contextualized RepresentationsConference of the European Chapter of the Association for Computational Linguistics (EACL), 2021
Laura Pérez-Mayos
Roberto Carlini
Miguel Ballesteros
Leo Wanner
203
10
0
27 Jan 2021
Regulatory Compliance through Doc2Doc Information Retrieval: A case
  study in EU/UK legislation where text similarity has limitations
Regulatory Compliance through Doc2Doc Information Retrieval: A case study in EU/UK legislation where text similarity has limitationsConference of the European Chapter of the Association for Computational Linguistics (EACL), 2021
Ilias Chalkidis
Manos Fergadiotis
Nikolaos Manginas
Eva Katakalou
Prodromos Malakasiotis
AILaw
154
30
0
26 Jan 2021
The heads hypothesis: A unifying statistical approach towards
  understanding multi-headed attention in BERT
The heads hypothesis: A unifying statistical approach towards understanding multi-headed attention in BERTAAAI Conference on Artificial Intelligence (AAAI), 2021
Madhura Pande
Aakriti Budhraja
Preksha Nema
Pratyush Kumar
Mitesh M. Khapra
193
20
0
22 Jan 2021
Of Non-Linearity and Commutativity in BERT
Of Non-Linearity and Commutativity in BERTIEEE International Joint Conference on Neural Network (IJCNN), 2021
Sumu Zhao
Damian Pascual
Gino Brunner
Roger Wattenhofer
316
18
0
12 Jan 2021
Superbizarre Is Not Superb: Derivational Morphology Improves BERT's
  Interpretation of Complex Words
Superbizarre Is Not Superb: Derivational Morphology Improves BERT's Interpretation of Complex WordsAnnual Meeting of the Association for Computational Linguistics (ACL), 2021
Valentin Hofmann
J. Pierrehumbert
Hinrich Schütze
494
79
0
02 Jan 2021
FiD-Ex: Improving Sequence-to-Sequence Models for Extractive Rationale
  Generation
FiD-Ex: Improving Sequence-to-Sequence Models for Extractive Rationale GenerationConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Kushal Lakhotia
Bhargavi Paranjape
Asish Ghoshal
Anuj Kumar
Yashar Mehdad
Srini Iyer
133
32
0
31 Dec 2020
Out of Order: How Important Is The Sequential Order of Words in a
  Sentence in Natural Language Understanding Tasks?
Out of Order: How Important Is The Sequential Order of Words in a Sentence in Natural Language Understanding Tasks?Findings (Findings), 2020
Thang M. Pham
Trung Bui
Long Mai
Anh Totti Nguyen
551
127
0
30 Dec 2020
Gender Bias in Multilingual Neural Machine Translation: The Architecture
  Matters
Gender Bias in Multilingual Neural Machine Translation: The Architecture Matters
Marta R. Costa-jussá
Carlos Escolano
Christine Basta
Javier Ferrando
Roser Batlle-Roca
Ksenia Kharitonova
161
20
0
24 Dec 2020
Pre-Training a Language Model Without Human Language
Pre-Training a Language Model Without Human Language
Cheng-Han Chiang
Hung-yi Lee
160
13
0
22 Dec 2020
Enhancing deep neural networks with morphological information
Enhancing deep neural networks with morphological informationNatural Language Engineering (NLE), 2020
Matej Klemen
Luka Krsnik
Marko Robnik-Šikonja
239
16
0
24 Nov 2020
Picking BERT's Brain: Probing for Linguistic Dependencies in
  Contextualized Embeddings Using Representational Similarity Analysis
Picking BERT's Brain: Probing for Linguistic Dependencies in Contextualized Embeddings Using Representational Similarity AnalysisInternational Conference on Computational Linguistics (COLING), 2020
Michael A. Lepori
R. Thomas McCoy
131
26
0
24 Nov 2020
Positional Artefacts Propagate Through Masked Language Model Embeddings
Positional Artefacts Propagate Through Masked Language Model Embeddings
Ziyang Luo
Artur Kulmizev
Xiaoxi Mao
305
41
0
09 Nov 2020
Influence Patterns for Explaining Information Flow in BERT
Influence Patterns for Explaining Information Flow in BERTNeural Information Processing Systems (NeurIPS), 2020
Kaiji Lu
Zifan Wang
Piotr (Peter) Mardziel
Anupam Datta
GNN
274
19
0
02 Nov 2020
Dynamic Contextualized Word Embeddings
Dynamic Contextualized Word EmbeddingsAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Valentin Hofmann
J. Pierrehumbert
Hinrich Schütze
417
57
0
23 Oct 2020
A Benchmark for Lease Contract Review
A Benchmark for Lease Contract Review
Spyretta Leivaditi
Julien Rossi
Evangelos Kanoulas
AILaw
294
43
0
20 Oct 2020
Layer-wise Guided Training for BERT: Learning Incrementally Refined
  Document Representations
Layer-wise Guided Training for BERT: Learning Incrementally Refined Document Representations
Nikolaos Manginas
Ilias Chalkidis
Prodromos Malakasiotis
114
5
0
12 Oct 2020
Unsupervised Distillation of Syntactic Information from Contextualized
  Word Representations
Unsupervised Distillation of Syntactic Information from Contextualized Word RepresentationsBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackboxNLP), 2020
Shauli Ravfogel
Yanai Elazar
Jacob Goldberger
Yoav Goldberg
221
13
0
11 Oct 2020
Recurrent babbling: evaluating the acquisition of grammar from limited
  input data
Recurrent babbling: evaluating the acquisition of grammar from limited input dataConference on Computational Natural Language Learning (CoNLL), 2020
Ludovica Pannitto
Aurélie Herbelot
146
16
0
09 Oct 2020
Intrinsic Probing through Dimension Selection
Intrinsic Probing through Dimension Selection
Lucas Torroba Hennigen
Adina Williams
Robert Bamler
212
61
0
06 Oct 2020
On the Interplay Between Fine-tuning and Sentence-level Probing for
  Linguistic Knowledge in Pre-trained Transformers
On the Interplay Between Fine-tuning and Sentence-level Probing for Linguistic Knowledge in Pre-trained TransformersBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackboxNLP), 2020
Marius Mosbach
A. Khokhlova
Michael A. Hedderich
Dietrich Klakow
169
51
0
06 Oct 2020
Guiding Attention for Self-Supervised Learning with Transformers
Guiding Attention for Self-Supervised Learning with TransformersFindings (Findings), 2020
Ameet Deshpande
Karthik Narasimhan
162
22
0
06 Oct 2020
Linguistic Profiling of a Neural Language Model
Linguistic Profiling of a Neural Language ModelInternational Conference on Computational Linguistics (COLING), 2020
Alessio Miaschi
D. Brunato
F. Dell’Orletta
Giulia Venturi
279
49
0
05 Oct 2020
Which *BERT? A Survey Organizing Contextualized Encoders
Which *BERT? A Survey Organizing Contextualized EncodersConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Patrick Xia
Shijie Wu
Benjamin Van Durme
227
53
0
02 Oct 2020
An information theoretic view on selecting linguistic probes
An information theoretic view on selecting linguistic probesConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Zining Zhu
Frank Rudzicz
169
22
0
15 Sep 2020
Bio-inspired Structure Identification in Language Embeddings
Bio-inspired Structure Identification in Language Embeddings
Hongwei Zhou
Zhou
Oskar Elek
P. Anand
A. Forbes
166
2
0
05 Sep 2020
Attention Flows: Analyzing and Comparing Attention Mechanisms in
  Language Models
Attention Flows: Analyzing and Comparing Attention Mechanisms in Language ModelsIEEE Transactions on Visualization and Computer Graphics (TVCG), 2020
Joseph F DeRose
Jiayao Wang
M. Berger
141
108
0
03 Sep 2020
Is Supervised Syntactic Parsing Beneficial for Language Understanding?
  An Empirical Investigation
Is Supervised Syntactic Parsing Beneficial for Language Understanding? An Empirical Investigation
Goran Glavaš
Ivan Vulić
264
71
0
15 Aug 2020
Deep Contextual Clinical Prediction with Reverse Distillation
Deep Contextual Clinical Prediction with Reverse DistillationAAAI Conference on Artificial Intelligence (AAAI), 2020
Rohan Kodialam
Rebecca Boiarsky
Justin Lim
Neil Dixit
Aditya Sai
David Sontag
240
26
0
10 Jul 2020
BERTology Meets Biology: Interpreting Attention in Protein Language
  Models
BERTology Meets Biology: Interpreting Attention in Protein Language Models
Jesse Vig
Ali Madani
Lav Varshney
Caiming Xiong
R. Socher
Nazneen Rajani
409
336
0
26 Jun 2020
A Pairwise Probe for Understanding BERT Fine-Tuning on Machine Reading
  Comprehension
A Pairwise Probe for Understanding BERT Fine-Tuning on Machine Reading ComprehensionAnnual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), 2020
Jie Cai
Zhengzhou Zhu
Ping Nie
Qian Liu
AAML
111
7
0
02 Jun 2020
Amnesic Probing: Behavioral Explanation with Amnesic Counterfactuals
Amnesic Probing: Behavioral Explanation with Amnesic Counterfactuals
Yanai Elazar
Shauli Ravfogel
Alon Jacovi
Yoav Goldberg
385
25
0
01 Jun 2020
Query Resolution for Conversational Search with Limited Supervision
Query Resolution for Conversational Search with Limited Supervision
Nikos Voskarides
Dan Li
Sudipta Singha Roy
Evangelos Kanoulas
Maarten de Rijke
186
140
0
24 May 2020
Weakly-Supervised Neural Response Selection from an Ensemble of
  Task-Specialised Dialogue Agents
Weakly-Supervised Neural Response Selection from an Ensemble of Task-Specialised Dialogue Agents
Asir Saeed
Khai Mai
Pham Quang Nhat Minh
Nguyen Tuan Duc
Danushka Bollegala
124
0
0
06 May 2020
The Sensitivity of Language Models and Humans to Winograd Schema
  Perturbations
The Sensitivity of Language Models and Humans to Winograd Schema PerturbationsAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Mostafa Abdou
Vinit Ravishankar
Maria Barrett
Yonatan Belinkov
Desmond Elliott
Anders Søgaard
ReLMLRM
209
36
0
04 May 2020
DagoBERT: Generating Derivational Morphology with a Pretrained Language
  Model
DagoBERT: Generating Derivational Morphology with a Pretrained Language Model
Valentin Hofmann
J. Pierrehumbert
Hinrich Schütze
254
2
0
02 May 2020
Intermediate-Task Transfer Learning with Pretrained Models for Natural
  Language Understanding: When and Why Does It Work?
Intermediate-Task Transfer Learning with Pretrained Models for Natural Language Understanding: When and Why Does It Work?Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Yada Pruksachatkun
Jason Phang
Haokun Liu
Phu Mon Htut
Xiaoyi Zhang
Richard Yuanzhe Pang
Clara Vania
Katharina Kann
Samuel R. Bowman
CLLLRM
229
204
0
01 May 2020
When BERT Plays the Lottery, All Tickets Are Winning
When BERT Plays the Lottery, All Tickets Are WinningConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Sai Prasanna
Anna Rogers
Anna Rumshisky
MILM
309
200
0
01 May 2020
Attribution Analysis of Grammatical Dependencies in LSTMs
Attribution Analysis of Grammatical Dependencies in LSTMs
Sophie Hao
250
3
0
30 Apr 2020
Representations of Syntax [MASK] Useful: Effects of Constituency and
  Dependency Structure in Recursive LSTMs
Representations of Syntax [MASK] Useful: Effects of Constituency and Dependency Structure in Recursive LSTMsAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Michael A. Lepori
Tal Linzen
R. Thomas McCoy
NAI
209
11
0
30 Apr 2020
A Matter of Framing: The Impact of Linguistic Formalism on Probing
  Results
A Matter of Framing: The Impact of Linguistic Formalism on Probing ResultsConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Ilia Kuznetsov
Iryna Gurevych
114
28
0
30 Apr 2020
Logic2Text: High-Fidelity Natural Language Generation from Logical Forms
Logic2Text: High-Fidelity Natural Language Generation from Logical FormsFindings (Findings), 2020
Zhiyu Zoey Chen
Wenhu Chen
Hanwen Zha
Xiyou Zhou
Yunkai Zhang
Sairam Sundaresan
William Yang Wang
NAI
204
72
0
30 Apr 2020
Quantifying the Contextualization of Word Representations with Semantic
  Class Probing
Quantifying the Contextualization of Word Representations with Semantic Class ProbingFindings (Findings), 2020
Mengjie Zhao
Philipp Dufter
Yadollah Yaghoobzadeh
Hinrich Schütze
276
28
0
25 Apr 2020
Attention is Not Only a Weight: Analyzing Transformers with Vector Norms
Attention is Not Only a Weight: Analyzing Transformers with Vector Norms
Goro Kobayashi
Tatsuki Kuribayashi
Sho Yokoi
Kentaro Inui
189
15
0
21 Apr 2020
Previous
1234
Next
Page 3 of 4