Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
1906.04284
Cited By
v1
v2 (latest)
Analyzing the Structure of Attention in a Transformer Language Model
7 June 2019
Jesse Vig
Yonatan Belinkov
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Analyzing the Structure of Attention in a Transformer Language Model"
26 / 226 papers shown
Linguistic Profiling of a Neural Language Model
International Conference on Computational Linguistics (COLING), 2020
Alessio Miaschi
D. Brunato
F. Dell’Orletta
Giulia Venturi
276
49
0
05 Oct 2020
My Body is a Cage: the Role of Morphology in Graph-Based Incompatible Control
International Conference on Learning Representations (ICLR), 2020
Vitaly Kurin
Maximilian Igl
Tim Rocktaschel
Wendelin Boehmer
Shimon Whiteson
AI4CE
355
99
0
05 Oct 2020
Syntax Representation in Word Embeddings and Neural Networks -- A Survey
Conference on Theory and Practice of Information Technologies (TPIT), 2020
Tomasz Limisiewicz
David Marecek
NAI
188
9
0
02 Oct 2020
Rethinking Attention with Performers
K. Choromanski
Valerii Likhosherstov
David Dohan
Xingyou Song
Andreea Gane
...
Afroz Mohiuddin
Lukasz Kaiser
David Belanger
Lucy J. Colwell
Adrian Weller
770
1,969
0
30 Sep 2020
Retrofitting Structure-aware Transformer Language Model for End Tasks
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Hao Fei
Yafeng Ren
Donghong Ji
118
46
0
16 Sep 2020
Do Syntax Trees Help Pre-trained Transformers Extract Information?
Devendra Singh Sachan
Yuhao Zhang
Peng Qi
William L. Hamilton
204
86
0
20 Aug 2020
The Language Interpretability Tool: Extensible, Interactive Visualizations and Analysis for NLP Models
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Ian Tenney
James Wexler
Jasmijn Bastings
Tolga Bolukbasi
Andy Coenen
...
Ellen Jiang
Mahima Pushkarna
Carey Radebaugh
Emily Reif
Ann Yuan
VLM
354
209
0
12 Aug 2020
BERTology Meets Biology: Interpreting Attention in Protein Language Models
Jesse Vig
Ali Madani
Lav Varshney
Caiming Xiong
R. Socher
Nazneen Rajani
408
335
0
26 Jun 2020
Masked Language Modeling for Proteins via Linearly Scalable Long-Context Transformers
K. Choromanski
Valerii Likhosherstov
David Dohan
Xingyou Song
Andreea Gane
...
Peter Hawkins
Jared Davis
David Belanger
Lucy J. Colwell
Adrian Weller
348
93
0
05 Jun 2020
Unsupervised Quality Estimation for Neural Machine Translation
M. Fomicheva
Shuo Sun
Lisa Yankovskaya
Frédéric Blain
Francisco Guzmán
Mark Fishel
Nikolaos Aletras
Vishrav Chaudhary
Lucia Specia
UQLM
296
249
0
21 May 2020
How Context Affects Language Models' Factual Predictions
Fabio Petroni
Patrick Lewis
Aleksandra Piktus
Tim Rocktaschel
Yuxiang Wu
Alexander H. Miller
Sebastian Riedel
KELM
214
252
0
10 May 2020
When BERT Plays the Lottery, All Tickets Are Winning
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Sai Prasanna
Anna Rogers
Anna Rumshisky
MILM
304
200
0
01 May 2020
Universal Dependencies according to BERT: both more specific and more general
Findings (Findings), 2020
Tomasz Limisiewicz
Rudolf Rosa
David Marevcek
147
18
0
30 Apr 2020
What Happens To BERT Embeddings During Fine-tuning?
BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackboxNLP), 2020
Amil Merchant
Elahe Rahimtoroghi
Ellie Pavlick
Ian Tenney
235
206
0
29 Apr 2020
Towards Transparent and Explainable Attention Models
Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Akash Kumar Mohankumar
Preksha Nema
Sharan Narasimhan
Mitesh M. Khapra
Balaji Vasan Srinivasan
Balaraman Ravindran
138
108
0
29 Apr 2020
Towards Faithfully Interpretable NLP Systems: How should we define and evaluate faithfulness?
Annual Meeting of the Association for Computational Linguistics (ACL), 2020
Alon Jacovi
Yoav Goldberg
XAI
538
699
0
07 Apr 2020
A Primer in BERTology: What we know about how BERT works
Transactions of the Association for Computational Linguistics (TACL), 2020
Anna Rogers
Olga Kovaleva
Anna Rumshisky
OffRL
470
1,717
0
27 Feb 2020
Fixed Encoder Self-Attention Patterns in Transformer-Based Machine Translation
Findings (Findings), 2020
Alessandro Raganato
Yves Scherrer
Jörg Tiedemann
364
96
0
24 Feb 2020
Are Transformers universal approximators of sequence-to-sequence functions?
International Conference on Learning Representations (ICLR), 2019
Chulhee Yun
Srinadh Bhojanapalli
A. S. Rawat
Sashank J. Reddi
Sanjiv Kumar
330
430
0
20 Dec 2019
Understanding Multi-Head Attention in Abstractive Summarization
Joris Baan
Maartje ter Hoeve
M. V. D. Wees
Anne Schuth
Maarten de Rijke
AAML
130
23
0
10 Nov 2019
exBERT: A Visual Analysis Tool to Explore Learned Representations in Transformers Models
Benjamin Hoover
Hendrik Strobelt
Sebastian Gehrmann
124
91
0
11 Oct 2019
Interrogating the Explanatory Power of Attention in Neural Machine Translation
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2019
Pooya Moradi
Nishant Kambhatla
Anoop Sarkar
245
17
0
30 Sep 2019
Attention Interpretability Across NLP Tasks
Shikhar Vashishth
Shyam Upadhyay
Gaurav Singh Tomar
Manaal Faruqui
XAI
MILM
201
186
0
24 Sep 2019
SANVis: Visual Analytics for Understanding Self-Attention Networks
Visual .. (VISUAL), 2019
Cheonbok Park
Inyoup Na
Yongjang Jo
Sungbok Shin
J. Yoo
Bum Chul Kwon
Jian Zhao
Hyungjong Noh
Yeonsoo Lee
Jaegul Choo
HAI
170
41
0
13 Sep 2019
Compositionality decomposed: how do neural networks generalise?
Journal of Artificial Intelligence Research (JAIR), 2019
Dieuwke Hupkes
Verna Dankers
Mathijs Mul
Elia Bruni
CoGe
412
370
0
22 Aug 2019
A Multiscale Visualization of Attention in the Transformer Model
Annual Meeting of the Association for Computational Linguistics (ACL), 2019
Jesse Vig
ViT
199
655
0
12 Jun 2019
Previous
1
2
3
4
5