Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1905.05950
Cited By
BERT Rediscovers the Classical NLP Pipeline
15 May 2019
Ian Tenney
Dipanjan Das
Ellie Pavlick
MILM
SSeg
Re-assign community
ArXiv
PDF
HTML
Papers citing
"BERT Rediscovers the Classical NLP Pipeline"
17 / 217 papers shown
Title
Pre-trained Models for Natural Language Processing: A Survey
Xipeng Qiu
Tianxiang Sun
Yige Xu
Yunfan Shao
Ning Dai
Xuanjing Huang
LM&MA
VLM
241
1,450
0
18 Mar 2020
Investigating Entity Knowledge in BERT with Simple Neural End-To-End Entity Linking
Samuel Broscheit
OCL
23
118
0
11 Mar 2020
Fixed Encoder Self-Attention Patterns in Transformer-Based Machine Translation
Alessandro Raganato
Yves Scherrer
Jörg Tiedemann
17
92
0
24 Feb 2020
oLMpics -- On what Language Model Pre-training Captures
Alon Talmor
Yanai Elazar
Yoav Goldberg
Jonathan Berant
LRM
17
300
0
31 Dec 2019
BERTje: A Dutch BERT Model
Wietse de Vries
Andreas van Cranenburgh
Arianna Bisazza
Tommaso Caselli
Gertjan van Noord
Malvina Nissim
VLM
SSeg
11
291
0
19 Dec 2019
Inducing Relational Knowledge from BERT
Zied Bouraoui
Jose Camacho-Collados
Steven Schockaert
13
166
0
28 Nov 2019
Generalizing Natural Language Analysis through Span-relation Representations
Zhengbao Jiang
W. Xu
Jun Araki
Graham Neubig
14
60
0
10 Nov 2019
HUBERT Untangles BERT to Improve Transfer across NLP Tasks
M. Moradshahi
Hamid Palangi
M. Lam
P. Smolensky
Jianfeng Gao
21
16
0
25 Oct 2019
On the use of BERT for Neural Machine Translation
S. Clinchant
K. Jung
Vassilina Nikoulina
19
89
0
27 Sep 2019
Visualizing and Understanding the Effectiveness of BERT
Y. Hao
Li Dong
Furu Wei
Ke Xu
22
181
0
15 Aug 2019
On Identifiability in Transformers
Gino Brunner
Yang Liu
Damian Pascual
Oliver Richter
Massimiliano Ciaramita
Roger Wattenhofer
ViT
8
186
0
12 Aug 2019
Leveraging Pre-trained Checkpoints for Sequence Generation Tasks
S. Rothe
Shashi Narayan
Aliaksei Severyn
SILM
57
433
0
29 Jul 2019
What Does BERT Look At? An Analysis of BERT's Attention
Kevin Clark
Urvashi Khandelwal
Omer Levy
Christopher D. Manning
MILM
8
1,578
0
11 Jun 2019
Analyzing the Structure of Attention in a Transformer Language Model
Jesse Vig
Yonatan Belinkov
14
357
0
07 Jun 2019
How multilingual is Multilingual BERT?
Telmo Pires
Eva Schlinger
Dan Garrette
LRM
VLM
10
1,369
0
04 Jun 2019
Taming Pretrained Transformers for Extreme Multi-label Text Classification
Wei-Cheng Chang
Hsiang-Fu Yu
Kai Zhong
Yiming Yang
Inderjit Dhillon
10
20
0
07 May 2019
What you can cram into a single vector: Probing sentence embeddings for linguistic properties
Alexis Conneau
Germán Kruszewski
Guillaume Lample
Loïc Barrault
Marco Baroni
199
882
0
03 May 2018
Previous
1
2
3
4
5