Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2010.01150
Cited By
Cost-effective Selection of Pretraining Data: A Case Study of Pretraining BERT on Social Media
2 October 2020
Xiang Dai
Sarvnaz Karimi
Ben Hachey
Cécile Paris
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Cost-effective Selection of Pretraining Data: A Case Study of Pretraining BERT on Social Media"
8 / 8 papers shown
Title
German FinBERT: A German Pre-trained Language Model
Moritz Scherrmann
30
0
0
15 Nov 2023
Better with Less: A Data-Active Perspective on Pre-Training Graph Neural Networks
Jiarong Xu
Renhong Huang
Xin Jiang
Yuxuan Cao
Carl Yang
Chunping Wang
Yang Yang
AI4CE
31
14
0
02 Nov 2023
The Effects of In-domain Corpus Size on pre-training BERT
Chris Sanchez
Zheyu Zhang
AI4CE
6
4
0
15 Dec 2022
Sort by Structure: Language Model Ranking as Dependency Probing
Max Müller-Eberstein
Rob van der Goot
Barbara Plank
33
3
0
10 Jun 2022
Extraction of Medication Names from Twitter Using Augmentation and an Ensemble of Language Models
I. Kulev
Berkay Köprü
Raul Rodriguez-Esteban
Diego Saldana Miranda
Yi Huang
Alessandro La Torraca
Elif Özkirimli
MedIm
15
4
0
12 Nov 2021
Discontinuous Named Entity Recognition as Maximal Clique Discovery
Yucheng Wang
Yu Bowen
Hongsong Zhu
Tingwen Liu
Nan Yu
Limin Sun
BDL
14
47
0
01 Jun 2021
An Analysis of Simple Data Augmentation for Named Entity Recognition
Xiang Dai
Heike Adel
32
194
0
22 Oct 2020
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
297
6,956
0
20 Apr 2018
1