ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2009.14124
  4. Cited By
Parsing with Multilingual BERT, a Small Corpus, and a Small Treebank

Parsing with Multilingual BERT, a Small Corpus, and a Small Treebank

29 September 2020
Ethan C. Chau
Lucy H. Lin
Noah A. Smith
ArXivPDFHTML

Papers citing "Parsing with Multilingual BERT, a Small Corpus, and a Small Treebank"

13 / 13 papers shown
Title
MaLA-500: Massive Language Adaptation of Large Language Models
MaLA-500: Massive Language Adaptation of Large Language Models
Peiqin Lin
Shaoxiong Ji
Jörg Tiedemann
André F. T. Martins
Hinrich Schütze
ELM
23
15
0
24 Jan 2024
mmT5: Modular Multilingual Pre-Training Solves Source Language
  Hallucinations
mmT5: Modular Multilingual Pre-Training Solves Source Language Hallucinations
Jonas Pfeiffer
Francesco Piccinno
Massimo Nicosia
Xinyi Wang
Machel Reid
Sebastian Ruder
VLM
LRM
29
27
0
23 May 2023
PrOnto: Language Model Evaluations for 859 Languages
PrOnto: Language Model Evaluations for 859 Languages
Luke Gessler
13
1
0
22 May 2023
Meeting the Needs of Low-Resource Languages: The Value of Automatic
  Alignments via Pretrained Models
Meeting the Needs of Low-Resource Languages: The Value of Automatic Alignments via Pretrained Models
Abteen Ebrahimi
Arya D. McCarthy
Arturo Oncevay
Luis Chiruzzo
J. Ortega
Gustavo A. Giménez-Lugo
Rolando A. Coto Solano
Katharina Kann
28
6
0
15 Feb 2023
You Can Have Your Data and Balance It Too: Towards Balanced and
  Efficient Multilingual Models
You Can Have Your Data and Balance It Too: Towards Balanced and Efficient Multilingual Models
Tomasz Limisiewicz
Daniel Malkin
Gabriel Stanovsky
6
4
0
13 Oct 2022
Pre-training Data Quality and Quantity for a Low-Resource Language: New
  Corpus and BERT Models for Maltese
Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and BERT Models for Maltese
Kurt Micallef
Albert Gatt
Marc Tanti
Lonneke van der Plas
Claudia Borg
18
28
0
21 May 2022
Lifting the Curse of Multilinguality by Pre-training Modular
  Transformers
Lifting the Curse of Multilinguality by Pre-training Modular Transformers
Jonas Pfeiffer
Naman Goyal
Xi Victoria Lin
Xian Li
James Cross
Sebastian Riedel
Mikel Artetxe
LRM
35
138
0
12 May 2022
IndoBERTweet: A Pretrained Language Model for Indonesian Twitter with
  Effective Domain-Specific Vocabulary Initialization
IndoBERTweet: A Pretrained Language Model for Indonesian Twitter with Effective Domain-Specific Vocabulary Initialization
Fajri Koto
Jey Han Lau
Timothy Baldwin
VLM
55
82
0
10 Sep 2021
DEMix Layers: Disentangling Domains for Modular Language Modeling
DEMix Layers: Disentangling Domains for Modular Language Modeling
Suchin Gururangan
Michael Lewis
Ari Holtzman
Noah A. Smith
Luke Zettlemoyer
KELM
MoE
6
127
0
11 Aug 2021
How to Adapt Your Pretrained Multilingual Model to 1600 Languages
How to Adapt Your Pretrained Multilingual Model to 1600 Languages
Abteen Ebrahimi
Katharina Kann
LRM
VLM
17
67
0
03 Jun 2021
How Good is Your Tokenizer? On the Monolingual Performance of
  Multilingual Language Models
How Good is Your Tokenizer? On the Monolingual Performance of Multilingual Language Models
Phillip Rust
Jonas Pfeiffer
Ivan Vulić
Sebastian Ruder
Iryna Gurevych
69
234
0
31 Dec 2020
UNKs Everywhere: Adapting Multilingual Language Models to New Scripts
UNKs Everywhere: Adapting Multilingual Language Models to New Scripts
Jonas Pfeiffer
Ivan Vulić
Iryna Gurevych
Sebastian Ruder
14
126
0
31 Dec 2020
Google's Neural Machine Translation System: Bridging the Gap between
  Human and Machine Translation
Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation
Yonghui Wu
M. Schuster
Z. Chen
Quoc V. Le
Mohammad Norouzi
...
Alex Rudnick
Oriol Vinyals
G. Corrado
Macduff Hughes
J. Dean
AIMat
716
6,740
0
26 Sep 2016
1