ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2002.07306
  4. Cited By
From English To Foreign Languages: Transferring Pre-trained Language
  Models
v1v2 (latest)

From English To Foreign Languages: Transferring Pre-trained Language Models

18 February 2020
Ke M. Tran
ArXiv (abs)PDFHTML

Papers citing "From English To Foreign Languages: Transferring Pre-trained Language Models"

32 / 32 papers shown
TokAlign: Efficient Vocabulary Adaptation via Token Alignment
TokAlign: Efficient Vocabulary Adaptation via Token AlignmentAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Chong Li
Jiajun Zhang
Chengqing Zong
VLM
250
5
0
04 Jun 2025
Token Distillation: Attention-aware Input Embeddings For New Tokens
Token Distillation: Attention-aware Input Embeddings For New Tokens
Konstantin Dobler
Desmond Elliott
Gerard de Melo
CLLVLM
540
1
0
26 May 2025
Cross-Lingual Optimization for Language Transfer in Large Language Models
Cross-Lingual Optimization for Language Transfer in Large Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Jungseob Lee
Seongtae Hong
Hyeonseok Moon
Heuiseok Lim
284
1
0
20 May 2025
Bielik v3 Small: Technical Report
Bielik v3 Small: Technical Report
Krzysztof Ociepa
Łukasz Flis
Remigiusz Kinas
Krzysztof Wróbel
Adrian Gwoździej
518
4
0
05 May 2025
HYPEROFA: Expanding LLM Vocabulary to New Languages via Hypernetwork-Based Embedding Initialization
HYPEROFA: Expanding LLM Vocabulary to New Languages via Hypernetwork-Based Embedding Initialization
Enes Özeren
Yihong Liu
Hinrich Schütze
315
1
0
21 Apr 2025
SuperBPE: Space Travel for Language Models
SuperBPE: Space Travel for Language Models
Alisa Liu
J. Hayase
Valentin Hofmann
Sewoong Oh
Noah A. Smith
Yejin Choi
594
38
0
17 Mar 2025
Adapters for Altering LLM Vocabularies: What Languages Benefit the Most?
Adapters for Altering LLM Vocabularies: What Languages Benefit the Most?International Conference on Learning Representations (ICLR), 2024
HyoJung Han
Akiko Eriguchi
Haoran Xu
Hieu T. Hoang
Marine Carpuat
Huda Khayrallah
VLM
329
9
0
12 Oct 2024
Gradient Routing: Masking Gradients to Localize Computation in Neural
  Networks
Gradient Routing: Masking Gradients to Localize Computation in Neural Networks
Alex Cloud
Jacob Goldman-Wetzler
Evžen Wybitul
Joseph Miller
Alexander Matt Turner
281
5
0
06 Oct 2024
An Empirical Comparison of Vocabulary Expansion and Initialization
  Approaches for Language Models
An Empirical Comparison of Vocabulary Expansion and Initialization Approaches for Language Models
Nandini Mundra
Aditya Nanda Kishore
Mary Dabre
Ratish Puduppully
Anoop Kunchukuttan
Mitesh Khapra
245
20
0
08 Jul 2024
Zero-Shot Tokenizer Transfer
Zero-Shot Tokenizer TransferNeural Information Processing Systems (NeurIPS), 2024
Benjamin Minixhofer
Edoardo Ponti
Ivan Vulić
VLM
376
29
0
13 May 2024
Bailong: Bilingual Transfer Learning based on QLoRA and Zip-tie
  Embedding
Bailong: Bilingual Transfer Learning based on QLoRA and Zip-tie Embedding
Lung-Chuan Chen
Zong-Ru Li
ALM
320
1
0
01 Apr 2024
Exploring Tokenization Strategies and Vocabulary Sizes for Enhanced
  Arabic Language Models
Exploring Tokenization Strategies and Vocabulary Sizes for Enhanced Arabic Language Models
M. Alrefaie
Nour Eldin Morsy
Nada Samir
247
9
0
17 Mar 2024
Transferring BERT Capabilities from High-Resource to Low-Resource
  Languages Using Vocabulary Matching
Transferring BERT Capabilities from High-Resource to Low-Resource Languages Using Vocabulary Matching
Piotr Rybak
176
5
0
22 Feb 2024
An Empirical Study on Cross-lingual Vocabulary Adaptation for Efficient
  Language Model Inference
An Empirical Study on Cross-lingual Vocabulary Adaptation for Efficient Language Model Inference
Atsuki Yamaguchi
Aline Villavicencio
Nikolaos Aletras
300
21
0
16 Feb 2024
OFA: A Framework of Initializing Unseen Subword Embeddings for Efficient
  Large-scale Multilingual Continued Pretraining
OFA: A Framework of Initializing Unseen Subword Embeddings for Efficient Large-scale Multilingual Continued Pretraining
Yihong Liu
Peiqin Lin
Mingyang Wang
Hinrich Schütze
305
38
0
15 Nov 2023
Extrapolating Large Language Models to Non-English by Aligning Languages
Extrapolating Large Language Models to Non-English by Aligning Languages
Wenhao Zhu
Yunzhe Lv
Qingxiu Dong
Fei Yuan
Jingjing Xu
Shujian Huang
Lingpeng Kong
Jiajun Chen
Lei Li
393
90
0
09 Aug 2023
Distilling Efficient Language-Specific Models for Cross-Lingual Transfer
Distilling Efficient Language-Specific Models for Cross-Lingual TransferAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Alan Ansell
Edoardo Ponti
Anna Korhonen
Ivan Vulić
272
6
0
02 Jun 2023
Mini-Model Adaptation: Efficiently Extending Pretrained Models to New
  Languages via Aligned Shallow Training
Mini-Model Adaptation: Efficiently Extending Pretrained Models to New Languages via Aligned Shallow TrainingAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Kelly Marchisio
Patrick Lewis
Yihong Chen
Mikel Artetxe
317
28
0
20 Dec 2022
GreenPLM: Cross-Lingual Transfer of Monolingual Pre-Trained Language
  Models at Almost No Cost
GreenPLM: Cross-Lingual Transfer of Monolingual Pre-Trained Language Models at Almost No CostInternational Joint Conference on Artificial Intelligence (IJCAI), 2022
Qingcheng Zeng
Lucas Garay
Peilin Zhou
Dading Chong
Yining Hua
Jiageng Wu
Yi-Cheng Pan
Han Zhou
Rob Voigt
Jie Yang
VLM
501
29
0
13 Nov 2022
Lifting the Curse of Multilinguality by Pre-training Modular
  Transformers
Lifting the Curse of Multilinguality by Pre-training Modular TransformersNorth American Chapter of the Association for Computational Linguistics (NAACL), 2022
Jonas Pfeiffer
Naman Goyal
Xi Lin
Xian Li
James Cross
Sebastian Riedel
Mikel Artetxe
LRM
300
169
0
12 May 2022
Language Contamination Helps Explain the Cross-lingual Capabilities of
  English Pretrained Models
Language Contamination Helps Explain the Cross-lingual Capabilities of English Pretrained ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Terra Blevins
Luke Zettlemoyer
391
108
0
17 Apr 2022
Oolong: Investigating What Makes Transfer Learning Hard with Controlled
  Studies
Oolong: Investigating What Makes Transfer Learning Hard with Controlled StudiesConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Zhengxuan Wu
Alex Tamkin
Isabel Papadimitriou
299
15
0
24 Feb 2022
WECHSEL: Effective initialization of subword embeddings for
  cross-lingual transfer of monolingual language models
WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models
Benjamin Minixhofer
Fabian Paischer
Navid Rekabsaz
397
110
0
13 Dec 2021
On the Universality of Deep Contextual Language Models
On the Universality of Deep Contextual Language Models
Shaily Bhatt
Poonam Goyal
Sandipan Dandapat
Monojit Choudhury
Sunayana Sitaram
ELM
346
5
0
15 Sep 2021
Subword Mapping and Anchoring across Languages
Subword Mapping and Anchoring across LanguagesConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
Giorgos Vernikos
Andrei Popescu-Belis
251
15
0
09 Sep 2021
Cross-lingual Transferring of Pre-trained Contextualized Language Models
Cross-lingual Transferring of Pre-trained Contextualized Language Models
Zuchao Li
Kevin Parnow
Hai Zhao
Zhuosheng Zhang
Rui Wang
Masao Utiyama
Eiichiro Sumita
196
8
0
27 Jul 2021
Identifying the Limits of Cross-Domain Knowledge Transfer for Pretrained
  Models
Identifying the Limits of Cross-Domain Knowledge Transfer for Pretrained ModelsWorkshop on Representation Learning for NLP (RepL4NLP), 2021
Zhengxuan Wu
Nelson F. Liu
Christopher Potts
152
5
0
17 Apr 2021
Graph Convolutional Network for Swahili News Classification
Graph Convolutional Network for Swahili News Classification
Alexandros Kastanos
Tyler Martin
GNN
177
4
0
16 Mar 2021
What makes multilingual BERT multilingual?
What makes multilingual BERT multilingual?
Chi-Liang Liu
Tsung-Yuan Hsu
Yung-Sung Chuang
Hung-yi Lee
233
6
0
20 Oct 2020
A Study of Cross-Lingual Ability and Language-specific Information in
  Multilingual BERT
A Study of Cross-Lingual Ability and Language-specific Information in Multilingual BERT
Chi-Liang Liu
Tsung-Yuan Hsu
Yung-Sung Chuang
Hung-yi Lee
218
16
0
20 Apr 2020
A Primer in BERTology: What we know about how BERT works
A Primer in BERTology: What we know about how BERT worksTransactions of the Association for Computational Linguistics (TACL), 2020
Anna Rogers
Olga Kovaleva
Anna Rumshisky
OffRL
626
1,818
0
27 Feb 2020
On the Cross-lingual Transferability of Monolingual Representations
On the Cross-lingual Transferability of Monolingual RepresentationsAnnual Meeting of the Association for Computational Linguistics (ACL), 2019
Mikel Artetxe
Sebastian Ruder
Dani Yogatama
735
851
0
25 Oct 2019
1
Page 1 of 1