ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.16508
  4. Cited By
Large Vocabulary Size Improves Large Language Models

Large Vocabulary Size Improves Large Language Models

24 June 2024
Sho Takase
Ryokan Ri
Shun Kiyono
Takuya Kato
ArXivPDFHTML

Papers citing "Large Vocabulary Size Improves Large Language Models"

7 / 7 papers shown
Title
The Rise of Small Language Models in Healthcare: A Comprehensive Survey
The Rise of Small Language Models in Healthcare: A Comprehensive Survey
Muskan Garg
Shaina Raza
Shebuti Rayana
Xingyi Liu
Sunghwan Sohn
LM&MA
AILaw
87
0
0
23 Apr 2025
Chem42: a Family of chemical Language Models for Target-aware Ligand Generation
Chem42: a Family of chemical Language Models for Target-aware Ligand Generation
A. Singh
Engin Tekin
Maryam Nadeem
Nancy A. ElNaker
Mohammad Amaan Sayeed
Natalia Vassilieva
Boulbaba Ben Amor
39
1
0
20 Mar 2025
The Fair Language Model Paradox
The Fair Language Model Paradox
Andrea Pinto
Tomer Galanti
Randall Balestriero
18
0
0
15 Oct 2024
What Language Model to Train if You Have One Million GPU Hours?
What Language Model to Train if You Have One Million GPU Hours?
Teven Le Scao
Thomas Wang
Daniel Hesslow
Lucile Saulnier
Stas Bekman
...
Lintang Sutawika
Jaesung Tae
Zheng-Xin Yong
Julien Launay
Iz Beltagy
MoE
AI4CE
215
103
0
27 Oct 2022
Why don't people use character-level machine translation?
Why don't people use character-level machine translation?
Jindrich Libovický
Helmut Schmid
Alexander M. Fraser
59
28
0
15 Oct 2021
Scaling Laws for Neural Language Models
Scaling Laws for Neural Language Models
Jared Kaplan
Sam McCandlish
T. Henighan
Tom B. Brown
B. Chess
R. Child
Scott Gray
Alec Radford
Jeff Wu
Dario Amodei
220
4,424
0
23 Jan 2020
Megatron-LM: Training Multi-Billion Parameter Language Models Using
  Model Parallelism
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
243
1,791
0
17 Sep 2019
1