ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2407.03963
  4. Cited By
LLM-jp: A Cross-organizational Project for the Research and Development of Fully Open Japanese LLMs

LLM-jp: A Cross-organizational Project for the Research and Development of Fully Open Japanese LLMs

31 December 2024
LLM-jp
Akiko Aizawa
Eiji Aramaki
Bowen Chen
Fei Cheng
Hiroyuki Deguchi
Rintaro Enomoto
Kazuki Fujii
Kensuke Fukumoto
Takuya Fukushima
Namgi Han
Yuto Harada
Chikara Hashimoto
Tatsuya Hiraoka
Shohei Hisada
Sosuke Hosokawa
Lu Jie
Keisuke Kamata
T. Kanazawa
H. Kanezashi
Hiroshi Kataoka
Satoru Katsumata
Daisuke Kawahara
Seiya Kawano
Atsushi Keyaki
Keisuke Kiryu
Hirokazu Kiyomaru
Takashi Kodama
Takahiro Kubo
Yohei Kuga
Ryoma Kumon
Shuhei Kurita
Sadao Kurohashi
Conglong Li
Taiki Maekawa
Hiroshi Matsuda
Yusuke Miyao
Kentaro Mizuki
Sakae Mizuki
Yugo Murawaki
Akim Mousterou
Ryo Nakamura
Taishi Nakamura
Kouta Nakayama
Tomoka Nakazato
Takuro Niitsuma
Jiro Nishitoba
Yusuke Oda
Hayato Ogawa
Takumi Okamoto
Naoaki Okazaki
Yohei Oseki
Shintaro Ozaki
Koki Ryu
Rafal Rzepka
Keisuke Sakaguchi
S. Sasaki
Satoshi Sekine
Kohei Suda
Saku Sugawara
Issa Sugiura
Hiroaki Sugiyama
Hisami Suzuki
Jun Suzuki
Toyotaro Suzumura
Kensuke Tachibana
Yu Takagi
Kyosuke Takami
Koichi Takeda
Masashi Takeshita
Masahiro Tanaka
Kenjiro Taura
A. Tolmachev
Nobuhiro Ueda
Zhen Wan
Shuntaro Yada
Sakiko Yahata
Yuya Yamamoto
Yusuke Yamauchi
Hitomi Yanaka
Rio Yokota
Koichiro Yoshino
ArXivPDFHTML

Papers citing "LLM-jp: A Cross-organizational Project for the Research and Development of Fully Open Japanese LLMs"

8 / 8 papers shown
Title
Drop-Upcycling: Training Sparse Mixture of Experts with Partial Re-initialization
Drop-Upcycling: Training Sparse Mixture of Experts with Partial Re-initialization
Taishi Nakamura
Takuya Akiba
Kazuki Fujii
Yusuke Oda
Rio Yokota
Jun Suzuki
MoMe
MoE
75
1
0
26 Feb 2025
Analyzing Social Biases in Japanese Large Language Models
Analyzing Social Biases in Japanese Large Language Models
Hitomi Yanaka
Namgi Han
Ryoma Kumon
Jie Lu
Masashi Takeshita
Ryo Sekizawa
Taisei Kato
Hiromi Arai
52
3
0
04 Jun 2024
Jamp: Controlled Japanese Temporal Inference Dataset for Evaluating
  Generalization Capacity of Language Models
Jamp: Controlled Japanese Temporal Inference Dataset for Evaluating Generalization Capacity of Language Models
Tomoki Sugimoto
Yasumasa Onoe
Hitomi Yanaka
53
4
0
19 Jun 2023
Training language models to follow instructions with human feedback
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
303
11,909
0
04 Mar 2022
BBQ: A Hand-Built Bias Benchmark for Question Answering
BBQ: A Hand-Built Bias Benchmark for Question Answering
Alicia Parrish
Angelica Chen
Nikita Nangia
Vishakh Padmakumar
Jason Phang
Jana Thompson
Phu Mon Htut
Sam Bowman
212
367
0
15 Oct 2021
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
Leo Gao
Stella Biderman
Sid Black
Laurence Golding
Travis Hoppe
...
Horace He
Anish Thite
Noa Nabeshima
Shawn Presser
Connor Leahy
AIMat
248
1,986
0
31 Dec 2020
Extracting Training Data from Large Language Models
Extracting Training Data from Large Language Models
Nicholas Carlini
Florian Tramèr
Eric Wallace
Matthew Jagielski
Ariel Herbert-Voss
...
Tom B. Brown
D. Song
Ulfar Erlingsson
Alina Oprea
Colin Raffel
MLAU
SILM
267
1,808
0
14 Dec 2020
Megatron-LM: Training Multi-Billion Parameter Language Models Using
  Model Parallelism
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
243
1,817
0
17 Sep 2019
1