ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2310.06266
21
12

CodeFuse-13B: A Pretrained Multi-lingual Code Large Language Model

10 October 2023
Peng Di
Jianguo Li
Hang Yu
Wei Jiang
Wenting Cai
Yang Cao
Chaoyu Chen
Dajun Chen
Hongwei Chen
Liang Chen
Gang Fan
Jie Gong
Zi Gong
Wen Hu
Tingting Guo
Zhichao Lei
Ting Li
Zheng Li
Ming Liang
Cong Liao
Bingchang Liu
Jiachen Liu
Zhiwei Liu
Shaojun Lu
Mingquan Shen
Guangpei Wang
Huan Wang
Z. Wang
Zhaogui Xu
Jiawei Yang
Qing Ye
Gehao Zhang
Yu Zhang
Zelin Zhao
Xunjin Zheng
Hailian Zhou
Lifu Zhu
Xianying Zhu
    ELM
    ALM
    AI4CE
ArXivPDFHTML
Abstract

Code Large Language Models (Code LLMs) have gained significant attention in the industry due to their wide applications in the full lifecycle of software engineering. However, the effectiveness of existing models in understanding non-English inputs for multi-lingual code-related tasks is still far from well studied. This paper introduces CodeFuse-13B, an open-sourced pre-trained code LLM. It is specifically designed for code-related tasks with both English and Chinese prompts and supports over 40 programming languages. CodeFuse achieves its effectiveness by utilizing a high quality pre-training dataset that is carefully filtered by program analyzers and optimized during the training process. Extensive experiments are conducted using real-world usage scenarios, the industry-standard benchmark HumanEval-x, and the specially designed CodeFuseEval for Chinese prompts. To assess the effectiveness of CodeFuse, we actively collected valuable human feedback from the AntGroup's software development process where CodeFuse has been successfully deployed. The results demonstrate that CodeFuse-13B achieves a HumanEval pass@1 score of 37.10%, positioning it as one of the top multi-lingual code LLMs with similar parameter sizes. In practical scenarios, such as code generation, code translation, code comments, and testcase generation, CodeFuse performs better than other models when confronted with Chinese prompts.

View on arXiv
Comments on this paper