89
v1v2 (latest)

GLM-5: from Vibe Coding to Agentic Engineering

GLM-5-Team
Aohan Zeng
Xin Lv
Zhenyu Hou
Zhengxiao Du
Qinkai Zheng
Bin Chen
Da Yin
Chendi Ge
Chenghua Huang
Chengxing Xie
Chenzheng Zhu
Congfeng Yin
Cunxiang Wang
Gengzheng Pan
Hao Zeng
Haoke Zhang
Haoran Wang
Huilong Chen
Jiajie Zhang
Jian Jiao
Jiaqi Guo
Jingsen Wang
Jingzhao Du
Jinzhu Wu
Kedong Wang
Lei Li
Lin Fan
Lucen Zhong
Mingdao Liu
Mingming Zhao
Pengfan Du
Qian Dong
Rui Lu
Shuang-Li
Shulin Cao
Song Liu
Ting Jiang
Xiaodong Chen
Xiaohan Zhang
Xuancheng Huang
Xuezhen Dong
Yabo Xu
Yao Wei
Yifan An
Yilin Niu
Yitong Zhu
Yuanhao Wen
Yukuo Cen
Yushi Bai
Zhongpei Qiao
Zihan Wang
Zikang Wang
Zilin Zhu
Ziqiang Liu
Zixuan Li
Bojie Wang
Bosi Wen
Can Huang
Changpeng Cai
Chao Yu
Chen Li
Chengwei Hu
Chenhui Zhang
Dan Zhang
Daoyan Lin
Dayong Yang
Di Wang
Ding Ai
Erle Zhu
Fangzhou Yi
Feiyu Chen
Guohong Wen
Hailong Sun
Haisha Zhao
Haiyi Hu
Hanchen Zhang
Hanrui Liu
Hanyu Zhang
Hao Peng
Hao Tai
Haobo Zhang
He Liu
Hongwei Wang
Hongxi Yan
Hongyu Ge
Huan Liu
Huanpeng Chu
Jiañi Zhao
Jiachen Wang
Jiajing Zhao
Jiamin Ren
Jiapeng Wang
Jiaxin Zhang
Jiayi Gui
Jiayue Zhao
Jijie Li
Jing An
Jing Li
Jingwei Yuan
Main:31 Pages
16 Figures
Bibliography:4 Pages
13 Tables
Appendix:5 Pages
Abstract

We present GLM-5, a next-generation foundation model designed to transition the paradigm of vibe coding to agentic engineering. Building upon the agentic, reasoning, and coding (ARC) capabilities of its predecessor, GLM-5 adopts DSA to significantly reduce training and inference costs while maintaining long-context fidelity. To advance model alignment and autonomy, we implement a new asynchronous reinforcement learning infrastructure that drastically improves post-training efficiency by decoupling generation from training. Furthermore, we propose novel asynchronous agent RL algorithms that further improve RL quality, enabling the model to learn from complex, long-horizon interactions more effectively. Through these innovations, GLM-5 achieves state-of-the-art performance on major open benchmarks. Most critically, GLM-5 demonstrates unprecedented capability in real-world coding tasks, surpassing previous baselines in handling end-to-end software engineering challenges. Code, models, and more information are available atthis https URL.

View on arXiv
Comments on this paper