228

Yuan3.0 Flash: An Open Multimodal Large Language Model for Enterprise Applications

YuanLab.ai
Shawn Wu
Sean Wang
Louie Li
Darcy Chen
Allen Wang
Jiangang Luo
Xudong Zhao
Joseph Shen
Gawain Ma
Jasper Jia
Marcus Mao
Claire Wang
Hunter He
Carol Wang
Zera Zhang
Jason Wang
Chonly Shen
Leo Zhang
Logan Chen
Qasim Meng
James Gong
Danied Zhao
Penn Zheng
Owen Zhu
Tong Yu
Main:14 Pages
11 Figures
Bibliography:4 Pages
12 Tables
Appendix:2 Pages
Abstract

We introduce Yuan3.0 Flash, an open-source Mixture-of-Experts (MoE) MultiModal Large Language Model featuring 3.7B activated parameters and 40B total parameters, specifically designed to enhance performance on enterprise-oriented tasks while maintaining competitive capabilities on general-purpose tasks. To address the overthinking phenomenon commonly observed in Large Reasoning Models (LRMs), we propose Reflection-aware Adaptive Policy Optimization (RAPO), a novel RL training algorithm that effectively regulates overthinking behaviors. In enterprise-oriented tasks such as retrieval-augmented generation (RAG), complex table understanding, and summarization, Yuan3.0 Flash consistently achieves superior performance. Moreover, it also demonstrates strong reasoning capabilities in domains such as mathematics, science, etc., attaining accuracy comparable to frontier model while requiring only approximately 1/4 to 1/2 of the average tokens. Yuan3.0 Flash has been fully open-sourced to facilitate further research and real-world deployment:this https URL.

View on arXiv
Comments on this paper