On Domain-Specific Post-Training for Multimodal Large Language Models

Adapting general multimodal large language models (MLLMs) to specific domains, such as scientific and industrial fields, is highly significant in promoting their practical applications. This paper systematically investigates domain adaptation of MLLMs through post-training, focusing on data synthesis, training pipelines, and task evaluation. (1) Data Synthesis: Using only open-source models, we develop a generate-then-filter pipeline that curates diverse visual instruction tasks based on domain-specific image-caption pairs. The resulting data surpass the data synthesized by manual rules or strong closed-source models (e.g., GPT-4V) in enhancing domain-specific performance. (2) Training Pipeline: While the two-stage training--initially on image-caption pairs followed by visual instruction tasks--is commonly adopted for developing general MLLMs, we apply a single-stage training pipeline to enhance task diversity for domain-specific post-training. (3) Task Evaluation: We conduct extensive experiments in high-impact domains such as biomedicine, food, and remote sensing, by post-training a variety of MLLMs and then evaluating MLLM performance on various domain-specific tasks. Furthermore, we fully open-source our models, code, and data to encourage future research in this area.
View on arXiv@article{cheng2025_2411.19930, title={ On Domain-Specific Post-Training for Multimodal Large Language Models }, author={ Daixuan Cheng and Shaohan Huang and Ziyu Zhu and Xintong Zhang and Wayne Xin Zhao and Zhongzhi Luan and Bo Dai and Zhenliang Zhang }, journal={arXiv preprint arXiv:2411.19930}, year={ 2025 } }