Fine-Grained Evaluation of Large Vision-Language Models in Autonomous Driving

Existing benchmarks for Vision-Language Model (VLM) on autonomous driving (AD) primarily assess interpretability through open-form visual question answering (QA) within coarse-grained tasks, which remain insufficient to assess capabilities in complex driving scenarios. To this end, we introduce , a challenging and fine-grained dataset featuring close-form QAs that progress from static foundational knowledge and elements to advanced reasoning for dynamic on-road situations. The elaborate spans 5 key domains: Traffic Knowledge Understanding, General Element Recognition, Traffic Graph Generation, Target Attribute Comprehension, and Ego Decision-Making and Planning. These domains are further broken down into 11 secondary aspects and 29 tertiary tasks for a granular evaluation. A thorough assessment of general and domain-specific (DS) VLMs on this benchmark reveals both their strengths and critical limitations in AD contexts. To further exploit the cognitive and reasoning interactions among the 5 domains for AD understanding, we start from a small-scale VLM and train the DS models on individual domain datasets (collected from 1.4M DS QAs across public sources). The experimental results demonstrate that the proposed benchmark provides a crucial step toward a more comprehensive assessment of VLMs in AD, paving the way for the development of more cognitively sophisticated and reasoning-capable AD systems.
View on arXiv@article{li2025_2503.21505, title={ Fine-Grained Evaluation of Large Vision-Language Models in Autonomous Driving }, author={ Yue Li and Meng Tian and Zhenyu Lin and Jiangtong Zhu and Dechang Zhu and Haiqiang Liu and Zining Wang and Yueyi Zhang and Zhiwei Xiong and Xinhai Zhao }, journal={arXiv preprint arXiv:2503.21505}, year={ 2025 } }