ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.00991
31
0

GOBench: Benchmarking Geometric Optics Generation and Understanding of MLLMs

1 June 2025
X. Zhu
Ziheng Jia
Jiarui Wang
Xiangyu Zhao
Haodong Duan
Xiongkuo Min
Jia Wang
Zicheng Zhang
Guangtao Zhai
    EGVMVLM
ArXiv (abs)PDFHTML
Main:6 Pages
5 Figures
Bibliography:2 Pages
3 Tables
Abstract

The rapid evolution of Multi-modality Large Language Models (MLLMs) is driving significant advancements in visual understanding and generation. Nevertheless, a comprehensive assessment of their capabilities, concerning the fine-grained physical principles especially in geometric optics, remains underexplored. To address this gap, we introduce GOBench, the first benchmark to systematically evaluate MLLMs' ability across two tasks: 1) Generating Optically Authentic Imagery and 2) Understanding Underlying Optical Phenomena. We curates high-quality prompts of geometric optical scenarios and use MLLMs to construct GOBench-Gen-1kthis http URLthen organize subjective experiments to assess the generated imagery based on Optical Authenticity, Aesthetic Quality, and Instruction Fidelity, revealing MLLMs' generation flaws that violate optical principles. For the understanding task, we apply crafted evaluation instructions to test optical understanding ability of eleven prominent MLLMs. The experimental results demonstrate that current models face significant challenges in both optical generation and understanding. The top-performing generative model, GPT-4o-Image, cannot perfectly complete all generation tasks, and the best-performing MLLM model, Gemini-2.5Pro, attains a mere 37.35\% accuracy in optical understanding.

View on arXiv
@article{zhu2025_2506.00991,
  title={ GOBench: Benchmarking Geometric Optics Generation and Understanding of MLLMs },
  author={ Xiaorong Zhu and Ziheng Jia and Jiarui Wang and Xiangyu Zhao and Haodong Duan and Xiongkuo Min and Jia Wang and Zicheng Zhang and Guangtao Zhai },
  journal={arXiv preprint arXiv:2506.00991},
  year={ 2025 }
}
Comments on this paper