ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2411.16508
105
8

All Languages Matter: Evaluating LMMs on Culturally Diverse 100 Languages

25 November 2024
Ashmal Vayani
Dinura Dissanayake
Hasindri Watawana
Noor Ahsan
Nevasini Sasikumar
Omkar Thawakar
H. Ademtew
Yahya Hmaiti
Amandeep Kumar
K. K.
Mykola Maslych
Wafa Al Ghallabi
M. Mihaylov
Chao Qin
Abdelrahman M. Shaker
Mike Zhang
Mahardika Krisna Ihsani
Amiel Esplana
Monil Gokani
Shachar Mirkin
Harsh Singh
Ashay Srivastava
Endre Hamerlik
Fathinah Asma Izzati
F. Maani
Sebastian Cavada
Jenny Chim
Rohit Gupta
Sanjay Manjunath
Kamila Zhumakhanova
Feno Heriniaina Rabevohitra
Azril Amirudin
Muhammad Ridzuan
Daniya Najiha Abdul Kareem
Ketan More
Kunyang Li
Pramesh Shakya
Muhammad Saad
Amirpouya Ghasemaghaei
Amirbek Djanibekov
Dilshod Azizov
Branislava Jankovic
Naman Bhatia
Alvaro Cabrera
Johan Obando-Ceron
Olympiah Otieno
Fabian Farestam
Muztoba Rabbani
Sanoojan Baliah
Santosh Sanjeev
Abduragim Shtanchaev
Maheen Fatima
Thao Nguyen
Amrin Kareem
Toluwani Aremu
Nathan Xavier
Amit Bhatkal
Hawau Olamide Toyin
Aman Chadha
Hisham Cholakkal
Rao Muhammad Anwer
M. Felsberg
Jorma T. Laaksonen
Thamar Solorio
Monojit Choudhury
Ivan Laptev
Mubarak Shah
Salman Khan
Fahad A Khan
ArXivPDFHTML
Abstract

Existing Large Multimodal Models (LMMs) generally focus on only a few regions and languages. As LMMs continue to improve, it is increasingly important to ensure they understand cultural contexts, respect local sensitivities, and support low-resource languages, all while effectively integrating corresponding visual cues. In pursuit of culturally diverse global multimodal models, our proposed All Languages Matter Benchmark (ALM-bench) represents the largest and most comprehensive effort to date for evaluating LMMs across 100 languages. ALM-bench challenges existing models by testing their ability to understand and reason about culturally diverse images paired with text in various languages, including many low-resource languages traditionally underrepresented in LMM research. The benchmark offers a robust and nuanced evaluation framework featuring various question formats, including true/false, multiple choice, and open-ended questions, which are further divided into short and long-answer categories. ALM-bench design ensures a comprehensive assessment of a model's ability to handle varied levels of difficulty in visual and linguistic reasoning. To capture the rich tapestry of global cultures, ALM-bench carefully curates content from 13 distinct cultural aspects, ranging from traditions and rituals to famous personalities and celebrations. Through this, ALM-bench not only provides a rigorous testing ground for state-of-the-art open and closed-source LMMs but also highlights the importance of cultural and linguistic inclusivity, encouraging the development of models that can serve diverse global populations effectively. Our benchmark is publicly available.

View on arXiv
@article{vayani2025_2411.16508,
  title={ All Languages Matter: Evaluating LMMs on Culturally Diverse 100 Languages },
  author={ Ashmal Vayani and Dinura Dissanayake and Hasindri Watawana and Noor Ahsan and Nevasini Sasikumar and Omkar Thawakar and Henok Biadglign Ademtew and Yahya Hmaiti and Amandeep Kumar and Kartik Kuckreja and Mykola Maslych and Wafa Al Ghallabi and Mihail Mihaylov and Chao Qin and Abdelrahman M Shaker and Mike Zhang and Mahardika Krisna Ihsani and Amiel Esplana and Monil Gokani and Shachar Mirkin and Harsh Singh and Ashay Srivastava and Endre Hamerlik and Fathinah Asma Izzati and Fadillah Adamsyah Maani and Sebastian Cavada and Jenny Chim and Rohit Gupta and Sanjay Manjunath and Kamila Zhumakhanova and Feno Heriniaina Rabevohitra and Azril Amirudin and Muhammad Ridzuan and Daniya Kareem and Ketan More and Kunyang Li and Pramesh Shakya and Muhammad Saad and Amirpouya Ghasemaghaei and Amirbek Djanibekov and Dilshod Azizov and Branislava Jankovic and Naman Bhatia and Alvaro Cabrera and Johan Obando-Ceron and Olympiah Otieno and Fabian Farestam and Muztoba Rabbani and Sanoojan Baliah and Santosh Sanjeev and Abduragim Shtanchaev and Maheen Fatima and Thao Nguyen and Amrin Kareem and Toluwani Aremu and Nathan Xavier and Amit Bhatkal and Hawau Toyin and Aman Chadha and Hisham Cholakkal and Rao Muhammad Anwer and Michael Felsberg and Jorma Laaksonen and Thamar Solorio and Monojit Choudhury and Ivan Laptev and Mubarak Shah and Salman Khan and Fahad Khan },
  journal={arXiv preprint arXiv:2411.16508},
  year={ 2025 }
}
Comments on this paper