ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.00748
21
0

Translate With Care: Addressing Gender Bias, Neutrality, and Reasoning in Large Language Model Translations

31 May 2025
Pardis Sadat Zahraei
Ali Emami
ArXiv (abs)PDFHTML
Main:9 Pages
2 Figures
Bibliography:4 Pages
24 Tables
Appendix:13 Pages
Abstract

Addressing gender bias and maintaining logical coherence in machine translation remains challenging, particularly when translating between natural gender languages, like English, and genderless languages, such as Persian, Indonesian, and Finnish. We introduce the Translate-with-Care (TWC) dataset, comprising 3,950 challenging scenarios across six low- to mid-resource languages, to assess translation systems' performance. Our analysis of diverse technologies, including GPT-4, mBART-50, NLLB-200, and Google Translate, reveals a universal struggle in translating genderless content, resulting in gender stereotyping and reasoning errors. All models preferred masculine pronouns when gender stereotypes could influence choices. Google Translate and GPT-4 showed particularly strong bias, favoring male pronouns 4-6 times more than feminine ones in leadership and professional success contexts. Fine-tuning mBART-50 on TWC substantially resolved these biases and errors, led to strong generalization, and surpassed proprietary LLMs while remaining open-source. This work emphasizes the need for targeted approaches to gender and semantic coherence in machine translation, particularly for genderless languages, contributing to more equitable and accurate translation systems.

View on arXiv
@article{zahraei2025_2506.00748,
  title={ Translate With Care: Addressing Gender Bias, Neutrality, and Reasoning in Large Language Model Translations },
  author={ Pardis Sadat Zahraei and Ali Emami },
  journal={arXiv preprint arXiv:2506.00748},
  year={ 2025 }
}
Comments on this paper