449
v1v2v3 (latest)

Auto-Search and Refinement: An Automated Framework for Gender Bias Mitigation in Large Language Models

Main:10 Pages
11 Figures
Bibliography:4 Pages
14 Tables
Appendix:7 Pages
Abstract

Pre-training large language models (LLMs) on vast text corpora enhances natural language processing capabilities but risks encoding social biases, particularly gender bias. While parameter-modification methods like fine-tuning mitigate bias, they are resource-intensive, unsuitable for closed-source models, and lack adaptability to evolving societal norms. Instruction-based approaches offer flexibility but often compromise task performance. To address these limitations, we propose FaIRMaker\textbf{FaIRMaker}, an automated and model-independent framework that employs an auto-search and refinement\textbf{auto-search and refinement} paradigm to adaptively generate Fairwords, which act as instructions integrated into input queries to reduce gender bias and enhance response quality. Extensive experiments demonstrate that FaIRMaker automatically searches for and dynamically refines Fairwords, effectively mitigating gender bias while preserving task integrity and ensuring compatibility with both API-based and open-source LLMs.

View on arXiv
Comments on this paper