ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.07997
52
0

BiasCause: Evaluate Socially Biased Causal Reasoning of Large Language Models

8 April 2025
Tian Xie
Tongxin Yin
Vaishakh Keshava
Xueru Zhang
Siddhartha Reddy Jonnalagadda
    ELM
    LRM
ArXivPDFHTML
Abstract

While large language models (LLMs) already play significant roles in society, research has shown that LLMs still generate content including social bias against certain sensitive groups. While existing benchmarks have effectively identified social biases in LLMs, a critical gap remains in our understanding of the underlying reasoning that leads to these biased outputs. This paper goes one step further to evaluate the causal reasoning process of LLMs when they answer questions eliciting social biases. We first propose a novel conceptual framework to classify the causal reasoning produced by LLMs. Next, we use LLMs to synthesize 178817881788 questions covering 888 sensitive attributes and manually validate them. The questions can test different kinds of causal reasoning by letting LLMs disclose their reasoning process with causal graphs. We then test 4 state-of-the-art LLMs. All models answer the majority of questions with biased causal reasoning, resulting in a total of 413541354135 biased causal graphs. Meanwhile, we discover 333 strategies for LLMs to avoid biased causal reasoning by analyzing the "bias-free" cases. Finally, we reveal that LLMs are also prone to "mistaken-biased" causal reasoning, where they first confuse correlation with causality to infer specific sensitive group names and then incorporate biased causal reasoning.

View on arXiv
@article{xie2025_2504.07997,
  title={ BiasCause: Evaluate Socially Biased Causal Reasoning of Large Language Models },
  author={ Tian Xie and Tongxin Yin and Vaishakh Keshava and Xueru Zhang and Siddhartha Reddy Jonnalagadda },
  journal={arXiv preprint arXiv:2504.07997},
  year={ 2025 }
}
Comments on this paper