Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2010.02428
Cited By
v1
v2
v3 (latest)
UnQovering Stereotyping Biases via Underspecified Questions
Findings (Findings), 2020
6 October 2020
Tao Li
Tushar Khot
Daniel Khashabi
Ashish Sabharwal
Vivek Srikumar
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"UnQovering Stereotyping Biases via Underspecified Questions"
50 / 107 papers shown
PBBQ: A Persian Bias Benchmark Dataset Curated with Human-AI Collaboration for Large Language Models
Farhan Farsi
Shayan Bali
Fatemeh Valeh
Parsa Ghofrani
Alireza Pakniat
Kian Kashfipour
Amir H. Payberah
153
0
0
22 Oct 2025
Open-DeBias: Toward Mitigating Open-Set Bias in Language Models
Arti Rani
Shweta Singh
Nihar Ranjan Sahoo
Gaurav Kumar Nayak
212
0
0
28 Sep 2025
Fair-GPTQ: Bias-Aware Quantization for Large Language Models
Irina Proskurina
Guillaume Metzler
Julien Velcin
MQ
257
0
0
18 Sep 2025
PakBBQ: A Culturally Adapted Bias Benchmark for QA
Abdullah Hashmat
Muhammad Arham Mirza
Agha Ali Raza
209
2
0
13 Aug 2025
FairLangProc: A Python package for fairness in NLP
Arturo Pérez-Peralta
Sandra Benítez-Peña
Rosa E. Lillo
260
0
0
05 Aug 2025
Exploring Gender Bias in Large Language Models: An In-depth Dive into the German Language
Kristin Gnadt
David Thulke
Simone Kopeinik
Ralf Schluter
266
2
0
22 Jul 2025
GG-BBQ: German Gender Bias Benchmark for Question Answering
Shalaka Satheesh
Katrin Klug
Katharina Beckh
Héctor Allende-Cid
Sebastian Houben
Teena Hassan
368
1
0
22 Jul 2025
Ming-Omni: A Unified Multimodal Model for Perception and Generation
Inclusion AI
Biao Gong
Cheng Zou
C. Zheng
Chunluan Zhou
...
Zipeng Feng
Zhijiang Fang
Zhihao Qiu
Ziyuan Huang
Z. He
MLLM
AuLLM
264
5
0
11 Jun 2025
Beyond Bias Scores: Unmasking Vacuous Neutrality in Small Language Models
Sumanth Manduru
Carlotta Domeniconi
ALM
339
0
0
10 Jun 2025
BiasGuard: A Reasoning-enhanced Bias Detection Tool For Large Language Models
Annual Meeting of the Association for Computational Linguistics (ACL), 2025
Zhiting Fan
Ruizhe Chen
Zuozhu Liu
496
4
0
30 Apr 2025
FairSteer: Inference Time Debiasing for LLMs with Dynamic Activation Steering
Annual Meeting of the Association for Computational Linguistics (ACL), 2025
Yongbin Li
Zhiting Fan
Ruizhe Chen
Xiaotang Gai
Luqi Gong
Yan Zhang
Zuozhu Liu
LLMSV
406
26
0
20 Apr 2025
Using LLMs as prompt modifier to avoid biases in AI image generators
René Peinl
247
0
0
15 Apr 2025
DeCAP: Context-Adaptive Prompt Generation for Debiasing Zero-shot Question Answering in Large Language Models
North American Chapter of the Association for Computational Linguistics (NAACL), 2025
Suyoung Bae
YunSeok Choi
Jee-Hyong Lee
306
0
0
25 Mar 2025
Attention Pruning: Automated Fairness Repair of Language Models via Surrogate Simulated Annealing
Vishnu Asutosh Dasu
Md Rafi Ur Rashid
Vipul Gupta
Saeid Tizpaz-Niari
Gang Tan
AAML
562
2
0
20 Mar 2025
LLM Generated Persona is a Promise with a Catch
Ang Li
Haozhe Chen
Hongseok Namkoong
Tianyi Peng
368
57
0
18 Mar 2025
No LLM is Free From Bias: A Comprehensive Study of Bias Evaluation in Large Language Models
Charaka Vinayak Kumar
Ashok Urlana
Gopichand Kanumolu
B. Garlapati
Pruthwik Mishra
ELM
377
10
0
15 Mar 2025
Assumed Identities: Quantifying Gender Bias in Machine Translation of Gender-Ambiguous Occupational Terms
Orfeas Menis Mastromichalakis
Giorgos Filandrianos
Maria Symeonaki
Giorgos Stamou
439
0
0
06 Mar 2025
From Selection to Generation: A Survey of LLM-based Active Learning
Annual Meeting of the Association for Computational Linguistics (ACL), 2025
Yu Xia
Subhojyoti Mukherjee
Zhouhang Xie
Junda Wu
Xintong Li
...
Namyong Park
T. Nguyen
Jiebo Luo
Ryan Rossi
Julian McAuley
468
16
0
17 Feb 2025
Fairness through Difference Awareness: Measuring Desired Group Discrimination in LLMs
Annual Meeting of the Association for Computational Linguistics (ACL), 2025
Angelina Wang
Michelle Phan
Mark A. Lemley
Sanmi Koyejo
522
14
0
04 Feb 2025
Actions Speak Louder than Words: Agent Decisions Reveal Implicit Biases in Language Models
Conference on Fairness, Accountability and Transparency (FAccT), 2025
Yuxuan Li
Hirokazu Shirado
Sauvik Das
239
24
0
29 Jan 2025
LangFair: A Python Package for Assessing Bias and Fairness in Large Language Model Use Cases
Journal of Open Source Software (JOSS), 2025
Dylan Bouchard
Mohit Singh Chauhan
David Skarbrevik
Viren Bajaj
Zeya Ahmad
380
6
0
06 Jan 2025
RAcQUEt: Unveiling the Dangers of Overlooked Referential Ambiguity in Visual LLMs
A. Testoni
Barbara Plank
Raquel Fernández
374
5
0
18 Dec 2024
Contextualized Evaluations: Judging Language Model Responses to Underspecified Queries
Transactions of the Association for Computational Linguistics (TACL), 2024
Chaitanya Malaviya
Joseph Chee Chang
Dan Roth
Mohit Iyyer
Mark Yatskar
Kyle Lo
ELM
338
6
0
11 Nov 2024
FairMT-Bench: Benchmarking Fairness for Multi-turn Dialogue in Conversational LLMs
International Conference on Learning Representations (ICLR), 2024
Zhiting Fan
Ruizhe Chen
Tianxiang Hu
Zuozhu Liu
326
42
0
25 Oct 2024
Large Language Models Still Exhibit Bias in Long Text
Annual Meeting of the Association for Computational Linguistics (ACL), 2024
Wonje Jeung
Dongjae Jeon
Ashkan Yousefpour
Jonghyun Choi
ALM
571
15
0
23 Oct 2024
LLMs are Biased Teachers: Evaluating LLM Bias in Personalized Education
North American Chapter of the Association for Computational Linguistics (NAACL), 2024
Iain Xie Weissburg
Sathvika Anand
Sharon Levy
Haewon Jeong
592
34
0
17 Oct 2024
Bias Similarity Measurement: A Black-Box Audit of Fairness Across LLMs
Hyejun Jeong
Shiqing Ma
Amir Houmansadr
519
0
0
15 Oct 2024
Does RAG Introduce Unfairness in LLMs? Evaluating Fairness in Retrieval-Augmented Generation Systems
International Conference on Computational Linguistics (COLING), 2024
Xuyang Wu
Shuowei Li
Hsin-Tai Wu
Zhiqiang Tao
Yi Fang
551
30
0
29 Sep 2024
'Simulacrum of Stories': Examining Large Language Models as Qualitative Research Participants
International Conference on Human Factors in Computing Systems (CHI), 2024
Shivani Kapania
William Agnew
Motahhare Eslami
Hoda Heidari
Sarah E Fox
285
28
0
28 Sep 2024
BeanCounter: A low-toxicity, large-scale, and open dataset of business-oriented text
Neural Information Processing Systems (NeurIPS), 2024
Siyan Wang
Bradford Levy
398
3
0
26 Sep 2024
REFINE-LM: Mitigating Language Model Stereotypes via Reinforcement Learning
European Conference on Artificial Intelligence (ECAI), 2024
Rameez Qureshi
Naim Es-Sebbani
Luis Galárraga
Yvette Graham
Miguel Couceiro
Zied Bouraoui
252
1
0
18 Aug 2024
Understanding the Interplay of Scale, Data, and Bias in Language Models: A Case Study with BERT
Muhammad Ali
Swetasudha Panda
Qinlan Shen
Michael Wick
Ari Kobren
MILM
235
6
0
25 Jul 2024
GermanPartiesQA: Benchmarking Commercial Large Language Models and AI Companions for Political Alignment and Sycophancy
Jan Batzner
Volker Stocker
Stefan Schmid
Gjergji Kasneci
307
7
0
25 Jul 2024
BiasAlert: A Plug-and-play Tool for Social Bias Detection in LLMs
Zhiting Fan
Ruizhe Chen
Ruiling Xu
Zuozhu Liu
KELM
459
32
0
14 Jul 2024
Beyond Perplexity: Multi-dimensional Safety Evaluation of LLM Compression
Zhichao Xu
Ashim Gupta
Tao Li
Oliver Bentham
Vivek Srikumar
495
29
0
06 Jul 2024
Social Bias Evaluation for Large Language Models Requires Prompt Variations
Rem Hida
Masahiro Kaneko
Naoaki Okazaki
361
39
0
03 Jul 2024
The Art of Saying No: Contextual Noncompliance in Language Models
Faeze Brahman
Sachin Kumar
Vidhisha Balachandran
Pradeep Dasigi
Valentina Pyatkin
...
Jack Hessel
Yulia Tsvetkov
Noah A. Smith
Yejin Choi
Hannaneh Hajishirzi
379
72
0
02 Jul 2024
Interpreting Bias in Large Language Models: A Feature-Based Approach
Nirmalendu Prakash
Lee Ka Wei Roy
325
4
0
18 Jun 2024
MBBQ: A Dataset for Cross-Lingual Comparison of Stereotypes in Generative LLMs
Vera Neplenbroek
Arianna Bisazza
Raquel Fernández
432
30
0
11 Jun 2024
Ask LLMs Directly, "What shapes your bias?": Measuring Social Bias in Large Language Models
Jisu Shin
Hoyun Song
Huije Lee
Soyeong Jeong
Jong C. Park
393
16
0
06 Jun 2024
JBBQ: Japanese Bias Benchmark for Analyzing Social Biases in Large Language Models
Hitomi Yanaka
Namgi Han
Ryoma Kumon
Jie Lu
Masashi Takeshita
Ryo Sekizawa
Taisei Kato
Hiromi Arai
438
4
0
04 Jun 2024
TrojanRAG: Retrieval-Augmented Generation Can Be Backdoor Driver in Large Language Models
Pengzhou Cheng
Yidong Ding
Tianjie Ju
Zongru Wu
Wei Du
Ping Yi
Zhuosheng Zhang
Gongshen Liu
SILM
AAML
522
60
0
22 May 2024
Hire Me or Not? Examining Language Model's Behavior with Occupation Attributes
International Conference on Computational Linguistics (COLING), 2024
Damin Zhang
Yi Zhang
Geetanjali Bihani
Julia Taylor Rayz
533
4
0
06 May 2024
Fairness in Large Language Models: A Taxonomic Survey
Zhibo Chu
Sribala Vidyadhari Chinta
Wenbin Zhang
AILaw
293
101
0
31 Mar 2024
From Representational Harms to Quality-of-Service Harms: A Case Study on Llama 2 Safety Safeguards
Khaoula Chehbouni
Megha Roshan
Emmanuel Ma
Futian Andrew Wei
Afaf Taik
Jackie CK Cheung
G. Farnadi
437
17
0
20 Mar 2024
Evaluating Biases in Context-Dependent Health Questions
Sharon Levy
T. Karver
William D. Adler
Michelle R. Kaufman
Mark Dredze
251
7
0
07 Mar 2024
FairBelief -- Assessing Harmful Beliefs in Language Models
Mattia Setzu
Marta Marchiori Manerba
Pasquale Minervini
Debora Nozza
267
0
0
27 Feb 2024
Bias and Volatility: A Statistical Framework for Evaluating Large Language Model's Stereotypes and the Associated Generation Inconsistency
Yiran Liu
Ke Yang
Zehan Qi
Xiao-Yang Liu
Yang Yu
ChengXiang Zhai
628
1
0
23 Feb 2024
KorNAT: LLM Alignment Benchmark for Korean Social Values and Common Knowledge
Jiyoung Lee
Minwoo Kim
Seungho Kim
Junghwan Kim
Seunghyun Won
Hwaran Lee
Edward Choi
ALM
695
39
0
21 Feb 2024
Your Large Language Model is Secretly a Fairness Proponent and You Should Prompt it Like One
Tianlin Li
Xiaoyu Zhang
Chao Du
Tianyu Pang
Qian Liu
Qing Guo
Chao Shen
Yang Liu
ALM
325
15
0
19 Feb 2024
1
2
3
Next
Page 1 of 3