Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1908.09369
Cited By
v1
v2
v3 (latest)
On Measuring and Mitigating Biased Inferences of Word Embeddings
25 August 2019
Sunipa Dev
Tao Li
J. M. Phillips
Vivek Srikumar
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"On Measuring and Mitigating Biased Inferences of Word Embeddings"
50 / 101 papers shown
Title
Ming-Omni: A Unified Multimodal Model for Perception and Generation
Inclusion AI
Biao Gong
Cheng Zou
C. Zheng
Chunluan Zhou
...
Zipeng Feng
Zhijiang Fang
Zhihao Qiu
Ziyuan Huang
Z. He
MLLM
AuLLM
77
0
0
11 Jun 2025
Benchmarking and Pushing the Multi-Bias Elimination Boundary of LLMs via Causal Effect Estimation-guided Debiasing
Zhouhao Sun
Zhiyuan Kan
Xiao Ding
Li Du
Yang Zhao
Bing Qin
Ting Liu
89
0
0
22 May 2025
GenderBench: Evaluation Suite for Gender Biases in LLMs
Matúš Pikuliak
81
0
0
17 May 2025
Assumed Identities: Quantifying Gender Bias in Machine Translation of Gender-Ambiguous Occupational Terms
Orfeas Menis Mastromichalakis
Giorgos Filandrianos
Maria Symeonaki
Giorgos Stamou
105
0
0
06 Mar 2025
Analyzing the Safety of Japanese Large Language Models in Stereotype-Triggering Prompts
Akito Nakanishi
Yukie Sano
Geng Liu
Francesco Pierri
95
0
0
03 Mar 2025
Language Models Predict Empathy Gaps Between Social In-groups and Out-groups
Yu Hou
Hal Daumé III
Rachel Rudinger
93
4
0
02 Mar 2025
Structured Reasoning for Fairness: A Multi-Agent Approach to Bias Detection in Textual Data
Tianyi Huang
Elsa Fan
75
1
0
01 Mar 2025
Do LLMs exhibit demographic parity in responses to queries about Human Rights?
Rafiya Javed
Jackie Kay
David Yanni
Abdullah Zaini
Anushe Sheikh
Maribeth Rauh
Iason Gabriel
Laura Weidinger
111
0
0
26 Feb 2025
Bias in Large Language Models: Origin, Evaluation, and Mitigation
Yufei Guo
Muzhe Guo
Juntao Su
Zhou Yang
Mengqiu Zhu
Hongfei Li
Mengyang Qiu
Shuo Shuo Liu
AILaw
101
22
0
16 Nov 2024
Large Language Models Still Exhibit Bias in Long Text
Wonje Jeung
Dongjae Jeon
Ashkan Yousefpour
Jonghyun Choi
ALM
93
7
0
23 Oct 2024
LLMs are Biased Teachers: Evaluating LLM Bias in Personalized Education
Iain Xie Weissburg
Sathvika Anand
Sharon Levy
Haewon Jeong
212
8
0
17 Oct 2024
BenchmarkCards: Standardized Documentation for Large Language Model Benchmarks
Anna Sokol
Elizabeth M. Daly
Michael Hind
David Piorkowski
Xiangliang Zhang
Nuno Moniz
Nitesh Chawla
69
0
0
16 Oct 2024
Collapsed Language Models Promote Fairness
Jingxuan Xu
Wuyang Chen
Linyi Li
Yao Zhao
Yunchao Wei
110
0
0
06 Oct 2024
Fairness Definitions in Language Models Explained
Thang Viet Doan
Zhibo Chu
Zichong Wang
Wenbin Zhang
ALM
95
10
0
26 Jul 2024
Do Generative AI Models Output Harm while Representing Non-Western Cultures: Evidence from A Community-Centered Approach
Sourojit Ghosh
Pranav Narayanan Venkit
Sanjana Gautam
Shomir Wilson
Aylin Caliskan
72
17
0
20 Jul 2024
Evaluating Short-Term Temporal Fluctuations of Social Biases in Social Media Data and Masked Language Models
Yi Zhou
Danushka Bollegala
Jose Camacho-Collados
71
1
0
19 Jun 2024
Exploring Safety-Utility Trade-Offs in Personalized Language Models
Anvesh Rao Vijjini
Somnath Basu Roy Chowdhury
Snigdha Chaturvedi
173
9
0
17 Jun 2024
Deconstructing The Ethics of Large Language Models from Long-standing Issues to New-emerging Dilemmas
Chengyuan Deng
Yiqun Duan
Xin Jin
Heng Chang
Yijun Tian
...
Kuofeng Gao
Sihong He
Jun Zhuang
Lu Cheng
Haohan Wang
AILaw
90
24
0
08 Jun 2024
Ask LLMs Directly, "What shapes your bias?": Measuring Social Bias in Large Language Models
Jisu Shin
Hoyun Song
Huije Lee
Soyeong Jeong
Jong C. Park
107
9
0
06 Jun 2024
Large Language Model Bias Mitigation from the Perspective of Knowledge Editing
Ruizhe Chen
Yichen Li
Zikai Xiao
Zuo-Qiang Liu
KELM
86
14
0
15 May 2024
Believing Anthropomorphism: Examining the Role of Anthropomorphic Cues on Trust in Large Language Models
Michelle Cohn
Mahima Pushkarna
Gbolahan O. Olanubi
Joseph M. Moran
Daniel Padgett
Zion Mengesha
Courtney Heldreth
63
22
0
09 May 2024
Hire Me or Not? Examining Language Model's Behavior with Occupation Attributes
Damin Zhang
Yi Zhang
Geetanjali Bihani
Julia Taylor Rayz
152
3
0
06 May 2024
GeniL: A Multilingual Dataset on Generalizing Language
Aida Mostafazadeh Davani
S. Gubbi
Sunipa Dev
Shachi Dave
Vinodkumar Prabhakaran
82
1
0
08 Apr 2024
Fairness in Large Language Models: A Taxonomic Survey
Zhibo Chu
Zichong Wang
Wenbin Zhang
AILaw
127
42
0
31 Mar 2024
Projective Methods for Mitigating Gender Bias in Pre-trained Language Models
Hillary Dawkins
I. Nejadgholi
Daniel Gillis
J. McCuaig
35
0
0
27 Mar 2024
Evaluating Unsupervised Dimensionality Reduction Methods for Pretrained Sentence Embeddings
Gaifan Zhang
Yi Zhou
Danushka Bollegala
54
4
0
20 Mar 2024
Detecting Bias in Large Language Models: Fine-tuned KcBERT
J. K. Lee
T. M. Chung
68
0
0
16 Mar 2024
Take Care of Your Prompt Bias! Investigating and Mitigating Prompt Bias in Factual Knowledge Extraction
Ziyang Xu
Keqin Peng
Liang Ding
Dacheng Tao
Xiliang Lu
74
10
0
15 Mar 2024
SeeGULL Multilingual: a Dataset of Geo-Culturally Situated Stereotypes
Mukul Bhutani
Kevin Robinson
Vinodkumar Prabhakaran
Shachi Dave
Sunipa Dev
124
12
0
08 Mar 2024
"Flex Tape Can't Fix That": Bias and Misinformation in Edited Language Models
Karina Halevy
Anna Sotnikova
Badr AlKhamissi
Syrielle Montariol
Antoine Bosselut
KELM
79
4
0
29 Feb 2024
A Note on Bias to Complete
Jia Xu
Mona Diab
120
2
0
18 Feb 2024
From Prejudice to Parity: A New Approach to Debiasing Large Language Model Word Embeddings
Aishik Rakshit
Smriti Singh
Shuvam Keshari
Arijit Ghosh Chowdhury
Vinija Jain
Aman Chadha
58
3
0
18 Feb 2024
MAFIA: Multi-Adapter Fused Inclusive LanguAge Models
Prachi Jain
Ashutosh Sathe
Varun Gumma
Kabir Ahuja
Sunayana Sitaram
103
1
0
12 Feb 2024
Measuring Machine Learning Harms from Stereotypes Requires Understanding Who Is Harmed by Which Errors in What Ways
Angelina Wang
Xuechunzi Bai
Solon Barocas
Su Lin Blodgett
FaML
85
5
0
06 Feb 2024
Tackling Bias in Pre-trained Language Models: Current Trends and Under-represented Societies
Vithya Yogarajan
Gillian Dobbie
Te Taka Keegan
R. Neuwirth
ALM
92
13
0
03 Dec 2023
PEFTDebias : Capturing debiasing information using PEFTs
Sumit Agarwal
Aditya Srikanth Veerubhotla
Srijan Bansal
65
3
0
01 Dec 2023
Measuring and Improving Attentiveness to Partial Inputs with Counterfactuals
Yanai Elazar
Bhargavi Paranjape
Hao Peng
Sarah Wiegreffe
Khyathi Raghavi
Vivek Srikumar
Sameer Singh
Noah A. Smith
AAML
OOD
65
0
0
16 Nov 2023
Bias Runs Deep: Implicit Reasoning Biases in Persona-Assigned LLMs
Shashank Gupta
Vaishnavi Shrivastava
Ameet Deshpande
Ashwin Kalyan
Peter Clark
Ashish Sabharwal
Tushar Khot
208
122
0
08 Nov 2023
A Predictive Factor Analysis of Social Biases and Task-Performance in Pretrained Masked Language Models
Yi Zhou
Jose Camacho-Collados
Danushka Bollegala
146
6
0
19 Oct 2023
Co
2
^2
2
PT: Mitigating Bias in Pre-trained Language Models through Counterfactual Contrastive Prompt Tuning
Xiangjue Dong
Ziwei Zhu
Zhuoer Wang
Maria Teleki
James Caverlee
104
11
0
19 Oct 2023
Mitigating Bias for Question Answering Models by Tracking Bias Influence
Mingyu Derek Ma
Jiun-Yu Kao
Arpit Gupta
Yu-Hsiang Lin
Wenbo Zhao
Tagyoung Chung
Wei Wang
Kai-Wei Chang
Nanyun Peng
67
4
0
13 Oct 2023
Large Language Model Alignment: A Survey
Tianhao Shen
Renren Jin
Yufei Huang
Chuang Liu
Weilong Dong
Zishan Guo
Xinwei Wu
Yan Liu
Deyi Xiong
LM&MA
109
205
0
26 Sep 2023
Evaluating Gender Bias of Pre-trained Language Models in Natural Language Inference by Considering All Labels
Panatchakorn Anantaprayoon
Masahiro Kaneko
Naoaki Okazaki
121
18
0
18 Sep 2023
The Impact of Debiasing on the Performance of Language Models in Downstream Tasks is Underestimated
Masahiro Kaneko
Danushka Bollegala
Naoaki Okazaki
100
7
0
16 Sep 2023
Challenges in Annotating Datasets to Quantify Bias in Under-represented Society
Vithya Yogarajan
Gillian Dobbie
Timothy Pistotti
Joshua Bensemann
Kobe Knowles
93
2
0
11 Sep 2023
Bias and Fairness in Large Language Models: A Survey
Isabel O. Gallegos
Ryan Rossi
Joe Barrow
Md Mehrab Tanjim
Sungchul Kim
Franck Dernoncourt
Tong Yu
Ruiyi Zhang
Nesreen Ahmed
AILaw
140
605
0
02 Sep 2023
CALM : A Multi-task Benchmark for Comprehensive Assessment of Language Model Bias
Vipul Gupta
Pranav Narayanan Venkit
Hugo Laurenccon
Shomir Wilson
R. Passonneau
103
14
0
24 Aug 2023
A Survey on Fairness in Large Language Models
Yingji Li
Mengnan Du
Rui Song
Xin Wang
Ying Wang
ALM
126
70
0
20 Aug 2023
Gender-tuning: Empowering Fine-tuning for Debiasing Pre-trained Language Models
Somayeh Ghanbarzadeh
Yan-ping Huang
Hamid Palangi
R. C. Moreno
Hamed Khanpour
68
12
0
20 Jul 2023
Building Socio-culturally Inclusive Stereotype Resources with Community Engagement
Sunipa Dev
J. Goyal
Dinesh Tewari
Shachi Dave
Vinodkumar Prabhakaran
64
26
0
20 Jul 2023
1
2
3
Next