Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1810.03611
Cited By
Understanding the Origins of Bias in Word Embeddings
8 October 2018
Marc-Etienne Brunet
Colleen Alkalay-Houlihan
Ashton Anderson
R. Zemel
FaML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Understanding the Origins of Bias in Word Embeddings"
34 / 34 papers shown
Title
Mechanistic Unveiling of Transformer Circuits: Self-Influence as a Key to Model Reasoning
L. Zhang
Lijie Hu
Di Wang
LRM
95
0
0
17 Feb 2025
Fine-Tuned LLMs are "Time Capsules" for Tracking Societal Bias Through Books
Sangmitra Madhusudan
Robert D Morabito
Skye Reid
Nikta Gohari Sadr
Ali Emami
58
0
0
07 Feb 2025
Data Debugging is NP-hard for Classifiers Trained with SGD
Zizheng Guo
Pengyu Chen
Yanzhang Fu
Xuelong Li
28
0
0
02 Aug 2024
Data Quality in Edge Machine Learning: A State-of-the-Art Survey
M. D. Belgoumri
Mohamed Reda Bouadjenek
Sunil Aryal
Hakim Hacid
41
1
0
01 Jun 2024
Best of Both Worlds: A Pliable and Generalizable Neuro-Symbolic Approach for Relation Classification
Robert Vacareanu
F. Alam
M. Islam
Haris Riaz
Mihai Surdeanu
NAI
27
2
0
05 Mar 2024
ConFit: Improving Resume-Job Matching using Data Augmentation and Contrastive Learning
Xiao Yu
Jinzhong Zhang
Zhou Yu
35
1
0
29 Jan 2024
Deeper Understanding of Black-box Predictions via Generalized Influence Functions
Hyeonsu Lyu
Jonggyu Jang
Sehyun Ryu
H. Yang
TDI
AI4CE
18
5
0
09 Dec 2023
A Survey on Fairness in Large Language Models
Yingji Li
Mengnan Du
Rui Song
Xin Wang
Ying Wang
ALM
49
59
0
20 Aug 2023
Gender-tuning: Empowering Fine-tuning for Debiasing Pre-trained Language Models
Somayeh Ghanbarzadeh
Yan-ping Huang
Hamid Palangi
R. C. Moreno
Hamed Khanpour
32
12
0
20 Jul 2023
Taught by the Internet, Exploring Bias in OpenAIs GPT3
Ali Ayaz
Aditya Nawalgaria
Ruilian Yin
18
0
0
04 Jun 2023
A Survey of Trustworthy Federated Learning with Perspectives on Security, Robustness, and Privacy
Yifei Zhang
Dun Zeng
Jinglong Luo
Zenglin Xu
Irwin King
FedML
84
47
0
21 Feb 2023
Influence Functions for Sequence Tagging Models
Sarthak Jain
Varun Manjunatha
Byron C. Wallace
A. Nenkova
TDI
30
8
0
25 Oct 2022
A methodology to characterize bias and harmful stereotypes in natural language processing in Latin America
Laura Alonso Alemany
Luciana Benotti
Hernán Maina
Lucía González
Mariela Rajngewerc
...
Guido Ivetta
Alexia Halvorsen
Amanda Rojo
M. Bordone
Beatriz Busaniche
29
3
0
14 Jul 2022
The Problem of Semantic Shift in Longitudinal Monitoring of Social Media: A Case Study on Mental Health During the COVID-19 Pandemic
Keith Harrigian
Mark Dredze
16
5
0
22 Jun 2022
Regional Negative Bias in Word Embeddings Predicts Racial Animus--but only via Name Frequency
Austin Van Loon
Salvatore Giorgi
Robb Willer
J. Eichstaedt
39
10
0
20 Jan 2022
Scaling Up Influence Functions
Andrea Schioppa
Polina Zablotskaia
David Vilar
Artem Sokolov
TDI
25
90
0
06 Dec 2021
Developing a novel fair-loan-predictor through a multi-sensitive debiasing pipeline: DualFair
Ashutosh Kumar Singh
Jashandeep Singh
Ariba Khan
Amar Gupta
FaML
19
3
0
17 Oct 2021
Low Frequency Names Exhibit Bias and Overfitting in Contextualizing Language Models
Robert Wolfe
Aylin Caliskan
85
51
0
01 Oct 2021
FairCanary: Rapid Continuous Explainable Fairness
Avijit Ghosh
Aalok Shanbhag
Christo Wilson
11
20
0
13 Jun 2021
Evaluating Gender Bias in Natural Language Inference
Shanya Sharma
Manan Dey
Koustuv Sinha
20
41
0
12 May 2021
On the Interpretability and Significance of Bias Metrics in Texts: a PMI-based Approach
Francisco Valentini
Germán Rosati
Damián E. Blasi
D. Slezak
Edgar Altszyler
22
3
0
13 Apr 2021
Probing Multimodal Embeddings for Linguistic Properties: the Visual-Semantic Case
Adam Dahlgren Lindström
Suna Bensch
Johanna Björklund
F. Drewes
16
20
0
22 Feb 2021
FastIF: Scalable Influence Functions for Efficient Model Interpretation and Debugging
Han Guo
Nazneen Rajani
Peter Hase
Mohit Bansal
Caiming Xiong
TDI
33
102
0
31 Dec 2020
Cross-Loss Influence Functions to Explain Deep Network Representations
Andrew Silva
Rohit Chopra
Matthew C. Gombolay
TDI
21
15
0
03 Dec 2020
Image Representations Learned With Unsupervised Pre-Training Contain Human-like Biases
Ryan Steed
Aylin Caliskan
SSL
19
156
0
28 Oct 2020
Cultural Cartography with Word Embeddings
Dustin S. Stoltz
Marshall A. Taylor
23
38
0
09 Jul 2020
Influence Functions in Deep Learning Are Fragile
S. Basu
Phillip E. Pope
S. Feizi
TDI
28
219
0
25 Jun 2020
Two Simple Ways to Learn Individual Fairness Metrics from Data
Debarghya Mukherjee
Mikhail Yurochkin
Moulinath Banerjee
Yuekai Sun
FaML
26
96
0
19 Jun 2020
Algorithmic Fairness
Dana Pessach
E. Shmueli
FaML
33
388
0
21 Jan 2020
Generating Interactive Worlds with Text
Angela Fan
Jack Urbanek
Pratik Ringshia
Emily Dinan
Emma Qian
...
Shrimai Prabhumoye
Douwe Kiela
Tim Rocktaschel
Arthur Szlam
Jason Weston
16
27
0
20 Nov 2019
Queens are Powerful too: Mitigating Gender Bias in Dialogue Generation
Emily Dinan
Angela Fan
Adina Williams
Jack Urbanek
Douwe Kiela
Jason Weston
22
205
0
10 Nov 2019
Assessing Social and Intersectional Biases in Contextualized Word Representations
Y. Tan
Elisa Celis
FaML
19
223
0
04 Nov 2019
Man is to Person as Woman is to Location: Measuring Gender Bias in Named Entity Recognition
Ninareh Mehrabi
Thamme Gowda
Fred Morstatter
Nanyun Peng
Aram Galstyan
12
57
0
24 Oct 2019
A Survey on Bias and Fairness in Machine Learning
Ninareh Mehrabi
Fred Morstatter
N. Saxena
Kristina Lerman
Aram Galstyan
SyDa
FaML
323
4,203
0
23 Aug 2019
1