ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.11861
  4. Cited By
Fair-PP: A Synthetic Dataset for Aligning LLM with Personalized Preferences of Social Equity

Fair-PP: A Synthetic Dataset for Aligning LLM with Personalized Preferences of Social Equity

17 May 2025
Qi Zhou
Jie Zhang
Dongxia Wang
Qiang Liu
Tianlin Li
Jin Song Dong
Wenhai Wang
Qing Guo
    SyDa
ArXiv (abs)PDFHTML

Papers citing "Fair-PP: A Synthetic Dataset for Aligning LLM with Personalized Preferences of Social Equity"

20 / 20 papers shown
MonoCLUE : Object-Aware Clustering Enhances Monocular 3D Object Detection
MonoCLUE : Object-Aware Clustering Enhances Monocular 3D Object DetectionAnnual Conference of the IEEE Industrial Electronics Society (IECON), 2022
Sunghun Yang
M. Lee
Jungho Lee
Sangyoun Lee
136
0
0
11 Nov 2025
SEA-LION: Southeast Asian Languages in One Network
SEA-LION: Southeast Asian Languages in One Network
Raymond Ng
Thanh Ngan Nguyen
Yuli Huang
Ngee Chia Tai
Wai Yi Leong
...
David Ong Tat-Wee
B. Liu
William-Chandra Tjhi
Xiaoshi Zhong
Leslie Teo
431
25
0
08 Apr 2025
SafeWorld: Geo-Diverse Safety Alignment
SafeWorld: Geo-Diverse Safety AlignmentNeural Information Processing Systems (NeurIPS), 2024
Da Yin
Haoyi Qiu
Kung-Hsiang Huang
Kai-Wei Chang
Nanyun Peng
339
11
0
09 Dec 2024
Interpretable Preferences via Multi-Objective Reward Modeling and
  Mixture-of-Experts
Interpretable Preferences via Multi-Objective Reward Modeling and Mixture-of-ExpertsConference on Empirical Methods in Natural Language Processing (EMNLP), 2024
Haoxiang Wang
Wei Xiong
Tengyang Xie
Han Zhao
Tong Zhang
316
307
0
18 Jun 2024
Safety Alignment Should Be Made More Than Just a Few Tokens Deep
Safety Alignment Should Be Made More Than Just a Few Tokens DeepInternational Conference on Learning Representations (ICLR), 2024
Xiangyu Qi
Ashwinee Panda
Kaifeng Lyu
Xiao Ma
Subhrajit Roy
Ahmad Beirami
Prateek Mittal
Peter Henderson
250
283
0
10 Jun 2024
Group Robust Preference Optimization in Reward-free RLHF
Group Robust Preference Optimization in Reward-free RLHF
Shyam Sundhar Ramesh
Yifan Hu
Iason Chaimalas
Viraj Mehta
Pier Giuseppe Sessa
Haitham Bou-Ammar
Ilija Bogunovic
329
87
0
30 May 2024
WorldValuesBench: A Large-Scale Benchmark Dataset for Multi-Cultural
  Value Awareness of Language Models
WorldValuesBench: A Large-Scale Benchmark Dataset for Multi-Cultural Value Awareness of Language Models
Wenlong Zhao
Debanjan Mondal
Niket Tandon
Danica Dillion
Kurt Gray
Yuling Gu
VLM
283
38
0
25 Apr 2024
Investigating Cultural Alignment of Large Language Models
Investigating Cultural Alignment of Large Language Models
Badr AlKhamissi
Muhammad N. ElNokrashy
Mai AlKhamissi
Mona T. Diab
407
119
0
20 Feb 2024
CultureLLM: Incorporating Cultural Differences into Large Language
  Models
CultureLLM: Incorporating Cultural Differences into Large Language Models
Cheng-rong Li
Mengzhou Chen
Yongfeng Zhang
Sunayana Sitaram
Xing Xie
VLM
315
51
0
09 Feb 2024
Cultural Bias and Cultural Alignment of Large Language Models
Cultural Bias and Cultural Alignment of Large Language ModelsPNAS Nexus (PNAS Nexus), 2023
Yan Tao
Olga Viberg
Ryan S. Baker
René F. Kizilcec
ELM
461
222
0
23 Nov 2023
Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights,
  and Duties
Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and DutiesAAAI Conference on Artificial Intelligence (AAAI), 2023
Taylor Sorensen
Liwei Jiang
Jena D. Hwang
Sydney Levine
Valentina Pyatkin
...
Kavel Rao
Chandra Bhagavatula
Maarten Sap
J. Tasioulas
Yejin Choi
SLR
509
92
0
02 Sep 2023
BeaverTails: Towards Improved Safety Alignment of LLM via a
  Human-Preference Dataset
BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference DatasetNeural Information Processing Systems (NeurIPS), 2023
Jiaming Ji
Mickel Liu
Juntao Dai
Xuehai Pan
Chi Zhang
Ce Bian
Chi Zhang
Ruiyang Sun
Yizhou Wang
Yaodong Yang
ALM
414
733
0
10 Jul 2023
Towards Measuring the Representation of Subjective Global Opinions in
  Language Models
Towards Measuring the Representation of Subjective Global Opinions in Language Models
Esin Durmus
Karina Nyugen
Thomas I. Liao
Nicholas Schiefer
Amanda Askell
...
Alex Tamkin
Janel Thamkul
Jared Kaplan
Jack Clark
Deep Ganguli
368
341
0
28 Jun 2023
Direct Preference Optimization: Your Language Model is Secretly a Reward
  Model
Direct Preference Optimization: Your Language Model is Secretly a Reward ModelNeural Information Processing Systems (NeurIPS), 2023
Rafael Rafailov
Archit Sharma
E. Mitchell
Stefano Ermon
Christopher D. Manning
Chelsea Finn
ALM
953
6,888
0
29 May 2023
AlpacaFarm: A Simulation Framework for Methods that Learn from Human
  Feedback
AlpacaFarm: A Simulation Framework for Methods that Learn from Human FeedbackNeural Information Processing Systems (NeurIPS), 2023
Yann Dubois
Xuechen Li
Rohan Taori
Tianyi Zhang
Ishaan Gulrajani
Jimmy Ba
Carlos Guestrin
Abigail Z. Jacobs
Tatsunori B. Hashimoto
ALM
506
774
0
22 May 2023
Whose Opinions Do Language Models Reflect?
Whose Opinions Do Language Models Reflect?International Conference on Machine Learning (ICML), 2023
Shibani Santurkar
Esin Durmus
Faisal Ladhak
Cinoo Lee
Abigail Z. Jacobs
Tatsunori Hashimoto
376
653
0
30 Mar 2023
GPT-4 Technical Report
GPT-4 Technical Report
OpenAI OpenAI
OpenAI Josh Achiam
Steven Adler
Sandhini Agarwal
Lama Ahmad
...
Shengjia Zhao
Tianhao Zheng
Juntang Zhuang
William Zhuk
Barret Zoph
LLMAGMLLM
4.7K
21,366
0
15 Mar 2023
Training a Helpful and Harmless Assistant with Reinforcement Learning
  from Human Feedback
Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback
Yuntao Bai
Andy Jones
Kamal Ndousse
Amanda Askell
Anna Chen
...
Jack Clark
Sam McCandlish
C. Olah
Benjamin Mann
Jared Kaplan
976
3,520
0
12 Apr 2022
Fine-Tuning Language Models from Human Preferences
Fine-Tuning Language Models from Human Preferences
Daniel M. Ziegler
Nisan Stiennon
Jeff Wu
Tom B. Brown
Alec Radford
Dario Amodei
Paul Christiano
G. Irving
ALM
1.8K
2,212
0
18 Sep 2019
Deep reinforcement learning from human preferences
Deep reinforcement learning from human preferencesNeural Information Processing Systems (NeurIPS), 2017
Paul Christiano
Jan Leike
Tom B. Brown
Miljan Martic
Shane Legg
Dario Amodei
1.6K
4,461
0
12 Jun 2017
1
Page 1 of 1