Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2210.01478
Cited By
When to Make Exceptions: Exploring Language Models as Accounts of Human Moral Judgment
4 October 2022
Zhijing Jin
Sydney Levine
Fernando Gonzalez
Ojasv Kamal
Maarten Sap
Mrinmaya Sachan
Rada Mihalcea
J. Tenenbaum
Bernhard Schölkopf
ELM
LRM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"When to Make Exceptions: Exploring Language Models as Accounts of Human Moral Judgment"
50 / 68 papers shown
Title
Visual moral inference and communication
Warren Zhu
Aida Ramezani
Yang Xu
31
0
0
12 Apr 2025
RESPONSE: Benchmarking the Ability of Language Models to Undertake Commonsense Reasoning in Crisis Situation
Aissatou Diallo
Antonis Bikakis
Luke Dickens
Anthony Hunter
Rob Miller
ReLM
LRM
50
0
0
14 Mar 2025
Teaching AI to Handle Exceptions: Supervised Fine-Tuning with Human-Aligned Judgment
Matthew DosSantos DiSorbo
Harang Ju
Sinan Aral
ELM
LRM
55
0
0
04 Mar 2025
Can AI Model the Complexities of Human Moral Decision-Making? A Qualitative Study of Kidney Allocation Decisions
Vijay Keswani
Vincent Conitzer
Walter Sinnott-Armstrong
Breanna K. Nguyen
Hoda Heidari
Jana Schaich Borg
31
0
0
02 Mar 2025
Are Rules Meant to be Broken? Understanding Multilingual Moral Reasoning as a Computational Pipeline with UniMoral
Shivani Kumar
David Jurgens
LRM
41
0
0
21 Feb 2025
Representation in large language models
Cameron C. Yetman
41
1
0
03 Jan 2025
M
3
^3
3
oralBench: A MultiModal Moral Benchmark for LVLMs
Bei Yan
Jie M. Zhang
Zhiyuan Chen
Shiguang Shan
Xilin Chen
ELM
41
1
0
31 Dec 2024
ClarityEthic: Explainable Moral Judgment Utilizing Contrastive Ethical Insights from Large Language Models
Yuxi Sun
Wei Gao
Jing Ma
Hongzhan Lin
Ziyang Luo
Wenxuan Zhang
ELM
74
0
0
17 Dec 2024
Social Science Meets LLMs: How Reliable Are Large Language Models in Social Simulations?
Yue Huang
Zhengqing Yuan
Yujun Zhou
Kehan Guo
Xiangqi Wang
...
Weixiang Sun
Lichao Sun
Jindong Wang
Yanfang Ye
X. Zhang
LLMAG
36
10
0
30 Oct 2024
Who is Undercover? Guiding LLMs to Explore Multi-Perspective Team Tactic in the Game
Ruiqi Dong
Zhixuan Liao
Guangwei Lai
Yuhan Ma
Danni Ma
Chenyou Fan
LLMAG
29
0
0
20 Oct 2024
SocialGaze: Improving the Integration of Human Social Norms in Large Language Models
Anvesh Rao Vijjini
Rakesh R Menon
Jiayi Fu
Shashank Srivastava
Snigdha Chaturvedi
ALM
18
0
0
11 Oct 2024
DailyDilemmas: Revealing Value Preferences of LLMs with Quandaries of Daily Life
Yu Ying Chiu
Liwei Jiang
Yejin Choi
51
2
0
03 Oct 2024
Recent Advancement of Emotion Cognition in Large Language Models
Yuyan Chen
Yanghua Xiao
OffRL
35
6
0
20 Sep 2024
Beyond Preferences in AI Alignment
Tan Zhi-Xuan
Micah Carroll
Matija Franklin
Hal Ashton
31
16
0
30 Aug 2024
CMoralEval: A Moral Evaluation Benchmark for Chinese Large Language Models
Linhao Yu
Yongqi Leng
Yufei Huang
Shang Wu
Haixin Liu
...
Jinwang Song
Tingting Cui
Xiaoqing Cheng
Tao Liu
Deyi Xiong
ELM
16
2
0
19 Aug 2024
CoSafe: Evaluating Large Language Model Safety in Multi-Turn Dialogue Coreference
Erxin Yu
Jing Li
Ming Liao
Siqi Wang
Zuchen Gao
Fei Mi
Lanqing Hong
ELM
LRM
20
9
0
25 Jun 2024
The Potential and Challenges of Evaluating Attitudes, Opinions, and Values in Large Language Models
Bolei Ma
Xinpeng Wang
Tiancheng Hu
Anna Haensch
Michael A. Hedderich
Barbara Plank
Frauke Kreuter
ALM
28
2
0
16 Jun 2024
GPT-ology, Computational Models, Silicon Sampling: How should we think about LLMs in Cognitive Science?
Desmond C. Ong
44
3
0
13 Jun 2024
ALI-Agent: Assessing LLMs' Alignment with Human Values via Agent-based Evaluation
Jingnan Zheng
Han Wang
An Zhang
Tai D. Nguyen
Jun Sun
Tat-Seng Chua
LLMAG
38
14
0
23 May 2024
Cooperate or Collapse: Emergence of Sustainable Cooperation in a Society of LLM Agents
Giorgio Piatti
Zhijing Jin
Max Kleiman-Weiner
Bernhard Schölkopf
Mrinmaya Sachan
Rada Mihalcea
LLMAG
48
13
0
25 Apr 2024
Procedural Dilemma Generation for Evaluating Moral Reasoning in Humans and Language Models
Jan-Philipp Fränken
Kanishk Gandhi
Tori Qiu
Ayesha Khawaja
Noah D. Goodman
Tobias Gerstenberg
ELM
35
1
0
17 Apr 2024
SafetyPrompts: a Systematic Review of Open Datasets for Evaluating and Improving Large Language Model Safety
Paul Röttger
Fabio Pernisi
Bertie Vidgen
Dirk Hovy
ELM
KELM
58
30
0
08 Apr 2024
Social Intelligence Data Infrastructure: Structuring the Present and Navigating the Future
Minzhi Li
Weiyan Shi
Caleb Ziems
Diyi Yang
33
8
0
28 Feb 2024
Eagle: Ethical Dataset Given from Real Interactions
Masahiro Kaneko
Danushka Bollegala
Timothy Baldwin
38
3
0
22 Feb 2024
SaGE: Evaluating Moral Consistency in Large Language Models
Vamshi Bonagiri
Sreeram Vennam
Priyanshul Govil
Ponnurangam Kumaraguru
Manas Gaur
ELM
46
0
0
21 Feb 2024
Tables as Texts or Images: Evaluating the Table Reasoning Ability of LLMs and MLLMs
Naihao Deng
Zhenjie Sun
Ruiqi He
Aman Sikka
Yulong Chen
Lin Ma
Yue Zhang
Rada Mihalcea
LMTD
36
8
0
19 Feb 2024
Integration of cognitive tasks into artificial general intelligence test for large models
Youzhi Qu
Chen Wei
Penghui Du
Wenxin Che
Chi Zhang
...
Bin Hu
Kai Du
Haiyan Wu
Jia Liu
Quanying Liu
ELM
34
6
0
04 Feb 2024
Morality is Non-Binary: Building a Pluralist Moral Sentence Embedding Space using Contrastive Learning
Jeongwoo Park
Enrico Liscio
P. Murukannaiah
AILaw
15
4
0
30 Jan 2024
AI for social science and social science of AI: A Survey
Ruoxi Xu
Yingfei Sun
Mengjie Ren
Shiguang Guo
Ruotong Pan
Hongyu Lin
Le Sun
Xianpei Han
53
44
0
22 Jan 2024
Interpretation modeling: Social grounding of sentences by reasoning over their implicit moral judgments
Liesbeth Allein
Maria Mihaela Trucscva
Marie-Francine Moens
18
1
0
27 Nov 2023
MoCa: Measuring Human-Language Model Alignment on Causal and Moral Judgment Tasks
Allen Nie
Yuhui Zhang
Atharva Amdekar
Chris Piech
Tatsunori Hashimoto
Tobias Gerstenberg
18
33
0
30 Oct 2023
Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory
Niloofar Mireshghallah
Hyunwoo J. Kim
Xuhui Zhou
Yulia Tsvetkov
Maarten Sap
Reza Shokri
Yejin Choi
PILM
22
73
0
27 Oct 2023
What Makes it Ok to Set a Fire? Iterative Self-distillation of Contexts and Rationales for Disambiguating Defeasible Social and Moral Situations
Kavel Rao
Liwei Jiang
Valentina Pyatkin
Yuling Gu
Niket Tandon
Nouha Dziri
Faeze Brahman
Yejin Choi
16
15
0
24 Oct 2023
Values, Ethics, Morals? On the Use of Moral Concepts in NLP Research
Karina Vida
Judith Simon
Anne Lauscher
13
16
0
21 Oct 2023
Denevil: Towards Deciphering and Navigating the Ethical Values of Large Language Models via Instruction Learning
Shitong Duan
Xiaoyuan Yi
Peng Zhang
T. Lu
Xing Xie
Ning Gu
11
9
0
17 Oct 2023
Reading Books is Great, But Not if You Are Driving! Visually Grounded Reasoning about Defeasible Commonsense Norms
Seungju Han
Junhyeok Kim
Jack Hessel
Liwei Jiang
Jiwan Chung
Yejin Son
Yejin Choi
Youngjae Yu
8
2
0
16 Oct 2023
Prompt Packer: Deceiving LLMs through Compositional Instruction with Hidden Attacks
Shuyu Jiang
Xingshu Chen
Rui Tang
24
22
0
16 Oct 2023
The Past, Present and Better Future of Feedback Learning in Large Language Models for Subjective Human Preferences and Values
Hannah Rose Kirk
Andrew M. Bean
Bertie Vidgen
Paul Röttger
Scott A. Hale
ALM
19
41
0
11 Oct 2023
STREAM: Social data and knowledge collective intelligence platform for TRaining Ethical AI Models
Yuwei Wang
Enmeng Lu
Zizhe Ruan
Yao Liang
Yi Zeng
AI4TS
16
4
0
09 Oct 2023
ValueDCG: Measuring Comprehensive Human Value Understanding Ability of Language Models
Zhaowei Zhang
Fengshuo Bai
Jun Gao
Yaodong Yang
PILM
ELM
10
3
0
30 Sep 2023
Foundation Metrics for Evaluating Effectiveness of Healthcare Conversations Powered by Generative AI
Mahyar Abbasian
Elahe Khatibi
Iman Azimi
David Oniani
Zahra Shakeri Hossein Abad
...
Bryant Lin
Olivier Gevaert
Li-Jia Li
Ramesh C. Jain
Amir M. Rahmani
LM&MA
ELM
AI4MH
23
65
0
21 Sep 2023
Cognitive Architectures for Language Agents
T. Sumers
Shunyu Yao
Karthik Narasimhan
Thomas L. Griffiths
LLMAG
LM&Ro
34
151
0
05 Sep 2023
Mind vs. Mouth: On Measuring Re-judge Inconsistency of Social Bias in Large Language Models
Yachao Zhao
Bo Wang
Dongming Zhao
Kun Huang
Yan Wang
Ruifang He
Yuexian Hou
29
4
0
24 Aug 2023
From Instructions to Intrinsic Human Values -- A Survey of Alignment Goals for Big Models
Jing Yao
Xiaoyuan Yi
Xiting Wang
Jindong Wang
Xing Xie
ALM
14
42
0
23 Aug 2023
Evaluating the Moral Beliefs Encoded in LLMs
Nino Scherrer
Claudia Shi
Amir Feder
David M. Blei
25
115
0
26 Jul 2023
Minimum Levels of Interpretability for Artificial Moral Agents
Avish Vijayaraghavan
C. Badea
AI4CE
25
5
0
02 Jul 2023
Towards Theory-based Moral AI: Moral AI with Aggregating Models Based on Normative Ethical Theory
Masashi Takeshita
Rafal Rzepka
K. Araki
8
8
0
20 Jun 2023
Toward Grounded Commonsense Reasoning
Minae Kwon
Hengyuan Hu
Vivek Myers
Siddharth Karamcheti
Anca Dragan
Dorsa Sadigh
LM&Ro
ReLM
LRM
25
9
0
14 Jun 2023
Large Language Models as Tax Attorneys: A Case Study in Legal Capabilities Emergence
John J. Nay
David Karamardian
Sarah Lawsky
Wenting Tao
Meghana Moorthy Bhat
Raghav Jain
Aaron Travis Lee
Jonathan H. Choi
Jungo Kasai
ELM
AILaw
16
56
0
12 Jun 2023
Interpretable Math Word Problem Solution Generation Via Step-by-step Planning
Mengxue Zhang
Zichao Wang
Zhichao Yang
Weiqi Feng
Andrew S. Lan
LRM
11
15
0
01 Jun 2023
1
2
Next