2022
DOI: 10.48550/arxiv.2210.01478
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

When to Make Exceptions: Exploring Language Models as Accounts of Human Moral Judgment

Abstract: AI systems are becoming increasingly intertwined with human life. In order to effectively collaborate with humans and ensure safety, AI systems need to be able to understand, interpret and predict human moral judgments and decisions. Human moral judgments are often guided by rules, but not always. A central challenge for AI safety is capturing the flexibility of the human moral mind -the ability to determine when a rule should be broken, especially in novel or unusual situations. In this paper, we present a no… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 22 publications
0
1
0
Order By: Relevance
“…Moral Judgment: Jin et al (2022) examine how LLMs answer moral puzzles about when rule breaking is permissible. They used chain-ofthought prompting method (Wei et al, 2022) to implement a 'contractualist' theory (Scanlon et al, 1982) of moral reasoning.…”
Section: Introductionmentioning
confidence: 99%
“…Moral Judgment: Jin et al (2022) examine how LLMs answer moral puzzles about when rule breaking is permissible. They used chain-ofthought prompting method (Wei et al, 2022) to implement a 'contractualist' theory (Scanlon et al, 1982) of moral reasoning.…”
Section: Introductionmentioning
confidence: 99%