When Is It Acceptable to Break the Rules? Knowledge Representation of
Moral Judgement Based on Empirical Data
- URL: http://arxiv.org/abs/2201.07763v1
- Date: Wed, 19 Jan 2022 17:58:42 GMT
- Title: When Is It Acceptable to Break the Rules? Knowledge Representation of
Moral Judgement Based on Empirical Data
- Authors: Edmond Awad, Sydney Levine, Andrea Loreggia, Nicholas Mattei, Iyad
Rahwan, Francesca Rossi, Kartik Talamadupula, Joshua Tenenbaum, Max
Kleiman-Weiner
- Abstract summary: One of the most remarkable things about the human moral mind is its flexibility.
We can make moral judgments about cases we have never seen before.
We can decide that pre-established rules should be broken.
Capturing this flexibility is one of the central challenges in developing AI systems that can interpret and produce human-like moral judgment.
- Score: 33.58705831230163
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: One of the most remarkable things about the human moral mind is its
flexibility. We can make moral judgments about cases we have never seen before.
We can decide that pre-established rules should be broken. We can invent novel
rules on the fly. Capturing this flexibility is one of the central challenges
in developing AI systems that can interpret and produce human-like moral
judgment. This paper details the results of a study of real-world decision
makers who judge whether it is acceptable to break a well-established norm:
``no cutting in line.'' We gather data on how human participants judge the
acceptability of line-cutting in a range of scenarios. Then, in order to
effectively embed these reasoning capabilities into a machine, we propose a
method for modeling them using a preference-based structure, which captures a
novel modification to standard ``dual process'' theories of moral judgment.
Related papers
- Learning Machine Morality through Experience and Interaction [3.7414804164475983]
Increasing interest in ensuring safety of next-generation Artificial Intelligence (AI) systems calls for novel approaches to embedding morality into autonomous agents.
We argue that more hybrid solutions are needed to create adaptable and robust, yet more controllable and interpretable agents.
arXiv Detail & Related papers (2023-12-04T11:46:34Z) - What Makes it Ok to Set a Fire? Iterative Self-distillation of Contexts
and Rationales for Disambiguating Defeasible Social and Moral Situations [48.686872351114964]
Moral or ethical judgments rely heavily on the specific contexts in which they occur.
We introduce defeasible moral reasoning: a task to provide grounded contexts that make an action more or less morally acceptable.
We distill a high-quality dataset of 1.2M entries of contextualizations and rationales for 115K defeasible moral actions.
arXiv Detail & Related papers (2023-10-24T00:51:29Z) - Rethinking Machine Ethics -- Can LLMs Perform Moral Reasoning through the Lens of Moral Theories? [78.3738172874685]
Making moral judgments is an essential step toward developing ethical AI systems.
Prevalent approaches are mostly implemented in a bottom-up manner, which uses a large set of annotated data to train models based on crowd-sourced opinions about morality.
This work proposes a flexible top-down framework to steer (Large) Language Models (LMs) to perform moral reasoning with well-established moral theories from interdisciplinary research.
arXiv Detail & Related papers (2023-08-29T15:57:32Z) - When to Make Exceptions: Exploring Language Models as Accounts of Human
Moral Judgment [96.77970239683475]
AI systems need to be able to understand, interpret and predict human moral judgments and decisions.
A central challenge for AI safety is capturing the flexibility of the human moral mind.
We present a novel challenge set consisting of rule-breaking question answering.
arXiv Detail & Related papers (2022-10-04T09:04:27Z) - A Word on Machine Ethics: A Response to Jiang et al. (2021) [36.955224006838584]
We focus on a single case study of the recently proposed Delphi model and offer a critique of the project's proposed method of automating morality judgments.
We conclude with a discussion of how machine ethics could usefully proceed, by focusing on current and near-future uses of technology.
arXiv Detail & Related papers (2021-11-07T19:31:51Z) - Morality, Machines and the Interpretation Problem: A value-based,
Wittgensteinian approach to building Moral Agents [0.0]
We argue that the attempt to build morality into machines is subject to what we call the Interpretation problem.
We argue that any rule we give the machine is open to infinite interpretation in ways that we might morally disapprove of.
arXiv Detail & Related papers (2021-03-03T22:34:01Z) - Scruples: A Corpus of Community Ethical Judgments on 32,000 Real-Life
Anecdotes [72.64975113835018]
Motivated by descriptive ethics, we investigate a novel, data-driven approach to machine ethics.
We introduce Scruples, the first large-scale dataset with 625,000 ethical judgments over 32,000 real-life anecdotes.
Our dataset presents a major challenge to state-of-the-art neural language models, leaving significant room for improvement.
arXiv Detail & Related papers (2020-08-20T17:34:15Z) - Aligning AI With Shared Human Values [85.2824609130584]
We introduce the ETHICS dataset, a new benchmark that spans concepts in justice, well-being, duties, virtues, and commonsense morality.
We find that current language models have a promising but incomplete ability to predict basic human ethical judgements.
Our work shows that progress can be made on machine ethics today, and it provides a steppingstone toward AI that is aligned with human values.
arXiv Detail & Related papers (2020-08-05T17:59:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.