Morality, Machines and the Interpretation Problem: A value-based,
Wittgensteinian approach to building Moral Agents
- URL: http://arxiv.org/abs/2103.02728v1
- Date: Wed, 3 Mar 2021 22:34:01 GMT
- Title: Morality, Machines and the Interpretation Problem: A value-based,
Wittgensteinian approach to building Moral Agents
- Authors: Cosmin Badea, Gregory Artus
- Abstract summary: We argue that the attempt to build morality into machines is subject to what we call the Interpretation problem.
We argue that any rule we give the machine is open to infinite interpretation in ways that we might morally disapprove of.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We argue that the attempt to build morality into machines is subject to what
we call the Interpretation problem, whereby any rule we give the machine is
open to infinite interpretation in ways that we might morally disapprove of,
and that the interpretation problem in Artificial Intelligence is an
illustration of Wittgenstein's general claim that no rule can contain the
criteria for its own application. Using games as an example, we attempt to
define the structure of normative spaces and argue that any rule-following
within a normative space is guided by values that are external to that space
and which cannot themselves be represented as rules. In light of this problem,
we analyse the types of mistakes an artificial moral agent could make and we
make suggestions about how to build morality into machines by getting them to
interpret the rules we give in accordance with these external values, through
explicit moral reasoning and the presence of structured values, the adjustment
of causal power assigned to the agent and interaction with human agents, such
that the machine develops a virtuous character and the impact of the
interpretation problem is minimised.
Related papers
- Technology as uncharted territory: Contextual integrity and the notion of AI as new ethical ground [55.2480439325792]
I argue that efforts to promote responsible and ethical AI can inadvertently contribute to and seemingly legitimize this disregard for established contextual norms.
I question the current narrow prioritization in AI ethics of moral innovation over moral preservation.
arXiv Detail & Related papers (2024-12-06T15:36:13Z) - Why Machines Can't Be Moral: Turing's Halting Problem and the Moral Limits of Artificial Intelligence [0.0]
I argue that explicit ethical machines, whose moral principles are inferred through a bottom-up approach, are unable to replicate human-like moral reasoning.
By utilizing Alan Turing's theory of computation, I demonstrate that moral reasoning is computationally intractable by these machines due to the halting problem.
arXiv Detail & Related papers (2024-07-24T17:50:24Z) - What Makes it Ok to Set a Fire? Iterative Self-distillation of Contexts
and Rationales for Disambiguating Defeasible Social and Moral Situations [48.686872351114964]
Moral or ethical judgments rely heavily on the specific contexts in which they occur.
We introduce defeasible moral reasoning: a task to provide grounded contexts that make an action more or less morally acceptable.
We distill a high-quality dataset of 1.2M entries of contextualizations and rationales for 115K defeasible moral actions.
arXiv Detail & Related papers (2023-10-24T00:51:29Z) - Phenomenal Yet Puzzling: Testing Inductive Reasoning Capabilities of Language Models with Hypothesis Refinement [92.61557711360652]
Language models (LMs) often fall short on inductive reasoning, despite achieving impressive success on research benchmarks.
We conduct a systematic study of the inductive reasoning capabilities of LMs through iterative hypothesis refinement.
We reveal several discrepancies between the inductive reasoning processes of LMs and humans, shedding light on both the potentials and limitations of using LMs in inductive reasoning tasks.
arXiv Detail & Related papers (2023-10-12T17:51:10Z) - Rethinking Machine Ethics -- Can LLMs Perform Moral Reasoning through the Lens of Moral Theories? [78.3738172874685]
Making moral judgments is an essential step toward developing ethical AI systems.
Prevalent approaches are mostly implemented in a bottom-up manner, which uses a large set of annotated data to train models based on crowd-sourced opinions about morality.
This work proposes a flexible top-down framework to steer (Large) Language Models (LMs) to perform moral reasoning with well-established moral theories from interdisciplinary research.
arXiv Detail & Related papers (2023-08-29T15:57:32Z) - When to Make Exceptions: Exploring Language Models as Accounts of Human
Moral Judgment [96.77970239683475]
AI systems need to be able to understand, interpret and predict human moral judgments and decisions.
A central challenge for AI safety is capturing the flexibility of the human moral mind.
We present a novel challenge set consisting of rule-breaking question answering.
arXiv Detail & Related papers (2022-10-04T09:04:27Z) - The Moral Integrity Corpus: A Benchmark for Ethical Dialogue Systems [36.90292508433193]
Moral deviations are difficult to mitigate because moral judgments are not universal.
Moral Integrity Corpus captures the moral assumptions of 38k prompt-reply pairs.
We show that current neural language models can automatically generate new RoTs that reasonably describe previously unseen interactions.
arXiv Detail & Related papers (2022-04-06T18:10:53Z) - When Is It Acceptable to Break the Rules? Knowledge Representation of
Moral Judgement Based on Empirical Data [33.58705831230163]
One of the most remarkable things about the human moral mind is its flexibility.
We can make moral judgments about cases we have never seen before.
We can decide that pre-established rules should be broken.
Capturing this flexibility is one of the central challenges in developing AI systems that can interpret and produce human-like moral judgment.
arXiv Detail & Related papers (2022-01-19T17:58:42Z) - Contextualized moral inference [12.574316678945195]
We present a text-based approach that predicts people's intuitive judgment of moral vignettes.
We show that a contextualized representation offers a substantial advantage over alternative representations.
arXiv Detail & Related papers (2020-08-25T00:34:28Z) - Aligning AI With Shared Human Values [85.2824609130584]
We introduce the ETHICS dataset, a new benchmark that spans concepts in justice, well-being, duties, virtues, and commonsense morality.
We find that current language models have a promising but incomplete ability to predict basic human ethical judgements.
Our work shows that progress can be made on machine ethics today, and it provides a steppingstone toward AI that is aligned with human values.
arXiv Detail & Related papers (2020-08-05T17:59:16Z) - Reinforcement Learning Under Moral Uncertainty [13.761051314923634]
An ambitious goal for machine learning is to create agents that behave ethically.
While ethical agents could be trained by rewarding correct behavior under a specific moral theory, there remains widespread disagreement about the nature of morality.
This paper proposes two training methods that realize different points among competing desiderata, and trains agents in simple environments to act under moral uncertainty.
arXiv Detail & Related papers (2020-06-08T16:40:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.