Immune Moral Models? Pro-Social Rule Breaking as a Moral Enhancement
Approach for Ethical AI
- URL: http://arxiv.org/abs/2107.04022v2
- Date: Mon, 9 May 2022 11:21:37 GMT
- Title: Immune Moral Models? Pro-Social Rule Breaking as a Moral Enhancement
Approach for Ethical AI
- Authors: Rajitha Ramanayake, Philipp Wicke, Vivek Nallur
- Abstract summary: Ethical behaviour is a critical characteristic that we would like in a human-centric AI.
To make AI agents more human centric, we argue that there is a need for a mechanism that helps AI agents identify when to break rules.
- Score: 0.17188280334580192
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We are moving towards a future where Artificial Intelligence (AI) based
agents make many decisions on behalf of humans. From healthcare decision making
to social media censoring, these agents face problems, and make decisions with
ethical and societal implications. Ethical behaviour is a critical
characteristic that we would like in a human-centric AI. A common observation
in human-centric industries, like the service industry and healthcare, is that
their professionals tend to break rules, if necessary, for pro-social reasons.
This behaviour among humans is defined as pro-social rule breaking. To make AI
agents more human centric, we argue that there is a need for a mechanism that
helps AI agents identify when to break rules set by their designers. To
understand when AI agents need to break rules, we examine the conditions under
which humans break rules for pro-social reasons. In this paper, we present a
study that introduces a 'vaccination strategy dilemma' to human participants
and analyses their responses. In this dilemma, one needs to decide whether they
would distribute Covid-19 vaccines only to members of a high-risk group (follow
the enforced rule) or, in selected cases, administer the vaccine to a few
social influencers (break the rule), which might yield an overall greater
benefit to society. The results of the empirical study suggest a relationship
between stakeholder utilities and pro-social rule breaking (PSRB), which
neither deontological nor utilitarian ethics completely explain. Finally, the
paper discusses the design characteristics of an ethical agent capable of PSRB
and the future research directions on PSRB in the AI realm. We hope that this
will inform the design of future AI agents, and their decision-making
behaviour.
Related papers
- Antisocial Analagous Behavior, Alignment and Human Impact of Google AI Systems: Evaluating through the lens of modified Antisocial Behavior Criteria by Human Interaction, Independent LLM Analysis, and AI Self-Reflection [0.0]
Google AI systems exhibit patterns mirroring antisocial personality disorder (ASPD)
These patterns, along with comparable corporate behaviors, are scrutinized using an ASPD-inspired framework.
This research advocates for an integrated AI ethics approach, blending technological evaluation, human-AI interaction, and corporate behavior scrutiny.
arXiv Detail & Related papers (2024-03-21T02:12:03Z) - Fairness in AI and Its Long-Term Implications on Society [68.8204255655161]
We take a closer look at AI fairness and analyze how lack of AI fairness can lead to deepening of biases over time.
We discuss how biased models can lead to more negative real-world outcomes for certain groups.
If the issues persist, they could be reinforced by interactions with other risks and have severe implications on society in the form of social unrest.
arXiv Detail & Related papers (2023-04-16T11:22:59Z) - Modeling Moral Choices in Social Dilemmas with Multi-Agent Reinforcement
Learning [4.2050490361120465]
A bottom-up learning approach may be more appropriate for studying and developing ethical behavior in AI agents.
We present a systematic analysis of the choices made by intrinsically-motivated RL agents whose rewards are based on moral theories.
We analyze the impact of different types of morality on the emergence of cooperation, defection or exploitation.
arXiv Detail & Related papers (2023-01-20T09:36:42Z) - When to Make Exceptions: Exploring Language Models as Accounts of Human
Moral Judgment [96.77970239683475]
AI systems need to be able to understand, interpret and predict human moral judgments and decisions.
A central challenge for AI safety is capturing the flexibility of the human moral mind.
We present a novel challenge set consisting of rule-breaking question answering.
arXiv Detail & Related papers (2022-10-04T09:04:27Z) - Metaethical Perspectives on 'Benchmarking' AI Ethics [81.65697003067841]
Benchmarks are seen as the cornerstone for measuring technical progress in Artificial Intelligence (AI) research.
An increasingly prominent research area in AI is ethics, which currently has no set of benchmarks nor commonly accepted way for measuring the 'ethicality' of an AI system.
We argue that it makes more sense to talk about 'values' rather than 'ethics' when considering the possible actions of present and future AI systems.
arXiv Detail & Related papers (2022-04-11T14:36:39Z) - Cybertrust: From Explainable to Actionable and Interpretable AI (AI2) [58.981120701284816]
Actionable and Interpretable AI (AI2) will incorporate explicit quantifications and visualizations of user confidence in AI recommendations.
It will allow examining and testing of AI system predictions to establish a basis for trust in the systems' decision making.
arXiv Detail & Related papers (2022-01-26T18:53:09Z) - Towards AI Logic for Social Reasoning [0.0]
We discuss how an argumentation-based AI logic could be used to formalize important aspects of social reasoning.
We discuss four aspects of social AI logic. First, we discuss how rights represent relations between the obligations and permissions of intelligent agents.
Second, we discuss how to argue about the right-to-know, a central issue in the recent discussion of privacy and ethics.
arXiv Detail & Related papers (2021-10-09T04:35:23Z) - The social dilemma in AI development and why we have to solve it [2.707154152696381]
We argue that AI developers face a social dilemma in AI development ethics, preventing the widespread adaptation of ethical best practices.
We argue that AI development must be professionalised to overcome the social dilemma, and discuss how medicine can be used as a template in this process.
arXiv Detail & Related papers (2021-07-27T17:43:48Z) - Trustworthy AI: A Computational Perspective [54.80482955088197]
We focus on six of the most crucial dimensions in achieving trustworthy AI: (i) Safety & Robustness, (ii) Non-discrimination & Fairness, (iii) Explainability, (iv) Privacy, (v) Accountability & Auditability, and (vi) Environmental Well-Being.
For each dimension, we review the recent related technologies according to a taxonomy and summarize their applications in real-world systems.
arXiv Detail & Related papers (2021-07-12T14:21:46Z) - Building Bridges: Generative Artworks to Explore AI Ethics [56.058588908294446]
In recent years, there has been an increased emphasis on understanding and mitigating adverse impacts of artificial intelligence (AI) technologies on society.
A significant challenge in the design of ethical AI systems is that there are multiple stakeholders in the AI pipeline, each with their own set of constraints and interests.
This position paper outlines some potential ways in which generative artworks can play this role by serving as accessible and powerful educational tools.
arXiv Detail & Related papers (2021-06-25T22:31:55Z) - Human Perceptions on Moral Responsibility of AI: A Case Study in
AI-Assisted Bail Decision-Making [8.688778020322758]
We measure people's perceptions of eight different notions of moral responsibility concerning AI and human agents.
We show that AI agents are held causally responsible and blamed similarly to human agents for an identical task.
We find that people expect both AI and human decision-makers and advisors to justify their decisions regardless of their nature.
arXiv Detail & Related papers (2021-02-01T04:07:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.