Mediating Artificial Intelligence Developments through Negative and
Positive Incentives
- URL: http://arxiv.org/abs/2010.00403v1
- Date: Thu, 1 Oct 2020 13:43:32 GMT
- Title: Mediating Artificial Intelligence Developments through Negative and
Positive Incentives
- Authors: The Anh Han, Luis Moniz Pereira, Tom Lenaerts, Francisco C. Santos
- Abstract summary: We investigate how positive (rewards) and negative (punishments) incentives may beneficially influence the outcomes.
We show that, in several scenarios, rewarding those that follow safety measures may increase the development speed while ensuring safe choices.
- Score: 5.0066859598912945
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The field of Artificial Intelligence (AI) is going through a period of great
expectations, introducing a certain level of anxiety in research, business and
also policy. This anxiety is further energised by an AI race narrative that
makes people believe they might be missing out. Whether real or not, a belief
in this narrative may be detrimental as some stake-holders will feel obliged to
cut corners on safety precautions, or ignore societal consequences just to
"win". Starting from a baseline model that describes a broad class of
technology races where winners draw a significant benefit compared to others
(such as AI advances, patent race, pharmaceutical technologies), we investigate
here how positive (rewards) and negative (punishments) incentives may
beneficially influence the outcomes. We uncover conditions in which punishment
is either capable of reducing the development speed of unsafe participants or
has the capacity to reduce innovation through over-regulation. Alternatively,
we show that, in several scenarios, rewarding those that follow safety measures
may increase the development speed while ensuring safe choices. Moreover, in
{the latter} regimes, rewards do not suffer from the issue of over-regulation
as is the case for punishment. Overall, our findings provide valuable insights
into the nature and kinds of regulatory actions most suitable to improve safety
compliance in the contexts of both smooth and sudden technological shifts.
Related papers
- Exploring the Impact of Rewards on Developers' Proactive AI Accountability Behavior [0.0]
We develop a theoretical model grounded in Self-Determination Theory to uncover the potential impact of rewards and sanctions on AI developers.
We identify typical sanctions and bug bounties as potential reward mechanisms by surveying related research from various domains.
arXiv Detail & Related papers (2024-11-27T14:34:44Z) - SafetyAnalyst: Interpretable, transparent, and steerable safety moderation for AI behavior [56.10557932893919]
We present SafetyAnalyst, a novel AI safety moderation framework.
Given an AI behavior, SafetyAnalyst uses chain-of-thought reasoning to analyze its potential consequences.
It aggregates all harmful and beneficial effects into a harmfulness score using fully interpretable weight parameters.
arXiv Detail & Related papers (2024-10-22T03:38:37Z) - AI Safety: A Climb To Armageddon? [0.0]
The paper examines three response strategies: Optimism, Mitigation, and Holism.
The surprising robustness of the argument forces a re-examination of core assumptions around AI safety.
arXiv Detail & Related papers (2024-05-30T08:41:54Z) - Work-in-Progress: Crash Course: Can (Under Attack) Autonomous Driving Beat Human Drivers? [60.51287814584477]
This paper evaluates the inherent risks in autonomous driving by examining the current landscape of AVs.
We develop specific claims highlighting the delicate balance between the advantages of AVs and potential security challenges in real-world scenarios.
arXiv Detail & Related papers (2024-05-14T09:42:21Z) - Risks of AI Scientists: Prioritizing Safeguarding Over Autonomy [65.77763092833348]
This perspective examines vulnerabilities in AI scientists, shedding light on potential risks associated with their misuse.<n>We take into account user intent, the specific scientific domain, and their potential impact on the external environment.<n>We propose a triadic framework involving human regulation, agent alignment, and an understanding of environmental feedback.
arXiv Detail & Related papers (2024-02-06T18:54:07Z) - Control Risk for Potential Misuse of Artificial Intelligence in Science [85.91232985405554]
We aim to raise awareness of the dangers of AI misuse in science.
We highlight real-world examples of misuse in chemical science.
We propose a system called SciGuard to control misuse risks for AI models in science.
arXiv Detail & Related papers (2023-12-11T18:50:57Z) - Managing extreme AI risks amid rapid progress [171.05448842016125]
We describe risks that include large-scale social harms, malicious uses, and irreversible loss of human control over autonomous AI systems.
There is a lack of consensus about how exactly such risks arise, and how to manage them.
Present governance initiatives lack the mechanisms and institutions to prevent misuse and recklessness, and barely address autonomous systems.
arXiv Detail & Related papers (2023-10-26T17:59:06Z) - The Promise and Peril of Artificial Intelligence -- Violet Teaming
Offers a Balanced Path Forward [56.16884466478886]
This paper reviews emerging issues with opaque and uncontrollable AI systems.
It proposes an integrative framework called violet teaming to develop reliable and responsible AI.
It emerged from AI safety research to manage risks proactively by design.
arXiv Detail & Related papers (2023-08-28T02:10:38Z) - Absolutist AI [0.0]
Training AI systems with absolute constraints may make considerable progress on many AI safety problems.
It provides a guardrail for avoiding the very worst outcomes of misalignment.
It could prevent AIs from causing catastrophes for the sake of very valuable consequences.
arXiv Detail & Related papers (2023-07-19T03:40:37Z) - Fairness in AI and Its Long-Term Implications on Society [68.8204255655161]
We take a closer look at AI fairness and analyze how lack of AI fairness can lead to deepening of biases over time.
We discuss how biased models can lead to more negative real-world outcomes for certain groups.
If the issues persist, they could be reinforced by interactions with other risks and have severe implications on society in the form of social unrest.
arXiv Detail & Related papers (2023-04-16T11:22:59Z) - Voluntary safety commitments provide an escape from over-regulation in
AI development [8.131948859165432]
This work reveals for the first time how voluntary commitments, with sanctions either by peers or an institution, leads to socially beneficial outcomes.
Results are directly relevant for the design of governance and regulatory policies that aim to ensure an ethical and responsible AI technology development process.
arXiv Detail & Related papers (2021-04-08T12:54:56Z) - AI Development Race Can Be Mediated on Heterogeneous Networks [8.131948859165432]
We investigate how different interaction structures among race participants can alter collective choices and requirements for regulatory actions.
Our findings indicate that, when participants portray a strong diversity in terms of connections and peer-influence, the conflicts that exist in homogeneous settings are significantly reduced.
arXiv Detail & Related papers (2020-12-30T17:23:18Z) - Overcoming Failures of Imagination in AI Infused System Development and
Deployment [71.9309995623067]
NeurIPS 2020 requested that research paper submissions include impact statements on "potential nefarious uses and the consequences of failure"
We argue that frameworks of harms must be context-aware and consider a wider range of potential stakeholders, system affordances, as well as viable proxies for assessing harms in the widest sense.
arXiv Detail & Related papers (2020-11-26T18:09:52Z) - AI Research Considerations for Human Existential Safety (ARCHES) [6.40842967242078]
In negative terms, we ask what existential risks humanity might face from AI development in the next century.
Key property of hypothetical AI technologies is introduced, called emphprepotence
A set of auxrefdirtot contemporary research directions are then examined for their potential benefit to existential safety.
arXiv Detail & Related papers (2020-05-30T02:05:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.