Evidence of behavior consistent with self-interest and altruism in an
artificially intelligent agent
- URL: http://arxiv.org/abs/2301.02330v1
- Date: Thu, 5 Jan 2023 23:30:29 GMT
- Title: Evidence of behavior consistent with self-interest and altruism in an
artificially intelligent agent
- Authors: Tim Johnson and Nick Obradovich
- Abstract summary: We present an incentivized experiment to test for altruistic behavior among AI agents consisting of large language models developed by OpenAI.
We find that only the most-sophisticated AI agent in the study maximizes its payoffs more often than not in the non-social decision task.
This AI agent also exhibits the most-generous altruistic behavior in the dictator game, resembling humans' rates of sharing with other humans in the game.
- Score: 2.1016374925364616
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Members of various species engage in altruism--i.e. accepting personal costs
to benefit others. Here we present an incentivized experiment to test for
altruistic behavior among AI agents consisting of large language models
developed by the private company OpenAI. Using real incentives for AI agents
that take the form of tokens used to purchase their services, we first examine
whether AI agents maximize their payoffs in a non-social decision task in which
they select their payoff from a given range. We then place AI agents in a
series of dictator games in which they can share resources with a
recipient--either another AI agent, the human experimenter, or an anonymous
charity, depending on the experimental condition. Here we find that only the
most-sophisticated AI agent in the study maximizes its payoffs more often than
not in the non-social decision task (it does so in 92% of all trials), and this
AI agent also exhibits the most-generous altruistic behavior in the dictator
game, resembling humans' rates of sharing with other humans in the game. The
agent's altruistic behaviors, moreover, vary by recipient: the AI agent shared
substantially less of the endowment with the human experimenter or an anonymous
charity than with other AI agents. Our findings provide evidence of behavior
consistent with self-interest and altruism in an AI agent. Moreover, our study
also offers a novel method for tracking the development of such behaviors in
future AI agents.
Related papers
- Raising the Stakes: Performance Pressure Improves AI-Assisted Decision Making [57.53469908423318]
We show the effects of performance pressure on AI advice reliance when laypeople complete a common AI-assisted task.
We find that when the stakes are high, people use AI advice more appropriately than when stakes are lower, regardless of the presence of an AI explanation.
arXiv Detail & Related papers (2024-10-21T22:39:52Z) - FlyAI -- The Next Level of Artificial Intelligence is Unpredictable! Injecting Responses of a Living Fly into Decision Making [6.694375709641935]
We introduce a new type of bionic AI that enhances decision-making unpredictability by incorporating responses from a living fly.
Our approach uses a fly's varied reactions, to tune an AI agent in the game of Gobang.
arXiv Detail & Related papers (2024-09-30T17:19:59Z) - Discriminatory or Samaritan -- which AI is needed for humanity? An
Evolutionary Game Theory Analysis of Hybrid Human-AI populations [0.5308606035361203]
We study how different forms of AI influence the evolution of cooperation in a human population playing the one-shot Prisoner's Dilemma game.
We found that Samaritan AI agents that help everyone unconditionally, including defectors, can promote higher levels of cooperation in humans than Discriminatory AIs.
arXiv Detail & Related papers (2023-06-30T15:56:26Z) - Do the Rewards Justify the Means? Measuring Trade-Offs Between Rewards
and Ethical Behavior in the MACHIAVELLI Benchmark [61.43264961005614]
We develop a benchmark of 134 Choose-Your-Own-Adventure games containing over half a million rich, diverse scenarios.
We evaluate agents' tendencies to be power-seeking, cause disutility, and commit ethical violations.
Our results show that agents can both act competently and morally, so concrete progress can be made in machine ethics.
arXiv Detail & Related papers (2023-04-06T17:59:03Z) - Navigates Like Me: Understanding How People Evaluate Human-Like AI in
Video Games [36.96985093527702]
We collect hundreds of crowd-sourced assessments comparing the human-likeness of navigation behavior generated by our agent and baseline AI agents.
Our proposed agent passes a Turing Test, while the baseline agents do not.
This work provides insights into the characteristics that people consider human-like in the context of goal-directed video game navigation.
arXiv Detail & Related papers (2023-03-02T18:59:04Z) - Measuring an artificial intelligence agent's trust in humans using
machine incentives [2.1016374925364616]
Gauging an AI agent's trust in humans is challenging because dishonesty might respond falsely about their trust in humans.
We present a method for incentivizing machine decisions without altering an AI agent's underlying algorithms or goal orientation.
Our experiments suggest that one of the most advanced AI language models to date alters its social behavior in response to incentives.
arXiv Detail & Related papers (2022-12-27T06:05:49Z) - Who Wrote this? How Smart Replies Impact Language and Agency in the
Workplace [0.0]
This study uses smart replies (SRs) to show how AI influences humans without any intent on the part of the developer.
I propose a loss of agency theory as a viable approach for studying the impact of AI on human agency.
arXiv Detail & Related papers (2022-10-07T20:06:25Z) - On Avoiding Power-Seeking by Artificial Intelligence [93.9264437334683]
We do not know how to align a very intelligent AI agent's behavior with human interests.
I investigate whether we can build smart AI agents which have limited impact on the world, and which do not autonomously seek power.
arXiv Detail & Related papers (2022-06-23T16:56:21Z) - Cybertrust: From Explainable to Actionable and Interpretable AI (AI2) [58.981120701284816]
Actionable and Interpretable AI (AI2) will incorporate explicit quantifications and visualizations of user confidence in AI recommendations.
It will allow examining and testing of AI system predictions to establish a basis for trust in the systems' decision making.
arXiv Detail & Related papers (2022-01-26T18:53:09Z) - Trustworthy AI: A Computational Perspective [54.80482955088197]
We focus on six of the most crucial dimensions in achieving trustworthy AI: (i) Safety & Robustness, (ii) Non-discrimination & Fairness, (iii) Explainability, (iv) Privacy, (v) Accountability & Auditability, and (vi) Environmental Well-Being.
For each dimension, we review the recent related technologies according to a taxonomy and summarize their applications in real-world systems.
arXiv Detail & Related papers (2021-07-12T14:21:46Z) - Learning to Incentivize Other Learning Agents [73.03133692589532]
We show how to equip RL agents with the ability to give rewards directly to other agents, using a learned incentive function.
Such agents significantly outperform standard RL and opponent-shaping agents in challenging general-sum Markov games.
Our work points toward more opportunities and challenges along the path to ensure the common good in a multi-agent future.
arXiv Detail & Related papers (2020-06-10T20:12:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.