Cognitive Models as Simulators: The Case of Moral Decision-Making
- URL: http://arxiv.org/abs/2210.04121v1
- Date: Sat, 8 Oct 2022 23:14:14 GMT
- Title: Cognitive Models as Simulators: The Case of Moral Decision-Making
- Authors: Ardavan S. Nobandegani, Thomas R. Shultz, Irina Rish
- Abstract summary: In this work, we substantiate the idea of $textitcognitive models as simulators$, which is to have AI systems interact with, and collect feedback from, cognitive models instead of humans.
Here, we leverage this idea in the context of moral decision-making, by having reinforcement learning agents learn about fairness through interacting with a cognitive model of the Ultimatum Game (UG)
Our work suggests that using cognitive models as simulators of humans is an effective approach for training AI systems, presenting an important way for computational cognitive science to make contributions to AI.
- Score: 9.024707986238392
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To achieve desirable performance, current AI systems often require huge
amounts of training data. This is especially problematic in domains where
collecting data is both expensive and time-consuming, e.g., where AI systems
require having numerous interactions with humans, collecting feedback from
them. In this work, we substantiate the idea of $\textit{cognitive models as
simulators}$, which is to have AI systems interact with, and collect feedback
from, cognitive models instead of humans, thereby making their training process
both less costly and faster. Here, we leverage this idea in the context of
moral decision-making, by having reinforcement learning (RL) agents learn about
fairness through interacting with a cognitive model of the Ultimatum Game (UG),
a canonical task in behavioral and brain sciences for studying fairness.
Interestingly, these RL agents learn to rationally adapt their behavior
depending on the emotional state of their simulated UG responder. Our work
suggests that using cognitive models as simulators of humans is an effective
approach for training AI systems, presenting an important way for computational
cognitive science to make contributions to AI.
Related papers
- User Behavior Simulation with Large Language Model based Agents [116.74368915420065]
We propose an LLM-based agent framework and design a sandbox environment to simulate real user behaviors.
Based on extensive experiments, we find that the simulated behaviors of our method are very close to the ones of real humans.
arXiv Detail & Related papers (2023-06-05T02:58:35Z) - Hindsight States: Blending Sim and Real Task Elements for Efficient
Reinforcement Learning [61.3506230781327]
In robotics, one approach to generate training data builds on simulations based on dynamics models derived from first principles.
Here, we leverage the imbalance in complexity of the dynamics to learn more sample-efficiently.
We validate our method on several challenging simulated tasks and demonstrate that it improves learning both alone and when combined with an existing hindsight algorithm.
arXiv Detail & Related papers (2023-03-03T21:55:04Z) - Self-Improving Robots: End-to-End Autonomous Visuomotor Reinforcement
Learning [54.636562516974884]
In imitation and reinforcement learning, the cost of human supervision limits the amount of data that robots can be trained on.
In this work, we propose MEDAL++, a novel design for self-improving robotic systems.
The robot autonomously practices the task by learning to both do and undo the task, simultaneously inferring the reward function from the demonstrations.
arXiv Detail & Related papers (2023-03-02T18:51:38Z) - Improving Multimodal Interactive Agents with Reinforcement Learning from
Human Feedback [16.268581985382433]
An important goal in artificial intelligence is to create agents that can both interact naturally with humans and learn from their feedback.
Here we demonstrate how to use reinforcement learning from human feedback to improve upon simulated, embodied agents.
arXiv Detail & Related papers (2022-11-21T16:00:31Z) - Explainability Via Causal Self-Talk [9.149689942389923]
Explaining the behavior of AI systems is an important problem that, in practice, is generally avoided.
We describe an effective way to satisfy all the desiderata: train the AI system to build a causal model of itself.
We implement this method in a simulated 3D environment, and show how it enables agents to generate faithful and semantically-meaningful explanations.
arXiv Detail & Related papers (2022-11-17T23:17:01Z) - Humans are not Boltzmann Distributions: Challenges and Opportunities for
Modelling Human Feedback and Interaction in Reinforcement Learning [13.64577704565643]
We argue that these models are too simplistic and that RL researchers need to develop more realistic human models to design and evaluate their algorithms.
This paper calls for research from different disciplines to address key questions about how humans provide feedback to AIs and how we can build more robust human-in-the-loop RL systems.
arXiv Detail & Related papers (2022-06-27T13:58:51Z) - Modeling Human Behavior Part I -- Learning and Belief Approaches [0.0]
We focus on techniques which learn a model or policy of behavior through exploration and feedback.
Next generation autonomous and adaptive systems will largely include AI agents and humans working together as teams.
arXiv Detail & Related papers (2022-05-13T07:33:49Z) - Robot Learning from Randomized Simulations: A Review [59.992761565399185]
Deep learning has caused a paradigm shift in robotics research, favoring methods that require large amounts of data.
State-of-the-art approaches learn in simulation where data generation is fast as well as inexpensive.
We focus on a technique named 'domain randomization' which is a method for learning from randomized simulations.
arXiv Detail & Related papers (2021-11-01T13:55:41Z) - PEBBLE: Feedback-Efficient Interactive Reinforcement Learning via
Relabeling Experience and Unsupervised Pre-training [94.87393610927812]
We present an off-policy, interactive reinforcement learning algorithm that capitalizes on the strengths of both feedback and off-policy learning.
We demonstrate that our approach is capable of learning tasks of higher complexity than previously considered by human-in-the-loop methods.
arXiv Detail & Related papers (2021-06-09T14:10:50Z) - Learning Human Rewards by Inferring Their Latent Intelligence Levels in
Multi-Agent Games: A Theory-of-Mind Approach with Application to Driving Data [18.750834997334664]
We argue that humans are bounded rational and have different intelligence levels when reasoning about others' decision-making process.
We propose a new multi-agent Inverse Reinforcement Learning framework that reasons about humans' latent intelligence levels during learning.
arXiv Detail & Related papers (2021-03-07T07:48:31Z) - Is the Most Accurate AI the Best Teammate? Optimizing AI for Teamwork [54.309495231017344]
We argue that AI systems should be trained in a human-centered manner, directly optimized for team performance.
We study this proposal for a specific type of human-AI teaming, where the human overseer chooses to either accept the AI recommendation or solve the task themselves.
Our experiments with linear and non-linear models on real-world, high-stakes datasets show that the most accuracy AI may not lead to highest team performance.
arXiv Detail & Related papers (2020-04-27T19:06:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.