Auditing Robot Learning for Safety and Compliance during Deployment
- URL: http://arxiv.org/abs/2110.05702v1
- Date: Tue, 12 Oct 2021 02:40:11 GMT
- Title: Auditing Robot Learning for Safety and Compliance during Deployment
- Authors: Homanga Bharadhwaj
- Abstract summary: We study how best to audit robot learning algorithms for checking their compatibility with humans.
We believe that this is a challenging problem that will require efforts from the entire robot learning community.
- Score: 4.742825811314168
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Robots of the future are going to exhibit increasingly human-like and
super-human intelligence in a myriad of different tasks. They are also likely
going to fail and be incompliant with human preferences in increasingly subtle
ways. Towards the goal of achieving autonomous robots, the robot learning
community has made rapid strides in applying machine learning techniques to
train robots through data and interaction. This makes the study of how best to
audit these algorithms for checking their compatibility with humans, pertinent
and urgent. In this paper, we draw inspiration from the AI Safety and Alignment
communities and make the case that we need to urgently consider ways in which
we can best audit our robot learning algorithms to check for failure modes, and
ensure that when operating autonomously, they are indeed behaving in ways that
the human algorithm designers intend them to. We believe that this is a
challenging problem that will require efforts from the entire robot learning
community, and do not attempt to provide a concrete framework for auditing.
Instead, we outline high-level guidance and a possible approach towards
formulating this framework which we hope will serve as a useful starting point
for thinking about auditing in the context of robot learning.
Related papers
- Commonsense Reasoning for Legged Robot Adaptation with Vision-Language Models [81.55156507635286]
Legged robots are physically capable of navigating a diverse variety of environments and overcoming a wide range of obstructions.
Current learning methods often struggle with generalization to the long tail of unexpected situations without heavy human supervision.
We propose a system, VLM-Predictive Control (VLM-PC), combining two key components that we find to be crucial for eliciting on-the-fly, adaptive behavior selection.
arXiv Detail & Related papers (2024-07-02T21:00:30Z) - HumanoidBench: Simulated Humanoid Benchmark for Whole-Body Locomotion and Manipulation [50.616995671367704]
We present a high-dimensional, simulated robot learning benchmark, HumanoidBench, featuring a humanoid robot equipped with dexterous hands.
Our findings reveal that state-of-the-art reinforcement learning algorithms struggle with most tasks, whereas a hierarchical learning approach achieves superior performance when supported by robust low-level policies.
arXiv Detail & Related papers (2024-03-15T17:45:44Z) - Growing from Exploration: A self-exploring framework for robots based on
foundation models [13.250831101705694]
We propose a framework named GExp, which enables robots to explore and learn autonomously without human intervention.
Inspired by the way that infants interact with the world, GExp encourages robots to understand and explore the environment with a series of self-generated tasks.
arXiv Detail & Related papers (2024-01-24T14:04:08Z) - Exploring AI-enhanced Shared Control for an Assistive Robotic Arm [4.999814847776098]
In particular, we explore how Artifical Intelligence (AI) can be integrated into a shared control paradigm.
In particular, we focus on the consequential requirements for the interface between human and robot.
arXiv Detail & Related papers (2023-06-23T14:19:56Z) - Stabilizing Contrastive RL: Techniques for Robotic Goal Reaching from
Offline Data [101.43350024175157]
Self-supervised learning has the potential to decrease the amount of human annotation and engineering effort required to learn control strategies.
Our work builds on prior work showing that the reinforcement learning (RL) itself can be cast as a self-supervised problem.
We demonstrate that a self-supervised RL algorithm based on contrastive learning can solve real-world, image-based robotic manipulation tasks.
arXiv Detail & Related papers (2023-06-06T01:36:56Z) - Self-Improving Robots: End-to-End Autonomous Visuomotor Reinforcement
Learning [54.636562516974884]
In imitation and reinforcement learning, the cost of human supervision limits the amount of data that robots can be trained on.
In this work, we propose MEDAL++, a novel design for self-improving robotic systems.
The robot autonomously practices the task by learning to both do and undo the task, simultaneously inferring the reward function from the demonstrations.
arXiv Detail & Related papers (2023-03-02T18:51:38Z) - Aligning Robot Representations with Humans [5.482532589225552]
Key question is how to best transfer knowledge learned in one environment to another, where shifting constraints and human preferences render adaptation challenging.
We postulate that because humans will be the ultimate evaluator of system success in the world, they are best suited to communicating the aspects of the tasks that matter to the robot.
We highlight three areas where we can use this approach to build interactive systems and offer future directions of work to better create advanced collaborative robots.
arXiv Detail & Related papers (2022-05-15T15:51:05Z) - Back to Reality for Imitation Learning [8.57914821832517]
Imitation learning, and robot learning in general, emerged due to breakthroughs in machine learning, rather than breakthroughs in robotics.
We believe that a better metric for real-world robot learning is time efficiency, which better models the true cost to humans.
arXiv Detail & Related papers (2021-11-25T02:03:52Z) - Dual-Arm Adversarial Robot Learning [0.6091702876917281]
We propose dual-arm settings as platforms for robot learning.
We will discuss the potential benefits of this setup as well as the challenges and research directions that can be pursued.
arXiv Detail & Related papers (2021-10-15T12:51:57Z) - Hierarchical Affordance Discovery using Intrinsic Motivation [69.9674326582747]
We propose an algorithm using intrinsic motivation to guide the learning of affordances for a mobile robot.
This algorithm is capable to autonomously discover, learn and adapt interrelated affordances without pre-programmed actions.
Once learned, these affordances may be used by the algorithm to plan sequences of actions in order to perform tasks of various difficulties.
arXiv Detail & Related papers (2020-09-23T07:18:21Z) - The Ingredients of Real-World Robotic Reinforcement Learning [71.92831985295163]
We discuss the elements that are needed for a robotic learning system that can continually and autonomously improve with data collected in the real world.
We propose a particular instantiation of such a system, using dexterous manipulation as our case study.
We demonstrate that our complete system can learn without any human intervention, acquiring a variety of vision-based skills with a real-world three-fingered hand.
arXiv Detail & Related papers (2020-04-27T03:36:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.