Learning from Incremental Directional Corrections
- URL: http://arxiv.org/abs/2011.15014v1
- Date: Mon, 30 Nov 2020 17:16:39 GMT
- Title: Learning from Incremental Directional Corrections
- Authors: Wanxin Jin, Todd D. Murphey, Shaoshuai Mou
- Abstract summary: We propose a technique which enables a robot to learn a control objective function incrementally from human user's corrections.
We only assume that each of the human's corrections, regardless of its magnitude, points in a direction that improves the robot's current motion.
The proposed method uses the direction of a correction to update the estimate of the objective function based on a cutting plane technique.
- Score: 9.45570271906093
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a technique which enables a robot to learn a control
objective function incrementally from human user's corrections. The human's
corrections can be as simple as directional corrections -- corrections that
indicate the direction of a control change without indicating its magnitude --
applied at some time instances during the robot's motion. We only assume that
each of the human's corrections, regardless of its magnitude, points in a
direction that improves the robot's current motion relative to an implicit
objective function. The proposed method uses the direction of a correction to
update the estimate of the objective function based on a cutting plane
technique. We establish the theoretical results to show that this process of
incremental correction and update guarantees convergence of the learned
objective function to the implicit one. The method is validated by both
simulations and two human-robot games, where human players teach a 2-link robot
arm and a 6-DoF quadrotor system for motion planning in environments with
obstacles.
Related papers
- Self-Supervised Learning of Grasping Arbitrary Objects On-the-Move [8.445514342786579]
This study introduces three fully convolutional neural network (FCN) models to predict static grasp primitive, dynamic grasp primitive, and residual moving velocity error from visual inputs.
The proposed method achieved the highest grasping accuracy and pick-and-place efficiency.
arXiv Detail & Related papers (2024-11-15T02:59:16Z) - Unsupervised Learning of Effective Actions in Robotics [0.9374652839580183]
Current state-of-the-art action representations in robotics lack proper effect-driven learning of the robot's actions.
We propose an unsupervised algorithm to discretize a continuous motion space and generate "action prototypes"
We evaluate our method on a simulated stair-climbing reinforcement learning task.
arXiv Detail & Related papers (2024-04-03T13:28:52Z) - "No, to the Right" -- Online Language Corrections for Robotic
Manipulation via Shared Autonomy [70.45420918526926]
We present LILAC, a framework for incorporating and adapting to natural language corrections online during execution.
Instead of discrete turn-taking between a human and robot, LILAC splits agency between the human and robot.
We show that our corrections-aware approach obtains higher task completion rates, and is subjectively preferred by users.
arXiv Detail & Related papers (2023-01-06T15:03:27Z) - Revisiting the Adversarial Robustness-Accuracy Tradeoff in Robot
Learning [121.9708998627352]
Recent work has shown that, in practical robot learning applications, the effects of adversarial training do not pose a fair trade-off.
This work revisits the robustness-accuracy trade-off in robot learning by analyzing if recent advances in robust training methods and theory can make adversarial training suitable for real-world robot applications.
arXiv Detail & Related papers (2022-04-15T08:12:15Z) - Correcting Robot Plans with Natural Language Feedback [88.92824527743105]
We explore natural language as an expressive and flexible tool for robot correction.
We show that these transformations enable users to correct goals, update robot motions, and recover from planning errors.
Our method makes it possible to compose multiple constraints and generalizes to unseen scenes, objects, and sentences in simulated environments and real-world environments.
arXiv Detail & Related papers (2022-04-11T15:22:43Z) - Goal-conditioned dual-action imitation learning for dexterous dual-arm robot manipulation [4.717749411286867]
Long-horizon dexterous robot manipulation of deformable objects, such as banana peeling, is a problematic task.
This paper presents a goal-conditioned dual-action deep imitation learning (DIL) approach that can learn dexterous manipulation skills.
arXiv Detail & Related papers (2022-03-18T05:17:00Z) - Active Uncertainty Learning for Human-Robot Interaction: An Implicit
Dual Control Approach [5.05828899601167]
We present an algorithmic approach to enable uncertainty learning for human-in-the-loop motion planning based on the implicit dual control paradigm.
Our approach relies on sampling-based approximation of dynamic programming model predictive control problem.
The resulting policy is shown to preserve the dual control effect for generic human predictive models with both continuous and categorical uncertainty.
arXiv Detail & Related papers (2022-02-15T20:40:06Z) - Domain Adaptive Robotic Gesture Recognition with Unsupervised
Kinematic-Visual Data Alignment [60.31418655784291]
We propose a novel unsupervised domain adaptation framework which can simultaneously transfer multi-modality knowledge, i.e., both kinematic and visual data, from simulator to real robot.
It remedies the domain gap with enhanced transferable features by using temporal cues in videos, and inherent correlations in multi-modal towards recognizing gesture.
Results show that our approach recovers the performance with great improvement gains, up to 12.91% in ACC and 20.16% in F1score without using any annotations in real robot.
arXiv Detail & Related papers (2021-03-06T09:10:03Z) - Online Body Schema Adaptation through Cost-Sensitive Active Learning [63.84207660737483]
The work was implemented in a simulation environment, using the 7DoF arm of the iCub robot simulator.
A cost-sensitive active learning approach is used to select optimal joint configurations.
The results show cost-sensitive active learning has similar accuracy to the standard active learning approach, while reducing in about half the executed movement.
arXiv Detail & Related papers (2021-01-26T16:01:02Z) - Rapidly Adaptable Legged Robots via Evolutionary Meta-Learning [65.88200578485316]
We present a new meta-learning method that allows robots to quickly adapt to changes in dynamics.
Our method significantly improves adaptation to changes in dynamics in high noise settings.
We validate our approach on a quadruped robot that learns to walk while subject to changes in dynamics.
arXiv Detail & Related papers (2020-03-02T22:56:27Z) - Exploiting Ergonomic Priors in Human-to-Robot Task Transfer [3.60953887026184]
A method based on programming by demonstration is proposed to learn null space policies from constrained motion data.
The effectiveness of the method has been demonstrated in a 3-link simulation and a real world experiment using a human subject as the demonstrator.
The approach is shown to outperform the current state-of-the-art approach in a simulated 3DoF robot manipulator control problem.
arXiv Detail & Related papers (2020-03-01T18:30:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.