Disturbance Injection under Partial Automation: Robust Imitation
Learning for Long-horizon Tasks
- URL: http://arxiv.org/abs/2303.12375v1
- Date: Wed, 22 Mar 2023 08:22:12 GMT
- Title: Disturbance Injection under Partial Automation: Robust Imitation
Learning for Long-horizon Tasks
- Authors: Hirotaka Tahara, Hikaru Sasaki, Hanbit Oh, Edgar Anarossi, and
Takamitsu Matsubara
- Abstract summary: Partial Automation (PA) with intelligent support systems has been introduced in industrial machinery and advanced automobiles.
This paper proposes Disturbance Injection under Partial Automation (DIPA) as a novel imitation learning framework.
We experimentally validated the effectiveness of our method for long-horizon tasks in two simulations and a real robot environment.
- Score: 11.554935619056819
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Partial Automation (PA) with intelligent support systems has been introduced
in industrial machinery and advanced automobiles to reduce the burden of long
hours of human operation. Under PA, operators perform manual operations
(providing actions) and operations that switch to automatic/manual mode
(mode-switching). Since PA reduces the total duration of manual operation,
these two action and mode-switching operations can be replicated by imitation
learning with high sample efficiency. To this end, this paper proposes
Disturbance Injection under Partial Automation (DIPA) as a novel imitation
learning framework. In DIPA, mode and actions (in the manual mode) are assumed
to be observables in each state and are used to learn both action and
mode-switching policies. The above learning is robustified by injecting
disturbances into the operator's actions to optimize the disturbance's level
for minimizing the covariate shift under PA. We experimentally validated the
effectiveness of our method for long-horizon tasks in two simulations and a
real robot environment and confirmed that our method outperformed the previous
methods and reduced the demonstration burden.
Related papers
- FAST: Efficient Action Tokenization for Vision-Language-Action Models [98.15494168962563]
We propose a new compression-based tokenization scheme for robot actions, based on the discrete cosine transform.
Based on FAST, we release FAST+, a universal robot action tokenizer, trained on 1M real robot action trajectories.
arXiv Detail & Related papers (2025-01-16T18:57:04Z) - LAMS: LLM-Driven Automatic Mode Switching for Assistive Teleoperation [4.22823627787465]
We introduce LLM-Driven Automatic Mode Switching (LAMS), a novel approach to automatically switch control modes based on task context.
Unlike existing methods, LAMS requires no prior task demonstrations and incrementally improves by integrating user-generated mode-switching examples.
We validate LAMS through an ablation study and a user study with 10 participants on complex, long-horizon tasks, demonstrating that LAMS effectively reduces manual mode switches, is preferred over alternative methods, and improves performance over time.
arXiv Detail & Related papers (2025-01-15T03:49:08Z) - CARP: Visuomotor Policy Learning via Coarse-to-Fine Autoregressive Prediction [28.761494362934087]
Coarse-to-Fine AutoRegressive Policy (CARP) is a novel paradigm for visuomotor policy learning.
It redefines the autoregressive action generation process as a coarse-to-fine, next-scale approach.
CARP achieves competitive success rates, with up to a 10% improvement, and delivers 10x faster inference compared to state-of-the-art policies.
arXiv Detail & Related papers (2024-12-09T18:59:18Z) - PIVOT-R: Primitive-Driven Waypoint-Aware World Model for Robotic Manipulation [68.17081518640934]
We propose a PrIrmitive-driVen waypOinT-aware world model for Robotic manipulation (PIVOT-R)
PIVOT-R consists of a Waypoint-aware World Model (WAWM) and a lightweight action prediction module.
Our PIVOT-R outperforms state-of-the-art open-source models on the SeaWave benchmark, achieving an average relative improvement of 19.45% across four levels of instruction tasks.
arXiv Detail & Related papers (2024-10-14T11:30:18Z) - BiKC: Keypose-Conditioned Consistency Policy for Bimanual Robotic Manipulation [48.08416841005715]
We introduce a novel keypose-conditioned consistency policy tailored for bimanual manipulation.
It is a hierarchical imitation learning framework that consists of a high-level keypose predictor and a low-level trajectory generator.
Simulated and real-world experimental results demonstrate that the proposed approach surpasses baseline methods in terms of success rate and operational efficiency.
arXiv Detail & Related papers (2024-06-14T14:49:12Z) - Unsupervised Learning of Effective Actions in Robotics [0.9374652839580183]
Current state-of-the-art action representations in robotics lack proper effect-driven learning of the robot's actions.
We propose an unsupervised algorithm to discretize a continuous motion space and generate "action prototypes"
We evaluate our method on a simulated stair-climbing reinforcement learning task.
arXiv Detail & Related papers (2024-04-03T13:28:52Z) - PRISE: LLM-Style Sequence Compression for Learning Temporal Action Abstractions in Control [55.81022882408587]
Temporal action abstractions, along with belief state representations, are a powerful knowledge sharing mechanism for sequential decision making.
We propose a novel view that treats inducing temporal action abstractions as a sequence compression problem.
We introduce an approach that combines continuous action quantization with byte pair encoding to learn powerful action abstractions.
arXiv Detail & Related papers (2024-02-16T04:55:09Z) - Persistent Reinforcement Learning via Subgoal Curricula [114.83989499740193]
Value-accelerated Persistent Reinforcement Learning (VaPRL) generates a curriculum of initial states.
VaPRL reduces the interventions required by three orders of magnitude compared to episodic reinforcement learning.
arXiv Detail & Related papers (2021-07-27T16:39:45Z) - Human-in-the-Loop Imitation Learning using Remote Teleoperation [72.2847988686463]
We build a data collection system tailored to 6-DoF manipulation settings.
We develop an algorithm to train the policy iteratively on new data collected by the system.
We demonstrate that agents trained on data collected by our intervention-based system and algorithm outperform agents trained on an equivalent number of samples collected by non-interventional demonstrators.
arXiv Detail & Related papers (2020-12-12T05:30:35Z) - SAFARI: Safe and Active Robot Imitation Learning with Imagination [16.967930721746676]
SAFARI is a novel active learning and control algorithm.
It allows an agent to request further human demonstrations when these out-of-distribution situations are met.
We show how this method enables the agent to autonomously predict failure rapidly and safely.
arXiv Detail & Related papers (2020-11-18T23:43:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.