Whole-Body Control of a Mobile Manipulator using End-to-End
Reinforcement Learning
- URL: http://arxiv.org/abs/2003.02637v1
- Date: Tue, 25 Feb 2020 21:21:57 GMT
- Title: Whole-Body Control of a Mobile Manipulator using End-to-End
Reinforcement Learning
- Authors: Julien Kindle, Fadri Furrer, Tonci Novkovic, Jen Jen Chung, Roland
Siegwart and Juan Nieto
- Abstract summary: We propose an end-to-end Reinforcement Learning (RL) approach to Whole-Body Control (WBC)
We compared our learned controller against a state-of-the-art sampling-based method in simulation and achieved faster overall mission times.
- Score: 31.150823782805283
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Mobile manipulation is usually achieved by sequentially executing base and
manipulator movements. This simplification, however, leads to a loss in
efficiency and in some cases a reduction of workspace size. Even though
different methods have been proposed to solve Whole-Body Control (WBC) online,
they are either limited by a kinematic model or do not allow for reactive,
online obstacle avoidance. In order to overcome these drawbacks, in this work,
we propose an end-to-end Reinforcement Learning (RL) approach to WBC. We
compared our learned controller against a state-of-the-art sampling-based
method in simulation and achieved faster overall mission times. In addition, we
validated the learned policy on our mobile manipulator RoyalPanda in
challenging narrow corridor environments.
Related papers
- Leveraging Constraint Violation Signals For Action-Constrained Reinforcement Learning [13.332006760984122]
Action-Constrained Reinforcement Learning (ACRL) employs a projection layer after the policy network to correct the action.
Recent methods were proposed to train generative models to learn a differentiable mapping between latent variables and feasible actions.
arXiv Detail & Related papers (2025-02-08T12:58:26Z) - Growing Q-Networks: Solving Continuous Control Tasks with Adaptive Control Resolution [51.83951489847344]
In robotics applications, smooth control signals are commonly preferred to reduce system wear and energy efficiency.
In this work, we aim to bridge this performance gap by growing discrete action spaces from coarse to fine control resolution.
Our work indicates that an adaptive control resolution in combination with value decomposition yields simple critic-only algorithms that yield surprisingly strong performance on continuous control tasks.
arXiv Detail & Related papers (2024-04-05T17:58:37Z) - Action-Quantized Offline Reinforcement Learning for Robotic Skill
Learning [68.16998247593209]
offline reinforcement learning (RL) paradigm provides recipe to convert static behavior datasets into policies that can perform better than the policy that collected the data.
In this paper, we propose an adaptive scheme for action quantization.
We show that several state-of-the-art offline RL methods such as IQL, CQL, and BRAC improve in performance on benchmarks when combined with our proposed discretization scheme.
arXiv Detail & Related papers (2023-10-18T06:07:10Z) - Collision Avoidance and Navigation for a Quadrotor Swarm Using End-to-end Deep Reinforcement Learning [8.864432196281268]
We propose an end-to-end DRL approach to control quadrotor swarms in environments with obstacles.
We provide our agents a curriculum and a replay buffer of the clipped collision episodes to improve performance in obstacle-rich environments.
Our work is the first work that demonstrates the possibility of learning neighbor-avoiding and obstacle-avoiding control policies trained with end-to-end DRL.
arXiv Detail & Related papers (2023-09-23T06:56:28Z) - Deep Whole-Body Control: Learning a Unified Policy for Manipulation and
Locomotion [25.35885216505385]
An attached arm can significantly increase the applicability of legged robots to mobile manipulation tasks.
Standard hierarchical control pipeline for such legged manipulators is to decouple the controller into that of manipulation and locomotion.
We learn a unified policy for whole-body control of a legged manipulator using reinforcement learning.
arXiv Detail & Related papers (2022-10-18T17:59:30Z) - Accelerated Policy Learning with Parallel Differentiable Simulation [59.665651562534755]
We present a differentiable simulator and a new policy learning algorithm (SHAC)
Our algorithm alleviates problems with local minima through a smooth critic function.
We show substantial improvements in sample efficiency and wall-clock time over state-of-the-art RL and differentiable simulation-based algorithms.
arXiv Detail & Related papers (2022-04-14T17:46:26Z) - Real-World Dexterous Object Manipulation based Deep Reinforcement
Learning [3.4493195428573613]
We show how to use deep reinforcement learning to control a robot.
Our framework reduces the disadvantage of low sample efficiency of deep reinforcement learning.
Our algorithm is trained in simulation and migrated to reality without fine-tuning.
arXiv Detail & Related papers (2021-11-22T02:48:05Z) - OSCAR: Data-Driven Operational Space Control for Adaptive and Robust
Robot Manipulation [50.59541802645156]
Operational Space Control (OSC) has been used as an effective task-space controller for manipulation.
We propose OSC for Adaptation and Robustness (OSCAR), a data-driven variant of OSC that compensates for modeling errors.
We evaluate our method on a variety of simulated manipulation problems, and find substantial improvements over an array of controller baselines.
arXiv Detail & Related papers (2021-10-02T01:21:38Z) - Reinforcement Learning for Robust Parameterized Locomotion Control of
Bipedal Robots [121.42930679076574]
We present a model-free reinforcement learning framework for training robust locomotion policies in simulation.
domain randomization is used to encourage the policies to learn behaviors that are robust across variations in system dynamics.
We demonstrate this on versatile walking behaviors such as tracking a target walking velocity, walking height, and turning yaw.
arXiv Detail & Related papers (2021-03-26T07:14:01Z) - Collision-Free Flocking with a Dynamic Squad of Fixed-Wing UAVs Using
Deep Reinforcement Learning [2.555094847583209]
We deal with the decentralized leader-follower flocking control problem through deep reinforcement learning (DRL)
We propose a novel reinforcement learning algorithm CACER-II for training a shared control policy for all the followers.
As a result, the variable-length system state can be encoded into a fixed-length embedding vector, which makes the learned DRL policies independent with the number or the order of followers.
arXiv Detail & Related papers (2021-01-20T11:23:35Z) - Continuous Transition: Improving Sample Efficiency for Continuous
Control Problems via MixUp [119.69304125647785]
This paper introduces a concise yet powerful method to construct Continuous Transition.
Specifically, we propose to synthesize new transitions for training by linearly interpolating the consecutive transitions.
To keep the constructed transitions authentic, we also develop a discriminator to guide the construction process automatically.
arXiv Detail & Related papers (2020-11-30T01:20:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.