Robust Multi-Modal Policies for Industrial Assembly via Reinforcement
Learning and Demonstrations: A Large-Scale Study
- URL: http://arxiv.org/abs/2103.11512v2
- Date: Tue, 23 Mar 2021 08:37:39 GMT
- Title: Robust Multi-Modal Policies for Industrial Assembly via Reinforcement
Learning and Demonstrations: A Large-Scale Study
- Authors: Jianlan Luo, Oleg Sushkov, Rugile Pevceviciute, Wenzhao Lian, Chang
Su, Mel Vecerik, Ning Ye, Stefan Schaal, Jon Scholz
- Abstract summary: We argue that it is the prohibitively large design space for Deep Reinforcement Learning (DRL) that are truly responsible for this lack of adoption.
This study suggests that DRL is capable of outperforming not only established engineered approaches, but the human motor system as well.
- Score: 14.696027001985554
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Over the past several years there has been a considerable research investment
into learning-based approaches to industrial assembly, but despite significant
progress these techniques have yet to be adopted by industry. We argue that it
is the prohibitively large design space for Deep Reinforcement Learning (DRL),
rather than algorithmic limitations per se, that are truly responsible for this
lack of adoption. Pushing these techniques into the industrial mainstream
requires an industry-oriented paradigm which differs significantly from the
academic mindset. In this paper we define criteria for industry-oriented DRL,
and perform a thorough comparison according to these criteria of one family of
learning approaches, DRL from demonstration, against a professional industrial
integrator on the recently established NIST assembly benchmark. We explain the
design choices, representing several years of investigation, which enabled our
DRL system to consistently outperform the integrator baseline in terms of both
speed and reliability. Finally, we conclude with a competition between our DRL
system and a human on a challenge task of insertion into a randomly moving
target. This study suggests that DRL is capable of outperforming not only
established engineered approaches, but the human motor system as well, and that
there remains significant room for improvement. Videos can be found on our
project website: https://sites.google.com/view/shield-nist.
Related papers
- Generative AI for Deep Reinforcement Learning: Framework, Analysis, and Use Cases [60.30995339585003]
Deep reinforcement learning (DRL) has been widely applied across various fields and has achieved remarkable accomplishments.
DRL faces certain limitations, including low sample efficiency and poor generalization.
We present how to leverage generative AI (GAI) to address these issues and enhance the performance of DRL algorithms.
arXiv Detail & Related papers (2024-05-31T01:25:40Z) - Aquatic Navigation: A Challenging Benchmark for Deep Reinforcement Learning [53.3760591018817]
We propose a new benchmarking environment for aquatic navigation using recent advances in the integration between game engines and Deep Reinforcement Learning.
Specifically, we focus on PPO, one of the most widely accepted algorithms, and we propose advanced training techniques.
Our empirical evaluation shows that a well-designed combination of these ingredients can achieve promising results.
arXiv Detail & Related papers (2024-05-30T23:20:23Z) - Deep reinforcement learning for machine scheduling: Methodology, the
state-of-the-art, and future directions [2.4541568670428915]
Machine scheduling aims to optimize job assignments to machines while adhering to manufacturing rules and job specifications.
Deep Reinforcement Learning (DRL), a key component of artificial general intelligence, has shown promise in various domains like gaming and robotics.
This paper offers a comprehensive review and comparison of DRL-based approaches, highlighting their methodology, applications, advantages, and limitations.
arXiv Detail & Related papers (2023-10-04T22:45:09Z) - RLLTE: Long-Term Evolution Project of Reinforcement Learning [48.181733263496746]
We present RLLTE: a long-term evolution, extremely modular, and open-source framework for reinforcement learning research and application.
Beyond delivering top-notch algorithm implementations, RLLTE also serves as a toolkit for developing algorithms.
RLLTE is expected to set standards for RL engineering practice and be highly stimulative for industry and academia.
arXiv Detail & Related papers (2023-09-28T12:30:37Z) - An Architecture for Deploying Reinforcement Learning in Industrial
Environments [3.18294468240512]
We present an OPC UA based Operational Technology (OT)-aware RL architecture.
We define an OPC UA information model allowing for a generalized plug-and-play like approach for exchanging the RL agent.
By means of solving a toy example, we show that this architecture can be used to determine the optimal policy.
arXiv Detail & Related papers (2023-06-02T10:22:01Z) - A Mini Review on the utilization of Reinforcement Learning with OPC UA [0.9208007322096533]
Reinforcement Learning (RL) is a powerful machine learning paradigm that has been applied in various fields such as robotics, natural language processing and game playing.
The key to fully exploiting this potential is the seamless integration of RL into existing industrial systems.
This work serves to bridge this gap by providing a brief technical overview of both technologies and carrying out a semi-exhaustive literature review.
arXiv Detail & Related papers (2023-05-24T13:03:48Z) - Reinforcement Learning Applied to Trading Systems: A Survey [5.118560450410779]
The recent achievements and the notoriety of Reinforcement Learning have increased its adoption in trading tasks.
This review attempts to promote the development of this field of study by researchers' commitment to standards adherence.
arXiv Detail & Related papers (2022-11-01T21:26:12Z) - Pessimistic Model Selection for Offline Deep Reinforcement Learning [56.282483586473816]
Deep Reinforcement Learning (DRL) has demonstrated great potentials in solving sequential decision making problems in many applications.
One main barrier is the over-fitting issue that leads to poor generalizability of the policy learned by DRL.
We propose a pessimistic model selection (PMS) approach for offline DRL with a theoretical guarantee.
arXiv Detail & Related papers (2021-11-29T06:29:49Z) - Towards Standardizing Reinforcement Learning Approaches for Stochastic
Production Scheduling [77.34726150561087]
reinforcement learning can be used to solve scheduling problems.
Existing studies rely on (sometimes) complex simulations for which the code is unavailable.
There is a vast array of RL designs to choose from.
standardization of model descriptions - both production setup and RL design - and validation scheme are a prerequisite.
arXiv Detail & Related papers (2021-04-16T16:07:10Z) - RL Unplugged: A Suite of Benchmarks for Offline Reinforcement Learning [108.9599280270704]
We propose a benchmark called RL Unplugged to evaluate and compare offline RL methods.
RL Unplugged includes data from a diverse range of domains including games and simulated motor control problems.
We will release data for all our tasks and open-source all algorithms presented in this paper.
arXiv Detail & Related papers (2020-06-24T17:14:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.