Pseudo-Labeling and Contextual Curriculum Learning for Online Grasp
Learning in Robotic Bin Picking
- URL: http://arxiv.org/abs/2403.02495v1
- Date: Mon, 4 Mar 2024 21:41:27 GMT
- Title: Pseudo-Labeling and Contextual Curriculum Learning for Online Grasp
Learning in Robotic Bin Picking
- Authors: Huy Le, Philipp Schillinger, Miroslav Gabriel, Alexander Qualmann, Ngo
Anh Vien
- Abstract summary: SSL-ConvSAC combines semi-supervised learning and reinforcement learning for online grasp learning.
We demonstrate promise for improving online grasp learning on bin picking tasks using a physical 7-DoF Franka Emika robot arm with a suction gripper.
- Score: 47.4409816260196
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The prevailing grasp prediction methods predominantly rely on offline
learning, overlooking the dynamic grasp learning that occurs during real-time
adaptation to novel picking scenarios. These scenarios may involve previously
unseen objects, variations in camera perspectives, and bin configurations,
among other factors. In this paper, we introduce a novel approach, SSL-ConvSAC,
that combines semi-supervised learning and reinforcement learning for online
grasp learning. By treating pixels with reward feedback as labeled data and
others as unlabeled, it efficiently exploits unlabeled data to enhance
learning. In addition, we address the imbalance between labeled and unlabeled
data by proposing a contextual curriculum-based method. We ablate the proposed
approach on real-world evaluation data and demonstrate promise for improving
online grasp learning on bin picking tasks using a physical 7-DoF Franka Emika
robot arm with a suction gripper. Video: https://youtu.be/OAro5pg8I9U
Related papers
- Exploiting the Semantic Knowledge of Pre-trained Text-Encoders for Continual Learning [70.64617500380287]
Continual learning allows models to learn from new data while retaining previously learned knowledge.
The semantic knowledge available in the label information of the images, offers important semantic information that can be related with previously acquired knowledge of semantic classes.
We propose integrating semantic guidance within and across tasks by capturing semantic similarity using text embeddings.
arXiv Detail & Related papers (2024-08-02T07:51:44Z) - Any-point Trajectory Modeling for Policy Learning [64.23861308947852]
We introduce Any-point Trajectory Modeling (ATM) to predict future trajectories of arbitrary points within a video frame.
ATM outperforms strong video pre-training baselines by 80% on average.
We show effective transfer learning of manipulation skills from human videos and videos from a different robot morphology.
arXiv Detail & Related papers (2023-12-28T23:34:43Z) - Uncertainty-driven Exploration Strategies for Online Grasp Learning [43.88491290121489]
We present an uncertainty-based approach for online learning of grasp predictions for robotic bin picking.
Specifically, the online learning algorithm with an effective exploration strategy can significantly improve its adaptation performance to unseen environment settings.
arXiv Detail & Related papers (2023-09-21T13:06:03Z) - Harnessing the Power of Text-image Contrastive Models for Automatic
Detection of Online Misinformation [50.46219766161111]
We develop a self-learning model to explore the constrastive learning in the domain of misinformation identification.
Our model shows the superior performance of non-matched image-text pair detection when the training data is insufficient.
arXiv Detail & Related papers (2023-04-19T02:53:59Z) - Brief Introduction to Contrastive Learning Pretext Tasks for Visual
Representation [0.0]
We introduce contrastive learning, a subset of unsupervised learning methods.
The purpose of contrastive learning is to embed augmented samples from the same sample near to each other while pushing away those that are not.
We offer some strategies from contrastive learning that have recently been published and are focused on pretext tasks for visual representation.
arXiv Detail & Related papers (2022-10-06T18:54:10Z) - An Exploration of Active Learning for Affective Digital Phenotyping [4.790279027864381]
Active learning is a paradigm for using algorithms to computationally select a useful subset of data points to label.
We explore active learning for naturalistic computer vision emotion data, a particularly heterogeneous and complex data space.
We find that active learning using information generated during gameplay slightly outperforms random selection of the same number of labeled frames.
arXiv Detail & Related papers (2022-04-05T01:01:32Z) - Online Continual Learning on a Contaminated Data Stream with Blurry Task
Boundaries [17.43350151320054]
A large body of continual learning (CL) methods assumes data streams with clean labels, and online learning scenarios under noisy data streams are yet underexplored.
We consider a more practical CL task setup of an online learning from blurry data stream with corrupted labels, where existing CL methods struggle.
We propose a novel strategy to manage and use the memory by a unified approach of label noise aware diverse sampling and robust learning with semi-supervised learning.
arXiv Detail & Related papers (2022-03-29T08:52:45Z) - Learning to Predict Gradients for Semi-Supervised Continual Learning [36.715712711431856]
Key challenge for machine intelligence is to learn new visual concepts without forgetting the previously acquired knowledge.
There is a gap between existing supervised continual learning and human-like intelligence, where human is able to learn from both labeled and unlabeled data.
We formulate a new semi-supervised continual learning method, which can be generically applied to existing continual learning models.
arXiv Detail & Related papers (2022-01-23T06:45:47Z) - Online Continual Learning with Natural Distribution Shifts: An Empirical
Study with Visual Data [101.6195176510611]
"Online" continual learning enables evaluating both information retention and online learning efficacy.
In online continual learning, each incoming small batch of data is first used for testing and then added to the training set, making the problem truly online.
We introduce a new benchmark for online continual visual learning that exhibits large scale and natural distribution shifts.
arXiv Detail & Related papers (2021-08-20T06:17:20Z) - Privileged Knowledge Distillation for Online Action Detection [114.5213840651675]
Online Action Detection (OAD) in videos is proposed as a per-frame labeling task to address the real-time prediction tasks.
This paper presents a novel learning-with-privileged based framework for online action detection where the future frames only observable at the training stages are considered as a form of privileged information.
arXiv Detail & Related papers (2020-11-18T08:52:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.