Learn Fast, Segment Well: Fast Object Segmentation Learning on the iCub
Robot
- URL: http://arxiv.org/abs/2206.13462v1
- Date: Mon, 27 Jun 2022 17:14:04 GMT
- Title: Learn Fast, Segment Well: Fast Object Segmentation Learning on the iCub
Robot
- Authors: Federico Ceola, Elisa Maiettini, Giulia Pasquale, Giacomo Meanti,
Lorenzo Rosasco, Lorenzo Natale
- Abstract summary: We study different techniques that allow adapting an object segmentation model in presence of novel objects or different domains.
We propose a pipeline for fast instance segmentation learning for robotic applications where data come in stream.
We benchmark the proposed pipeline on two datasets and we deploy it on a real robot, iCub humanoid.
- Score: 20.813028212068424
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The visual system of a robot has different requirements depending on the
application: it may require high accuracy or reliability, be constrained by
limited resources or need fast adaptation to dynamically changing environments.
In this work, we focus on the instance segmentation task and provide a
comprehensive study of different techniques that allow adapting an object
segmentation model in presence of novel objects or different domains. We
propose a pipeline for fast instance segmentation learning designed for robotic
applications where data come in stream. It is based on an hybrid method
leveraging on a pre-trained CNN for feature extraction and fast-to-train
Kernel-based classifiers. We also propose a training protocol that allows to
shorten the training time by performing feature extraction during the data
acquisition. We benchmark the proposed pipeline on two robotics datasets and we
deploy it on a real robot, i.e. the iCub humanoid. To this aim, we adapt our
method to an incremental setting in which novel objects are learned on-line by
the robot. The code to reproduce the experiments is publicly available on
GitHub.
Related papers
- Scaling Proprioceptive-Visual Learning with Heterogeneous Pre-trained Transformers [41.069074375686164]
We propose Heterogeneous Pre-trained Transformers (HPT), which pre-train a trunk of a policy neural network to learn a task and embodiment shared representation.
We conduct experiments to investigate the scaling behaviors of training objectives, to the extent of 52 datasets.
HPTs outperform several baselines and enhance the fine-tuned policy performance by over 20% on unseen tasks.
arXiv Detail & Related papers (2024-09-30T17:39:41Z) - Scaling Manipulation Learning with Visual Kinematic Chain Prediction [32.99644520625179]
We propose the visual kinematics chain as a precise and universal representation of quasi-static actions for robot learning over diverse environments.
We demonstrate the superior performance of VKT over BC transformers as a general agent on Calvin, RLBench, Open-X, and real robot manipulation tasks.
arXiv Detail & Related papers (2024-06-12T03:10:27Z) - SERL: A Software Suite for Sample-Efficient Robotic Reinforcement
Learning [85.21378553454672]
We develop a library containing a sample efficient off-policy deep RL method, together with methods for computing rewards and resetting the environment.
We find that our implementation can achieve very efficient learning, acquiring policies for PCB board assembly, cable routing, and object relocation.
These policies achieve perfect or near-perfect success rates, extreme robustness even under perturbations, and exhibit emergent robustness recovery and correction behaviors.
arXiv Detail & Related papers (2024-01-29T10:01:10Z) - Robot Fine-Tuning Made Easy: Pre-Training Rewards and Policies for
Autonomous Real-World Reinforcement Learning [58.3994826169858]
We introduce RoboFuME, a reset-free fine-tuning system for robotic reinforcement learning.
Our insights are to utilize offline reinforcement learning techniques to ensure efficient online fine-tuning of a pre-trained policy.
Our method can incorporate data from an existing robot dataset and improve on a target task within as little as 3 hours of autonomous real-world experience.
arXiv Detail & Related papers (2023-10-23T17:50:08Z) - ALBench: A Framework for Evaluating Active Learning in Object Detection [102.81795062493536]
This paper contributes an active learning benchmark framework named as ALBench for evaluating active learning in object detection.
Developed on an automatic deep model training system, this ALBench framework is easy-to-use, compatible with different active learning algorithms, and ensures the same training and testing protocols.
arXiv Detail & Related papers (2022-07-27T07:46:23Z) - One to Many: Adaptive Instrument Segmentation via Meta Learning and
Dynamic Online Adaptation in Robotic Surgical Video [71.43912903508765]
MDAL is a dynamic online adaptive learning scheme for instrument segmentation in robot-assisted surgery.
It learns the general knowledge of instruments and the fast adaptation ability through the video-specific meta-learning paradigm.
It outperforms other state-of-the-art methods on two datasets.
arXiv Detail & Related papers (2021-03-24T05:02:18Z) - Few-Shot Visual Grounding for Natural Human-Robot Interaction [0.0]
We propose a software architecture that segments a target object from a crowded scene, indicated verbally by a human user.
At the core of our system, we employ a multi-modal deep neural network for visual grounding.
We evaluate the performance of the proposed model on real RGB-D data collected from public scene datasets.
arXiv Detail & Related papers (2021-03-17T15:24:02Z) - Bayesian Meta-Learning for Few-Shot Policy Adaptation Across Robotic
Platforms [60.59764170868101]
Reinforcement learning methods can achieve significant performance but require a large amount of training data collected on the same robotic platform.
We formulate it as a few-shot meta-learning problem where the goal is to find a model that captures the common structure shared across different robotic platforms.
We experimentally evaluate our framework on a simulated reaching and a real-robot picking task using 400 simulated robots.
arXiv Detail & Related papers (2021-03-05T14:16:20Z) - Where is my hand? Deep hand segmentation for visual self-recognition in
humanoid robots [129.46920552019247]
We propose the use of a Convolution Neural Network (CNN) to segment the robot hand from an image in an egocentric view.
We fine-tuned the Mask-RCNN network for the specific task of segmenting the hand of the humanoid robot Vizzy.
arXiv Detail & Related papers (2021-02-09T10:34:32Z) - Fast Object Segmentation Learning with Kernel-based Methods for Robotics [21.48920421574167]
Object segmentation is a key component in the visual system of a robot that performs tasks like grasping and object manipulation.
We propose a novel architecture for object segmentation, that overcomes this problem and provides comparable performance in a fraction of the time required by the state-of-the-art methods.
Our approach is validated on the YCB-Video dataset which is widely adopted in the computer vision and robotics community.
arXiv Detail & Related papers (2020-11-25T15:07:39Z) - Fast Region Proposal Learning for Object Detection for Robotics [21.48920421574167]
We propose an architecture that leverages on the powerful representation of deep learning descriptors, while permitting fast adaptation time.
In this paper, we demonstrate that a further boost in accuracy can be obtained by adapting, in addition to the regions candidate generation on the task at hand.
arXiv Detail & Related papers (2020-11-25T14:51:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.