Remote Task-oriented Grasp Area Teaching By Non-Experts through
Interactive Segmentation and Few-Shot Learning
- URL: http://arxiv.org/abs/2303.10195v1
- Date: Fri, 17 Mar 2023 18:09:01 GMT
- Title: Remote Task-oriented Grasp Area Teaching By Non-Experts through
Interactive Segmentation and Few-Shot Learning
- Authors: Furkan Kaynar, Sudarshan Rajagopalan, Shaobo Zhou, Eckehard Steinbach
- Abstract summary: A robot must be able to discriminate between different grasping styles depending on the prospective manipulation task.
We propose a novel two-step framework towards this aim.
We receive grasp area demonstrations for a new task via interactive segmentation.
We learn from these few demonstrations to estimate the required grasp area on an unseen scene for the given task.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A robot operating in unstructured environments must be able to discriminate
between different grasping styles depending on the prospective manipulation
task. Having a system that allows learning from remote non-expert
demonstrations can very feasibly extend the cognitive skills of a robot for
task-oriented grasping. We propose a novel two-step framework towards this aim.
The first step involves grasp area estimation by segmentation. We receive grasp
area demonstrations for a new task via interactive segmentation, and learn from
these few demonstrations to estimate the required grasp area on an unseen scene
for the given task. The second step is autonomous grasp estimation in the
segmented region. To train the segmentation network for few-shot learning, we
built a grasp area segmentation (GAS) dataset with 10089 images grouped into
1121 segmentation tasks. We benefit from an efficient meta learning algorithm
for training for few-shot adaptation. Experimental evaluation showed that our
method successfully detects the correct grasp area on the respective objects in
unseen test scenes and effectively allows remote teaching of new grasp
strategies by non-experts.
Related papers
- Self-supervised Learning via Cluster Distance Prediction for Operating Room Context Awareness [44.15562068190958]
In the Operating Room, semantic segmentation is at the core of creating robots aware of clinical surroundings.
State-of-the-art semantic segmentation and activity recognition approaches are fully supervised, which is not scalable.
We propose a new 3D self-supervised task for OR scene understanding utilizing OR scene images captured with ToF cameras.
arXiv Detail & Related papers (2024-07-07T17:17:52Z) - Auxiliary Tasks Enhanced Dual-affinity Learning for Weakly Supervised
Semantic Segmentation [79.05949524349005]
We propose AuxSegNet+, a weakly supervised auxiliary learning framework to explore the rich information from saliency maps.
We also propose a cross-task affinity learning mechanism to learn pixel-level affinities from the saliency and segmentation feature maps.
arXiv Detail & Related papers (2024-03-02T10:03:21Z) - Language-Conditioned Change-point Detection to Identify Sub-Tasks in
Robotics Domains [43.96051384180866]
We identify sub-tasks within a demonstrated robot trajectory using language instructions.
We propose a language-conditioned change-point detection method to identify sub-tasks in a problem.
arXiv Detail & Related papers (2023-09-01T21:40:34Z) - Towards Deeply Unified Depth-aware Panoptic Segmentation with
Bi-directional Guidance Learning [63.63516124646916]
We propose a deeply unified framework for depth-aware panoptic segmentation.
We propose a bi-directional guidance learning approach to facilitate cross-task feature learning.
Our method sets the new state of the art for depth-aware panoptic segmentation on both Cityscapes-DVPS and SemKITTI-DVPS datasets.
arXiv Detail & Related papers (2023-07-27T11:28:33Z) - GenCo: An Auxiliary Generator from Contrastive Learning for Enhanced
Few-Shot Learning in Remote Sensing [9.504503675097137]
We introduce a generator-based contrastive learning framework (GenCo) that pre-trains backbones and simultaneously explores variants of feature samples.
In fine-tuning, the auxiliary generator can be used to enrich limited labeled data samples in feature space.
We demonstrate the effectiveness of our method in improving few-shot learning performance on two key remote sensing datasets.
arXiv Detail & Related papers (2023-07-27T03:59:19Z) - AIMS: All-Inclusive Multi-Level Segmentation [93.5041381700744]
We propose a new task, All-Inclusive Multi-Level (AIMS), which segments visual regions into three levels: part, entity, and relation.
We also build a unified AIMS model through multi-dataset multi-task training to address the two major challenges of annotation inconsistency and task correlation.
arXiv Detail & Related papers (2023-05-28T16:28:49Z) - A Multi-Task Deep Learning Approach for Sensor-based Human Activity
Recognition and Segmentation [4.987833356397567]
We propose a new deep neural network to solve the two tasks simultaneously.
The proposed network adopts selective convolution and features multiscale windows to segment activities of long or short time durations.
Our proposed method outperforms the state-of-the-art methods both for activity recognition and segmentation.
arXiv Detail & Related papers (2023-03-20T13:34:28Z) - Self-supervised Pre-training for Semantic Segmentation in an Indoor
Scene [8.357801312689622]
We propose RegConsist, a method for self-supervised pre-training of a semantic segmentation model.
We use a variant of contrastive learning to train a DCNN model for predicting semantic segmentation from RGB views in the target environment.
The proposed method outperforms models pre-trained on ImageNet and achieves competitive performance when using models that are trained for exactly the same task but on a different dataset.
arXiv Detail & Related papers (2022-10-04T20:10:14Z) - Leveraging Auxiliary Tasks with Affinity Learning for Weakly Supervised
Semantic Segmentation [88.49669148290306]
We propose a novel weakly supervised multi-task framework called AuxSegNet to leverage saliency detection and multi-label image classification as auxiliary tasks.
Inspired by their similar structured semantics, we also propose to learn a cross-task global pixel-level affinity map from the saliency and segmentation representations.
The learned cross-task affinity can be used to refine saliency predictions and propagate CAM maps to provide improved pseudo labels for both tasks.
arXiv Detail & Related papers (2021-07-25T11:39:58Z) - Learning to Relate Depth and Semantics for Unsupervised Domain
Adaptation [87.1188556802942]
We present an approach for encoding visual task relationships to improve model performance in an Unsupervised Domain Adaptation (UDA) setting.
We propose a novel Cross-Task Relation Layer (CTRL), which encodes task dependencies between the semantic and depth predictions.
Furthermore, we propose an Iterative Self-Learning (ISL) training scheme, which exploits semantic pseudo-labels to provide extra supervision on the target domain.
arXiv Detail & Related papers (2021-05-17T13:42:09Z) - Self-supervised Segmentation via Background Inpainting [96.10971980098196]
We introduce a self-supervised detection and segmentation approach that can work with single images captured by a potentially moving camera.
We exploit a self-supervised loss function that we exploit to train a proposal-based segmentation network.
We apply our method to human detection and segmentation in images that visually depart from those of standard benchmarks and outperform existing self-supervised methods.
arXiv Detail & Related papers (2020-11-11T08:34:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.