Segmenting Transparent Objects in the Wild
- URL: http://arxiv.org/abs/2003.13948v3
- Date: Sun, 2 Aug 2020 03:32:48 GMT
- Title: Segmenting Transparent Objects in the Wild
- Authors: Enze Xie, Wenjia Wang, Wenhai Wang, Mingyu Ding, Chunhua Shen, Ping
Luo
- Abstract summary: This work proposes a large-scale dataset for transparent object segmentation, named Trans10K, consisting of 10,428 images of real scenarios with carefully manual annotations.
To evaluate the effectiveness of Trans10K, we propose a novel boundary-aware segmentation method, termed TransLab, which exploits boundary as the clue to improve segmentation of transparent objects.
- Score: 98.80906604285163
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transparent objects such as windows and bottles made by glass widely exist in
the real world. Segmenting transparent objects is challenging because these
objects have diverse appearance inherited from the image background, making
them had similar appearance with their surroundings. Besides the technical
difficulty of this task, only a few previous datasets were specially designed
and collected to explore this task and most of the existing datasets have major
drawbacks. They either possess limited sample size such as merely a thousand of
images without manual annotations, or they generate all images by using
computer graphics method (i.e. not real image). To address this important
problem, this work proposes a large-scale dataset for transparent object
segmentation, named Trans10K, consisting of 10,428 images of real scenarios
with carefully manual annotations, which are 10 times larger than the existing
datasets. The transparent objects in Trans10K are extremely challenging due to
high diversity in scale, viewpoint and occlusion as shown in Fig. 1. To
evaluate the effectiveness of Trans10K, we propose a novel boundary-aware
segmentation method, termed TransLab, which exploits boundary as the clue to
improve segmentation of transparent objects. Extensive experiments and ablation
studies demonstrate the effectiveness of Trans10K and validate the practicality
of learning object boundary in TransLab. For example, TransLab significantly
outperforms 20 recent object segmentation methods based on deep learning,
showing that this task is largely unsolved. We believe that both Trans10K and
TransLab have important contributions to both the academia and industry,
facilitating future researches and applications.
Related papers
- LAC-Net: Linear-Fusion Attention-Guided Convolutional Network for Accurate Robotic Grasping Under the Occlusion [79.22197702626542]
This paper introduces a framework that explores amodal segmentation for robotic grasping in cluttered scenes.
We propose a Linear-fusion Attention-guided Convolutional Network (LAC-Net)
The results on different datasets show that our method achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-08-06T14:50:48Z) - A New Dataset and a Distractor-Aware Architecture for Transparent Object
Tracking [34.08943612955157]
Performance of modern trackers degrades substantially on transparent objects compared to opaque objects.
We propose the first transparent object tracking training dataset Trans2k that consists of over 2k sequences with 104,343 images overall.
We also present a new distractor-aware transparent object tracker (DiTra) that treats localization accuracy and target identification as separate tasks.
arXiv Detail & Related papers (2024-01-08T13:04:28Z) - Transparent Object Tracking with Enhanced Fusion Module [56.403878717170784]
We propose a new tracker architecture that uses our fusion techniques to achieve superior results for transparent object tracking.
Our results and the implementation of code will be made publicly available at https://github.com/kalyan05TOTEM.
arXiv Detail & Related papers (2023-09-13T03:52:09Z) - TransNet: Transparent Object Manipulation Through Category-Level Pose
Estimation [6.844391823478345]
We propose a two-stage pipeline that estimates category-level transparent object pose using localized depth completion and surface normal estimation.
Results show that TransNet achieves improved pose estimation accuracy on transparent objects.
We use TransNet to build an autonomous transparent object manipulation system for robotic pick-and-place and pouring tasks.
arXiv Detail & Related papers (2023-07-23T18:38:42Z) - Trans2k: Unlocking the Power of Deep Models for Transparent Object
Tracking [41.039837388154]
We propose the first transparent object tracking training dataset Trans2k that consists of over 2k sequences with 104,343 images overall.
We quantify domain-specific attributes and render the dataset containing visual attributes and tracking situations not covered in the existing object training datasets.
The dataset and the rendering engine will be publicly released to unlock the power of modern learning-based trackers and foster new designs in transparent object tracking.
arXiv Detail & Related papers (2022-10-07T10:08:13Z) - TransCG: A Large-Scale Real-World Dataset for Transparent Object Depth
Completion and Grasping [46.6058840385155]
We contribute a large-scale real-world dataset for transparent object depth completion.
Our dataset contains 57,715 RGB-D images from 130 different scenes.
We propose an end-to-end depth completion network, which takes the RGB image and the inaccurate depth map as inputs and outputs a refined depth map.
arXiv Detail & Related papers (2022-02-17T06:50:20Z) - Seeing Glass: Joint Point Cloud and Depth Completion for Transparent
Objects [16.714074893209713]
TranspareNet is a joint point cloud and depth completion method.
It can complete the depth of transparent objects in cluttered and complex scenes.
TranspareNet outperforms existing state-of-the-art depth completion methods on multiple datasets.
arXiv Detail & Related papers (2021-09-30T21:09:09Z) - FakeMix Augmentation Improves Transparent Object Detection [24.540569928274984]
We propose a novel content-dependent data augmentation method termed FakeMix to overcome the boundary-related imbalance problem.
We also present AdaptiveASPP, an enhanced version of ASPP, that can capture multi-scale and cross-modality features dynamically.
arXiv Detail & Related papers (2021-03-24T15:51:37Z) - Segmenting Transparent Object in the Wild with Transformer [47.97930429998238]
This work presents a new fine-grained transparent object segmentation dataset, termed Trans10K-v2, extending Trans10K-v1.
It has 11 fine-grained categories of transparent objects, commonly occurring in the human domestic environment.
A novel transformer-based segmentation pipeline termed Trans2Seg is proposed.
arXiv Detail & Related papers (2021-01-21T06:41:00Z) - Refined Plane Segmentation for Cuboid-Shaped Objects by Leveraging Edge
Detection [63.942632088208505]
We propose a post-processing algorithm to align the segmented plane masks with edges detected in the image.
This allows us to increase the accuracy of state-of-the-art approaches, while limiting ourselves to cuboid-shaped objects.
arXiv Detail & Related papers (2020-03-28T18:51:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.