SpectralWaste Dataset: Multimodal Data for Waste Sorting Automation
- URL: http://arxiv.org/abs/2403.18033v1
- Date: Tue, 26 Mar 2024 18:39:38 GMT
- Title: SpectralWaste Dataset: Multimodal Data for Waste Sorting Automation
- Authors: Sara Casao, Fernando Peña, Alberto Sabater, Rosa Castillón, Darío Suárez, Eduardo Montijano, Ana C. Murillo,
- Abstract summary: We present SpectralWaste, the first dataset collected from an operational plastic waste sorting facility.
This dataset contains labels for several categories of objects that commonly appear in sorting plants.
We propose a pipeline employing different object segmentation architectures and evaluate the alternatives on our dataset.
- Score: 46.178512739789426
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The increase in non-biodegradable waste is a worldwide concern. Recycling facilities play a crucial role, but their automation is hindered by the complex characteristics of waste recycling lines like clutter or object deformation. In addition, the lack of publicly available labeled data for these environments makes developing robust perception systems challenging. Our work explores the benefits of multimodal perception for object segmentation in real waste management scenarios. First, we present SpectralWaste, the first dataset collected from an operational plastic waste sorting facility that provides synchronized hyperspectral and conventional RGB images. This dataset contains labels for several categories of objects that commonly appear in sorting plants and need to be detected and separated from the main trash flow for several reasons, such as security in the management line or reuse. Additionally, we propose a pipeline employing different object segmentation architectures and evaluate the alternatives on our dataset, conducting an extensive analysis for both multimodal and unimodal alternatives. Our evaluation pays special attention to efficiency and suitability for real-time processing and demonstrates how HSI can bring a boost to RGB-only perception in these realistic industrial settings without much computational overhead.
Related papers
- COSNet: A Novel Semantic Segmentation Network using Enhanced Boundaries in Cluttered Scenes [9.970265640589966]
We introduce an efficacious segmentation network, named COSNet, that uses boundary cues along with multi-contextual information to accurately segment the objects in cluttered scenes.
Our COSNet achieves a significant gain of 1.8% on ZeroWaste-f and 2.1% on SpectralWaste datasets respectively in terms of mIoU metric.
arXiv Detail & Related papers (2024-10-31T17:03:38Z) - WasteGAN: Data Augmentation for Robotic Waste Sorting through Generative Adversarial Networks [7.775894876221921]
We introduce a data augmentation method based on a novel GAN architecture called wasteGAN.
The proposed method allows to increase the performance of semantic segmentation models, starting from a very limited bunch of labeled examples.
We then leverage the higher-quality segmentation masks predicted from models trained on the wasteGAN synthetic data to compute semantic-aware grasp poses.
arXiv Detail & Related papers (2024-09-25T15:04:21Z) - VisDA 2022 Challenge: Domain Adaptation for Industrial Waste Sorting [61.52419223232737]
In industrial waste sorting, one of the biggest challenges is the extreme diversity of the input stream.
We present the VisDA 2022 Challenge on Domain Adaptation for Industrial Waste Sorting.
arXiv Detail & Related papers (2023-03-26T21:38:38Z) - MMRNet: Improving Reliability for Multimodal Object Detection and
Segmentation for Bin Picking via Multimodal Redundancy [68.7563053122698]
We propose a reliable object detection and segmentation system with MultiModal Redundancy (MMRNet)
This is the first system that introduces the concept of multimodal redundancy to address sensor failure issues during deployment.
We present a new label-free multi-modal consistency (MC) score that utilizes the output from all modalities to measure the overall system output reliability and uncertainty.
arXiv Detail & Related papers (2022-10-19T19:15:07Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Scene-Aware
Ambidextrous Bin Picking via Physics-based Metaverse Synthesis [72.85526892440251]
We introduce MetaGraspNet, a large-scale photo-realistic bin picking dataset constructed via physics-based metaverse synthesis.
The proposed dataset contains 217k RGBD images across 82 different article types, with full annotations for object detection, amodal perception, keypoint detection, manipulation order and ambidextrous grasp labels for a parallel-jaw and vacuum gripper.
We also provide a real dataset consisting of over 2.3k fully annotated high-quality RGBD images, divided into 5 levels of difficulties and an unseen object set to evaluate different object and layout properties.
arXiv Detail & Related papers (2022-08-08T08:15:34Z) - ZeroWaste Dataset: Towards Automated Waste Recycling [51.053682077915546]
We present the first in-the-wild industrial-grade waste detection and segmentation dataset, ZeroWaste.
This dataset contains over1800fully segmented video frames collected from a real waste sorting plant.
We show that state-of-the-art segmentation methods struggle to correctly detect and classify target objects.
arXiv Detail & Related papers (2021-06-04T22:17:09Z) - Segmenting Unseen Industrial Components in a Heavy Clutter Using RGB-D
Fusion and Synthetic Data [0.4724825031148411]
Industrial components are texture-less, reflective, and often found in cluttered and unstructured environments.
We present a synthetic data generation pipeline that randomizes textures via domain randomization to focus on the shape information.
We also propose an RGB-D Fusion Mask R-CNN with a confidence map estimator, which exploits reliable depth information in multiple feature levels.
arXiv Detail & Related papers (2020-02-10T02:33:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.