The Garbage Dataset (GD): A Multi-Class Image Benchmark for Automated Waste Segregation
- URL: http://arxiv.org/abs/2602.10500v1
- Date: Wed, 11 Feb 2026 04:01:12 GMT
- Title: The Garbage Dataset (GD): A Multi-Class Image Benchmark for Automated Waste Segregation
- Authors: Suman Kunwar,
- Abstract summary: The dataset comprises 13,348 labeled images collected through multiple methods, including DWaste mobile app and curated web sources.<n>The dataset was benchmarked using state-of-the-art deep learning models.<n>Experiment results indicate EfficientNetV2S achieved the highest performance with 96.19% accuracy and a 0.96 F1-score, though with a moderate carbon cost.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This study introduces the Garbage Dataset (GD), a publicly available image dataset designed to advance automated waste segregation through machine learning and computer vision. It's a diverse dataset covering 10 common household waste categories: metal, glass, biological, paper, battery, trash, cardboard, shoes, clothes, and plastic. The dataset comprises 13,348 labeled images collected through multiple methods, including DWaste mobile app and curated web sources. Methods included rigorous validation through checksums and outlier detection, analysis of class imbalance and visual separability via PCA/t-SNE, and assessment of background complexity using entropy and saliency measures. The dataset was benchmarked using state-of-the-art deep learning models (EfficientNetV2M, EfficientNetV2S, MobileNet, ResNet50, ResNet101) evaluated on performance metrics and operational carbon emissions. Experiment results indicate EfficientNetV2S achieved the highest performance with 96.19% accuracy and a 0.96 F1-score, though with a moderate carbon cost. Analysis revealed inherent dataset characteristics including class imbalance, a skew toward high-outlier classes (plastic, cardboard, paper), and brightness variations that require consideration. The main conclusion is that GD provides a valuable, real-world benchmark for waste classification research while highlighting important challenges such as class imbalance, background complexity, and environmental trade-offs in model selection that must be addressed for practical deployment. The dataset is publicly released to support further research in environmental sustainability applications.
Related papers
- GlobalWasteData: A Large-Scale, Integrated Dataset for Robust Waste Classification and Environmental Monitoring [5.4998857381465465]
We introduce the GlobalWasteData (GWD) archive, a large scale dataset of 89,807 images across 14 main categories.<n>This GWD archive offers consistent labeling, improved domain diversity, and more balanced class representation.<n>Overall, this dataset offers a strong foundation for Machine Learning (ML) applications in environmental monitoring, recycling automation, and waste identification.
arXiv Detail & Related papers (2026-02-07T09:36:39Z) - Hybrid Synthetic Data Generation with Domain Randomization Enables Zero-Shot Vision-Based Part Inspection Under Extreme Class Imbalance [3.7696918637188817]
Training robust machine learning models requires large volumes of high-quality labeled data.<n>Defective samples are intrinsically rare, leading to severe class imbalance that degrades model performance.<n>Synthetic data generation offers a promising solution by enabling the creation of large, balanced, and fully annotated datasets.
arXiv Detail & Related papers (2025-11-28T05:30:49Z) - Textile Analysis for Recycling Automation using Transfer Learning and Zero-Shot Foundation Models [6.932314083921248]
This paper investigates the use of standard RGB imagery, a cost-effective sensing modality, for key pre-processing tasks in an automated system.<n>We present computer vision components designed for a conveyor belt setup to perform (a) classification of four common textile types and (b) segmentation of non- textile features such as buttons and zippers.
arXiv Detail & Related papers (2025-06-06T22:49:53Z) - Unsupervised Waste Classification By Dual-Encoder Contrastive Learning and Multi-Clustering Voting (DECMCV) [9.828020457690688]
This study presents a novel unsupervised method, Dual-Encoder Contrastive Learning with Multi-Clustering Voting (DECMCV)<n>On a real-world dataset of 4,169 waste images, only 50 labeled samples were needed to accurately label thousands, improving classification accuracy by 29.85% compared to supervised models.
arXiv Detail & Related papers (2025-03-04T03:31:01Z) - Advancing Recycling Efficiency: A Comparative Analysis of Deep Learning Models in Waste Classification [0.0]
The research tackles the pressing issue of waste classification for recycling by analyzing various deep learning models.
The results indicate the method significantly boosts accuracy in complex waste categories.
The research paves the way for future advancements in multi-category waste recycling.
arXiv Detail & Related papers (2024-11-05T03:44:54Z) - Benchmarking Zero-Shot Robustness of Multimodal Foundation Models: A Pilot Study [61.65123150513683]
multimodal foundation models, such as CLIP, produce state-of-the-art zero-shot results.
It is reported that these models close the robustness gap by matching the performance of supervised models trained on ImageNet.
We show that CLIP leads to a significant robustness drop compared to supervised ImageNet models on our benchmark.
arXiv Detail & Related papers (2024-03-15T17:33:49Z) - Innovative Horizons in Aerial Imagery: LSKNet Meets DiffusionDet for
Advanced Object Detection [55.2480439325792]
We present an in-depth evaluation of an object detection model that integrates the LSKNet backbone with the DiffusionDet head.
The proposed model achieves a mean average precision (MAP) of approximately 45.7%, which is a significant improvement.
This advancement underscores the effectiveness of the proposed modifications and sets a new benchmark in aerial image analysis.
arXiv Detail & Related papers (2023-11-21T19:49:13Z) - Intrinsic Self-Supervision for Data Quality Audits [35.69673085324971]
Benchmark datasets in computer vision often contain off-topic images, near duplicates, and label errors.
In this paper, we revisit the task of data cleaning and formalize it as either a ranking problem, or a scoring problem.
We find that a specific combination of context-aware self-supervised representation learning and distance-based indicators is effective in finding issues without annotation biases.
arXiv Detail & Related papers (2023-05-26T15:57:04Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Scene-Aware
Ambidextrous Bin Picking via Physics-based Metaverse Synthesis [72.85526892440251]
We introduce MetaGraspNet, a large-scale photo-realistic bin picking dataset constructed via physics-based metaverse synthesis.
The proposed dataset contains 217k RGBD images across 82 different article types, with full annotations for object detection, amodal perception, keypoint detection, manipulation order and ambidextrous grasp labels for a parallel-jaw and vacuum gripper.
We also provide a real dataset consisting of over 2.3k fully annotated high-quality RGBD images, divided into 5 levels of difficulties and an unseen object set to evaluate different object and layout properties.
arXiv Detail & Related papers (2022-08-08T08:15:34Z) - ZeroWaste Dataset: Towards Automated Waste Recycling [51.053682077915546]
We present the first in-the-wild industrial-grade waste detection and segmentation dataset, ZeroWaste.
This dataset contains over1800fully segmented video frames collected from a real waste sorting plant.
We show that state-of-the-art segmentation methods struggle to correctly detect and classify target objects.
arXiv Detail & Related papers (2021-06-04T22:17:09Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - GraspNet: A Large-Scale Clustered and Densely Annotated Dataset for
Object Grasping [49.777649953381676]
We contribute a large-scale grasp pose detection dataset with a unified evaluation system.
Our dataset contains 87,040 RGBD images with over 370 million grasp poses.
arXiv Detail & Related papers (2019-12-31T18:15:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.