ConstScene: Dataset and Model for Advancing Robust Semantic Segmentation
in Construction Environments
- URL: http://arxiv.org/abs/2312.16516v2
- Date: Fri, 19 Jan 2024 12:29:47 GMT
- Title: ConstScene: Dataset and Model for Advancing Robust Semantic Segmentation
in Construction Environments
- Authors: Maghsood Salimi, Mohammad Loni, Sara Afshar, Antonio Cicchetti, Marjan
Sirjani
- Abstract summary: This paper introduces a new semantic segmentation dataset specifically tailored for construction sites.
The dataset is designed to enhance the training and evaluation of object detection models.
- Score: 1.4070907500169874
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The increasing demand for autonomous machines in construction environments
necessitates the development of robust object detection algorithms that can
perform effectively across various weather and environmental conditions. This
paper introduces a new semantic segmentation dataset specifically tailored for
construction sites, taking into account the diverse challenges posed by adverse
weather and environmental conditions. The dataset is designed to enhance the
training and evaluation of object detection models, fostering their
adaptability and reliability in real-world construction applications. Our
dataset comprises annotated images captured under a wide range of different
weather conditions, including but not limited to sunny days, rainy periods,
foggy atmospheres, and low-light situations. Additionally, environmental
factors such as the existence of dirt/mud on the camera lens are integrated
into the dataset through actual captures and synthetic generation to simulate
the complex conditions prevalent in construction sites. We also generate
synthetic images of the annotations including precise semantic segmentation
masks for various objects commonly found in construction environments, such as
wheel loader machines, personnel, cars, and structural elements. To demonstrate
the dataset's utility, we evaluate state-of-the-art object detection algorithms
on our proposed benchmark. The results highlight the dataset's success in
adversarial training models across diverse conditions, showcasing its efficacy
compared to existing datasets that lack such environmental variability.
Related papers
- Relighting from a Single Image: Datasets and Deep Intrinsic-based Architecture [0.7499722271664147]
Single image scene relighting aims to generate a realistic new version of an input image so that it appears to be illuminated by a new target light condition.
We propose two new datasets: a synthetic dataset with the ground truth of intrinsic components and a real dataset collected under laboratory conditions.
Our method outperforms the state-of-the-art methods in performance, as tested on both existing datasets and our newly developed datasets.
arXiv Detail & Related papers (2024-09-27T14:15:02Z) - BEHAVIOR Vision Suite: Customizable Dataset Generation via Simulation [57.40024206484446]
We introduce the BEHAVIOR Vision Suite (BVS), a set of tools and assets to generate fully customized synthetic data for systematic evaluation of computer vision models.
BVS supports a large number of adjustable parameters at the scene level.
We showcase three example application scenarios.
arXiv Detail & Related papers (2024-05-15T17:57:56Z) - CarPatch: A Synthetic Benchmark for Radiance Field Evaluation on Vehicle
Components [77.33782775860028]
We introduce CarPatch, a novel synthetic benchmark of vehicles.
In addition to a set of images annotated with their intrinsic and extrinsic camera parameters, the corresponding depth maps and semantic segmentation masks have been generated for each view.
Global and part-based metrics have been defined and used to evaluate, compare, and better characterize some state-of-the-art techniques.
arXiv Detail & Related papers (2023-07-24T11:59:07Z) - AVOIDDS: Aircraft Vision-based Intruder Detection Dataset and Simulator [37.579437595742995]
We introduce AVOIDDS, a realistic object detection benchmark for the vision-based aircraft detect-and-avoid problem.
We provide a labeled dataset consisting of 72,000 photorealistic images of intruder aircraft with various lighting conditions.
We also provide an interface that evaluates trained models on slices of this dataset to identify changes in performance with respect to changing environmental conditions.
arXiv Detail & Related papers (2023-06-19T23:58:07Z) - Persistent Homology Meets Object Unity: Object Recognition in Clutter [2.356908851188234]
Recognition of occluded objects in unseen and unstructured indoor environments is a challenging problem for mobile robots.
We propose a new descriptor, TOPS, for point clouds generated from depth images and an accompanying recognition framework, THOR, inspired by human reasoning.
THOR outperforms state-of-the-art methods on both the datasets and achieves substantially higher recognition accuracy for all the scenarios of the UW-IS Occluded dataset.
arXiv Detail & Related papers (2023-05-05T19:42:39Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Scene-Aware
Ambidextrous Bin Picking via Physics-based Metaverse Synthesis [72.85526892440251]
We introduce MetaGraspNet, a large-scale photo-realistic bin picking dataset constructed via physics-based metaverse synthesis.
The proposed dataset contains 217k RGBD images across 82 different article types, with full annotations for object detection, amodal perception, keypoint detection, manipulation order and ambidextrous grasp labels for a parallel-jaw and vacuum gripper.
We also provide a real dataset consisting of over 2.3k fully annotated high-quality RGBD images, divided into 5 levels of difficulties and an unseen object set to evaluate different object and layout properties.
arXiv Detail & Related papers (2022-08-08T08:15:34Z) - A Multi-purpose Real Haze Benchmark with Quantifiable Haze Levels and
Ground Truth [61.90504318229845]
This paper introduces the first paired real image benchmark dataset with hazy and haze-free images, and in-situ haze density measurements.
This dataset was produced in a controlled environment with professional smoke generating machines that covered the entire scene.
A subset of this dataset has been used for the Object Detection in Haze Track of CVPR UG2 2022 challenge.
arXiv Detail & Related papers (2022-06-13T19:14:06Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - SVIRO: Synthetic Vehicle Interior Rear Seat Occupancy Dataset and
Benchmark [11.101588888002045]
We release SVIRO, a synthetic dataset for sceneries in the passenger compartment of ten different vehicles.
We analyze machine learning-based approaches for their generalization capacities and reliability when trained on a limited number of variations.
arXiv Detail & Related papers (2020-01-10T14:44:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.