Panoptic Segmentation using Synthetic and Real Data
- URL: http://arxiv.org/abs/2204.07069v1
- Date: Thu, 14 Apr 2022 16:01:48 GMT
- Title: Panoptic Segmentation using Synthetic and Real Data
- Authors: Camillo Quattrocchi, Daniele Di Mauro, Antonino Furnari, Giovanni
Maria Farinella
- Abstract summary: We propose a pipeline which allows to generate synthetic images from 3D models of real environments and real objects.
Experiments show that the use of synthetic images allows to drastically reduce the number of real images needed to obtain reasonable panoptic segmentation performance.
- Score: 21.226582774172062
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Being able to understand the relations between the user and the surrounding
environment is instrumental to assist users in a worksite. For instance,
understanding which objects a user is interacting with from images and video
collected through a wearable device can be useful to inform the worker on the
usage of specific objects in order to improve productivity and prevent
accidents. Despite modern vision systems can rely on advanced algorithms for
object detection, semantic and panoptic segmentation, these methods still
require large quantities of domain-specific labeled data, which can be
difficult to obtain in industrial scenarios. Motivated by this observation, we
propose a pipeline which allows to generate synthetic images from 3D models of
real environments and real objects. The generated images are automatically
labeled and hence effortless to obtain. Exploiting the proposed pipeline, we
generate a dataset comprising synthetic images automatically labeled for
panoptic segmentation. This set is complemented by a small number of manually
labeled real images for fine-tuning. Experiments show that the use of synthetic
images allows to drastically reduce the number of real images needed to obtain
reasonable panoptic segmentation performance.
Related papers
- Visual Car Brand Classification by Implementing a Synthetic Image Dataset Creation Pipeline [3.524869467682149]
We propose an automatic pipeline for generating synthetic image datasets using Stable Diffusion.
We leverage YOLOv8 for automatic bounding box detection and quality assessment of synthesized images.
arXiv Detail & Related papers (2024-06-03T07:44:08Z) - Investigation of the Impact of Synthetic Training Data in the Industrial
Application of Terminal Strip Object Detection [4.327763441385371]
In this paper, we investigate the sim-to-real generalization performance of standard object detectors on the complex industrial application of terminal strip object detection.
We manually annotated 300 real images of terminal strips for the evaluation. The results show the cruciality of the objects of interest to have the same scale in either domain.
arXiv Detail & Related papers (2024-03-06T18:33:27Z) - Perceptual Artifacts Localization for Image Synthesis Tasks [59.638307505334076]
We introduce a novel dataset comprising 10,168 generated images, each annotated with per-pixel perceptual artifact labels.
A segmentation model, trained on our proposed dataset, effectively localizes artifacts across a range of tasks.
We propose an innovative zoom-in inpainting pipeline that seamlessly rectifies perceptual artifacts in the generated images.
arXiv Detail & Related papers (2023-10-09T10:22:08Z) - Structure-Guided Image Completion with Image-level and Object-level Semantic Discriminators [97.12135238534628]
We propose a learning paradigm that consists of semantic discriminators and object-level discriminators for improving the generation of complex semantics and objects.
Specifically, the semantic discriminators leverage pretrained visual features to improve the realism of the generated visual concepts.
Our proposed scheme significantly improves the generation quality and achieves state-of-the-art results on various tasks.
arXiv Detail & Related papers (2022-12-13T01:36:56Z) - Label-Free Synthetic Pretraining of Object Detectors [67.17371526567325]
We propose a new approach, Synthetic optimized layout with Instance Detection (SOLID), to pretrain object detectors with synthetic images.
Our "SOLID" approach consists of two main components: (1) generating synthetic images using a collection of unlabelled 3D models with optimized scene arrangement; (2) pretraining an object detector on "instance detection" task.
Our approach does not need any semantic labels for pretraining and allows the use of arbitrary, diverse 3D models.
arXiv Detail & Related papers (2022-08-08T16:55:17Z) - RICE: Refining Instance Masks in Cluttered Environments with Graph
Neural Networks [53.15260967235835]
We propose a novel framework that refines the output of such methods by utilizing a graph-based representation of instance masks.
We train deep networks capable of sampling smart perturbations to the segmentations, and a graph neural network, which can encode relations between objects, to evaluate the segmentations.
We demonstrate an application that uses uncertainty estimates generated by our method to guide a manipulator, leading to efficient understanding of cluttered scenes.
arXiv Detail & Related papers (2021-06-29T20:29:29Z) - Image Compositing for Segmentation of Surgical Tools without Manual
Annotations [10.05087029666444]
We propose to automate the creation of a realistic training dataset by exploiting techniques stemming from special effects.
Foreground data is captured by placing sample surgical instruments over a chroma key.
Background data is captured by collecting videos that do not contain instruments.
We show that by training a vanilla U-Net on semi-synthetic data only and applying a simple post-processing, we are able to match the results of the same network trained on a publicly available manually labeled real dataset.
arXiv Detail & Related papers (2021-02-18T18:14:43Z) - Detection and Segmentation of Custom Objects using High Distraction
Photorealistic Synthetic Data [0.5076419064097732]
We show a straightforward and useful methodology for performing instance segmentation using synthetic data.
The goal is to achieve high performance on manually-gathered and annotated real-world data of custom objects.
This white-paper provides strong evidence that photorealistic simulated data can be used in practical real world applications.
arXiv Detail & Related papers (2020-07-28T16:33:42Z) - OpenRooms: An End-to-End Open Framework for Photorealistic Indoor Scene
Datasets [103.54691385842314]
We propose a novel framework for creating large-scale photorealistic datasets of indoor scenes.
Our goal is to make the dataset creation process widely accessible.
This enables important applications in inverse rendering, scene understanding and robotics.
arXiv Detail & Related papers (2020-07-25T06:48:47Z) - Intrinsic Autoencoders for Joint Neural Rendering and Intrinsic Image
Decomposition [67.9464567157846]
We propose an autoencoder for joint generation of realistic images from synthetic 3D models while simultaneously decomposing real images into their intrinsic shape and appearance properties.
Our experiments confirm that a joint treatment of rendering and decomposition is indeed beneficial and that our approach outperforms state-of-the-art image-to-image translation baselines both qualitatively and quantitatively.
arXiv Detail & Related papers (2020-06-29T12:53:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.