Sim2Real Instance-Level Style Transfer for 6D Pose Estimation
- URL: http://arxiv.org/abs/2203.02069v1
- Date: Thu, 3 Mar 2022 23:46:47 GMT
- Title: Sim2Real Instance-Level Style Transfer for 6D Pose Estimation
- Authors: Takuya Ikeda, Suomi Tanishige, Ayako Amma, Michael Sudano, Herv\'e
Audren, Koichi Nishiwaki
- Abstract summary: We introduce a simulation to reality (sim2real) instance-level style transfer for 6D pose estimation network training.
Our approach transfers the style of target objects individually, from synthetic to real, without human intervention.
- Score: 0.4893345190925177
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, synthetic data has been widely used in the training of 6D
pose estimation networks, in part because it automatically provides perfect
annotation at low cost. However, there are still non-trivial domain gaps, such
as differences in textures/materials, between synthetic and real data. These
gaps have a measurable impact on performance. To solve this problem, we
introduce a simulation to reality (sim2real) instance-level style transfer for
6D pose estimation network training. Our approach transfers the style of target
objects individually, from synthetic to real, without human intervention. This
improves the quality of synthetic data for training pose estimation networks.
We also propose a complete pipeline from data collection to the training of a
pose estimation network and conduct extensive evaluation on a real-world
robotic platform. Our evaluation shows significant improvement achieved by our
method in both pose estimation performance and the realism of images adapted by
the style transfer.
Related papers
- Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - Exploring the Effectiveness of Dataset Synthesis: An application of
Apple Detection in Orchards [68.95806641664713]
We explore the usability of Stable Diffusion 2.1-base for generating synthetic datasets of apple trees for object detection.
We train a YOLOv5m object detection model to predict apples in a real-world apple detection dataset.
Results demonstrate that the model trained on generated data is slightly underperforming compared to a baseline model trained on real-world images.
arXiv Detail & Related papers (2023-06-20T09:46:01Z) - Robust Category-Level 3D Pose Estimation from Synthetic Data [17.247607850702558]
We introduce SyntheticP3D, a new synthetic dataset for object pose estimation generated from CAD models.
We propose a novel approach (CC3D) for training neural mesh models that perform pose estimation via inverse rendering.
arXiv Detail & Related papers (2023-05-25T14:56:03Z) - TexPose: Neural Texture Learning for Self-Supervised 6D Object Pose
Estimation [55.94900327396771]
We introduce neural texture learning for 6D object pose estimation from synthetic data.
We learn to predict realistic texture of objects from real image collections.
We learn pose estimation from pixel-perfect synthetic data.
arXiv Detail & Related papers (2022-12-25T13:36:32Z) - Estimating the Pose of a Euro Pallet with an RGB Camera based on
Synthetic Training Data [0.0]
In this work, the generation of synthetic training data for 6D pose estimation of pallets is presented.
The data is then used to train the Deep Object Pose Estimation (DOPE) algorithm.
The experimental validation of the algorithm proves that the 6D pose estimation of a standardized Euro pallet with a Red-Green-Blue (RGB) camera is feasible.
arXiv Detail & Related papers (2022-10-12T08:13:11Z) - Sim-to-Real 6D Object Pose Estimation via Iterative Self-training for
Robotic Bin-picking [98.5984733963713]
We propose an iterative self-training framework for sim-to-real 6D object pose estimation to facilitate cost-effective robotic grasping.
We establish a photo-realistic simulator to synthesize abundant virtual data, and use this to train an initial pose estimation network.
This network then takes the role of a teacher model, which generates pose predictions for unlabeled real data.
arXiv Detail & Related papers (2022-04-14T15:54:01Z) - Occlusion-Aware Self-Supervised Monocular 6D Object Pose Estimation [88.8963330073454]
We propose a novel monocular 6D pose estimation approach by means of self-supervised learning.
We leverage current trends in noisy student training and differentiable rendering to further self-supervise the model.
Our proposed self-supervision outperforms all other methods relying on synthetic data.
arXiv Detail & Related papers (2022-03-19T15:12:06Z) - Towards Scale Consistent Monocular Visual Odometry by Learning from the
Virtual World [83.36195426897768]
We propose VRVO, a novel framework for retrieving the absolute scale from virtual data.
We first train a scale-aware disparity network using both monocular real images and stereo virtual data.
The resulting scale-consistent disparities are then integrated with a direct VO system.
arXiv Detail & Related papers (2022-03-11T01:51:54Z) - Generating Annotated Training Data for 6D Object Pose Estimation in
Operational Environments with Minimal User Interaction [1.0044401320520304]
We present a proof of concept for a novel approach of autonomously generating annotated training data for 6D object pose estimation.
This approach is designed for learning new objects in operational environments while requiring little interaction and no expertise on the part of the user.
arXiv Detail & Related papers (2021-03-17T14:46:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.