CAD2Real: Deep learning with domain randomization of CAD data for 3D
pose estimation of electronic control unit housings
- URL: http://arxiv.org/abs/2009.12312v1
- Date: Fri, 25 Sep 2020 16:08:16 GMT
- Title: CAD2Real: Deep learning with domain randomization of CAD data for 3D
pose estimation of electronic control unit housings
- Authors: Simon Baeuerle, Jonas Barth, Elton Renato Tavares de Menezes, Andreas
Steimer, Ralf Mikut
- Abstract summary: We train state-of-the-art artificial neural networks (ANNs) on purely synthetic training data.
By randomizing parameters during rendering of training images, we enable inference on RGB images of a real sample part.
- Score: 0.20999222360659608
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Electronic control units (ECUs) are essential for many automobile components,
e.g. engine, anti-lock braking system (ABS), steering and airbags. For some
products, the 3D pose of each single ECU needs to be determined during series
production. Deep learning approaches can not easily be applied to this problem,
because labeled training data is not available in sufficient numbers. Thus, we
train state-of-the-art artificial neural networks (ANNs) on purely synthetic
training data, which is automatically created from a single CAD file. By
randomizing parameters during rendering of training images, we enable inference
on RGB images of a real sample part. In contrast to classic image processing
approaches, this data-driven approach poses only few requirements regarding the
measurement setup and transfers to related use cases with little development
effort.
Related papers
- Automatic Reverse Engineering: Creating computer-aided design (CAD)
models from multi-view images [0.0]
We present a novel network for an automated reverse engineering task.
A proof-of-concept is demonstrated by successfully reconstructing a number of valid CAD models.
It is shown that some of the capabilities of our network can be transferred to this domain.
arXiv Detail & Related papers (2023-09-23T06:42:09Z) - 3D Data Augmentation for Driving Scenes on Camera [50.41413053812315]
We propose a 3D data augmentation approach termed Drive-3DAug, aiming at augmenting the driving scenes on camera in the 3D space.
We first utilize Neural Radiance Field (NeRF) to reconstruct the 3D models of background and foreground objects.
Then, augmented driving scenes can be obtained by placing the 3D objects with adapted location and orientation at the pre-defined valid region of backgrounds.
arXiv Detail & Related papers (2023-03-18T05:51:05Z) - Generalized Few-Shot 3D Object Detection of LiDAR Point Cloud for
Autonomous Driving [91.39625612027386]
We propose a novel task, called generalized few-shot 3D object detection, where we have a large amount of training data for common (base) objects, but only a few data for rare (novel) classes.
Specifically, we analyze in-depth differences between images and point clouds, and then present a practical principle for the few-shot setting in the 3D LiDAR dataset.
To solve this task, we propose an incremental fine-tuning method to extend existing 3D detection models to recognize both common and rare objects.
arXiv Detail & Related papers (2023-02-08T07:11:36Z) - HUM3DIL: Semi-supervised Multi-modal 3D Human Pose Estimation for
Autonomous Driving [95.42203932627102]
3D human pose estimation is an emerging technology, which can enable the autonomous vehicle to perceive and understand the subtle and complex behaviors of pedestrians.
Our method efficiently makes use of these complementary signals, in a semi-supervised fashion and outperforms existing methods with a large margin.
Specifically, we embed LiDAR points into pixel-aligned multi-modal features, which we pass through a sequence of Transformer refinement stages.
arXiv Detail & Related papers (2022-12-15T11:15:14Z) - A Novel Neural Network Training Method for Autonomous Driving Using
Semi-Pseudo-Labels and 3D Data Augmentations [0.0]
Training neural networks to perform 3D object detection for autonomous driving requires a large amount of diverse annotated data.
We have designed a convolutional neural network for 3D object detection which can significantly increase the detection range.
arXiv Detail & Related papers (2022-07-20T13:04:08Z) - Simplified Learning of CAD Features Leveraging a Deep Residual
Autoencoder [3.567248644184455]
In computer vision, deep residual neural networks like EfficientNet have set new standards in terms of robustness and accuracy.
One key problem underlying the training of deep neural networks is the immanent lack of a sufficient amount of training data.
We present a deep residual 3D autoencoder based on the EfficientNet architecture, intended for transfer learning tasks related to 3D CAD model assessment.
arXiv Detail & Related papers (2022-02-21T10:27:55Z) - 3D-VField: Learning to Adversarially Deform Point Clouds for Robust 3D
Object Detection [111.32054128362427]
In safety-critical settings, robustness on out-of-distribution and long-tail samples is fundamental to circumvent dangerous issues.
We substantially improve the generalization of 3D object detectors to out-of-domain data by taking into account deformed point clouds during training.
We propose and share open source CrashD: a synthetic dataset of realistic damaged and rare cars.
arXiv Detail & Related papers (2021-12-09T08:50:54Z) - CAN3D: Fast 3D Medical Image Segmentation via Compact Context
Aggregation [6.188937569449575]
We present a compact convolutional neural network with a shallow memory footprint to efficiently reduce the number of model parameters required for state-of-art performance.
The proposed network can maintain data integrity by directly processing large full-size 3D input volumes with no patches required.
arXiv Detail & Related papers (2021-09-12T06:27:59Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - PerMO: Perceiving More at Once from a Single Image for Autonomous
Driving [76.35684439949094]
We present a novel approach to detect, segment, and reconstruct complete textured 3D models of vehicles from a single image.
Our approach combines the strengths of deep learning and the elegance of traditional techniques.
We have integrated these algorithms with an autonomous driving system.
arXiv Detail & Related papers (2020-07-16T05:02:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.