Leveraging Multi-view Data for Improved Detection Performance: An
Industrial Use Case
- URL: http://arxiv.org/abs/2304.08111v1
- Date: Mon, 17 Apr 2023 09:41:37 GMT
- Title: Leveraging Multi-view Data for Improved Detection Performance: An
Industrial Use Case
- Authors: Faranak Shamsafar, Sunil Jaiswal, Benjamin Kelkel, Kireeti Bodduna,
Klaus Illgner-Fehns
- Abstract summary: We present a multi-view object detection framework that offers a fast and precise solution.
We introduce a novel multi-view dataset with semi-automatic ground-truth data, which results in significant labeling resource savings.
Our experiments demonstrate a 15% improvement in mAP for detecting components that range in size from 0.5 to 27.0 mm.
- Score: 0.5249805590164901
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Printed circuit boards (PCBs) are essential components of electronic devices,
and ensuring their quality is crucial in their production. However, the vast
variety of components and PCBs manufactured by different companies makes it
challenging to adapt to production lines with speed demands. To address this
challenge, we present a multi-view object detection framework that offers a
fast and precise solution. We introduce a novel multi-view dataset with
semi-automatic ground-truth data, which results in significant labeling
resource savings. Labeling PCB boards for object detection is a challenging
task due to the high density of components and the small size of the objects,
which makes it difficult to identify and label them accurately. By training an
object detector model with multi-view data, we achieve improved performance
over single-view images. To further enhance the accuracy, we develop a
multi-view inference method that aggregates results from different viewpoints.
Our experiments demonstrate a 15% improvement in mAP for detecting components
that range in size from 0.5 to 27.0 mm.
Related papers
- Boosting 3D Object Detection with Semantic-Aware Multi-Branch Framework [44.44329455757931]
In autonomous driving, LiDAR sensors are vital for acquiring 3D point clouds, providing reliable geometric information.
Traditional sampling methods of preprocessing often ignore semantic features, leading to detail loss and ground point interference.
We propose a multi-branch two-stage 3D object detection framework using a Semantic-aware Multi-branch Sampling (SMS) module and multi-view constraints.
arXiv Detail & Related papers (2024-07-08T09:25:45Z) - SOOD++: Leveraging Unlabeled Data to Boost Oriented Object Detection [59.868772767818975]
We propose a simple yet effective Semi-supervised Oriented Object Detection method termed SOOD++.
Specifically, we observe that objects from aerial images are usually arbitrary orientations, small scales, and aggregation.
Extensive experiments conducted on various multi-oriented object datasets under various labeled settings demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2024-07-01T07:03:51Z) - Better Sampling, towards Better End-to-end Small Object Detection [7.7473020808686694]
Small object detection remains unsatisfactory due to limited characteristics and high density and mutual overlap.
We propose methods enhancing sampling within an end-to-end framework.
Our model demonstrates a significant enhancement, achieving a 2.9% increase in average precision (AP) over the state-of-the-art (SOTA) on the VisDrone dataset.
arXiv Detail & Related papers (2024-05-17T04:37:44Z) - Efficient Meta-Learning Enabled Lightweight Multiscale Few-Shot Object Detection in Remote Sensing Images [15.12889076965307]
YOLOv7 one-stage detector is subjected to a novel meta-learning training framework.
This transformation allows the detector to adeptly address FSOD tasks while capitalizing on its inherent advantage of lightweight.
To validate the effectiveness of our proposed detector, we conducted performance comparisons with current state-of-the-art detectors.
arXiv Detail & Related papers (2024-04-29T04:56:52Z) - Towards Automatic Power Battery Detection: New Challenge, Benchmark
Dataset and Baseline [70.30473488226093]
We conduct a comprehensive study on a new task named power battery detection (PBD)
It aims to localize the dense cathode and anode plates endpoints from X-ray images to evaluate the quality of power batteries.
We propose a novel segmentation-based solution for PBD, termed multi-dimensional collaborative network (MDCNet)
arXiv Detail & Related papers (2023-12-05T06:18:38Z) - SimPLR: A Simple and Plain Transformer for Scaling-Efficient Object Detection and Segmentation [49.65221743520028]
We show that a transformer-based detector with scale-aware attention enables the plain detector SimPLR' whose backbone and detection head are both non-hierarchical and operate on single-scale features.
Compared to the multi-scale and single-scale state-of-the-art, our model scales much better with bigger capacity (self-supervised) models and more pre-training data.
arXiv Detail & Related papers (2023-10-09T17:59:26Z) - A novel Multi to Single Module for small object detection [2.920753968664803]
The performance of small object detectors is often compromised by a lack of pixels and less significant features.
This paper proposes a novel the Multi to Single Module (M2S), which enhances a specific layer through improving feature extraction and refining features.
The effectiveness of the proposed method is evaluated on two datasets, VisDrone2021-DET and SeaDronesSeeV2.
arXiv Detail & Related papers (2023-03-27T08:17:22Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Scene-Aware
Ambidextrous Bin Picking via Physics-based Metaverse Synthesis [72.85526892440251]
We introduce MetaGraspNet, a large-scale photo-realistic bin picking dataset constructed via physics-based metaverse synthesis.
The proposed dataset contains 217k RGBD images across 82 different article types, with full annotations for object detection, amodal perception, keypoint detection, manipulation order and ambidextrous grasp labels for a parallel-jaw and vacuum gripper.
We also provide a real dataset consisting of over 2.3k fully annotated high-quality RGBD images, divided into 5 levels of difficulties and an unseen object set to evaluate different object and layout properties.
arXiv Detail & Related papers (2022-08-08T08:15:34Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Vision-driven Robotic
Grasping via Physics-based Metaverse Synthesis [78.26022688167133]
We present a large-scale benchmark dataset for vision-driven robotic grasping via physics-based metaverse synthesis.
The proposed dataset contains 100,000 images and 25 different object types.
We also propose a new layout-weighted performance metric alongside the dataset for evaluating object detection and segmentation performance.
arXiv Detail & Related papers (2021-12-29T17:23:24Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.