VISION Datasets: A Benchmark for Vision-based InduStrial InspectiON
- URL: http://arxiv.org/abs/2306.07890v2
- Date: Sun, 18 Jun 2023 01:11:04 GMT
- Title: VISION Datasets: A Benchmark for Vision-based InduStrial InspectiON
- Authors: Haoping Bai, Shancong Mou, Tatiana Likhomanenko, Ramazan Gokberk
Cinbis, Oncel Tuzel, Ping Huang, Jiulong Shan, Jianjun Shi, Meng Cao
- Abstract summary: VISION datasets are diverse collection of 14 industrial inspection datasets.
With a total of 18k images encompassing 44 defect types, VISION strives to mirror a wide range of real-world production scenarios.
- Score: 28.511625423590605
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Despite progress in vision-based inspection algorithms, real-world industrial
challenges -- specifically in data availability, quality, and complex
production requirements -- often remain under-addressed. We introduce the
VISION Datasets, a diverse collection of 14 industrial inspection datasets,
uniquely poised to meet these challenges. Unlike previous datasets, VISION
brings versatility to defect detection, offering annotation masks across all
splits and catering to various detection methodologies. Our datasets also
feature instance-segmentation annotation, enabling precise defect
identification. With a total of 18k images encompassing 44 defect types, VISION
strives to mirror a wide range of real-world production scenarios. By
supporting two ongoing challenge competitions on the VISION Datasets, we hope
to foster further advancements in vision-based industrial inspection.
Related papers
- A PRISMA Driven Systematic Review of Publicly Available Datasets for Benchmark and Model Developments for Industrial Defect Detection [0.0]
A critical barrier to progress is the scarcity of comprehensive datasets featuring annotated defects.
This systematic review, spanning from 2015 to 2023, identifies 15 publicly available datasets.
The goal of this systematic review is to consolidate these datasets in a single location, providing researchers with a comprehensive reference.
arXiv Detail & Related papers (2024-06-11T20:14:59Z) - Video Anomaly Detection in 10 Years: A Survey and Outlook [10.143205531474907]
Video anomaly detection (VAD) holds immense importance across diverse domains such as surveillance, healthcare, and environmental monitoring.
This survey explores deep learning-based VAD, expanding beyond traditional supervised training paradigms to encompass emerging weakly supervised, self-supervised, and unsupervised approaches.
arXiv Detail & Related papers (2024-05-29T17:56:31Z) - BEHAVIOR Vision Suite: Customizable Dataset Generation via Simulation [57.40024206484446]
We introduce the BEHAVIOR Vision Suite (BVS), a set of tools and assets to generate fully customized synthetic data for systematic evaluation of computer vision models.
BVS supports a large number of adjustable parameters at the scene level.
We showcase three example application scenarios.
arXiv Detail & Related papers (2024-05-15T17:57:56Z) - Advancing Video Anomaly Detection: A Concise Review and a New Dataset [8.822253683273841]
Video Anomaly Detection (VAD) finds widespread applications in security surveillance, traffic monitoring, industrial monitoring, and healthcare.
Despite extensive research efforts, there remains a lack of concise reviews that provide insightful guidance for researchers.
We present such a review, examining models and datasets from various perspectives.
arXiv Detail & Related papers (2024-02-07T13:54:56Z) - A Survey on Visual Anomaly Detection: Challenge, Approach, and Prospect [29.006716009327032]
Visual Anomaly Detection (VAD) endeavors to pinpoint deviations from the concept of normality in visual data, widely applied across diverse domains, e.g., industrial defect inspection, and medical lesion detection.
This survey comprehensively examines recent advancements in VAD by identifying three primary challenges: 1) scarcity of training data, 2) diversity of visual modalities, and 3) complexity of hierarchical anomalies.
arXiv Detail & Related papers (2024-01-29T18:41:21Z) - InsPLAD: A Dataset and Benchmark for Power Line Asset Inspection in UAV
Images [1.8524180288472398]
This paper introduces InsPLAD, a Power Line Asset Inspection dataset and Benchmark containing 10,607 high-resolution Unmanned Aerial Vehicles colour images.
The dataset contains seventeen unique power line assets captured from real-world operating power lines.
We thoroughly evaluate state-of-the-art and popular methods for three image-level computer vision tasks covered by InsPLAD: object detection, through the AP metric; defect classification, through Balanced Accuracy; and anomaly detection, through the AUROC metric.
arXiv Detail & Related papers (2023-11-02T22:06:23Z) - Wild Face Anti-Spoofing Challenge 2023: Benchmark and Results [73.98594459933008]
Face anti-spoofing (FAS) is an essential mechanism for safeguarding the integrity of automated face recognition systems.
This limitation can be attributed to the scarcity and lack of diversity in publicly available FAS datasets.
We introduce the Wild Face Anti-Spoofing dataset, a large-scale, diverse FAS dataset collected in unconstrained settings.
arXiv Detail & Related papers (2023-04-12T10:29:42Z) - Universal Object Detection with Large Vision Model [79.06618136217142]
This study focuses on the large-scale, multi-domain universal object detection problem.
To address these challenges, we introduce our approach to label handling, hierarchy-aware design, and resource-efficient model training.
Our method has demonstrated remarkable performance, securing a prestigious second-place ranking in the object detection track of the Robust Vision Challenge 2022.
arXiv Detail & Related papers (2022-12-19T12:40:13Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Scene-Aware
Ambidextrous Bin Picking via Physics-based Metaverse Synthesis [72.85526892440251]
We introduce MetaGraspNet, a large-scale photo-realistic bin picking dataset constructed via physics-based metaverse synthesis.
The proposed dataset contains 217k RGBD images across 82 different article types, with full annotations for object detection, amodal perception, keypoint detection, manipulation order and ambidextrous grasp labels for a parallel-jaw and vacuum gripper.
We also provide a real dataset consisting of over 2.3k fully annotated high-quality RGBD images, divided into 5 levels of difficulties and an unseen object set to evaluate different object and layout properties.
arXiv Detail & Related papers (2022-08-08T08:15:34Z) - Uncertainty-Aware Multi-View Representation Learning [53.06828186507994]
We devise a novel unsupervised multi-view learning approach, termed as Dynamic Uncertainty-Aware Networks (DUA-Nets)
Guided by the uncertainty of data estimated from the generation perspective, intrinsic information from multiple views is integrated to obtain noise-free representations.
Our model achieves superior performance in extensive experiments and shows the robustness to noisy data.
arXiv Detail & Related papers (2022-01-15T07:16:20Z) - A Variational Information Bottleneck Approach to Multi-Omics Data
Integration [98.6475134630792]
We propose a deep variational information bottleneck (IB) approach for incomplete multi-view observations.
Our method applies the IB framework on marginal and joint representations of the observed views to focus on intra-view and inter-view interactions that are relevant for the target.
Experiments on real-world datasets show that our method consistently achieves gain from data integration and outperforms state-of-the-art benchmarks.
arXiv Detail & Related papers (2021-02-05T06:05:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.