ViViD++: Vision for Visibility Dataset
- URL: http://arxiv.org/abs/2204.06183v2
- Date: Thu, 14 Apr 2022 00:38:12 GMT
- Title: ViViD++: Vision for Visibility Dataset
- Authors: Alex Junho Lee, Younggun Cho, Young-sik Shin, Ayoung Kim, Hyun Myung
- Abstract summary: We present a dataset capturing diverse visual data formats that target varying luminance conditions.
Despite the alternative sensors' potential, there still are few datasets with alternative vision sensors.
We provide these measurements along with inertial sensors and ground-truth for developing robust visual SLAM under poor illumination.
- Score: 14.839450468199457
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we present a dataset capturing diverse visual data formats
that target varying luminance conditions. While RGB cameras provide nourishing
and intuitive information, changes in lighting conditions potentially result in
catastrophic failure for robotic applications based on vision sensors.
Approaches overcoming illumination problems have included developing more
robust algorithms or other types of visual sensors, such as thermal and event
cameras. Despite the alternative sensors' potential, there still are few
datasets with alternative vision sensors. Thus, we provided a dataset recorded
from alternative vision sensors, by handheld or mounted on a car, repeatedly in
the same space but in different conditions. We aim to acquire visible
information from co-aligned alternative vision sensors. Our sensor system
collects data more independently from visible light intensity by measuring the
amount of infrared dissipation, depth by structured reflection, and
instantaneous temporal changes in luminance. We provide these measurements
along with inertial sensors and ground-truth for developing robust visual SLAM
under poor illumination. The full dataset is available at:
https://visibilitydataset.github.io/
Related papers
- SPARK: Multi-Vision Sensor Perception and Reasoning Benchmark for Large-scale Vision-Language Models [43.79587815909473]
This paper aims to establish a multi-vision Sensor Perception And Reasoning benchmarK called SPARK.
We generated 6,248 vision-language test samples to investigate multi-vision sensory perception and multi-vision sensory reasoning on physical sensor knowledge proficiency.
Results showed that most models displayed deficiencies in multi-vision sensory reasoning to varying extents.
arXiv Detail & Related papers (2024-08-22T03:59:48Z) - DIDLM:A Comprehensive Multi-Sensor Dataset with Infrared Cameras, Depth Cameras, LiDAR, and 4D Millimeter-Wave Radar in Challenging Scenarios for 3D Mapping [7.050468075029598]
This study presents a comprehensive multi-sensor dataset designed for 3D mapping in challenging indoor and outdoor environments.
The dataset comprises data from infrared cameras, depth cameras, LiDAR, and 4D millimeter-wave radar.
Various SLAM algorithms are employed to process the dataset, revealing performance differences among algorithms in different scenarios.
arXiv Detail & Related papers (2024-04-15T09:49:33Z) - Dataset and Benchmark: Novel Sensors for Autonomous Vehicle Perception [7.474695739346621]
This paper introduces the Novel Sensors for Autonomous Vehicle Perception dataset to facilitate future research on this topic.
The data was collected by repeatedly driving two 8 km routes and includes varied lighting conditions and opposing viewpoint perspectives.
To our knowledge, the NSAVP dataset is the first to include stereo thermal cameras together with stereo event and monochrome cameras.
arXiv Detail & Related papers (2024-01-24T23:25:23Z) - On the Importance of Accurate Geometry Data for Dense 3D Vision Tasks [61.74608497496841]
Training on inaccurate or corrupt data induces model bias and hampers generalisation capabilities.
This paper investigates the effect of sensor errors for the dense 3D vision tasks of depth estimation and reconstruction.
arXiv Detail & Related papers (2023-03-26T22:32:44Z) - Drone Detection and Tracking in Real-Time by Fusion of Different Sensing
Modalities [66.4525391417921]
We design and evaluate a multi-sensor drone detection system.
Our solution integrates a fish-eye camera as well to monitor a wider part of the sky and steer the other cameras towards objects of interest.
The thermal camera is shown to be a feasible solution as good as the video camera, even if the camera employed here has a lower resolution.
arXiv Detail & Related papers (2022-07-05T10:00:58Z) - Learning Enriched Illuminants for Cross and Single Sensor Color
Constancy [182.4997117953705]
We propose cross-sensor self-supervised training to train the network.
We train the network by randomly sampling the artificial illuminants in a sensor-independent manner.
Experiments show that our cross-sensor model and single-sensor model outperform other state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2022-03-21T15:45:35Z) - TUM-VIE: The TUM Stereo Visual-Inertial Event Dataset [50.8779574716494]
Event cameras are bio-inspired vision sensors which measure per pixel brightness changes.
They offer numerous benefits over traditional, frame-based cameras, including low latency, high dynamic range, high temporal resolution and low power consumption.
To foster the development of 3D perception and navigation algorithms with event cameras, we present the TUM-VIE dataset.
arXiv Detail & Related papers (2021-08-16T19:53:56Z) - Radar Voxel Fusion for 3D Object Detection [0.0]
This paper develops a low-level sensor fusion network for 3D object detection.
The radar sensor fusion proves especially beneficial in inclement conditions such as rain and night scenes.
arXiv Detail & Related papers (2021-06-26T20:34:12Z) - Semantics-aware Adaptive Knowledge Distillation for Sensor-to-Vision
Action Recognition [131.6328804788164]
We propose a framework, named Semantics-aware Adaptive Knowledge Distillation Networks (SAKDN), to enhance action recognition in vision-sensor modality (videos)
The SAKDN uses multiple wearable-sensors as teacher modalities and uses RGB videos as student modality.
arXiv Detail & Related papers (2020-09-01T03:38:31Z) - Learning Camera Miscalibration Detection [83.38916296044394]
This paper focuses on a data-driven approach to learn the detection of miscalibration in vision sensors, specifically RGB cameras.
Our contributions include a proposed miscalibration metric for RGB cameras and a novel semi-synthetic dataset generation pipeline based on this metric.
By training a deep convolutional neural network, we demonstrate the effectiveness of our pipeline to identify whether a recalibration of the camera's intrinsic parameters is required or not.
arXiv Detail & Related papers (2020-05-24T10:32:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.