LLVIP: A Visible-infrared Paired Dataset for Low-light Vision
- URL: http://arxiv.org/abs/2108.10831v4
- Date: Wed, 14 Jun 2023 12:14:17 GMT
- Title: LLVIP: A Visible-infrared Paired Dataset for Low-light Vision
- Authors: Xinyu Jia, Chuang Zhu, Minzhen Li, Wenqi Tang, Shengjie Liu, Wenli
Zhou
- Abstract summary: We present LLVIP, a visible-infrared paired dataset for low-light vision.
This dataset contains 30976 images, or 15488 pairs, most of which were taken at very dark scenes.
We compare the dataset with other visible-infrared datasets and evaluate the performance of some popular visual algorithms.
- Score: 4.453060631960743
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is very challenging for various visual tasks such as image fusion,
pedestrian detection and image-to-image translation in low light conditions due
to the loss of effective target areas. In this case, infrared and visible
images can be used together to provide both rich detail information and
effective target areas. In this paper, we present LLVIP, a visible-infrared
paired dataset for low-light vision. This dataset contains 30976 images, or
15488 pairs, most of which were taken at very dark scenes, and all of the
images are strictly aligned in time and space. Pedestrians in the dataset are
labeled. We compare the dataset with other visible-infrared datasets and
evaluate the performance of some popular visual algorithms including image
fusion, pedestrian detection and image-to-image translation on the dataset. The
experimental results demonstrate the complementary effect of fusion on image
information, and find the deficiency of existing algorithms of the three visual
tasks in very low-light conditions. We believe the LLVIP dataset will
contribute to the community of computer vision by promoting image fusion,
pedestrian detection and image-to-image translation in very low-light
applications. The dataset is being released in
https://bupt-ai-cz.github.io/LLVIP. Raw data is also provided for further
research such as image registration.
Related papers
- LoLI-Street: Benchmarking Low-Light Image Enhancement and Beyond [37.47964043913622]
We introduce a new dataset LoLI-Street (Low-Light Images of Streets) with 33k paired low-light and well-exposed images from street scenes in developed cities.
LoLI-Street dataset also features 1,000 real low-light test images for testing LLIE models under real-life conditions.
arXiv Detail & Related papers (2024-10-13T13:11:56Z) - Enhancing Low-Light Images Using Infrared-Encoded Images [81.8710581927427]
Previous arts mainly focus on the low-light images captured in the visible spectrum using pixel-wise loss.
We propose a novel approach to increase the visibility of images captured under low-light environments by removing the in-camera infrared (IR) cut-off filter.
arXiv Detail & Related papers (2023-07-09T08:29:19Z) - Application of image-to-image translation in improving pedestrian
detection [0.0]
In this study we are going to use advanced deep learning models like pix2pixGAN and YOLOv7 on LLVIP dataset, containing visible-infrared image pairs for low light vision.
This dataset contains 33672 images and most of the images were captured in dark scenes, tightly synchronized with time and location.
arXiv Detail & Related papers (2022-09-08T08:07:01Z) - Enhancing Low-Light Images in Real World via Cross-Image Disentanglement [58.754943762945864]
We propose a new low-light image enhancement dataset consisting of misaligned training images with real-world corruptions.
Our model achieves state-of-the-art performances on both the newly proposed dataset and other popular low-light datasets.
arXiv Detail & Related papers (2022-01-10T03:12:52Z) - Cross-Modal Object Tracking: Modality-Aware Representations and A
Unified Benchmark [8.932487291107812]
In many visual systems, visual tracking often bases on RGB image sequences, in which some targets are invalid in low-light conditions.
We propose a new algorithm, which learns the modality-aware target representation to mitigate the appearance gap between RGB and NIR modalities in the tracking process.
We will release the dataset for free academic usage, dataset download link and code will be released soon.
arXiv Detail & Related papers (2021-11-08T03:58:55Z) - Refer-it-in-RGBD: A Bottom-up Approach for 3D Visual Grounding in RGBD
Images [69.5662419067878]
Grounding referring expressions in RGBD image has been an emerging field.
We present a novel task of 3D visual grounding in single-view RGBD image where the referred objects are often only partially scanned due to occlusion.
Our approach first fuses the language and the visual features at the bottom level to generate a heatmap that localizes the relevant regions in the RGBD image.
Then our approach conducts an adaptive feature learning based on the heatmap and performs the object-level matching with another visio-linguistic fusion to finally ground the referred object.
arXiv Detail & Related papers (2021-03-14T11:18:50Z) - Scaling Up Visual and Vision-Language Representation Learning With Noisy
Text Supervision [57.031588264841]
We leverage a noisy dataset of over one billion image alt-text pairs, obtained without expensive filtering or post-processing steps.
A simple dual-encoder architecture learns to align visual and language representations of the image and text pairs using a contrastive loss.
We show that the scale of our corpus can make up for its noise and leads to state-of-the-art representations even with such a simple learning scheme.
arXiv Detail & Related papers (2021-02-11T10:08:12Z) - The Cube++ Illumination Estimation Dataset [50.58610459038332]
A new illumination estimation dataset is proposed in this paper.
It consists of 4890 images with known illumination colors as well as with additional semantic data.
The dataset can be used for training and testing of methods that perform single or two-illuminant estimation.
arXiv Detail & Related papers (2020-11-19T18:50:08Z) - Real-MFF: A Large Realistic Multi-focus Image Dataset with Ground Truth [58.226535803985804]
We introduce a large and realistic multi-focus dataset called Real-MFF.
The dataset contains 710 pairs of source images with corresponding ground truth images.
We evaluate 10 typical multi-focus algorithms on this dataset for the purpose of illustration.
arXiv Detail & Related papers (2020-03-28T12:33:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.