Local Contrast and Global Contextual Information Make Infrared Small
Object Salient Again
- URL: http://arxiv.org/abs/2301.12093v2
- Date: Tue, 31 Jan 2023 04:02:40 GMT
- Title: Local Contrast and Global Contextual Information Make Infrared Small
Object Salient Again
- Authors: Chenyi Wang, Huan Wang, Peiwen Pan
- Abstract summary: Infrared small object detection (ISOS) aims to segment small objects only covered with several pixels from clutter background in infrared images.
It's of great challenge due to: 1) small objects lack of sufficient intensity, shape and texture information; 2) small objects are easily lost in the process where detection models, say deep neural networks, obtain high-level semantic features and image-level receptive fields through successive downsampling.
This paper proposes a reliable detection model for ISOS, dubbed UCFNet, which can handle well the two issues.
Experiments on several public datasets demonstrate that our method significantly outperforms the state-of-the
- Score: 5.324958606516871
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Infrared small object detection (ISOS) aims to segment small objects only
covered with several pixels from clutter background in infrared images. It's of
great challenge due to: 1) small objects lack of sufficient intensity, shape
and texture information; 2) small objects are easily lost in the process where
detection models, say deep neural networks, obtain high-level semantic features
and image-level receptive fields through successive downsampling. This paper
proposes a reliable detection model for ISOS, dubbed UCFNet, which can handle
well the two issues. It builds upon central difference convolution (CDC) and
fast Fourier convolution (FFC). On one hand, CDC can effectively guide the
network to learn the contrast information between small objects and the
background, as the contrast information is very essential in human visual
system dealing with the ISOS task. On the other hand, FFC can gain image-level
receptive fields and extract global information while preventing small objects
from being overwhelmed.Experiments on several public datasets demonstrate that
our method significantly outperforms the state-of-the-art ISOS models, and can
provide useful guidelines for designing better ISOS deep models. Codes will be
available soon.
Related papers
- SFA-UNet: More Attention to Multi-Scale Contrast and Contextual Information in Infrared Small Object Segmentation [0.0]
Infrared Small Object (ISOS) remains a major focus due to several challenges.
We propose a modified U-Net architecture, named SFA-UNet, by combining Scharr Convolution (SC) and Fast Fourier Convolution (FFC) in addition to vertical and horizontal Attention gates (AG) into UNet.
SC helps to learn the foreground-to-background contrast information whereas FFC provide multi-scale contextual information while mitigating the small objects vanishing problem.
arXiv Detail & Related papers (2024-10-30T10:21:23Z) - SIRST-5K: Exploring Massive Negatives Synthesis with Self-supervised
Learning for Robust Infrared Small Target Detection [53.19618419772467]
Single-frame infrared small target (SIRST) detection aims to recognize small targets from clutter backgrounds.
With the development of Transformer, the scale of SIRST models is constantly increasing.
With a rich diversity of infrared small target data, our algorithm significantly improves the model performance and convergence speed.
arXiv Detail & Related papers (2024-03-08T16:14:54Z) - ILNet: Low-level Matters for Salient Infrared Small Target Detection [5.248337726304453]
Infrared small target detection is a technique for finding small targets from infrared clutter background.
Due to the dearth of high-level semantic information, small infrared target features are weakened in the deep layers of the CNN.
We propose an infrared low-level network (ILNet) that considers infrared small targets as salient areas with little semantic information.
arXiv Detail & Related papers (2023-09-24T14:09:37Z) - UIU-Net: U-Net in U-Net for Infrared Small Object Detection [36.72184013409837]
We propose a simple and effective U-Net in U-Net'' framework, UIU-Net for short, and detect small objects in infrared images.
As the name suggests, UIU-Net embeds a tiny U-Net into a larger U-Net backbone, enabling the multi-level and multi-scale representation learning of objects.
The proposed UIU-Net also produces powerful generalization performance for video sequence infrared small object datasets.
arXiv Detail & Related papers (2022-12-02T04:52:26Z) - Learning Enriched Features for Fast Image Restoration and Enhancement [166.17296369600774]
This paper presents a holistic goal of maintaining spatially-precise high-resolution representations through the entire network.
We learn an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
Our approach achieves state-of-the-art results for a variety of image processing tasks, including defocus deblurring, image denoising, super-resolution, and image enhancement.
arXiv Detail & Related papers (2022-04-19T17:59:45Z) - High-resolution Iterative Feedback Network for Camouflaged Object
Detection [128.893782016078]
Spotting camouflaged objects that are visually assimilated into the background is tricky for object detection algorithms.
We aim to extract the high-resolution texture details to avoid the detail degradation that causes blurred vision in edges and boundaries.
We introduce a novel HitNet to refine the low-resolution representations by high-resolution features in an iterative feedback manner.
arXiv Detail & Related papers (2022-03-22T11:20:21Z) - Infrared Small-Dim Target Detection with Transformer under Complex
Backgrounds [155.388487263872]
We propose a new infrared small-dim target detection method with the transformer.
We adopt the self-attention mechanism of the transformer to learn the interaction information of image features in a larger range.
We also design a feature enhancement module to learn more features of small-dim targets.
arXiv Detail & Related papers (2021-09-29T12:23:41Z) - You Better Look Twice: a new perspective for designing accurate
detectors with reduced computations [56.34005280792013]
BLT-net is a new low-computation two-stage object detection architecture.
It reduces computations by separating objects from background using a very lite first-stage.
Resulting image proposals are then processed in the second-stage by a highly accurate model.
arXiv Detail & Related papers (2021-07-21T12:39:51Z) - Efficient Classification of Very Large Images with Tiny Objects [15.822654320750054]
We present an end-to-end CNN model termed Zoom-In network for classification of large images with tiny objects.
We evaluate our method on two large-image datasets and one gigapixel dataset.
arXiv Detail & Related papers (2021-06-04T20:13:04Z) - Dense Multiscale Feature Fusion Pyramid Networks for Object Detection in
UAV-Captured Images [0.09065034043031667]
We propose a novel method called Dense Multiscale Feature Fusion Pyramid Networks(DMFFPN), which is aimed at obtaining rich features as much as possible.
Specifically, the dense connection is designed to fully utilize the representation from the different convolutional layers.
Experiments on the drone-based datasets named VisDrone-DET suggest a competitive performance of our method.
arXiv Detail & Related papers (2020-12-19T10:05:31Z) - Ventral-Dorsal Neural Networks: Object Detection via Selective Attention [51.79577908317031]
We propose a new framework called Ventral-Dorsal Networks (VDNets)
Inspired by the structure of the human visual system, we propose the integration of a "Ventral Network" and a "Dorsal Network"
Our experimental results reveal that the proposed method outperforms state-of-the-art object detection approaches.
arXiv Detail & Related papers (2020-05-15T23:57:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.