Visible and Clear: Finding Tiny Objects in Difference Map
- URL: http://arxiv.org/abs/2405.11276v3
- Date: Mon, 30 Sep 2024 11:54:01 GMT
- Title: Visible and Clear: Finding Tiny Objects in Difference Map
- Authors: Bing Cao, Haiyu Yao, Pengfei Zhu, Qinghua Hu,
- Abstract summary: We introduce a self-reconstruction mechanism in the detection model, and discover the strong correlation between it and the tiny objects.
Specifically, we impose a reconstruction head in-between the neck of a detector, constructing a difference map of the reconstructed image and the input, which shows high sensitivity to tiny objects.
We further develop a Difference Map Guided Feature Enhancement (DGFE) module to make the tiny feature representation more clear.
- Score: 50.54061010335082
- License:
- Abstract: Tiny object detection is one of the key challenges in the field of object detection. The performance of most generic detectors dramatically decreases in tiny object detection tasks. The main challenge lies in extracting effective features of tiny objects. Existing methods usually perform generation-based feature enhancement, which is seriously affected by spurious textures and artifacts, making it difficult to make the tiny-object-specific features visible and clear for detection. To address this issue, we propose a self-reconstructed tiny object detection (SR-TOD) framework. We for the first time introduce a self-reconstruction mechanism in the detection model, and discover the strong correlation between it and the tiny objects. Specifically, we impose a reconstruction head in-between the neck of a detector, constructing a difference map of the reconstructed image and the input, which shows high sensitivity to tiny objects. This inspires us to enhance the weak representations of tiny objects under the guidance of the difference maps. Thus, improving the visibility of tiny objects for the detectors. Building on this, we further develop a Difference Map Guided Feature Enhancement (DGFE) module to make the tiny feature representation more clear. In addition, we further propose a new multi-instance anti-UAV dataset, which is called DroneSwarms dataset and contains a large number of tiny drones with the smallest average size to date. Extensive experiments on the DroneSwarms dataset and other datasets demonstrate the effectiveness of the proposed method. The code and dataset will be publicly available.
Related papers
- FADE: A Dataset for Detecting Falling Objects around Buildings in Video [75.48118923174712]
Falling objects from buildings can cause severe injuries to pedestrians due to the great impact force they exert.
FADE contains 1,881 videos from 18 scenes, featuring 8 falling object categories, 4 weather conditions, and 4 video resolutions.
We develop a new object detection method called FADE-Net, which effectively leverages motion information.
arXiv Detail & Related papers (2024-08-11T11:43:56Z) - ESOD: Efficient Small Object Detection on High-Resolution Images [36.80623357577051]
Small objects are usually sparsely distributed and locally clustered.
Massive feature extraction computations are wasted on the non-target background area of images.
We propose to reuse the detector's backbone to conduct feature-level object-seeking and patch-slicing.
arXiv Detail & Related papers (2024-07-23T12:21:23Z) - Small Object Detection by DETR via Information Augmentation and Adaptive
Feature Fusion [4.9860018132769985]
The RT-DETR model performs well in real-time object detection, but performs poorly in small object detection accuracy.
We propose an adaptive feature fusion algorithm that assigns learnable parameters to each feature map from different levels.
This enhances the model's ability to capture object features at different scales, thereby improving the accuracy of detecting small objects.
arXiv Detail & Related papers (2024-01-16T00:01:23Z) - Improving the Detection of Small Oriented Objects in Aerial Images [0.0]
We propose a method to accurately detect small oriented objects in aerial images by enhancing the classification and regression tasks of the oriented object detection model.
We designed the Attention-Points Network consisting of two losses: Guided-Attention Loss (GALoss) and Box-Points Loss (BPLoss)
Experimental results show the effectiveness of our Attention-Points Network on a standard oriented aerial dataset with small object instances.
arXiv Detail & Related papers (2024-01-12T11:00:07Z) - Innovative Horizons in Aerial Imagery: LSKNet Meets DiffusionDet for
Advanced Object Detection [55.2480439325792]
We present an in-depth evaluation of an object detection model that integrates the LSKNet backbone with the DiffusionDet head.
The proposed model achieves a mean average precision (MAP) of approximately 45.7%, which is a significant improvement.
This advancement underscores the effectiveness of the proposed modifications and sets a new benchmark in aerial image analysis.
arXiv Detail & Related papers (2023-11-21T19:49:13Z) - Object-Centric Multiple Object Tracking [124.30650395969126]
This paper proposes a video object-centric model for multiple-object tracking pipelines.
It consists of an index-merge module that adapts the object-centric slots into detection outputs and an object memory module.
Benefited from object-centric learning, we only require sparse detection labels for object localization and feature binding.
arXiv Detail & Related papers (2023-09-01T03:34:12Z) - Rethinking the backbone architecture for tiny object detection [0.0]
Existing tiny object detection methods use standard deep neural networks as their backbone architecture.
We argue that such backbones are inappropriate for detecting tiny objects as they are designed for the classification of larger objects, and do not have the spatial resolution to identify small targets.
We design 'bottom-heavy' versions of backbones that allocate more resources to processing higher-resolution features without introducing any additional computational burden overall.
arXiv Detail & Related papers (2023-03-20T16:50:29Z) - Embracing Single Stride 3D Object Detector with Sparse Transformer [63.179720817019096]
In LiDAR-based 3D object detection for autonomous driving, the ratio of the object size to input scene size is significantly smaller compared to 2D detection cases.
Many 3D detectors directly follow the common practice of 2D detectors, which downsample the feature maps even after quantizing the point clouds.
We propose Single-stride Sparse Transformer (SST) to maintain the original resolution from the beginning to the end of the network.
arXiv Detail & Related papers (2021-12-13T02:12:02Z) - MultiResolution Attention Extractor for Small Object Detection [40.74232149130456]
Small objects are difficult to detect because of their low resolution and small size.
Inspired by human vision "attention" mechanism, we exploit two feature extraction methods to mine the most useful information of small objects.
arXiv Detail & Related papers (2020-06-10T16:47:56Z) - NETNet: Neighbor Erasing and Transferring Network for Better Single Shot
Object Detection [170.30694322460045]
We propose a new Neighbor Erasing and Transferring (NET) mechanism to reconfigure the pyramid features and explore scale-aware features.
A single-shot network called NETNet is constructed for scale-aware object detection.
arXiv Detail & Related papers (2020-01-18T15:21:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.