ARUBA: An Architecture-Agnostic Balanced Loss for Aerial Object
Detection
- URL: http://arxiv.org/abs/2210.04574v3
- Date: Sat, 18 Nov 2023 05:36:37 GMT
- Title: ARUBA: An Architecture-Agnostic Balanced Loss for Aerial Object
Detection
- Authors: Rebbapragada V C Sairam, Monish Keswani, Uttaran Sinha, Nishit Shah,
Vineeth N Balasubramanian
- Abstract summary: We denote size of an object as the number of pixels it covers in an image and size imbalance as the over-representation of certain sizes of objects in a dataset.
We propose a novel ARchitectUre-agnostic BAlanced Loss (ARUBA) that can be applied as a plugin on top of any object detection model.
- Score: 24.085715205081385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks tend to reciprocate the bias of their training dataset.
In object detection, the bias exists in the form of various imbalances such as
class, background-foreground, and object size. In this paper, we denote size of
an object as the number of pixels it covers in an image and size imbalance as
the over-representation of certain sizes of objects in a dataset. We aim to
address the problem of size imbalance in drone-based aerial image datasets.
Existing methods for solving size imbalance are based on architectural changes
that utilize multiple scales of images or feature maps for detecting objects of
different sizes. We, on the other hand, propose a novel ARchitectUre-agnostic
BAlanced Loss (ARUBA) that can be applied as a plugin on top of any object
detection model. It follows a neighborhood-driven approach inspired by the
ordinality of object size. We evaluate the effectiveness of our approach
through comprehensive experiments on aerial datasets such as HRSC2016,
DOTAv1.0, DOTAv1.5 and VisDrone and obtain consistent improvement in
performance.
Related papers
- SOOD++: Leveraging Unlabeled Data to Boost Oriented Object Detection [59.868772767818975]
We propose a simple yet effective Semi-supervised Oriented Object Detection method termed SOOD++.
Specifically, we observe that objects from aerial images are usually arbitrary orientations, small scales, and aggregation.
Extensive experiments conducted on various multi-oriented object datasets under various labeled settings demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2024-07-01T07:03:51Z) - DASSF: Dynamic-Attention Scale-Sequence Fusion for Aerial Object Detection [6.635903943457569]
The original YOLO algorithm has low overall detection accuracy due to its weak ability to perceive targets of different scales.
This paper proposes a dynamic-attention scale-sequence fusion algorithm (DASSF) for small target detection in aerial images.
Experimental results show that when the DASSF method is applied to YOLOv8, compared to YOLOv8n, the model shows an increase of 9.2% and 2.4% in the mean average precision (mAP)
arXiv Detail & Related papers (2024-06-18T05:26:44Z) - YOLC: You Only Look Clusters for Tiny Object Detection in Aerial Images [33.80392696735718]
YOLC (You Only Look Clusters) is an efficient and effective framework that builds on an anchor-free object detector, CenterNet.
To overcome the challenges posed by large-scale images and non-uniform object distribution, we introduce a Local Scale Module (LSM) that adaptively searches cluster regions for zooming in for accurate detection.
We perform extensive experiments on two aerial image datasets, including Visdrone 2019 and UAVDT, to demonstrate the effectiveness and superiority of our proposed approach.
arXiv Detail & Related papers (2024-04-09T10:03:44Z) - Improving the Detection of Small Oriented Objects in Aerial Images [0.0]
We propose a method to accurately detect small oriented objects in aerial images by enhancing the classification and regression tasks of the oriented object detection model.
We designed the Attention-Points Network consisting of two losses: Guided-Attention Loss (GALoss) and Box-Points Loss (BPLoss)
Experimental results show the effectiveness of our Attention-Points Network on a standard oriented aerial dataset with small object instances.
arXiv Detail & Related papers (2024-01-12T11:00:07Z) - Innovative Horizons in Aerial Imagery: LSKNet Meets DiffusionDet for
Advanced Object Detection [55.2480439325792]
We present an in-depth evaluation of an object detection model that integrates the LSKNet backbone with the DiffusionDet head.
The proposed model achieves a mean average precision (MAP) of approximately 45.7%, which is a significant improvement.
This advancement underscores the effectiveness of the proposed modifications and sets a new benchmark in aerial image analysis.
arXiv Detail & Related papers (2023-11-21T19:49:13Z) - On the Importance of Large Objects in CNN Based Object Detection
Algorithms [0.0]
We highlight the importance of large objects in learning features that are critical for all sizes.
We show that giving more weight to large objects leads to improved detection scores across all object sizes.
arXiv Detail & Related papers (2023-11-20T12:32:32Z) - On the Robustness of Object Detection Models in Aerial Images [37.50307094643692]
We introduce two novel benchmarks based on DOTA-v1.0.
The first benchmark encompasses 19 prevalent corruptions, while the second focuses on cloud-corrupted images.
We find that enhanced model architectures, larger networks, well-crafted modules, and judicious data augmentation strategies collectively enhance the robustness of aerial object detection models.
arXiv Detail & Related papers (2023-08-29T15:16:51Z) - Object Detection in Aerial Images with Uncertainty-Aware Graph Network [61.02591506040606]
We propose a novel uncertainty-aware object detection framework with a structured-graph, where nodes and edges are denoted by objects.
We refer to our model as Uncertainty-Aware Graph network for object DETection (UAGDet)
arXiv Detail & Related papers (2022-08-23T07:29:03Z) - Towards Model Generalization for Monocular 3D Object Detection [57.25828870799331]
We present an effective unified camera-generalized paradigm (CGP) for Mono3D object detection.
We also propose the 2D-3D geometry-consistent object scaling strategy (GCOS) to bridge the gap via an instance-level augment.
Our method called DGMono3D achieves remarkable performance on all evaluated datasets and surpasses the SoTA unsupervised domain adaptation scheme.
arXiv Detail & Related papers (2022-05-23T23:05:07Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Object Detection in Aerial Images: A Large-Scale Benchmark and
Challenges [124.48654341780431]
We present a large-scale dataset of Object deTection in Aerial images (DOTA) and comprehensive baselines for ODAI.
The proposed DOTA dataset contains 1,793,658 object instances of 18 categories of oriented-bounding-box annotations collected from 11,268 aerial images.
We build baselines covering 10 state-of-the-art algorithms with over 70 configurations, where the speed and accuracy performances of each model have been evaluated.
arXiv Detail & Related papers (2021-02-24T11:20:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.