GOOD: Towards Domain Generalized Orientated Object Detection
- URL: http://arxiv.org/abs/2402.12765v1
- Date: Tue, 20 Feb 2024 07:12:22 GMT
- Title: GOOD: Towards Domain Generalized Orientated Object Detection
- Authors: Qi Bi, Beichen Zhou, Jingjun Yi, Wei Ji, Haolan Zhan, Gui-Song Xia
- Abstract summary: Oriented object detection has been rapidly developed in the past few years, but most of these methods assume the training and testing images are under the same statistical distribution.
We propose the task of domain generalized oriented object detection, which intends to explore the generalization of oriented object detectors on arbitrary unseen target domains.
- Score: 39.76969237020444
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Oriented object detection has been rapidly developed in the past few years,
but most of these methods assume the training and testing images are under the
same statistical distribution, which is far from reality. In this paper, we
propose the task of domain generalized oriented object detection, which intends
to explore the generalization of oriented object detectors on arbitrary unseen
target domains. Learning domain generalized oriented object detectors is
particularly challenging, as the cross-domain style variation not only
negatively impacts the content representation, but also leads to unreliable
orientation predictions. To address these challenges, we propose a generalized
oriented object detector (GOOD). After style hallucination by the emerging
contrastive language-image pre-training (CLIP), it consists of two key
components, namely, rotation-aware content consistency learning (RAC) and style
consistency learning (SEC). The proposed RAC allows the oriented object
detector to learn stable orientation representation from style-diversified
samples. The proposed SEC further stabilizes the generalization ability of
content representation from different image styles. Extensive experiments on
multiple cross-domain settings show the state-of-the-art performance of GOOD.
Source code will be publicly available.
Related papers
- Detecting Every Object from Events [24.58024539462497]
We propose Detecting Every Object in Events (DEOE), an approach tailored for achieving high-speed, class-agnostic open-world object detection in event-based vision.
Our code is available at https://github.com/Hatins/DEOE.
arXiv Detail & Related papers (2024-04-08T08:20:53Z) - Cross-Domain Few-Shot Object Detection via Enhanced Open-Set Object Detector [72.05791402494727]
This paper studies the challenging cross-domain few-shot object detection (CD-FSOD)
It aims to develop an accurate object detector for novel domains with minimal labeled examples.
arXiv Detail & Related papers (2024-02-05T15:25:32Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - CLIP the Gap: A Single Domain Generalization Approach for Object
Detection [60.20931827772482]
Single Domain Generalization tackles the problem of training a model on a single source domain so that it generalizes to any unseen target domain.
We propose to leverage a pre-trained vision-language model to introduce semantic domain concepts via textual prompts.
We achieve this via a semantic augmentation strategy acting on the features extracted by the detector backbone, as well as a text-based classification loss.
arXiv Detail & Related papers (2023-01-13T12:01:18Z) - Contrastive View Design Strategies to Enhance Robustness to Domain
Shifts in Downstream Object Detection [37.06088084592779]
We conduct an empirical study of contrastive learning and out-of-domain object detection.
We propose strategies to augment views and enhance robustness in appearance-shifted and context-shifted scenarios.
Our results and insights show how to ensure robustness through the choice of views in contrastive learning.
arXiv Detail & Related papers (2022-12-09T00:34:50Z) - Robust Region Feature Synthesizer for Zero-Shot Object Detection [87.79902339984142]
We build a novel zero-shot object detection framework that contains an Intra-class Semantic Diverging component and an Inter-class Structure Preserving component.
It is the first study to carry out zero-shot object detection in remote sensing imagery.
arXiv Detail & Related papers (2022-01-01T03:09:15Z) - Robust Object Detection via Instance-Level Temporal Cycle Confusion [89.1027433760578]
We study the effectiveness of auxiliary self-supervised tasks to improve the out-of-distribution generalization of object detectors.
Inspired by the principle of maximum entropy, we introduce a novel self-supervised task, instance-level temporal cycle confusion (CycConf)
For each object, the task is to find the most different object proposals in the adjacent frame in a video and then cycle back to itself for self-supervision.
arXiv Detail & Related papers (2021-04-16T21:35:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.