Universal-RCNN: Universal Object Detector via Transferable Graph R-CNN
- URL: http://arxiv.org/abs/2002.07417v1
- Date: Tue, 18 Feb 2020 07:57:45 GMT
- Title: Universal-RCNN: Universal Object Detector via Transferable Graph R-CNN
- Authors: Hang Xu, Linpu Fang, Xiaodan Liang, Wenxiong Kang, Zhenguo Li
- Abstract summary: We present a novel universal object detector called Universal-RCNN.
We first generate a global semantic pool by integrating all high-level semantic representation of all the categories.
An Intra-Domain Reasoning Module learns and propagates the sparse graph representation within one dataset guided by a spatial-aware GCN.
- Score: 117.80737222754306
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The dominant object detection approaches treat each dataset separately and
fit towards a specific domain, which cannot adapt to other domains without
extensive retraining. In this paper, we address the problem of designing a
universal object detection model that exploits diverse category granularity
from multiple domains and predict all kinds of categories in one system.
Existing works treat this problem by integrating multiple detection branches
upon one shared backbone network. However, this paradigm overlooks the crucial
semantic correlations between multiple domains, such as categories hierarchy,
visual similarity, and linguistic relationship. To address these drawbacks, we
present a novel universal object detector called Universal-RCNN that
incorporates graph transfer learning for propagating relevant semantic
information across multiple datasets to reach semantic coherency. Specifically,
we first generate a global semantic pool by integrating all high-level semantic
representation of all the categories. Then an Intra-Domain Reasoning Module
learns and propagates the sparse graph representation within one dataset guided
by a spatial-aware GCN. Finally, an InterDomain Transfer Module is proposed to
exploit diverse transfer dependencies across all domains and enhance the
regional feature representation by attending and transferring semantic contexts
globally. Extensive experiments demonstrate that the proposed method
significantly outperforms multiple-branch models and achieves the
state-of-the-art results on multiple object detection benchmarks (mAP: 49.1% on
COCO).
Related papers
- Generalize or Detect? Towards Robust Semantic Segmentation Under Multiple Distribution Shifts [56.57141696245328]
In open-world scenarios, where both novel classes and domains may exist, an ideal segmentation model should detect anomaly classes for safety.
Existing methods often struggle to distinguish between domain-level and semantic-level distribution shifts.
arXiv Detail & Related papers (2024-11-06T11:03:02Z) - Improving Anomaly Segmentation with Multi-Granularity Cross-Domain
Alignment [17.086123737443714]
Anomaly segmentation plays a pivotal role in identifying atypical objects in images, crucial for hazard detection in autonomous driving systems.
While existing methods demonstrate noteworthy results on synthetic data, they often fail to consider the disparity between synthetic and real-world data domains.
We introduce the Multi-Granularity Cross-Domain Alignment framework, tailored to harmonize features across domains at both the scene and individual sample levels.
arXiv Detail & Related papers (2023-08-16T22:54:49Z) - Mix and Reason: Reasoning over Semantic Topology with Data Mixing for
Domain Generalization [48.90173060487124]
Domain generalization (DG) enables a learning machine from multiple seen source domains to an unseen target one.
mire consists of two key components, namely, Category-aware Data Mixing (CDM) and Adaptive Semantic Topology Refinement (ASTR)
experiments on multiple DG benchmarks validate the effectiveness and robustness of the proposed mire.
arXiv Detail & Related papers (2022-10-14T06:52:34Z) - Multi-Granularity Alignment Domain Adaptation for Object Detection [33.32519045960187]
Domain adaptive object detection is challenging due to distinctive data distribution between source domain and target domain.
We propose a unified multi-granularity alignment based object detection framework towards domain-invariant feature learning.
arXiv Detail & Related papers (2022-03-31T09:05:06Z) - Compound Domain Generalization via Meta-Knowledge Encoding [55.22920476224671]
We introduce Style-induced Domain-specific Normalization (SDNorm) to re-normalize the multi-modal underlying distributions.
We harness the prototype representations, the centroids of classes, to perform relational modeling in the embedding space.
Experiments on four standard Domain Generalization benchmarks reveal that COMEN exceeds the state-of-the-art performance without the need of domain supervision.
arXiv Detail & Related papers (2022-03-24T11:54:59Z) - Domain Generalisation for Object Detection under Covariate and Concept Shift [10.32461766065764]
Domain generalisation aims to promote the learning of domain-invariant features while suppressing domain-specific features.
An approach to domain generalisation for object detection is proposed, the first such approach applicable to any object detection architecture.
arXiv Detail & Related papers (2022-03-10T11:14:18Z) - AFAN: Augmented Feature Alignment Network for Cross-Domain Object
Detection [90.18752912204778]
Unsupervised domain adaptation for object detection is a challenging problem with many real-world applications.
We propose a novel augmented feature alignment network (AFAN) which integrates intermediate domain image generation and domain-adversarial training.
Our approach significantly outperforms the state-of-the-art methods on standard benchmarks for both similar and dissimilar domain adaptations.
arXiv Detail & Related papers (2021-06-10T05:01:20Z) - Learning to Combine: Knowledge Aggregation for Multi-Source Domain
Adaptation [56.694330303488435]
We propose a Learning to Combine for Multi-Source Domain Adaptation (LtC-MSDA) framework.
In the nutshell, a knowledge graph is constructed on the prototypes of various domains to realize the information propagation among semantically adjacent representations.
Our approach outperforms existing methods with a remarkable margin.
arXiv Detail & Related papers (2020-07-17T07:52:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.