Knowledge Combination to Learn Rotated Detection Without Rotated
Annotation
- URL: http://arxiv.org/abs/2304.02199v2
- Date: Thu, 4 May 2023 10:10:10 GMT
- Title: Knowledge Combination to Learn Rotated Detection Without Rotated
Annotation
- Authors: Tianyu Zhu, Bryce Ferenczi, Pulak Purkait, Tom Drummond, Hamid
Rezatofighi, Anton van den Hengel
- Abstract summary: Rotated bounding boxes drastically reduce output ambiguity of elongated objects.
Despite the effectiveness, rotated detectors are not widely employed.
We propose a framework that allows the model to predict precise rotated boxes.
- Score: 53.439096583978504
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Rotated bounding boxes drastically reduce output ambiguity of elongated
objects, making it superior to axis-aligned bounding boxes. Despite the
effectiveness, rotated detectors are not widely employed. Annotating rotated
bounding boxes is such a laborious process that they are not provided in many
detection datasets where axis-aligned annotations are used instead. In this
paper, we propose a framework that allows the model to predict precise rotated
boxes only requiring cheaper axis-aligned annotation of the target dataset 1.
To achieve this, we leverage the fact that neural networks are capable of
learning richer representation of the target domain than what is utilized by
the task. The under-utilized representation can be exploited to address a more
detailed task. Our framework combines task knowledge of an out-of-domain source
dataset with stronger annotation and domain knowledge of the target dataset
with weaker annotation. A novel assignment process and projection loss are used
to enable the co-training on the source and target datasets. As a result, the
model is able to solve the more detailed task in the target domain, without
additional computation overhead during inference. We extensively evaluate the
method on various target datasets including fresh-produce dataset, HRSC2016 and
SSDD. Results show that the proposed method consistently performs on par with
the fully supervised approach.
Related papers
- Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - infoVerse: A Universal Framework for Dataset Characterization with
Multidimensional Meta-information [68.76707843019886]
infoVerse is a universal framework for dataset characterization.
infoVerse captures multidimensional characteristics of datasets by incorporating various model-driven meta-information.
In three real-world applications (data pruning, active learning, and data annotation), the samples chosen on infoVerse space consistently outperform strong baselines.
arXiv Detail & Related papers (2023-05-30T18:12:48Z) - SalienDet: A Saliency-based Feature Enhancement Algorithm for Object
Detection for Autonomous Driving [160.57870373052577]
We propose a saliency-based OD algorithm (SalienDet) to detect unknown objects.
Our SalienDet utilizes a saliency-based algorithm to enhance image features for object proposal generation.
We design a dataset relabeling approach to differentiate the unknown objects from all objects in training sample set to achieve Open-World Detection.
arXiv Detail & Related papers (2023-05-11T16:19:44Z) - INoD: Injected Noise Discriminator for Self-Supervised Representation
Learning in Agricultural Fields [6.891600948991265]
We propose an Injected Noise Discriminator (INoD) which exploits principles of feature replacement and dataset discrimination for self-supervised representation learning.
INoD interleaves feature maps from two disjoint datasets during their convolutional encoding and predicts the dataset affiliation of the resultant feature map as a pretext task.
Our approach enables the network to learn unequivocal representations of objects seen in one dataset while observing them in conjunction with similar features from the disjoint dataset.
arXiv Detail & Related papers (2023-03-31T14:46:31Z) - Coreset Sampling from Open-Set for Fine-Grained Self-Supervised Learning [10.57079240576682]
We introduce a novel Open-Set Self-Supervised Learning problem under the assumption that a large-scale unlabeled open-set is available.
In our problem setup, it is crucial to consider the distribution mismatch between the open-set and target dataset.
We demonstrate that SimCore significantly improves representation learning performance through extensive experimental settings.
arXiv Detail & Related papers (2023-03-20T13:38:29Z) - Cross Domain Object Detection by Target-Perceived Dual Branch
Distillation [49.68119030818388]
Cross domain object detection is a realistic and challenging task in the wild.
We propose a novel Target-perceived Dual-branch Distillation (TDD) framework.
Our TDD significantly outperforms the state-of-the-art methods on all the benchmarks.
arXiv Detail & Related papers (2022-05-03T03:51:32Z) - Clustering augmented Self-Supervised Learning: Anapplication to Land
Cover Mapping [10.720852987343896]
We introduce a new method for land cover mapping by using a clustering based pretext task for self-supervised learning.
We demonstrate the effectiveness of the method on two societally relevant applications.
arXiv Detail & Related papers (2021-08-16T19:35:43Z) - Learning to Count in the Crowd from Limited Labeled Data [109.2954525909007]
We focus on reducing the annotation efforts by learning to count in the crowd from limited number of labeled samples.
Specifically, we propose a Gaussian Process-based iterative learning mechanism that involves estimation of pseudo-ground truth for the unlabeled data.
arXiv Detail & Related papers (2020-07-07T04:17:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.