Many-shot from Low-shot: Learning to Annotate using Mixed Supervision
for Object Detection
- URL: http://arxiv.org/abs/2008.09694v2
- Date: Wed, 26 Aug 2020 17:26:13 GMT
- Title: Many-shot from Low-shot: Learning to Annotate using Mixed Supervision
for Object Detection
- Authors: Carlo Biffi, Steven McDonagh, Philip Torr, Ales Leonardis, Sarah
Parisot
- Abstract summary: Online annotation module (OAM) learns to generate a many-shot set of emphreliable annotations from a larger volume of weakly labelled images.
Our OAM can be jointly trained with any fully supervised two-stage object detection method, providing additional training annotations on the fly.
The integration of the OAM with Fast(er) R-CNN improves their performance by $17%$ mAP, $9%$ AP50 on PASCAL VOC 2007 and MS-COCO benchmarks, and significantly outperforms competing methods using mixed supervision.
- Score: 27.354492072251492
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object detection has witnessed significant progress by relying on large,
manually annotated datasets. Annotating such datasets is highly time consuming
and expensive, which motivates the development of weakly supervised and
few-shot object detection methods. However, these methods largely underperform
with respect to their strongly supervised counterpart, as weak training signals
\emph{often} result in partial or oversized detections. Towards solving this
problem we introduce, for the first time, an online annotation module (OAM)
that learns to generate a many-shot set of \emph{reliable} annotations from a
larger volume of weakly labelled images. Our OAM can be jointly trained with
any fully supervised two-stage object detection method, providing additional
training annotations on the fly. This results in a fully end-to-end strategy
that only requires a low-shot set of fully annotated images. The integration of
the OAM with Fast(er) R-CNN improves their performance by $17\%$ mAP, $9\%$
AP50 on PASCAL VOC 2007 and MS-COCO benchmarks, and significantly outperforms
competing methods using mixed supervision.
Related papers
- Efficient Meta-Learning Enabled Lightweight Multiscale Few-Shot Object Detection in Remote Sensing Images [15.12889076965307]
YOLOv7 one-stage detector is subjected to a novel meta-learning training framework.
This transformation allows the detector to adeptly address FSOD tasks while capitalizing on its inherent advantage of lightweight.
To validate the effectiveness of our proposed detector, we conducted performance comparisons with current state-of-the-art detectors.
arXiv Detail & Related papers (2024-04-29T04:56:52Z) - USD: Unknown Sensitive Detector Empowered by Decoupled Objectness and
Segment Anything Model [14.080744645704751]
Open World Object Detection (OWOD) is a novel and challenging computer vision task.
We propose a simple yet effective learning strategy, namely Decoupled Objectness Learning (DOL), which divides the learning of these two boundaries into decoder layers.
We also introduce an Auxiliary Supervision Framework (ASF) that uses a pseudo-labeling and a soft-weighting strategies to alleviate the negative impact of noise.
arXiv Detail & Related papers (2023-06-04T06:42:09Z) - WSSOD: A New Pipeline for Weakly- and Semi-Supervised Object Detection [75.80075054706079]
We propose a weakly- and semi-supervised object detection framework (WSSOD)
An agent detector is first trained on a joint dataset and then used to predict pseudo bounding boxes on weakly-annotated images.
The proposed framework demonstrates remarkable performance on PASCAL-VOC and MSCOCO benchmark, achieving a high performance comparable to those obtained in fully-supervised settings.
arXiv Detail & Related papers (2021-05-21T11:58:50Z) - Few-Cost Salient Object Detection with Adversarial-Paced Learning [95.0220555274653]
This paper proposes to learn the effective salient object detection model based on the manual annotation on a few training images only.
We name this task as the few-cost salient object detection and propose an adversarial-paced learning (APL)-based framework to facilitate the few-cost learning scenario.
arXiv Detail & Related papers (2021-04-05T14:15:49Z) - Weakly-Supervised Saliency Detection via Salient Object Subitizing [57.17613373230722]
We introduce saliency subitizing as the weak supervision since it is class-agnostic.
This allows the supervision to be aligned with the property of saliency detection.
We conduct extensive experiments on five benchmark datasets.
arXiv Detail & Related papers (2021-01-04T12:51:45Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z) - Towards Using Count-level Weak Supervision for Crowd Counting [55.58468947486247]
This paper studies the problem of weakly-supervised crowd counting which learns a model from only a small amount of location-level annotations (fully-supervised) but a large amount of count-level annotations (weakly-supervised)
We devise a simple-yet-effective training strategy, namely Multiple Auxiliary Tasks Training (MATT), to construct regularizes for restricting the freedom of the generated density maps.
arXiv Detail & Related papers (2020-02-29T02:58:36Z) - EHSOD: CAM-Guided End-to-end Hybrid-Supervised Object Detection with
Cascade Refinement [53.69674636044927]
We present EHSOD, an end-to-end hybrid-supervised object detection system.
It can be trained in one shot on both fully and weakly-annotated data.
It achieves comparable results on multiple object detection benchmarks with only 30% fully-annotated data.
arXiv Detail & Related papers (2020-02-18T08:04:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.