SIOD: Single Instance Annotated Per Category Per Image for Object
Detection
- URL: http://arxiv.org/abs/2203.15353v2
- Date: Wed, 30 Mar 2022 02:24:36 GMT
- Title: SIOD: Single Instance Annotated Per Category Per Image for Object
Detection
- Authors: Hanjun Li, Xingjia Pan, Ke Yan, Fan Tang, Wei-Shi Zheng
- Abstract summary: We propose the Single Instance annotated Object Detection (SIOD), requiring only one instance annotation for each existing category in an image.
Degraded from inter-task (WSOD) or inter-image (SSOD) discrepancies to the intra-image discrepancy, SIOD provides more reliable and rich prior knowledge for mining the rest of unlabeled instances.
Under the SIOD setting, we propose a simple yet effective framework, termed Dual-Mining (DMiner), which consists of a Similarity-based Pseudo Label Generating module (SPLG) and a Pixel-level Group Contrastive Learning module (PGCL)
- Score: 67.64774488115299
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object detection under imperfect data receives great attention recently.
Weakly supervised object detection (WSOD) suffers from severe localization
issues due to the lack of instance-level annotation, while semi-supervised
object detection (SSOD) remains challenging led by the inter-image discrepancy
between labeled and unlabeled data. In this study, we propose the Single
Instance annotated Object Detection (SIOD), requiring only one instance
annotation for each existing category in an image. Degraded from inter-task
(WSOD) or inter-image (SSOD) discrepancies to the intra-image discrepancy, SIOD
provides more reliable and rich prior knowledge for mining the rest of
unlabeled instances and trades off the annotation cost and performance. Under
the SIOD setting, we propose a simple yet effective framework, termed
Dual-Mining (DMiner), which consists of a Similarity-based Pseudo Label
Generating module (SPLG) and a Pixel-level Group Contrastive Learning module
(PGCL). SPLG firstly mines latent instances from feature representation space
to alleviate the annotation missing problem. To avoid being misled by
inaccurate pseudo labels, we propose PGCL to boost the tolerance to false
pseudo labels. Extensive experiments on MS COCO verify the feasibility of the
SIOD setting and the superiority of the proposed method, which obtains
consistent and significant improvements compared to baseline methods and
achieves comparable results with fully supervised object detection (FSOD)
methods with only 40% instances annotated.
Related papers
- Semi-supervised Open-World Object Detection [74.95267079505145]
We introduce a more realistic formulation, named semi-supervised open-world detection (SS-OWOD)
We demonstrate that the performance of the state-of-the-art OWOD detector dramatically deteriorates in the proposed SS-OWOD setting.
Our experiments on 4 datasets including MS COCO, PASCAL, Objects365 and DOTA demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2024-02-25T07:12:51Z) - Open-Set Semi-Supervised Object Detection [43.464223594166654]
Recent developments for Semi-Supervised Object Detection (SSOD) have shown the promise of leveraging unlabeled data to improve an object detector.
We consider a more practical yet challenging problem, Open-Set Semi-Supervised Object Detection (OSSOD)
Our proposed framework effectively addresses the semantic expansion issue and shows consistent improvements on many OSSOD benchmarks.
arXiv Detail & Related papers (2022-08-29T17:04:30Z) - Learning to Detect Instance-level Salient Objects Using Complementary
Image Labels [55.049347205603304]
We present the first weakly-supervised approach to the salient instance detection problem.
We propose a novel weakly-supervised network with three branches: a Saliency Detection Branch leveraging class consistency information to locate candidate objects; a Boundary Detection Branch exploiting class discrepancy information to delineate object boundaries; and a Centroid Detection Branch using subitizing information to detect salient instance centroids.
arXiv Detail & Related papers (2021-11-19T10:15:22Z) - WSSOD: A New Pipeline for Weakly- and Semi-Supervised Object Detection [75.80075054706079]
We propose a weakly- and semi-supervised object detection framework (WSSOD)
An agent detector is first trained on a joint dataset and then used to predict pseudo bounding boxes on weakly-annotated images.
The proposed framework demonstrates remarkable performance on PASCAL-VOC and MSCOCO benchmark, achieving a high performance comparable to those obtained in fully-supervised settings.
arXiv Detail & Related papers (2021-05-21T11:58:50Z) - Object Detection Made Simpler by Eliminating Heuristic NMS [70.93004137521946]
We show a simple NMS-free, end-to-end object detection framework.
We attain on par or even improved detection accuracy compared with the original one-stage detector.
arXiv Detail & Related papers (2021-01-28T02:38:29Z) - Co-mining: Self-Supervised Learning for Sparsely Annotated Object
Detection [29.683119976550007]
We propose a simple but effective mechanism, called Co-mining, for sparsely annotated object detection.
In our Co-mining, two branches of a Siamese network predict the pseudo-label sets for each other.
Experiments are performed on MS dataset with three different sparsely annotated settings.
arXiv Detail & Related papers (2020-12-03T14:23:43Z) - Temporal Action Detection with Multi-level Supervision [116.55596693897388]
We introduce the Semi-supervised Action Detection (SSAD) task with a mixture of labeled and unlabeled data.
We analyze different types of errors in the proposed SSAD baselines which are directly adapted from the semi-supervised classification task.
We incorporate weakly-labeled data into SSAD and propose Omni-supervised Action Detection (OSAD) with three levels of supervision.
arXiv Detail & Related papers (2020-11-24T04:45:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.