Learning Object Scale With Click Supervision for Object Detection
- URL: http://arxiv.org/abs/2002.08555v1
- Date: Thu, 20 Feb 2020 03:59:46 GMT
- Title: Learning Object Scale With Click Supervision for Object Detection
- Authors: Liao Zhang, Yan Yan, Lin Cheng, and Hanzi Wang
- Abstract summary: We propose a simple yet effective method which incorporatesCNN visualization with click supervision to generate the pseudoground-truths.
These pseudo ground-truthscans be used to train a fully-supervised detector.
Experimental results on the PASCAL VOC2007 and VOC 2012 datasets show that the proposed methodcan obtain much higher accuracy for estimating the object scale.
- Score: 29.421113887739413
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Weakly-supervised object detection has recently attracted increasing
attention since it only requires image-levelannotations. However, the
performance obtained by existingmethods is still far from being satisfactory
compared with fully-supervised object detection methods. To achieve a good
trade-off between annotation cost and object detection performance,we propose a
simple yet effective method which incorporatesCNN visualization with click
supervision to generate the pseudoground-truths (i.e., bounding boxes). These
pseudo ground-truthscan be used to train a fully-supervised detector. To
estimatethe object scale, we firstly adopt a proposal selection algorithmto
preserve high-quality proposals, and then generate ClassActivation Maps (CAMs)
for these preserved proposals by theproposed CNN visualization algorithm called
Spatial AttentionCAM. Finally, we fuse these CAMs together to generate
pseudoground-truths and train a fully-supervised object detector withthese
ground-truths. Experimental results on the PASCAL VOC2007 and VOC 2012 datasets
show that the proposed methodcan obtain much higher accuracy for estimating the
object scale,compared with the state-of-the-art image-level based methodsand
the center-click based method
Related papers
- SeMoLi: What Moves Together Belongs Together [51.72754014130369]
We tackle semi-supervised object detection based on motion cues.
Recent results suggest that motion-based clustering methods can be used to pseudo-label instances of moving objects.
We re-think this approach and suggest that both, object detection, as well as motion-inspired pseudo-labeling, can be tackled in a data-driven manner.
arXiv Detail & Related papers (2024-02-29T18:54:53Z) - Towards Few-Annotation Learning for Object Detection: Are
Transformer-based Models More Efficient ? [11.416621957617334]
In this paper, we propose a semi-supervised method tailored for the current state-of-the-art object detector Deformable DETR.
We evaluate our method on the semi-supervised object detection benchmarks COCO and Pascal VOC, and it outperforms previous methods, especially when annotations are scarce.
arXiv Detail & Related papers (2023-10-30T18:51:25Z) - Improving Online Lane Graph Extraction by Object-Lane Clustering [106.71926896061686]
We propose an architecture and loss formulation to improve the accuracy of local lane graph estimates.
The proposed method learns to assign the objects to centerlines by considering the centerlines as cluster centers.
We show that our method can achieve significant performance improvements by using the outputs of existing 3D object detection methods.
arXiv Detail & Related papers (2023-07-20T15:21:28Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Learning Remote Sensing Object Detection with Single Point Supervision [17.12725535531483]
Pointly Supervised Object Detection (PSOD) has attracted considerable interests due to its lower labeling cost as compared to box-level supervised object detection.
We make the first attempt to achieve RS object detection with single point supervision, and propose a PSOD method tailored for RS images.
Our method can achieve significantly better performance as compared to state-of-the-art image-level and point-level supervised detection methods, and reduce the performance gap between PSOD and box-level supervised object detection.
arXiv Detail & Related papers (2023-05-23T15:06:04Z) - Instance Localization for Self-supervised Detection Pretraining [68.24102560821623]
We propose a new self-supervised pretext task, called instance localization.
We show that integration of bounding boxes into pretraining promotes better task alignment and architecture alignment for transfer learning.
Experimental results demonstrate that our approach yields state-of-the-art transfer learning results for object detection.
arXiv Detail & Related papers (2021-02-16T17:58:57Z) - Slender Object Detection: Diagnoses and Improvements [74.40792217534]
In this paper, we are concerned with the detection of a particular type of objects with extreme aspect ratios, namely textbfslender objects.
For a classical object detection method, a drastic drop of $18.9%$ mAP on COCO is observed, if solely evaluated on slender objects.
arXiv Detail & Related papers (2020-11-17T09:39:42Z) - A Self-Training Approach for Point-Supervised Object Detection and
Counting in Crowds [54.73161039445703]
We propose a novel self-training approach that enables a typical object detector trained only with point-level annotations.
During training, we utilize the available point annotations to supervise the estimation of the center points of objects.
Experimental results show that our approach significantly outperforms state-of-the-art point-supervised methods under both detection and counting tasks.
arXiv Detail & Related papers (2020-07-25T02:14:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.