Bagging Regional Classification Activation Maps for Weakly Supervised
Object Localization
- URL: http://arxiv.org/abs/2207.07818v1
- Date: Sat, 16 Jul 2022 03:03:01 GMT
- Title: Bagging Regional Classification Activation Maps for Weakly Supervised
Object Localization
- Authors: Lei Zhu, Qian Chen, Lujia Jin, Yunfei You, and Yanye Lu
- Abstract summary: BagCAMs is a plug-and-play mechanism to better project a well-trained classifier for the localization task.
Our BagCAMs adopts a proposed regional localizer generation strategy to define a set of regional localizers.
Experiments indicate that adopting our proposed BagCAMs can improve the performance of baseline WSOL methods.
- Score: 11.25759292976175
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Classification activation map (CAM), utilizing the classification structure
to generate pixel-wise localization maps, is a crucial mechanism for weakly
supervised object localization (WSOL). However, CAM directly uses the
classifier trained on image-level features to locate objects, making it prefers
to discern global discriminative factors rather than regional object cues. Thus
only the discriminative locations are activated when feeding pixel-level
features into this classifier. To solve this issue, this paper elaborates a
plug-and-play mechanism called BagCAMs to better project a well-trained
classifier for the localization task without refining or re-training the
baseline structure. Our BagCAMs adopts a proposed regional localizer generation
(RLG) strategy to define a set of regional localizers and then derive them from
a well-trained classifier. These regional localizers can be viewed as the base
learner that only discerns region-wise object factors for localization tasks,
and their results can be effectively weighted by our BagCAMs to form the final
localization map. Experiments indicate that adopting our proposed BagCAMs can
improve the performance of baseline WSOL methods to a great extent and obtains
state-of-the-art performance on three WSOL benchmarks. Code are released at
https://github.com/zh460045050/BagCAMs.
Related papers
- Background Activation Suppression for Weakly Supervised Object
Localization and Semantic Segmentation [84.62067728093358]
Weakly supervised object localization and semantic segmentation aim to localize objects using only image-level labels.
New paradigm has emerged by generating a foreground prediction map to achieve pixel-level localization.
This paper presents two astonishing experimental observations on the object localization learning process.
arXiv Detail & Related papers (2023-09-22T15:44:10Z) - DETR Doesn't Need Multi-Scale or Locality Design [69.56292005230185]
This paper presents an improved DETR detector that maintains a "plain" nature.
It uses a single-scale feature map and global cross-attention calculations without specific locality constraints.
We show that two simple technologies are surprisingly effective within a plain design to compensate for the lack of multi-scale feature maps and locality constraints.
arXiv Detail & Related papers (2023-08-03T17:59:04Z) - LCPFormer: Towards Effective 3D Point Cloud Analysis via Local Context
Propagation in Transformers [60.51925353387151]
We propose a novel module named Local Context Propagation (LCP) to exploit the message passing between neighboring local regions.
We use the overlap points of adjacent local regions as intermediaries, then re-weight the features of these shared points from different local regions before passing them to the next layers.
The proposed method is applicable to different tasks and outperforms various transformer-based methods in benchmarks including 3D shape classification and dense prediction tasks.
arXiv Detail & Related papers (2022-10-23T15:43:01Z) - Constrained Sampling for Class-Agnostic Weakly Supervised Object
Localization [10.542859578763068]
Self-supervised vision transformers can generate accurate localization maps of the objects in an image.
We propose leveraging the multiple maps generated by the different transformer heads to acquire pseudo-labels for training a weakly-supervised object localization model.
arXiv Detail & Related papers (2022-09-09T19:58:38Z) - CREAM: Weakly Supervised Object Localization via Class RE-Activation
Mapping [18.67907876709536]
Class RE-Activation Mapping (CREAM) is a clustering-based approach to boost the activation values of the integral object regions.
CREAM achieves the state-of-the-art performance on CUB, ILSVRC and OpenImages benchmark datasets.
arXiv Detail & Related papers (2022-05-27T11:57:41Z) - Spatial Likelihood Voting with Self-Knowledge Distillation for Weakly
Supervised Object Detection [54.24966006457756]
We propose a WSOD framework called the Spatial Likelihood Voting with Self-knowledge Distillation Network (SLV-SD Net)
SLV-SD Net converges region proposal localization without bounding box annotations.
Experiments on the PASCAL VOC 2007/2012 and MS-COCO datasets demonstrate the excellent performance of SLV-SD Net.
arXiv Detail & Related papers (2022-04-14T11:56:19Z) - Bridging the Gap between Classification and Localization for Weakly
Supervised Object Localization [39.63778214094173]
Weakly supervised object localization aims to find a target object region in a given image with only weak supervision, such as image-level labels.
We find the gap between classification and localization in terms of the misalignment of the directions between an input feature and a class-specific weight.
We propose a method to align feature directions with a class-specific weight to bridge the gap.
arXiv Detail & Related papers (2022-04-01T05:49:22Z) - Weakly Supervised Object Localization as Domain Adaption [19.854125742336688]
Weakly supervised object localization (WSOL) focuses on localizing objects only with the supervision of image-level classification masks.
Most previous WSOL methods follow the classification activation map (CAM) that localizes objects based on the classification structure with the multi-instance learning (MIL) mechanism.
This work provides a novel perspective that models WSOL as a domain adaption (DA) task, where the score estimator trained on the source/image domain is tested on the target/pixel domain to locate objects.
arXiv Detail & Related papers (2022-03-03T13:50:22Z) - Unveiling the Potential of Structure-Preserving for Weakly Supervised
Object Localization [71.79436685992128]
We propose a two-stage approach, termed structure-preserving activation (SPA), towards fully leveraging the structure information incorporated in convolutional features for WSOL.
In the first stage, a restricted activation module (RAM) is designed to alleviate the structure-missing issue caused by the classification network.
In the second stage, we propose a post-process approach, termed self-correlation map generating (SCG) module to obtain structure-preserving localization maps.
arXiv Detail & Related papers (2021-03-08T03:04:14Z) - Local Context Attention for Salient Object Segmentation [5.542044768017415]
We propose a novel Local Context Attention Network (LCANet) to generate locally reinforcement feature maps in a uniform representational architecture.
The proposed network introduces an Attentional Correlation Filter (ACF) module to generate explicit local attention by calculating the correlation feature map between coarse prediction and global context.
Comprehensive experiments are conducted on several salient object segmentation datasets, demonstrating the superior performance of the proposed LCANet against the state-of-the-art methods.
arXiv Detail & Related papers (2020-09-24T09:20:06Z) - Rethinking Localization Map: Towards Accurate Object Perception with
Self-Enhancement Maps [78.2581910688094]
This work introduces a novel self-enhancement method to harvest accurate object localization maps and object boundaries with only category labels as supervision.
In particular, the proposed Self-Enhancement Maps achieve the state-of-the-art localization accuracy of 54.88% on ILSVRC.
arXiv Detail & Related papers (2020-06-09T12:35:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.