Background Activation Suppression for Weakly Supervised Object
Localization
- URL: http://arxiv.org/abs/2112.00580v1
- Date: Wed, 1 Dec 2021 15:53:40 GMT
- Title: Background Activation Suppression for Weakly Supervised Object
Localization
- Authors: Pingyu Wu, Wei Zhai, Yang Cao
- Abstract summary: We argue for using activation value to achieve more efficient learning.
In this paper, we propose a Background Activation Suppression (BAS) method.
BAS achieves significant and consistent improvement over the baseline methods on the CUB-200-2011 and ILSVRC datasets.
- Score: 11.31345656299108
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Weakly supervised object localization (WSOL) aims to localize the object
region using only image-level labels as supervision. Recently a new paradigm
has emerged by generating a foreground prediction map (FPM) to achieve the
localization task. Existing FPM-based methods use cross-entropy (CE) to
evaluate the foreground prediction map and to guide the learning of generator.
We argue for using activation value to achieve more efficient learning. It is
based on the experimental observation that, for a trained network, CE converges
to zero when the foreground mask covers only part of the object region. While
activation value increases until the mask expands to the object boundary, which
indicates that more object areas can be learned by using activation value. In
this paper, we propose a Background Activation Suppression (BAS) method.
Specifically, an Activation Map Constraint module (AMC) is designed to
facilitate the learning of generator by suppressing the background activation
values. Meanwhile, by using the foreground region guidance and the area
constraint, BAS can learn the whole region of the object. Furthermore, in the
inference phase, we consider the prediction maps of different categories
together to obtain the final localization results. Extensive experiments show
that BAS achieves significant and consistent improvement over the baseline
methods on the CUB-200-2011 and ILSVRC datasets.
Related papers
- Background Activation Suppression for Weakly Supervised Object
Localization and Semantic Segmentation [84.62067728093358]
Weakly supervised object localization and semantic segmentation aim to localize objects using only image-level labels.
New paradigm has emerged by generating a foreground prediction map to achieve pixel-level localization.
This paper presents two astonishing experimental observations on the object localization learning process.
arXiv Detail & Related papers (2023-09-22T15:44:10Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - MOST: Multiple Object localization with Self-supervised Transformers for
object discovery [97.47075050779085]
We present Multiple Object localization with Self-supervised Transformers (MOST)
MOST uses features of transformers trained using self-supervised learning to localize multiple objects in real world images.
We show MOST can be used for self-supervised pre-training of object detectors, and yields consistent improvements on fully, semi-supervised object detection and unsupervised region proposal generation.
arXiv Detail & Related papers (2023-04-11T17:57:27Z) - Constrained Sampling for Class-Agnostic Weakly Supervised Object
Localization [10.542859578763068]
Self-supervised vision transformers can generate accurate localization maps of the objects in an image.
We propose leveraging the multiple maps generated by the different transformer heads to acquire pseudo-labels for training a weakly-supervised object localization model.
arXiv Detail & Related papers (2022-09-09T19:58:38Z) - CREAM: Weakly Supervised Object Localization via Class RE-Activation
Mapping [18.67907876709536]
Class RE-Activation Mapping (CREAM) is a clustering-based approach to boost the activation values of the integral object regions.
CREAM achieves the state-of-the-art performance on CUB, ILSVRC and OpenImages benchmark datasets.
arXiv Detail & Related papers (2022-05-27T11:57:41Z) - Spatial Likelihood Voting with Self-Knowledge Distillation for Weakly
Supervised Object Detection [54.24966006457756]
We propose a WSOD framework called the Spatial Likelihood Voting with Self-knowledge Distillation Network (SLV-SD Net)
SLV-SD Net converges region proposal localization without bounding box annotations.
Experiments on the PASCAL VOC 2007/2012 and MS-COCO datasets demonstrate the excellent performance of SLV-SD Net.
arXiv Detail & Related papers (2022-04-14T11:56:19Z) - Online Refinement of Low-level Feature Based Activation Map for Weakly
Supervised Object Localization [15.665479740413229]
We present a two-stage learning framework for weakly supervised object localization (WSOL)
In the first stage, an activation map generator produces activation maps based on the low-level feature maps in the classifier.
In the second stage, we employ an evaluator to evaluate the activation maps predicted by the activation map generator.
Based on the low-level object information preserved in the first stage, the second stage model gradually generates a well-separated, complete, and compact activation map of object in the image.
arXiv Detail & Related papers (2021-10-12T05:09:21Z) - Unveiling the Potential of Structure-Preserving for Weakly Supervised
Object Localization [71.79436685992128]
We propose a two-stage approach, termed structure-preserving activation (SPA), towards fully leveraging the structure information incorporated in convolutional features for WSOL.
In the first stage, a restricted activation module (RAM) is designed to alleviate the structure-missing issue caused by the classification network.
In the second stage, we propose a post-process approach, termed self-correlation map generating (SCG) module to obtain structure-preserving localization maps.
arXiv Detail & Related papers (2021-03-08T03:04:14Z) - Rethinking Localization Map: Towards Accurate Object Perception with
Self-Enhancement Maps [78.2581910688094]
This work introduces a novel self-enhancement method to harvest accurate object localization maps and object boundaries with only category labels as supervision.
In particular, the proposed Self-Enhancement Maps achieve the state-of-the-art localization accuracy of 54.88% on ILSVRC.
arXiv Detail & Related papers (2020-06-09T12:35:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.