Learning a Weight Map for Weakly-Supervised Localization
- URL: http://arxiv.org/abs/2111.14131v1
- Date: Sun, 28 Nov 2021 12:45:23 GMT
- Title: Learning a Weight Map for Weakly-Supervised Localization
- Authors: Tal Shaharabany and Lior Wolf
- Abstract summary: We train a generative network $g$ that outputs, given the input image, a per-pixel weight map that indicates the location of the object within the image.
Our results indicate that the method outperforms existing localization methods by a sizable margin on the challenging fine-grained classification datasets.
- Score: 93.91375268580806
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the weakly supervised localization setting, supervision is given as an
image-level label. We propose to employ an image classifier $f$ and to train a
generative network $g$ that outputs, given the input image, a per-pixel weight
map that indicates the location of the object within the image. Network $g$ is
trained by minimizing the discrepancy between the output of the classifier $f$
on the original image and its output given the same image weighted by the
output of $g$. The scheme requires a regularization term that ensures that $g$
does not provide a uniform weight, and an early stopping criterion in order to
prevent $g$ from over-segmenting the image. Our results indicate that the
method outperforms existing localization methods by a sizable margin on the
challenging fine-grained classification datasets, as well as a generic image
recognition dataset. Additionally, the obtained weight map is also
state-of-the-art in weakly supervised segmentation in fine-grained
categorization datasets.
Related papers
- Rethinking Generalization in Few-Shot Classification [28.809141478504532]
Single image-level annotations only correctly describe an often small subset of an image's content.
In this paper, we take a closer look at the implications in the context of $textitfew-shot learning$.
We build on recent advances in unsupervised training of networks via masked image modelling to overcome the lack of fine-grained labels.
arXiv Detail & Related papers (2022-06-15T03:05:21Z) - LEAD: Self-Supervised Landmark Estimation by Aligning Distributions of
Feature Similarity [49.84167231111667]
Existing works in self-supervised landmark detection are based on learning dense (pixel-level) feature representations from an image.
We introduce an approach to enhance the learning of dense equivariant representations in a self-supervised fashion.
We show that having such a prior in the feature extractor helps in landmark detection, even under drastically limited number of annotations.
arXiv Detail & Related papers (2022-04-06T17:48:18Z) - Local contrastive loss with pseudo-label based self-training for
semi-supervised medical image segmentation [13.996217500923413]
Semi/self-supervised learning-based approaches exploit unlabeled data along with limited annotated data.
Recent self-supervised learning methods use contrastive loss to learn good global level representations from unlabeled images.
We propose a local contrastive loss to learn good pixel level features useful for segmentation by exploiting semantic label information.
arXiv Detail & Related papers (2021-12-17T17:38:56Z) - Global and Local Alignment Networks for Unpaired Image-to-Image
Translation [170.08142745705575]
The goal of unpaired image-to-image translation is to produce an output image reflecting the target domain's style.
Due to the lack of attention to the content change in existing methods, semantic information from source images suffers from degradation during translation.
We introduce a novel approach, Global and Local Alignment Networks (GLA-Net)
Our method effectively generates sharper and more realistic images than existing approaches.
arXiv Detail & Related papers (2021-11-19T18:01:54Z) - Region-level Active Learning for Cluttered Scenes [60.93811392293329]
We introduce a new strategy that subsumes previous Image-level and Object-level approaches into a generalized, Region-level approach.
We show that this approach significantly decreases labeling effort and improves rare object search on realistic data with inherent class-imbalance and cluttered scenes.
arXiv Detail & Related papers (2021-08-20T14:02:38Z) - Semi-weakly Supervised Contrastive Representation Learning for Retinal
Fundus Images [0.2538209532048867]
We propose a semi-weakly supervised contrastive learning framework for representation learning using semi-weakly annotated images.
We empirically validate the transfer learning performance of SWCL on seven public retinal fundus datasets.
arXiv Detail & Related papers (2021-08-04T15:50:09Z) - Instance Localization for Self-supervised Detection Pretraining [68.24102560821623]
We propose a new self-supervised pretext task, called instance localization.
We show that integration of bounding boxes into pretraining promotes better task alignment and architecture alignment for transfer learning.
Experimental results demonstrate that our approach yields state-of-the-art transfer learning results for object detection.
arXiv Detail & Related papers (2021-02-16T17:58:57Z) - Grafit: Learning fine-grained image representations with coarse labels [114.17782143848315]
This paper tackles the problem of learning a finer representation than the one provided by training labels.
By jointly leveraging the coarse labels and the underlying fine-grained latent space, it significantly improves the accuracy of category-level retrieval methods.
arXiv Detail & Related papers (2020-11-25T19:06:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.