Region Comparison Network for Interpretable Few-shot Image
Classification
- URL: http://arxiv.org/abs/2009.03558v1
- Date: Tue, 8 Sep 2020 07:29:05 GMT
- Title: Region Comparison Network for Interpretable Few-shot Image
Classification
- Authors: Zhiyu Xue, Lixin Duan, Wen Li, Lin Chen and Jiebo Luo
- Abstract summary: Few-shot image classification has been proposed to effectively use only a limited number of labeled examples to train models for new classes.
We propose a metric learning based method named Region Comparison Network (RCN), which is able to reveal how few-shot learning works.
We also present a new way to generalize the interpretability from the level of tasks to categories.
- Score: 97.97902360117368
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While deep learning has been successfully applied to many real-world computer
vision tasks, training robust classifiers usually requires a large amount of
well-labeled data. However, the annotation is often expensive and
time-consuming. Few-shot image classification has thus been proposed to
effectively use only a limited number of labeled examples to train models for
new classes. Recent works based on transferable metric learning methods have
achieved promising classification performance through learning the similarity
between the features of samples from the query and support sets. However, rare
of them explicitly considers the model interpretability, which can actually be
revealed during the training phase.
For that, in this work, we propose a metric learning based method named
Region Comparison Network (RCN), which is able to reveal how few-shot learning
works as in a neural network as well as to find out specific regions that are
related to each other in images coming from the query and support sets.
Moreover, we also present a visualization strategy named Region Activation
Mapping (RAM) to intuitively explain what our method has learned by visualizing
intermediate variables in our network. We also present a new way to generalize
the interpretability from the level of tasks to categories, which can also be
viewed as a method to find the prototypical parts for supporting the final
decision of our RCN. Extensive experiments on four benchmark datasets clearly
show the effectiveness of our method over existing baselines.
Related papers
- Semantic Representation and Dependency Learning for Multi-Label Image
Recognition [76.52120002993728]
We propose a novel and effective semantic representation and dependency learning (SRDL) framework to learn category-specific semantic representation for each category.
Specifically, we design a category-specific attentional regions (CAR) module to generate channel/spatial-wise attention matrices to guide model.
We also design an object erasing (OE) module to implicitly learn semantic dependency among categories by erasing semantic-aware regions.
arXiv Detail & Related papers (2022-04-08T00:55:15Z) - Experience feedback using Representation Learning for Few-Shot Object
Detection on Aerial Images [2.8560476609689185]
The performance of our method is assessed on DOTA, a large-scale remote sensing images dataset.
It highlights in particular some intrinsic weaknesses for the few-shot object detection task.
arXiv Detail & Related papers (2021-09-27T13:04:53Z) - Dynamic Relevance Learning for Few-Shot Object Detection [6.550840743803705]
We propose a dynamic relevance learning model, which utilizes the relationship between all support images and Region of Interest (RoI) on the query images to construct a dynamic graph convolutional network (GCN)
The proposed model achieves the best overall performance, which shows its effectiveness of learning more generalized features.
arXiv Detail & Related papers (2021-08-04T18:29:42Z) - Exploiting the relationship between visual and textual features in
social networks for image classification with zero-shot deep learning [0.0]
In this work, we propose a classifier ensemble based on the transferable learning capabilities of the CLIP neural network architecture.
Our experiments, based on image classification tasks according to the labels of the Places dataset, are performed by first considering only the visual part.
Considering the associated texts to the images can help to improve the accuracy depending on the goal.
arXiv Detail & Related papers (2021-07-08T10:54:59Z) - Distribution Alignment: A Unified Framework for Long-tail Visual
Recognition [52.36728157779307]
We propose a unified distribution alignment strategy for long-tail visual recognition.
We then introduce a generalized re-weight method in the two-stage learning to balance the class prior.
Our approach achieves the state-of-the-art results across all four recognition tasks with a simple and unified framework.
arXiv Detail & Related papers (2021-03-30T14:09:53Z) - Visualization of Supervised and Self-Supervised Neural Networks via
Attribution Guided Factorization [87.96102461221415]
We develop an algorithm that provides per-class explainability.
In an extensive battery of experiments, we demonstrate the ability of our methods to class-specific visualization.
arXiv Detail & Related papers (2020-12-03T18:48:39Z) - Few-shot Classification via Adaptive Attention [93.06105498633492]
We propose a novel few-shot learning method via optimizing and fast adapting the query sample representation based on very few reference samples.
As demonstrated experimentally, the proposed model achieves state-of-the-art classification results on various benchmark few-shot classification and fine-grained recognition datasets.
arXiv Detail & Related papers (2020-08-06T05:52:59Z) - ReMarNet: Conjoint Relation and Margin Learning for Small-Sample Image
Classification [49.87503122462432]
We introduce a novel neural network termed Relation-and-Margin learning Network (ReMarNet)
Our method assembles two networks of different backbones so as to learn the features that can perform excellently in both of the aforementioned two classification mechanisms.
Experiments on four image datasets demonstrate that our approach is effective in learning discriminative features from a small set of labeled samples.
arXiv Detail & Related papers (2020-06-27T13:50:20Z) - Interpretable Time-series Classification on Few-shot Samples [27.05851877375113]
This paper proposes an interpretable neural-based framework, namely textitDual Prototypical Shapelet Networks (DPSN) for few-shot time-series classification.
DPSN interprets the model from dual granularity: 1) global overview using representative time series samples, and 2) local highlights using discriminative shapelets.
We have derived 18 few-shot TSC datasets from public benchmark datasets and evaluated the proposed method by comparing with baselines.
arXiv Detail & Related papers (2020-06-03T03:47:14Z) - Weakly-supervised Object Localization for Few-shot Learning and
Fine-grained Few-shot Learning [0.5156484100374058]
Few-shot learning aims to learn novel visual categories from very few samples.
We propose a Self-Attention Based Complementary Module (SAC Module) to fulfill the weakly-supervised object localization.
We also produce the activated masks for selecting discriminative deep descriptors for few-shot classification.
arXiv Detail & Related papers (2020-03-02T14:07:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.