Knowledge Transfer Based Fine-grained Visual Classification
- URL: http://arxiv.org/abs/2012.11389v1
- Date: Mon, 21 Dec 2020 14:41:08 GMT
- Title: Knowledge Transfer Based Fine-grained Visual Classification
- Authors: Siqing Zhang, Ruoyi Du, Dongliang Chang, Zhanyu Ma, Jun Guo
- Abstract summary: Fine-grained visual classification (FGVC) aims to distinguish the sub-classes of the same category.
Its essential solution is to mine the subtle and discriminative regions.
CNNs, which employ the cross entropy loss (CE-loss) as the loss function, show poor performance.
- Score: 19.233180617535492
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Fine-grained visual classification (FGVC) aims to distinguish the sub-classes
of the same category and its essential solution is to mine the subtle and
discriminative regions. Convolution neural networks (CNNs), which employ the
cross entropy loss (CE-loss) as the loss function, show poor performance since
the model can only learn the most discriminative part and ignore other
meaningful regions. Some existing works try to solve this problem by mining
more discriminative regions by some detection techniques or attention
mechanisms. However, most of them will meet the background noise problem when
trying to find more discriminative regions. In this paper, we address it in a
knowledge transfer learning manner. Multiple models are trained one by one, and
all previously trained models are regarded as teacher models to supervise the
training of the current one. Specifically, a orthogonal loss (OR-loss) is
proposed to encourage the network to find diverse and meaningful regions. In
addition, the first model is trained with only CE-Loss. Finally, all models'
outputs with complementary knowledge are combined together for the final
prediction result. We demonstrate the superiority of the proposed method and
obtain state-of-the-art (SOTA) performances on three popular FGVC datasets.
Related papers
- Towards Seamless Adaptation of Pre-trained Models for Visual Place Recognition [72.35438297011176]
We propose a novel method to realize seamless adaptation of pre-trained models for visual place recognition (VPR)
Specifically, to obtain both global and local features that focus on salient landmarks for discriminating places, we design a hybrid adaptation method.
Experimental results show that our method outperforms the state-of-the-art methods with less training data and training time.
arXiv Detail & Related papers (2024-02-22T12:55:01Z) - Fine-grained Recognition with Learnable Semantic Data Augmentation [68.48892326854494]
Fine-grained image recognition is a longstanding computer vision challenge.
We propose diversifying the training data at the feature-level to alleviate the discriminative region loss problem.
Our method significantly improves the generalization performance on several popular classification networks.
arXiv Detail & Related papers (2023-09-01T11:15:50Z) - Few-shot Open-set Recognition Using Background as Unknowns [58.04165813493666]
Few-shot open-set recognition aims to classify both seen and novel images given only limited training data of seen classes.
Our proposed method not only outperforms multiple baselines but also sets new results on three popular benchmarks.
arXiv Detail & Related papers (2022-07-19T04:19:29Z) - Discriminative Feature Learning through Feature Distance Loss [0.0]
This work proposes a novel method that combines variant rich base models to concentrate on different important image regions for classification.
Experiments on benchmark convolutional neural networks (VGG16, ResNet, AlexNet), popular datasets (Cifar10, Cifar100, miniImageNet, NEU, BSD, TEX) show our methods effectiveness and generalization ability.
arXiv Detail & Related papers (2022-05-23T20:01:32Z) - R2-Trans:Fine-Grained Visual Categorization with Redundancy Reduction [21.11038841356125]
Fine-grained visual categorization (FGVC) aims to discriminate similar subcategories, whose main challenge is the large intraclass diversities and subtle inter-class differences.
We present a novel approach for FGVC, which can simultaneously make use of partial yet sufficient discriminative information in environmental cues and also compress the redundant information in class-token with respect to the target.
arXiv Detail & Related papers (2022-04-21T13:35:38Z) - Learning What Not to Segment: A New Perspective on Few-Shot Segmentation [63.910211095033596]
Recently few-shot segmentation (FSS) has been extensively developed.
This paper proposes a fresh and straightforward insight to alleviate the problem.
In light of the unique nature of the proposed approach, we also extend it to a more realistic but challenging setting.
arXiv Detail & Related papers (2022-03-15T03:08:27Z) - Calibrating Class Activation Maps for Long-Tailed Visual Recognition [60.77124328049557]
We present two effective modifications of CNNs to improve network learning from long-tailed distribution.
First, we present a Class Activation Map (CAMC) module to improve the learning and prediction of network classifiers.
Second, we investigate the use of normalized classifiers for representation learning in long-tailed problems.
arXiv Detail & Related papers (2021-08-29T05:45:03Z) - Interpretable Attention Guided Network for Fine-grained Visual
Classification [36.657203916383594]
Fine-grained visual classification (FGVC) is challenging but more critical than traditional classification tasks.
We propose an Interpretable Attention Guided Network (IAGN) for fine-grained visual classification.
arXiv Detail & Related papers (2021-03-08T12:27:51Z) - Learning Granularity-Aware Convolutional Neural Network for Fine-Grained
Visual Classification [0.0]
We propose a novel Granularity-Aware Congrainedal Neural Network (GA-CNN) that progressively explores discriminative features.
GA-CNN does not need bounding boxes/part annotations and can be trained end-to-end.
Our approach achieves state-of-the-art performances on three benchmark datasets.
arXiv Detail & Related papers (2021-03-04T02:18:07Z) - Playing to distraction: towards a robust training of CNN classifiers
through visual explanation techniques [1.2321022105220707]
We present a novel and robust training scheme that integrates visual explanation techniques in the learning process.
In particular, we work on the challenging EgoFoodPlaces dataset, achieving state-of-the-art results with a lower level of complexity.
arXiv Detail & Related papers (2020-12-28T10:24:32Z) - Fine-Grained Visual Classification via Progressive Multi-Granularity
Training of Jigsaw Patches [67.51747235117]
Fine-grained visual classification (FGVC) is much more challenging than traditional classification tasks.
Recent works mainly tackle this problem by focusing on how to locate the most discriminative parts.
We propose a novel framework for fine-grained visual classification to tackle these problems.
arXiv Detail & Related papers (2020-03-08T19:27:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.