Channel DropBlock: An Improved Regularization Method for Fine-Grained
Visual Classification
- URL: http://arxiv.org/abs/2106.03432v1
- Date: Mon, 7 Jun 2021 09:03:02 GMT
- Title: Channel DropBlock: An Improved Regularization Method for Fine-Grained
Visual Classification
- Authors: Yifeng Ding, Shuwei Dong, Yujun Tong, Zhanyu Ma, Bo Xiao, and Haibin
Ling
- Abstract summary: Existing approaches mainly tackle this problem by introducing attention mechanisms to locate the discriminative parts or feature encoding approaches to extract the highly parameterized features in a weakly-supervised fashion.
In this work, we propose a lightweight yet effective regularization method named Channel DropBlock (CDB) in combination with two alternative correlation metrics, to address this problem.
- Score: 58.07257910065007
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Classifying the sub-categories of an object from the same super-category
(e.g., bird) in a fine-grained visual classification (FGVC) task highly relies
on mining multiple discriminative features. Existing approaches mainly tackle
this problem by introducing attention mechanisms to locate the discriminative
parts or feature encoding approaches to extract the highly parameterized
features in a weakly-supervised fashion. In this work, we propose a lightweight
yet effective regularization method named Channel DropBlock (CDB), in
combination with two alternative correlation metrics, to address this problem.
The key idea is to randomly mask out a group of correlated channels during
training to destruct features from co-adaptations and thus enhance feature
representations. Extensive experiments on three benchmark FGVC datasets show
that CDB effectively improves the performance.
Related papers
- Data-free Knowledge Distillation for Fine-grained Visual Categorization [9.969720644789781]
We propose an approach called DFKD-FGVC that extends DFKD to fine-grained visual categorization(FGVC) tasks.
We evaluate our approach on three widely-used FGVC benchmarks (Aircraft, Cars196, and CUB200) and demonstrate its superior performance.
arXiv Detail & Related papers (2024-04-18T09:44:56Z) - Boosting Few-Shot Learning via Attentive Feature Regularization [35.4031662352264]
Few-shot learning (FSL) based on manifold regularization aims to improve the recognition capacity of novel objects with limited training samples.
This paper proposes feature regularization (AFR) which aims to improve the feature representativeness and discriminability.
arXiv Detail & Related papers (2024-03-23T14:36:48Z) - Revisiting Foreground and Background Separation in Weakly-supervised
Temporal Action Localization: A Clustering-based Approach [48.684550829098534]
Weakly-supervised temporal action localization aims to localize action instances in videos with only video-level action labels.
We propose a novel clustering-based F&B separation algorithm.
We evaluate our method on three benchmarks: THUMOS14, ActivityNet v1.2 and v1.3.
arXiv Detail & Related papers (2023-12-21T18:57:12Z) - Efficient Bilateral Cross-Modality Cluster Matching for Unsupervised Visible-Infrared Person ReID [56.573905143954015]
We propose a novel bilateral cluster matching-based learning framework to reduce the modality gap by matching cross-modality clusters.
Under such a supervisory signal, a Modality-Specific and Modality-Agnostic (MSMA) contrastive learning framework is proposed to align features jointly at a cluster-level.
Experiments on the public SYSU-MM01 and RegDB datasets demonstrate the effectiveness of the proposed method.
arXiv Detail & Related papers (2023-05-22T03:27:46Z) - Dynamic Clustering and Cluster Contrastive Learning for Unsupervised
Person Re-identification [29.167783500369442]
Unsupervised Re-ID methods aim at learning robust and discriminative features from unlabeled data.
We propose a dynamic clustering and cluster contrastive learning (DCCC) method.
Experiments on several widely used public datasets validate the effectiveness of our proposed DCCC.
arXiv Detail & Related papers (2023-03-13T01:56:53Z) - Weakly Supervised Video Anomaly Detection Based on Cross-Batch
Clustering Guidance [39.43891080713327]
Weakly supervised video anomaly detection (WSVAD) is a challenging task since only video-level labels are available for training.
We propose a novel WSVAD method based on cross-batch clustering guidance.
arXiv Detail & Related papers (2022-12-16T14:38:30Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - Uncertainty-aware Clustering for Unsupervised Domain Adaptive Object
Re-identification [123.75412386783904]
State-of-the-art object Re-ID approaches adopt clustering algorithms to generate pseudo-labels for the unlabeled target domain.
We propose an uncertainty-aware clustering framework (UCF) for UDA tasks.
Our UCF method consistently achieves state-of-the-art performance in multiple UDA tasks for object Re-ID.
arXiv Detail & Related papers (2021-08-22T09:57:14Z) - Adversarial Feature Augmentation and Normalization for Visual
Recognition [109.6834687220478]
Recent advances in computer vision take advantage of adversarial data augmentation to ameliorate the generalization ability of classification models.
Here, we present an effective and efficient alternative that advocates adversarial augmentation on intermediate feature embeddings.
We validate the proposed approach across diverse visual recognition tasks with representative backbone networks.
arXiv Detail & Related papers (2021-03-22T20:36:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.