Beyond Prototypes: Semantic Anchor Regularization for Better
Representation Learning
- URL: http://arxiv.org/abs/2312.11872v2
- Date: Sun, 4 Feb 2024 14:12:38 GMT
- Title: Beyond Prototypes: Semantic Anchor Regularization for Better
Representation Learning
- Authors: Yanqi Ge, Qiang Nie, Ye Huang, Yong Liu, Chengjie Wang, Feng Zheng,
Wen Li, Lixin Duan
- Abstract summary: One of the ultimate goals of representation learning is to achieve compactness within a class and well-separability between classes.
We propose a novel perspective to use pre-defined class anchors serving as feature centroid to unidirectionally guide feature learning.
The proposed Semantic Anchor Regularization (SAR) can be used in a plug-and-play manner in the existing models.
- Score: 82.29761875805369
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One of the ultimate goals of representation learning is to achieve
compactness within a class and well-separability between classes. Many
outstanding metric-based and prototype-based methods following the
Expectation-Maximization paradigm, have been proposed for this objective.
However, they inevitably introduce biases into the learning process,
particularly with long-tail distributed training data. In this paper, we reveal
that the class prototype is not necessarily to be derived from training
features and propose a novel perspective to use pre-defined class anchors
serving as feature centroid to unidirectionally guide feature learning.
However, the pre-defined anchors may have a large semantic distance from the
pixel features, which prevents them from being directly applied. To address
this issue and generate feature centroid independent from feature learning, a
simple yet effective Semantic Anchor Regularization (SAR) is proposed. SAR
ensures the interclass separability of semantic anchors in the semantic space
by employing a classifier-aware auxiliary cross-entropy loss during training
via disentanglement learning. By pulling the learned features to these semantic
anchors, several advantages can be attained: 1) the intra-class compactness and
naturally inter-class separability, 2) induced bias or errors from feature
learning can be avoided, and 3) robustness to the long-tailed problem. The
proposed SAR can be used in a plug-and-play manner in the existing models.
Extensive experiments demonstrate that the SAR performs better than previous
sophisticated prototype-based methods. The implementation is available at
https://github.com/geyanqi/SAR.
Related papers
- SEER-ZSL: Semantic Encoder-Enhanced Representations for Generalized
Zero-Shot Learning [0.7420433640907689]
Generalized Zero-Shot Learning (GZSL) recognizes unseen classes by transferring knowledge from the seen classes.
This paper introduces a dual strategy to address the generalization gap.
arXiv Detail & Related papers (2023-12-20T15:18:51Z) - Enhancing Few-shot CLIP with Semantic-Aware Fine-Tuning [61.902254546858465]
Methods based on Contrastive Language-Image Pre-training have exhibited promising performance in few-shot adaptation tasks.
We propose fine-tuning the parameters of the attention pooling layer during the training process to encourage the model to focus on task-specific semantics.
arXiv Detail & Related papers (2023-11-08T05:18:57Z) - FeCAM: Exploiting the Heterogeneity of Class Distributions in
Exemplar-Free Continual Learning [21.088762527081883]
Exemplar-free class-incremental learning (CIL) poses several challenges since it prohibits the rehearsal of data from previous tasks.
Recent approaches to incrementally learning the classifier by freezing the feature extractor after the first task have gained much attention.
We explore prototypical networks for CIL, which generate new class prototypes using the frozen feature extractor and classify the features based on the Euclidean distance to the prototypes.
arXiv Detail & Related papers (2023-09-25T11:54:33Z) - RanPAC: Random Projections and Pre-trained Models for Continual Learning [59.07316955610658]
Continual learning (CL) aims to learn different tasks (such as classification) in a non-stationary data stream without forgetting old ones.
We propose a concise and effective approach for CL with pre-trained models.
arXiv Detail & Related papers (2023-07-05T12:49:02Z) - Improving Deep Representation Learning via Auxiliary Learnable Target Coding [69.79343510578877]
This paper introduces a novel learnable target coding as an auxiliary regularization of deep representation learning.
Specifically, a margin-based triplet loss and a correlation consistency loss on the proposed target codes are designed to encourage more discriminative representations.
arXiv Detail & Related papers (2023-05-30T01:38:54Z) - Dual Prototypical Contrastive Learning for Few-shot Semantic
Segmentation [55.339405417090084]
We propose a dual prototypical contrastive learning approach tailored to the few-shot semantic segmentation (FSS) task.
The main idea is to encourage the prototypes more discriminative by increasing inter-class distance while reducing intra-class distance in prototype feature space.
We demonstrate that the proposed dual contrastive learning approach outperforms state-of-the-art FSS methods on PASCAL-5i and COCO-20i datasets.
arXiv Detail & Related papers (2021-11-09T08:14:50Z) - SSA: Semantic Structure Aware Inference for Weakly Pixel-Wise Dense
Predictions without Cost [36.27226683586425]
The semantic structure aware inference (SSA) is proposed to explore the semantic structure information hidden in different stages of the CNN-based network to generate high-quality CAM in the model inference.
The proposed method has the advantage of no parameters and does not need to be trained. Therefore, it can be applied to a wide range of weakly-supervised pixel-wise dense prediction tasks.
arXiv Detail & Related papers (2021-11-05T11:07:21Z) - Prototypical Contrastive Learning of Unsupervised Representations [171.3046900127166]
Prototypical Contrastive Learning (PCL) is an unsupervised representation learning method.
PCL implicitly encodes semantic structures of the data into the learned embedding space.
PCL outperforms state-of-the-art instance-wise contrastive learning methods on multiple benchmarks.
arXiv Detail & Related papers (2020-05-11T09:53:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.