MHFC: Multi-Head Feature Collaboration for Few-Shot Learning
- URL: http://arxiv.org/abs/2109.07785v1
- Date: Thu, 16 Sep 2021 08:09:35 GMT
- Title: MHFC: Multi-Head Feature Collaboration for Few-Shot Learning
- Authors: Shuai Shao, Lei Xing, Yan Wang, Rui Xu, Chunyan Zhao, Yan-Jiang Wang,
Bao-Di Liu
- Abstract summary: Few-shot learning aims to address the data-scarce problem.
We propose Multi-Head Feature Collaboration (MHFC) algorithm, which attempts to project the multi-head features to a unified space.
We evaluate the proposed method on five benchmark datasets and achieve significant improvements of 2.1%-7.8% compared with state-of-the-arts.
- Score: 17.699793591135904
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Few-shot learning (FSL) aims to address the data-scarce problem. A standard
FSL framework is composed of two components: (1) Pre-train. Employ the base
data to generate a CNN-based feature extraction model (FEM). (2) Meta-test.
Apply the trained FEM to acquire the novel data's features and recognize them.
FSL relies heavily on the design of the FEM. However, various FEMs have
distinct emphases. For example, several may focus more attention on the contour
information, whereas others may lay particular emphasis on the texture
information. The single-head feature is only a one-sided representation of the
sample. Besides the negative influence of cross-domain (e.g., the trained FEM
can not adapt to the novel class flawlessly), the distribution of novel data
may have a certain degree of deviation compared with the ground truth
distribution, which is dubbed as distribution-shift-problem (DSP). To address
the DSP, we propose Multi-Head Feature Collaboration (MHFC) algorithm, which
attempts to project the multi-head features (e.g., multiple features extracted
from a variety of FEMs) to a unified space and fuse them to capture more
discriminative information. Typically, first, we introduce a subspace learning
method to transform the multi-head features to aligned low-dimensional
representations. It corrects the DSP via learning the feature with more
powerful discrimination and overcomes the problem of inconsistent measurement
scales from different head features. Then, we design an attention block to
update combination weights for each head feature automatically. It
comprehensively considers the contribution of various perspectives and further
improves the discrimination of features. We evaluate the proposed method on
five benchmark datasets (including cross-domain experiments) and achieve
significant improvements of 2.1%-7.8% compared with state-of-the-arts.
Related papers
- Multi-scale Quaternion CNN and BiGRU with Cross Self-attention Feature Fusion for Fault Diagnosis of Bearing [5.3598912592106345]
Deep learning has led to significant advances in bearing fault diagnosis (FD)
We propose a novel FD model by integrating multiscale quaternion convolutional neural network (MQCNN), bidirectional gated recurrent unit (BiG), and cross self-attention feature fusion (CSAFF)
arXiv Detail & Related papers (2024-05-25T07:55:02Z) - Class-Imbalanced Semi-Supervised Learning for Large-Scale Point Cloud
Semantic Segmentation via Decoupling Optimization [64.36097398869774]
Semi-supervised learning (SSL) has been an active research topic for large-scale 3D scene understanding.
The existing SSL-based methods suffer from severe training bias due to class imbalance and long-tail distributions of the point cloud data.
We introduce a new decoupling optimization framework, which disentangles feature representation learning and classifier in an alternative optimization manner to shift the bias decision boundary effectively.
arXiv Detail & Related papers (2024-01-13T04:16:40Z) - Dual Compensation Residual Networks for Class Imbalanced Learning [98.35401757647749]
We propose Dual Compensation Residual Networks to better fit both tail and head classes.
An important factor causing overfitting is that there is severe feature drift between training and test data on tail classes.
We also propose a Residual Balanced Multi-Proxies classifier to alleviate the under-fitting issue.
arXiv Detail & Related papers (2023-08-25T04:06:30Z) - FF2: A Feature Fusion Two-Stream Framework for Punctuation Restoration [27.14686854704104]
We propose a Feature Fusion two-stream framework (FF2) for punctuation restoration.
Specifically, one stream leverages a pre-trained language model to capture the semantic feature, while another auxiliary module captures the feature at hand.
Without additional data, the experimental results on the popular benchmark IWSLT demonstrate that FF2 achieves new SOTA performance.
arXiv Detail & Related papers (2022-11-09T06:18:17Z) - GCT: Graph Co-Training for Semi-Supervised Few-Shot Learning [46.07923908746946]
We propose a novel label prediction method, Isolated Graph Learning (IGL)
IGL introduces the Laplacian operator to encode the raw data to graph space, which helps reduce the dependence on features when classifying.
Second, we propose Graph Co-Training (GCT) to tackle this challenge from a multi-modal fusion perspective.
arXiv Detail & Related papers (2022-03-15T09:13:35Z) - Transformers Can Do Bayesian Inference [56.99390658880008]
We present Prior-Data Fitted Networks (PFNs)
PFNs leverage in-context learning in large-scale machine learning techniques to approximate a large set of posteriors.
We demonstrate that PFNs can near-perfectly mimic Gaussian processes and also enable efficient Bayesian inference for intractable problems.
arXiv Detail & Related papers (2021-12-20T13:07:39Z) - MDFM: Multi-Decision Fusing Model for Few-Shot Learning [16.47647579893923]
We propose a novel method Multi-Decision Fusing Model (MDFM) to enhance the efficacy and robustness of the model.
We evaluate the proposed method on five benchmark datasets and achieve significant improvements of 3.4%-7.3% compared with state-of-the-arts.
arXiv Detail & Related papers (2021-12-01T18:13:09Z) - Exploring Complementary Strengths of Invariant and Equivariant
Representations for Few-Shot Learning [96.75889543560497]
In many real-world problems, collecting a large number of labeled samples is infeasible.
Few-shot learning is the dominant approach to address this issue, where the objective is to quickly adapt to novel categories in presence of a limited number of samples.
We propose a novel training mechanism that simultaneously enforces equivariance and invariance to a general set of geometric transformations.
arXiv Detail & Related papers (2021-03-01T21:14:33Z) - Deep F-measure Maximization for End-to-End Speech Understanding [52.36496114728355]
We propose a differentiable approximation to the F-measure and train the network with this objective using standard backpropagation.
We perform experiments on two standard fairness datasets, Adult, Communities and Crime, and also on speech-to-intent detection on the ATIS dataset and speech-to-image concept classification on the Speech-COCO dataset.
In all four of these tasks, F-measure results in improved micro-F1 scores, with absolute improvements of up to 8% absolute, as compared to models trained with the cross-entropy loss function.
arXiv Detail & Related papers (2020-08-08T03:02:27Z) - TAFSSL: Task-Adaptive Feature Sub-Space Learning for few-shot
classification [50.358839666165764]
We show that the Task-Adaptive Feature Sub-Space Learning (TAFSSL) can significantly boost the performance in Few-Shot Learning scenarios.
Specifically, we show that on the challenging miniImageNet and tieredImageNet benchmarks, TAFSSL can improve the current state-of-the-art in both transductive and semi-supervised FSL settings by more than $5%$.
arXiv Detail & Related papers (2020-03-14T16:59:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.