MDFM: Multi-Decision Fusing Model for Few-Shot Learning
- URL: http://arxiv.org/abs/2112.00690v2
- Date: Fri, 3 Dec 2021 08:41:40 GMT
- Title: MDFM: Multi-Decision Fusing Model for Few-Shot Learning
- Authors: Shuai Shao, Lei Xing, Rui Xu, Weifeng Liu, Yan-Jiang Wang, Bao-Di Liu
- Abstract summary: We propose a novel method Multi-Decision Fusing Model (MDFM) to enhance the efficacy and robustness of the model.
We evaluate the proposed method on five benchmark datasets and achieve significant improvements of 3.4%-7.3% compared with state-of-the-arts.
- Score: 16.47647579893923
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, researchers pay growing attention to the few-shot learning
(FSL) task to address the data-scarce problem. A standard FSL framework is
composed of two components: i) Pre-train. Employ the base data to generate a
CNN-based feature extraction model (FEM). ii) Meta-test. Apply the trained FEM
to the novel data (category is different from base data) to acquire the feature
embeddings and recognize them. Although researchers have made remarkable
breakthroughs in FSL, there still exists a fundamental problem. Since the
trained FEM with base data usually cannot adapt to the novel class flawlessly,
the novel data's feature may lead to the distribution shift problem. To address
this challenge, we hypothesize that even if most of the decisions based on
different FEMs are viewed as weak decisions, which are not available for all
classes, they still perform decently in some specific categories. Inspired by
this assumption, we propose a novel method Multi-Decision Fusing Model (MDFM),
which comprehensively considers the decisions based on multiple FEMs to enhance
the efficacy and robustness of the model. MDFM is a simple, flexible,
non-parametric method that can directly apply to the existing FEMs. Besides, we
extend the proposed MDFM to two FSL settings (i.e., supervised and
semi-supervised settings). We evaluate the proposed method on five benchmark
datasets and achieve significant improvements of 3.4%-7.3% compared with
state-of-the-arts.
Related papers
- Enhancing Information Maximization with Distance-Aware Contrastive
Learning for Source-Free Cross-Domain Few-Shot Learning [55.715623885418815]
Cross-Domain Few-Shot Learning methods require access to source domain data to train a model in the pre-training phase.
Due to increasing concerns about data privacy and the desire to reduce data transmission and training costs, it is necessary to develop a CDFSL solution without accessing source data.
This paper proposes an Enhanced Information Maximization with Distance-Aware Contrastive Learning method to address these challenges.
arXiv Detail & Related papers (2024-03-04T12:10:24Z) - Continual Learning with Pre-Trained Models: A Survey [61.97613090666247]
Continual Learning aims to overcome the catastrophic forgetting of former knowledge when learning new ones.
This paper presents a comprehensive survey of the latest advancements in PTM-based CL.
arXiv Detail & Related papers (2024-01-29T18:27:52Z) - Learning-Based Difficulty Calibration for Enhanced Membership Inference Attacks [3.470379197911889]
Membership Inference Attacks (MIA) allows adversaries to determine whether a specific data point was part of a model's training dataset.
We present a novel approach to MIA that is aimed at significantly improving TPR at low False Positive Rate (FPR)
Experiment results show that LDC-MIA can improve TPR at low FPR by up to 4x compared to the other difficulty calibration based MIAs.
arXiv Detail & Related papers (2024-01-10T04:58:17Z) - A Survey on Efficient Federated Learning Methods for Foundation Model Training [62.473245910234304]
Federated Learning (FL) has become an established technique to facilitate privacy-preserving collaborative training across a multitude of clients.
In the wake of Foundation Models (FM), the reality is different for many deep learning applications.
We discuss the benefits and drawbacks of parameter-efficient fine-tuning (PEFT) for FL applications.
arXiv Detail & Related papers (2024-01-09T10:22:23Z) - FedDAT: An Approach for Foundation Model Finetuning in Multi-Modal
Heterogeneous Federated Learning [37.96957782129352]
We propose a finetuning framework tailored to heterogeneous multi-modal foundation models, called Federated Dual-Aadapter Teacher (Fed DAT)
Fed DAT addresses data heterogeneity by regularizing the client local updates and applying Mutual Knowledge Distillation (MKD) for an efficient knowledge transfer.
To demonstrate its effectiveness, we conduct extensive experiments on four multi-modality FL benchmarks with different types of data heterogeneity.
arXiv Detail & Related papers (2023-08-21T21:57:01Z) - GCT: Graph Co-Training for Semi-Supervised Few-Shot Learning [46.07923908746946]
We propose a novel label prediction method, Isolated Graph Learning (IGL)
IGL introduces the Laplacian operator to encode the raw data to graph space, which helps reduce the dependence on features when classifying.
Second, we propose Graph Co-Training (GCT) to tackle this challenge from a multi-modal fusion perspective.
arXiv Detail & Related papers (2022-03-15T09:13:35Z) - A Strong Baseline for Semi-Supervised Incremental Few-Shot Learning [54.617688468341704]
Few-shot learning aims to learn models that generalize to novel classes with limited training samples.
We propose a novel paradigm containing two parts: (1) a well-designed meta-training algorithm for mitigating ambiguity between base and novel classes caused by unreliable pseudo labels and (2) a model adaptation mechanism to learn discriminative features for novel classes while preserving base knowledge using few labeled and all the unlabeled data.
arXiv Detail & Related papers (2021-10-21T13:25:52Z) - MHFC: Multi-Head Feature Collaboration for Few-Shot Learning [17.699793591135904]
Few-shot learning aims to address the data-scarce problem.
We propose Multi-Head Feature Collaboration (MHFC) algorithm, which attempts to project the multi-head features to a unified space.
We evaluate the proposed method on five benchmark datasets and achieve significant improvements of 2.1%-7.8% compared with state-of-the-arts.
arXiv Detail & Related papers (2021-09-16T08:09:35Z) - Exploring Complementary Strengths of Invariant and Equivariant
Representations for Few-Shot Learning [96.75889543560497]
In many real-world problems, collecting a large number of labeled samples is infeasible.
Few-shot learning is the dominant approach to address this issue, where the objective is to quickly adapt to novel categories in presence of a limited number of samples.
We propose a novel training mechanism that simultaneously enforces equivariance and invariance to a general set of geometric transformations.
arXiv Detail & Related papers (2021-03-01T21:14:33Z) - FedSemi: An Adaptive Federated Semi-Supervised Learning Framework [23.90642104477983]
Federated learning (FL) has emerged as an effective technique to co-training machine learning models without actually sharing data and leaking privacy.
Most existing FL methods focus on the supervised setting and ignore the utilization of unlabeled data.
We propose FedSemi, a novel, adaptive, and general framework, which firstly introduces the consistency regularization into FL using a teacher-student model.
arXiv Detail & Related papers (2020-12-06T15:46:04Z) - Deep F-measure Maximization for End-to-End Speech Understanding [52.36496114728355]
We propose a differentiable approximation to the F-measure and train the network with this objective using standard backpropagation.
We perform experiments on two standard fairness datasets, Adult, Communities and Crime, and also on speech-to-intent detection on the ATIS dataset and speech-to-image concept classification on the Speech-COCO dataset.
In all four of these tasks, F-measure results in improved micro-F1 scores, with absolute improvements of up to 8% absolute, as compared to models trained with the cross-entropy loss function.
arXiv Detail & Related papers (2020-08-08T03:02:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.