Out-of-distribution Few-shot Learning For Edge Devices without Model
Fine-tuning
- URL: http://arxiv.org/abs/2304.06309v1
- Date: Thu, 13 Apr 2023 07:33:22 GMT
- Title: Out-of-distribution Few-shot Learning For Edge Devices without Model
Fine-tuning
- Authors: Xinyun Zhang and Lanqing Hong
- Abstract summary: Few-shot learning is a promising technique to achieve personalized user experiences on edge devices.
This paper proposes a plug-and-play module called Task-aware Normalization (TANO) that enables efficient and task-aware adaptation of a deep neural network without backpropagation.
TANO provides stable but task-specific estimations of the normalization statistics to close the distribution gaps and achieve efficient model adaptation.
- Score: 10.422316867474681
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot learning (FSL) via customization of a deep learning network with
limited data has emerged as a promising technique to achieve personalized user
experiences on edge devices. However, existing FSL methods primarily assume
independent and identically distributed (IID) data and utilize either
computational backpropagation updates for each task or a common model with
task-specific prototypes. Unfortunately, the former solution is infeasible for
edge devices that lack on-device backpropagation capabilities, while the latter
often struggles with limited generalization ability, especially for
out-of-distribution (OOD) data. This paper proposes a lightweight,
plug-and-play FSL module called Task-aware Normalization (TANO) that enables
efficient and task-aware adaptation of a deep neural network without
backpropagation. TANO covers the properties of multiple user groups by
coordinating the updates of several groups of the normalization statistics
during meta-training and automatically identifies the appropriate normalization
group for a downstream few-shot task. Consequently, TANO provides stable but
task-specific estimations of the normalization statistics to close the
distribution gaps and achieve efficient model adaptation. Results on both
intra-domain and out-of-domain generalization experiments demonstrate that TANO
outperforms recent methods in terms of accuracy, inference speed, and model
size. Moreover, TANO achieves promising results on widely-used FSL benchmarks
and data from real applications.
Related papers
- Uncertainty Aware Learning for Language Model Alignment [97.36361196793929]
We propose uncertainty-aware learning (UAL) to improve the model alignment of different task scenarios.
We implement UAL in a simple fashion -- adaptively setting the label smoothing value of training according to the uncertainty of individual samples.
Experiments on widely used benchmarks demonstrate that our UAL significantly and consistently outperforms standard supervised fine-tuning.
arXiv Detail & Related papers (2024-06-07T11:37:45Z) - PeFAD: A Parameter-Efficient Federated Framework for Time Series Anomaly Detection [51.20479454379662]
We propose a.
Federated Anomaly Detection framework named PeFAD with the increasing privacy concerns.
We conduct extensive evaluations on four real datasets, where PeFAD outperforms existing state-of-the-art baselines by up to 28.74%.
arXiv Detail & Related papers (2024-06-04T13:51:08Z) - Dual-Personalizing Adapter for Federated Foundation Models [35.863585349109385]
We propose a new setting, termed test-time personalization, which concentrates on the targeted local task and extends to other tasks that exhibit test-time distribution shifts.
Specifically, a dual-personalizing adapter architecture (FedDPA) is proposed, comprising a global adapter and a local adapter for addressing test-time distribution shifts and personalization.
The effectiveness of the proposed method has been evaluated on benchmark datasets across different NLP tasks.
arXiv Detail & Related papers (2024-03-28T08:19:33Z) - Adaptive Test-Time Personalization for Federated Learning [51.25437606915392]
We introduce a novel setting called test-time personalized federated learning (TTPFL)
In TTPFL, clients locally adapt a global model in an unsupervised way without relying on any labeled data during test-time.
We propose a novel algorithm called ATP to adaptively learn the adaptation rates for each module in the model from distribution shifts among source domains.
arXiv Detail & Related papers (2023-10-28T20:42:47Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - Exploring Complementary Strengths of Invariant and Equivariant
Representations for Few-Shot Learning [96.75889543560497]
In many real-world problems, collecting a large number of labeled samples is infeasible.
Few-shot learning is the dominant approach to address this issue, where the objective is to quickly adapt to novel categories in presence of a limited number of samples.
We propose a novel training mechanism that simultaneously enforces equivariance and invariance to a general set of geometric transformations.
arXiv Detail & Related papers (2021-03-01T21:14:33Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.