AMF: Adaptable Weighting Fusion with Multiple Fine-tuning for Image
Classification
- URL: http://arxiv.org/abs/2207.12944v1
- Date: Tue, 26 Jul 2022 14:50:03 GMT
- Title: AMF: Adaptable Weighting Fusion with Multiple Fine-tuning for Image
Classification
- Authors: Xuyang Shen, Jo Plested, Sabrina Caldwell, Yiran Zhong and Tom Gedeon
- Abstract summary: We propose the Adaptable Multi-tuning method, which adaptively determines each data sample's fine-tuning strategy.
Our method outperforms the standard fine-tuning approach by 1.69%, 2.79% on the datasets FGVC-Aircraft, and Describable Texture.
- Score: 14.05052135034412
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Fine-tuning is widely applied in image classification tasks as a transfer
learning approach. It re-uses the knowledge from a source task to learn and
obtain a high performance in target tasks. Fine-tuning is able to alleviate the
challenge of insufficient training data and expensive labelling of new data.
However, standard fine-tuning has limited performance in complex data
distributions. To address this issue, we propose the Adaptable Multi-tuning
method, which adaptively determines each data sample's fine-tuning strategy. In
this framework, multiple fine-tuning settings and one policy network are
defined. The policy network in Adaptable Multi-tuning can dynamically adjust to
an optimal weighting to feed different samples into models that are trained
using different fine-tuning strategies. Our method outperforms the standard
fine-tuning approach by 1.69%, 2.79% on the datasets FGVC-Aircraft, and
Describable Texture, yielding comparable performance on the datasets Stanford
Cars, CIFAR-10, and Fashion-MNIST.
Related papers
- Adapt-$\infty$: Scalable Lifelong Multimodal Instruction Tuning via Dynamic Data Selection [89.42023974249122]
Adapt-$infty$ is a new multi-way and adaptive data selection approach for Lifelong Instruction Tuning.
We construct pseudo-skill clusters by grouping gradient-based sample vectors.
We select the best-performing data selector for each skill cluster from a pool of selector experts.
arXiv Detail & Related papers (2024-10-14T15:48:09Z) - AdaNAT: Exploring Adaptive Policy for Token-Based Image Generation [65.01527698201956]
Non-autoregressive Transformers (NATs) are able to synthesize images with decent quality in a small number of steps.
We propose AdaNAT, a learnable approach that automatically configures a suitable policy tailored for every sample to be generated.
arXiv Detail & Related papers (2024-08-31T03:53:57Z) - Comparative Analysis of Different Efficient Fine Tuning Methods of Large Language Models (LLMs) in Low-Resource Setting [0.0]
We try to push the understanding of different fine-tuning strategies for large language models (LLMs)
We compare state-of-the-art methods like vanilla fine-tuning and Pattern-Based Fine-Tuning (PBFT) on pre-trained models across two datasets, COLA and MNLI.
Our findings suggest that these alternative strategies can exhibit out-of-domain generalization comparable to that of vanilla FT and PBFT.
arXiv Detail & Related papers (2024-05-21T20:08:52Z) - AutoFT: Learning an Objective for Robust Fine-Tuning [60.641186718253735]
Foundation models encode rich representations that can be adapted to downstream tasks by fine-tuning.
Current approaches to robust fine-tuning use hand-crafted regularization techniques.
We propose AutoFT, a data-driven approach for robust fine-tuning.
arXiv Detail & Related papers (2024-01-18T18:58:49Z) - SLoRA: Federated Parameter Efficient Fine-Tuning of Language Models [28.764782216513037]
Federated Learning (FL) can benefit from distributed and private data of the FL edge clients for fine-tuning.
We propose a method called SLoRA, which overcomes the key limitations of LoRA in high heterogeneous data scenarios.
Our experimental results demonstrate that SLoRA achieves performance comparable to full fine-tuning.
arXiv Detail & Related papers (2023-08-12T10:33:57Z) - Adaptable Text Matching via Meta-Weight Regulator [14.619068650513917]
Meta-Weight Regulator (MWR) is a meta-learning approach that learns to assign weights to the source examples based on their relevance to the target loss.
MWR first trains the model on uniformly weighted source examples, and measures the efficacy of the model on the target examples via a loss function.
As MWR is model-agnostic, it can be applied to any backbone neural model.
arXiv Detail & Related papers (2022-04-27T02:28:40Z) - Task Adaptive Parameter Sharing for Multi-Task Learning [114.80350786535952]
Adaptive Task Adapting Sharing (TAPS) is a method for tuning a base model to a new task by adaptively modifying a small, task-specific subset of layers.
Compared to other methods, TAPS retains high accuracy on downstream tasks while introducing few task-specific parameters.
We evaluate our method on a suite of fine-tuning tasks and architectures (ResNet, DenseNet, ViT) and show that it achieves state-of-the-art performance while being simple to implement.
arXiv Detail & Related papers (2022-03-30T23:16:07Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - Towards Efficient and Data Agnostic Image Classification Training
Pipeline for Embedded Systems [0.0]
This work is focusing on reviewing the latest augmentation and regularization methods for the image classification.
We can achieve a reasonable performance on a variety of downstream image classification tasks without manual tuning of parameters to each particular task.
Resulting models are computationally efficient and can be deployed to CPU using the OpenVINO toolkit.
arXiv Detail & Related papers (2021-08-16T12:38:05Z) - ANIMC: A Soft Framework for Auto-weighted Noisy and Incomplete
Multi-view Clustering [59.77141155608009]
We propose a novel Auto-weighted Noisy and Incomplete Multi-view Clustering framework (ANIMC) via a soft auto-weighted strategy and a doubly soft regular regression model.
ANIMC has three unique advantages: 1) it is a soft algorithm to adjust our framework in different scenarios, thereby improving its generalization ability; 2) it automatically learns a proper weight for each view, thereby reducing the influence of noises; and 3) it aligns the same instances in different views, thereby decreasing the impact of missing instances.
arXiv Detail & Related papers (2020-11-20T10:37:27Z) - Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning
in NLP Using Fewer Parameters & Less Data [5.689320790746046]
Multi-Task Learning (MTL) networks have emerged as a promising method for transferring learned knowledge across different tasks.
However, MTL must deal with challenges such as: overfitting to low resource tasks, catastrophic forgetting, and negative task transfer.
We propose a novel Transformer architecture consisting of a new conditional attention mechanism and a set of task-conditioned modules.
arXiv Detail & Related papers (2020-09-19T02:04:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.