ProtoDA: Efficient Transfer Learning for Few-Shot Intent Classification
- URL: http://arxiv.org/abs/2101.11753v1
- Date: Thu, 28 Jan 2021 00:19:13 GMT
- Title: ProtoDA: Efficient Transfer Learning for Few-Shot Intent Classification
- Authors: Manoj Kumar, Varun Kumar, Hadrien Glaude, Cyprien delichy, Aman Alok
and Rahul Gupta
- Abstract summary: We adopt an alternative approach by transfer learning on an ensemble of related tasks using prototypical networks under the meta-learning paradigm.
Using intent classification as a case study, we demonstrate that increasing variability in training tasks can significantly improve classification performance.
- Score: 21.933876113300897
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Practical sequence classification tasks in natural language processing often
suffer from low training data availability for target classes. Recent works
towards mitigating this problem have focused on transfer learning using
embeddings pre-trained on often unrelated tasks, for instance, language
modeling. We adopt an alternative approach by transfer learning on an ensemble
of related tasks using prototypical networks under the meta-learning paradigm.
Using intent classification as a case study, we demonstrate that increasing
variability in training tasks can significantly improve classification
performance. Further, we apply data augmentation in conjunction with
meta-learning to reduce sampling bias. We make use of a conditional generator
for data augmentation that is trained directly using the meta-learning
objective and simultaneously with prototypical networks, hence ensuring that
data augmentation is customized to the task. We explore augmentation in the
sentence embedding space as well as prototypical embedding space. Combining
meta-learning with augmentation provides upto 6.49% and 8.53% relative F1-score
improvements over the best performing systems in the 5-shot and 10-shot
learning, respectively.
Related papers
- Transfer Learning with Informative Priors: Simple Baselines Better than Previously Reported [4.453137996095194]
We compare transfer learning with and without source task informed priors across 5 datasets.
For the scenario of 5-300 examples per class, we find negative or negligible gains on 2 datasets, modest gains on 2 other datasets, and substantial gains on one dataset.
arXiv Detail & Related papers (2024-05-24T14:12:23Z) - Enhancing Visual Continual Learning with Language-Guided Supervision [76.38481740848434]
Continual learning aims to empower models to learn new tasks without forgetting previously acquired knowledge.
We argue that the scarce semantic information conveyed by the one-hot labels hampers the effective knowledge transfer across tasks.
Specifically, we use PLMs to generate semantic targets for each class, which are frozen and serve as supervision signals.
arXiv Detail & Related papers (2024-03-24T12:41:58Z) - Adversarial Auto-Augment with Label Preservation: A Representation
Learning Principle Guided Approach [95.74102207187545]
We show that a prior-free autonomous data augmentation's objective can be derived from a representation learning principle.
We then propose a practical surrogate to the objective that can be efficiently optimized and integrated seamlessly into existing methods.
arXiv Detail & Related papers (2022-11-02T02:02:51Z) - Generating meta-learning tasks to evolve parametric loss for
classification learning [1.1355370218310157]
In existing meta-learning approaches, learning tasks for training meta-models are usually collected from public datasets.
We propose a meta-learning approach based on randomly generated meta-learning tasks to obtain a parametric loss for classification learning based on big data.
arXiv Detail & Related papers (2021-11-20T13:07:55Z) - Meta-Regularization by Enforcing Mutual-Exclusiveness [0.8057006406834467]
We propose a regularization technique for meta-learning models that gives the model designer more control over the information flow during meta-training.
Our proposed regularization function shows an accuracy boost of $sim$ $36%$ on the Omniglot dataset.
arXiv Detail & Related papers (2021-01-24T22:57:19Z) - Meta-learning the Learning Trends Shared Across Tasks [123.10294801296926]
Gradient-based meta-learning algorithms excel at quick adaptation to new tasks with limited data.
Existing meta-learning approaches only depend on the current task information during the adaptation.
We propose a 'Path-aware' model-agnostic meta-learning approach.
arXiv Detail & Related papers (2020-10-19T08:06:47Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - Few Is Enough: Task-Augmented Active Meta-Learning for Brain Cell
Classification [8.998976678920236]
We propose a tAsk-auGmented actIve meta-LEarning (AGILE) method to efficiently adapt Deep Neural Networks to new tasks.
AGILE combines a meta-learning algorithm with a novel task augmentation technique which we use to generate an initial adaptive model.
We show that the proposed task-augmented meta-learning framework can learn to classify new cell types after a single gradient step.
arXiv Detail & Related papers (2020-07-09T18:03:12Z) - Uniform Priors for Data-Efficient Transfer [65.086680950871]
We show that features that are most transferable have high uniformity in the embedding space.
We evaluate the regularization on its ability to facilitate adaptation to unseen tasks and data.
arXiv Detail & Related papers (2020-06-30T04:39:36Z) - iTAML: An Incremental Task-Agnostic Meta-learning Approach [123.10294801296926]
Humans can continuously learn new knowledge as their experience grows.
Previous learning in deep neural networks can quickly fade out when they are trained on a new task.
We introduce a novel meta-learning approach that seeks to maintain an equilibrium between all encountered tasks.
arXiv Detail & Related papers (2020-03-25T21:42:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.