Task-similarity Aware Meta-learning through Nonparametric Kernel
Regression
- URL: http://arxiv.org/abs/2006.07212v2
- Date: Mon, 12 Oct 2020 06:57:56 GMT
- Title: Task-similarity Aware Meta-learning through Nonparametric Kernel
Regression
- Authors: Arun Venkitaraman, Anders Hansson, and Bo Wahlberg
- Abstract summary: This paper investigates the use of nonparametric kernel-regression to obtain a tasksimilarity aware meta-learning algorithm.
Our hypothesis is that the use of tasksimilarity helps meta-learning when the available tasks are limited and may contain outlier/ dissimilar tasks.
- Score: 8.801367758434335
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper investigates the use of nonparametric kernel-regression to obtain
a tasksimilarity aware meta-learning algorithm. Our hypothesis is that the use
of tasksimilarity helps meta-learning when the available tasks are limited and
may contain outlier/ dissimilar tasks. While existing meta-learning approaches
implicitly assume the tasks as being similar, it is generally unclear how this
task-similarity could be quantified and used in the learning. As a result, most
popular metalearning approaches do not actively use the
similarity/dissimilarity between the tasks, but rely on availability of huge
number of tasks for their working. Our contribution is a novel framework for
meta-learning that explicitly uses task-similarity in the form of kernels and
an associated meta-learning algorithm. We model the task-specific parameters to
belong to a reproducing kernel Hilbert space where the kernel function captures
the similarity across tasks. The proposed algorithm iteratively learns a
meta-parameter which is used to assign a task-specific descriptor for every
task. The task descriptors are then used to quantify the task-similarity
through the kernel function. We show how our approach conceptually generalizes
the popular meta-learning approaches of model-agnostic meta-learning (MAML) and
Meta-stochastic gradient descent (Meta-SGD) approaches. Numerical experiments
with regression tasks show that our algorithm outperforms these approaches when
the number of tasks is limited, even in the presence of outlier or dissimilar
tasks. This supports our hypothesis that task-similarity helps improve the
metalearning performance in task-limited and adverse settings.
Related papers
- Meta-Learning with Heterogeneous Tasks [42.695853959923625]
Heterogeneous Tasks Robust Meta-learning (HeTRoM)
An efficient iterative optimization algorithm based on bi-level optimization.
Results demonstrate that our method provides flexibility, enabling users to adapt to diverse task settings.
arXiv Detail & Related papers (2024-10-24T16:32:23Z) - Data-Efficient and Robust Task Selection for Meta-Learning [1.4557421099695473]
We propose the Data-Efficient and Robust Task Selection (DERTS) algorithm, which can be incorporated into both gradient and metric-based meta-learning algorithms.
DERTS selects weighted subsets of tasks from task pools by minimizing the approximation error of the full gradient of task pools in the meta-training stage.
Unlike existing algorithms, DERTS does not require any architecture modification for training and can handle noisy label data in both the support and query sets.
arXiv Detail & Related papers (2024-05-11T19:47:27Z) - Set-based Meta-Interpolation for Few-Task Meta-Learning [79.4236527774689]
We propose a novel domain-agnostic task augmentation method, Meta-Interpolation, to densify the meta-training task distribution.
We empirically validate the efficacy of Meta-Interpolation on eight datasets spanning across various domains.
arXiv Detail & Related papers (2022-05-20T06:53:03Z) - The Effect of Diversity in Meta-Learning [79.56118674435844]
Few-shot learning aims to learn representations that can tackle novel tasks given a small number of examples.
Recent studies show that task distribution plays a vital role in the model's performance.
We study different task distributions on a myriad of models and datasets to evaluate the effect of task diversity on meta-learning algorithms.
arXiv Detail & Related papers (2022-01-27T19:39:07Z) - Does MAML Only Work via Feature Re-use? A Data Centric Perspective [19.556093984142418]
We provide empirical results that shed some light on how meta-learned MAML representations function.
We show that it is possible to define a family of synthetic benchmarks that result in a low degree of feature re-use.
We conjecture the core challenge of re-thinking meta-learning is in the design of few-shot learning data sets and benchmarks.
arXiv Detail & Related papers (2021-12-24T20:18:38Z) - Learning an Explicit Hyperparameter Prediction Function Conditioned on
Tasks [62.63852372239708]
Meta learning aims to learn the learning methodology for machine learning from observed tasks, so as to generalize to new query tasks.
We interpret such learning methodology as learning an explicit hyper- parameter prediction function shared by all training tasks.
Such setting guarantees that the meta-learned learning methodology is able to flexibly fit diverse query tasks.
arXiv Detail & Related papers (2021-07-06T04:05:08Z) - Meta-Learning with Fewer Tasks through Task Interpolation [67.03769747726666]
Current meta-learning algorithms require a large number of meta-training tasks, which may not be accessible in real-world scenarios.
By meta-learning with task gradient (MLTI), our approach effectively generates additional tasks by randomly sampling a pair of tasks and interpolating the corresponding features and labels.
Empirically, in our experiments on eight datasets from diverse domains, we find that the proposed general MLTI framework is compatible with representative meta-learning algorithms and consistently outperforms other state-of-the-art strategies.
arXiv Detail & Related papers (2021-06-04T20:15:34Z) - MetaKernel: Learning Variational Random Features with Limited Labels [120.90737681252594]
Few-shot learning deals with the fundamental and challenging problem of learning from a few annotated samples, while being able to generalize well on new tasks.
We propose meta-learning kernels with random Fourier features for few-shot learning, we call Meta Kernel.
arXiv Detail & Related papers (2021-05-08T21:24:09Z) - An Information-Theoretic Analysis of the Impact of Task Similarity on
Meta-Learning [44.320945743871285]
We present novel information-theoretic bounds on the average absolute value of the meta-generalization gap.
Our bounds explicitly capture the impact of task relatedness, the number of tasks, and the number of data samples per task on the meta-generalization gap.
arXiv Detail & Related papers (2021-01-21T01:38:16Z) - Adaptive Task Sampling for Meta-Learning [79.61146834134459]
Key idea of meta-learning for few-shot classification is to mimic the few-shot situations faced at test time.
We propose an adaptive task sampling method to improve the generalization performance.
arXiv Detail & Related papers (2020-07-17T03:15:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.