CMVAE: Causal Meta VAE for Unsupervised Meta-Learning
- URL: http://arxiv.org/abs/2302.09731v1
- Date: Mon, 20 Feb 2023 02:49:35 GMT
- Title: CMVAE: Causal Meta VAE for Unsupervised Meta-Learning
- Authors: Guodong Qi, Huimin Yu
- Abstract summary: Unsupervised meta-learning aims to learn the meta knowledge from unlabeled data and rapidly adapt to novel tasks.
Existing approaches may be misled by the context-bias from the training data.
We propose Causal Meta VAE (CMVAE) that encodes the priors into latent codes in the causal space and learns their relationships simultaneously to achieve the downstream few-shot image classification task.
- Score: 3.0839245814393728
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised meta-learning aims to learn the meta knowledge from unlabeled
data and rapidly adapt to novel tasks. However, existing approaches may be
misled by the context-bias (e.g. background) from the training data. In this
paper, we abstract the unsupervised meta-learning problem into a Structural
Causal Model (SCM) and point out that such bias arises due to hidden
confounders. To eliminate the confounders, we define the priors are
\textit{conditionally} independent, learn the relationships between priors and
intervene on them with casual factorization. Furthermore, we propose Causal
Meta VAE (CMVAE) that encodes the priors into latent codes in the causal space
and learns their relationships simultaneously to achieve the downstream
few-shot image classification task. Results on toy datasets and three benchmark
datasets demonstrate that our method can remove the context-bias and it
outperforms other state-of-the-art unsupervised meta-learning algorithms
because of bias-removal. Code is available at
\url{https://github.com/GuodongQi/CMVAE}
Related papers
- Privacy Challenges in Meta-Learning: An Investigation on Model-Agnostic Meta-Learning [2.8948274245812327]
In current meta-learning methods, task learners locally learn models from sensitive data, termed support sets.
Despite the absence of explicit data sharing, privacy concerns persist.
This paper examines potential data leakage in a prominent metalearning algorithm, specifically Model-Agnostic Meta-Learning (MAML)
arXiv Detail & Related papers (2024-06-01T01:10:35Z) - Gromov-Wasserstein Autoencoders [36.656435006076975]
We propose a novel representation learning method, Gromov-Wasserstein Autoencoders (GWAE)
Instead of a likelihood-based objective, GWAE models have a trainable prior optimized by minimizing the Gromov-Wasserstein (GW) metric.
By restricting the family of the trainable prior, we can introduce meta-priors to control latent representations for downstream tasks.
arXiv Detail & Related papers (2022-09-15T02:34:39Z) - Unpaired Referring Expression Grounding via Bidirectional Cross-Modal
Matching [53.27673119360868]
Referring expression grounding is an important and challenging task in computer vision.
We propose a novel bidirectional cross-modal matching (BiCM) framework to address these challenges.
Our framework outperforms previous works by 6.55% and 9.94% on two popular grounding datasets.
arXiv Detail & Related papers (2022-01-18T01:13:19Z) - Transfer Meta-Learning: Information-Theoretic Bounds and Information
Meta-Risk Minimization [47.7605527786164]
Meta-learning automatically infers an inductive bias by observing data from a number of related tasks.
We introduce the problem of transfer meta-learning, in which tasks are drawn from a target task environment during meta-testing.
arXiv Detail & Related papers (2020-11-04T12:55:43Z) - Meta-Learning with Context-Agnostic Initialisations [86.47040878540139]
We introduce a context-adversarial component into the meta-learning process.
This produces an initialisation for fine-tuning to target which is context-agnostic and task-generalised.
We evaluate our approach on three commonly used meta-learning algorithms and two problems.
arXiv Detail & Related papers (2020-07-29T08:08:38Z) - Semi-Supervised Learning with Meta-Gradient [123.26748223837802]
We propose a simple yet effective meta-learning algorithm in semi-supervised learning.
We find that the proposed algorithm performs favorably against state-of-the-art methods.
arXiv Detail & Related papers (2020-07-08T08:48:56Z) - Rethinking Few-Shot Image Classification: a Good Embedding Is All You
Need? [72.00712736992618]
We show that a simple baseline: learning a supervised or self-supervised representation on the meta-training set, outperforms state-of-the-art few-shot learning methods.
An additional boost can be achieved through the use of self-distillation.
We believe that our findings motivate a rethinking of few-shot image classification benchmarks and the associated role of meta-learning algorithms.
arXiv Detail & Related papers (2020-03-25T17:58:42Z) - Meta Cyclical Annealing Schedule: A Simple Approach to Avoiding
Meta-Amortization Error [50.83356836818667]
We develop a novel meta-regularization objective using it cyclical annealing schedule and it maximum mean discrepancy (MMD) criterion.
The experimental results show that our approach substantially outperforms standard meta-learning algorithms.
arXiv Detail & Related papers (2020-03-04T04:43:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.