Disentangled Generation with Information Bottleneck for Few-Shot
Learning
- URL: http://arxiv.org/abs/2211.16185v1
- Date: Tue, 29 Nov 2022 13:29:36 GMT
- Title: Disentangled Generation with Information Bottleneck for Few-Shot
Learning
- Authors: Zhuohang Dang, Jihong Wang, Minnan Luo, Chengyou Jia, Caixia Yan,
Qinghua Zheng
- Abstract summary: Few-shot learning, which aims to classify unseen classes with few samples, is challenging due to data scarcity.
We propose a novel Information Bottleneck (IB) based Disentangled Generation Framework (DisGenIB)
DisGenIB can simultaneously guarantee the discrimination and diversity of generated samples.
- Score: 21.131911207010376
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot learning (FSL), which aims to classify unseen classes with few
samples, is challenging due to data scarcity. Although various generative
methods have been explored for FSL, the entangled generation process of these
methods exacerbates the distribution shift in FSL, thus greatly limiting the
quality of generated samples. To these challenges, we propose a novel
Information Bottleneck (IB) based Disentangled Generation Framework for FSL,
termed as DisGenIB, that can simultaneously guarantee the discrimination and
diversity of generated samples. Specifically, we formulate a novel framework
with information bottleneck that applies for both disentangled representation
learning and sample generation. Different from existing IB-based methods that
can hardly exploit priors, we demonstrate our DisGenIB can effectively utilize
priors to further facilitate disentanglement. We further prove in theory that
some previous generative and disentanglement methods are special cases of our
DisGenIB, which demonstrates the generality of the proposed DisGenIB. Extensive
experiments on challenging FSL benchmarks confirm the effectiveness and
superiority of DisGenIB, together with the validity of our theoretical
analyses. Our codes will be open-source upon acceptance.
Related papers
- Adaptive Destruction Processes for Diffusion Samplers [12.446080077998834]
This paper explores the challenges and benefits of a trainable destruction process in diffusion samplers.<n>We show that, when the number of steps is limited, training both generation and destruction processes results in faster convergence and improved sampling quality.
arXiv Detail & Related papers (2025-06-02T11:07:27Z) - Joint-stochastic-approximation Random Fields with Application to Semi-supervised Learning [10.293017518216908]
We present joint-stochastic-approximation random fields (JRFs) -- a new family of algorithms for building deep undirected generative models.<n>It is found through synthetic experiments that JRFs work well in balancing mode covering and mode missing, and match the empirical data distribution well.
arXiv Detail & Related papers (2025-05-24T07:04:32Z) - Dirichlet-Based Coarse-to-Fine Example Selection For Open-Set Annotation [37.33424244520009]
We propose a Dirichlet-based Coarse-to-Fine Example Selection (DCFS) strategy.
Our method introduces simplex-based evidential deep learning (EDL) to break translation invariance.
Experiments on various openness ratio datasets demonstrate that DCFS achieves state-of-art performance.
arXiv Detail & Related papers (2024-09-26T07:47:50Z) - Enhancing Unsupervised Sentence Embeddings via Knowledge-Driven Data Augmentation and Gaussian-Decayed Contrastive Learning [37.54523122932728]
We propose a pipeline-based data augmentation method via large language models (LLMs)
To tackle the issue of low data diversity, our pipeline utilizes knowledge graphs (KGs) to extract entities and quantities.
To address high data noise, the GCSE model uses a Gaussian-decayed function to limit the impact of false hard negative samples.
arXiv Detail & Related papers (2024-09-19T16:29:58Z) - Data-Free Federated Class Incremental Learning with Diffusion-Based Generative Memory [27.651921957220004]
We introduce a novel data-free federated class incremental learning framework with diffusion-based generative memory (DFedDGM)
We design a new balanced sampler to help train the diffusion models to alleviate the common non-IID problem in FL.
We also introduce an entropy-based sample filtering technique from an information theory perspective to enhance the quality of generative samples.
arXiv Detail & Related papers (2024-05-22T20:59:18Z) - BECLR: Batch Enhanced Contrastive Few-Shot Learning [1.450405446885067]
Unsupervised few-shot learning aspires to bridge this gap by discarding the reliance on annotations at training time.
We propose a novel Dynamic Clustered mEmory (DyCE) module to promote a highly separable latent representation space.
We then tackle the, somehow overlooked yet critical, issue of sample bias at the few-shot inference stage.
arXiv Detail & Related papers (2024-02-04T10:52:43Z) - Learning Invariant Molecular Representation in Latent Discrete Space [52.13724532622099]
We propose a new framework for learning molecular representations that exhibit invariance and robustness against distribution shifts.
Our model achieves stronger generalization against state-of-the-art baselines in the presence of various distribution shifts.
arXiv Detail & Related papers (2023-10-22T04:06:44Z) - Collaborative Intelligence Orchestration: Inconsistency-Based Fusion of
Semi-Supervised Learning and Active Learning [60.26659373318915]
Active learning (AL) and semi-supervised learning (SSL) are two effective, but often isolated, means to alleviate the data-hungry problem.
We propose an innovative Inconsistency-based virtual aDvErial algorithm to further investigate SSL-AL's potential superiority.
Two real-world case studies visualize the practical industrial value of applying and deploying the proposed data sampling algorithm.
arXiv Detail & Related papers (2022-06-07T13:28:43Z) - Exploring Complementary Strengths of Invariant and Equivariant
Representations for Few-Shot Learning [96.75889543560497]
In many real-world problems, collecting a large number of labeled samples is infeasible.
Few-shot learning is the dominant approach to address this issue, where the objective is to quickly adapt to novel categories in presence of a limited number of samples.
We propose a novel training mechanism that simultaneously enforces equivariance and invariance to a general set of geometric transformations.
arXiv Detail & Related papers (2021-03-01T21:14:33Z) - On Data-Augmentation and Consistency-Based Semi-Supervised Learning [77.57285768500225]
Recently proposed consistency-based Semi-Supervised Learning (SSL) methods have advanced the state of the art in several SSL tasks.
Despite these advances, the understanding of these methods is still relatively limited.
arXiv Detail & Related papers (2021-01-18T10:12:31Z) - Open Set Recognition with Conditional Probabilistic Generative Models [51.40872765917125]
We propose Conditional Probabilistic Generative Models (CPGM) for open set recognition.
CPGM can detect unknown samples but also classify known classes by forcing different latent features to approximate conditional Gaussian distributions.
Experiment results on multiple benchmark datasets reveal that the proposed method significantly outperforms the baselines.
arXiv Detail & Related papers (2020-08-12T06:23:49Z) - Adversarial Feature Hallucination Networks for Few-Shot Learning [84.31660118264514]
Adversarial Feature Hallucination Networks (AFHN) is based on conditional Wasserstein Generative Adversarial networks (cWGAN)
Two novel regularizers are incorporated into AFHN to encourage discriminability and diversity of the synthesized features.
arXiv Detail & Related papers (2020-03-30T02:43:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.