Domain segmentation and adjustment for generalized zero-shot learning
- URL: http://arxiv.org/abs/2002.00226v1
- Date: Sat, 1 Feb 2020 15:00:56 GMT
- Title: Domain segmentation and adjustment for generalized zero-shot learning
- Authors: Xinsheng Wang, Shanmin Pang, Jihua Zhu
- Abstract summary: In zero-shot learning, synthesizing unseen data with generative models has been the most popular method to address the imbalance of training data between seen and unseen classes.
We argue that synthesizing unseen data may not be an ideal approach for addressing the domain shift caused by the imbalance of the training data.
In this paper, we propose to realize the generalized zero-shot recognition in different domains.
- Score: 22.933463036413624
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the generalized zero-shot learning, synthesizing unseen data with
generative models has been the most popular method to address the imbalance of
training data between seen and unseen classes. However, this method requires
that the unseen semantic information is available during the training stage,
and training generative models is not trivial. Given that the generator of
these models can only be trained with seen classes, we argue that synthesizing
unseen data may not be an ideal approach for addressing the domain shift caused
by the imbalance of the training data. In this paper, we propose to realize the
generalized zero-shot recognition in different domains. Thus, unseen (seen)
classes can avoid the effect of the seen (unseen) classes. In practice, we
propose a threshold and probabilistic distribution joint method to segment the
testing instances into seen, unseen and uncertain domains. Moreover, the
uncertain domain is further adjusted to alleviate the domain shift. Extensive
experiments on five benchmark datasets show that the proposed method exhibits
competitive performance compared with that based on generative models.
Related papers
- Online Performance Estimation with Unlabeled Data: A Bayesian Application of the Hui-Walter Paradigm [0.0]
We adapt the Hui-Walter paradigm, a method traditionally applied in epidemiology and medicine, to the field of machine learning.
We estimate key performance metrics such as false positive rate, false negative rate, and priors in scenarios where no ground truth is available.
We extend this paradigm for handling online data, opening up new possibilities for dynamic data environments.
arXiv Detail & Related papers (2024-01-17T17:46:10Z) - Class Distribution Shifts in Zero-Shot Learning: Learning Robust Representations [3.8980564330208662]
We propose a model that assumes that the attribute responsible for the shift is unknown in advance.
We show that our approach improves generalization on diverse class distributions in both simulations and real-world datasets.
arXiv Detail & Related papers (2023-11-30T14:14:31Z) - Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning [92.89846887298852]
Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
arXiv Detail & Related papers (2022-10-11T10:20:31Z) - Predicting Out-of-Domain Generalization with Neighborhood Invariance [59.05399533508682]
We propose a measure of a classifier's output invariance in a local transformation neighborhood.
Our measure is simple to calculate, does not depend on the test point's true label, and can be applied even in out-of-domain (OOD) settings.
In experiments on benchmarks in image classification, sentiment analysis, and natural language inference, we demonstrate a strong and robust correlation between our measure and actual OOD generalization.
arXiv Detail & Related papers (2022-07-05T14:55:16Z) - Self Training with Ensemble of Teacher Models [8.257085583227695]
In order to train robust deep learning models, large amounts of labelled data is required.
In the absence of such large repositories of labelled data, unlabeled data can be exploited for the same.
Semi-Supervised learning aims to utilize such unlabeled data for training classification models.
arXiv Detail & Related papers (2021-07-17T09:44:09Z) - A Curriculum-style Self-training Approach for Source-Free Semantic Segmentation [91.13472029666312]
We propose a curriculum-style self-training approach for source-free domain adaptive semantic segmentation.
Our method yields state-of-the-art performance on source-free semantic segmentation tasks for both synthetic-to-real and adverse conditions.
arXiv Detail & Related papers (2021-06-22T10:21:39Z) - Neural Networks for Learning Counterfactual G-Invariances from Single
Environments [13.848760376470038]
neural networks are believed to have difficulties extrapolating beyond training data distribution.
This work shows that, for extrapolations based on finite transformation groups, a model's inability to extrapolate is unrelated to its capacity.
arXiv Detail & Related papers (2021-04-20T16:35:35Z) - Entropy-Based Uncertainty Calibration for Generalized Zero-Shot Learning [49.04790688256481]
The goal of generalized zero-shot learning (GZSL) is to recognise both seen and unseen classes.
Most GZSL methods typically learn to synthesise visual representations from semantic information on the unseen classes.
We propose a novel framework that leverages dual variational autoencoders with a triplet loss to learn discriminative latent features.
arXiv Detail & Related papers (2021-01-09T05:21:27Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z) - Rethinking Class-Balanced Methods for Long-Tailed Visual Recognition
from a Domain Adaptation Perspective [98.70226503904402]
Object frequency in the real world often follows a power law, leading to a mismatch between datasets with long-tailed class distributions.
We propose to augment the classic class-balanced learning by explicitly estimating the differences between the class-conditioned distributions with a meta-learning approach.
arXiv Detail & Related papers (2020-03-24T11:28:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.