Low-Budget Label Query through Domain Alignment Enforcement
- URL: http://arxiv.org/abs/2001.00238v2
- Date: Sun, 29 Mar 2020 11:43:05 GMT
- Title: Low-Budget Label Query through Domain Alignment Enforcement
- Authors: Jurandy Almeida, Cristiano Saltori, Paolo Rota, and Nicu Sebe
- Abstract summary: We tackle a new problem named low-budget label query.
We first improve an Unsupervised Domain Adaptation (UDA) method to better align source and target domains.
We then propose a simple yet effective selection method based on uniform sampling of the prediction consistency distribution.
- Score: 48.06803561387064
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning revolution happened thanks to the availability of a massive
amount of labelled data which have contributed to the development of models
with extraordinary inference capabilities. Despite the public availability of a
large quantity of datasets, to address specific requirements it is often
necessary to generate a new set of labelled data. Quite often, the production
of labels is costly and sometimes it requires specific know-how to be
fulfilled. In this work, we tackle a new problem named low-budget label query
that consists in suggesting to the user a small (low budget) set of samples to
be labelled, from a completely unlabelled dataset, with the final goal of
maximizing the classification accuracy on that dataset. In this work we first
improve an Unsupervised Domain Adaptation (UDA) method to better align source
and target domains using consistency constraints, reaching the state of the art
on a few UDA tasks. Finally, using the previously trained model as reference,
we propose a simple yet effective selection method based on uniform sampling of
the prediction consistency distribution, which is deterministic and steadily
outperforms other baselines as well as competing models on a large variety of
publicly available datasets.
Related papers
- Probabilistic Test-Time Generalization by Variational Neighbor-Labeling [62.158807685159736]
This paper strives for domain generalization, where models are trained exclusively on source domains before being deployed on unseen target domains.
Probability pseudo-labeling of target samples to generalize the source-trained model to the target domain at test time.
Variational neighbor labels that incorporate the information of neighboring target samples to generate more robust pseudo labels.
arXiv Detail & Related papers (2023-07-08T18:58:08Z) - Combining Public Human Activity Recognition Datasets to Mitigate Labeled
Data Scarcity [1.274578243851308]
We propose a novel strategy to combine publicly available datasets with the goal of learning a generalized HAR model.
Our experimental evaluation, which includes experimenting with different state-of-the-art neural network architectures, shows that combining public datasets can significantly reduce the number of labeled samples.
arXiv Detail & Related papers (2023-06-23T18:51:22Z) - Combating Label Distribution Shift for Active Domain Adaptation [16.270897459117755]
We consider the problem of active domain adaptation (ADA) to unlabeled target data.
Inspired by recent analysis on a critical issue from label distribution mismatch between source and target in domain adaptation, we devise a method that addresses the issue for the first time in ADA.
arXiv Detail & Related papers (2022-08-13T09:06:45Z) - S$^3$VAADA: Submodular Subset Selection for Virtual Adversarial Active
Domain Adaptation [49.01925978752677]
In the real-world scenario's it might be feasible to get labels for a small proportion of target data.
We propose S$3$VAADA which i) introduces a novel submodular criterion to select a maximally informative subset to label and ii) enhances a cluster-based DA procedure.
Our approach consistently outperforms the competing state-of-the-art approaches on datasets with varying degrees of domain shifts.
arXiv Detail & Related papers (2021-09-18T10:53:57Z) - Instance Level Affinity-Based Transfer for Unsupervised Domain
Adaptation [74.71931918541748]
We propose an instance affinity based criterion for source to target transfer during adaptation, called ILA-DA.
We first propose a reliable and efficient method to extract similar and dissimilar samples across source and target, and utilize a multi-sample contrastive loss to drive the domain alignment process.
We verify the effectiveness of ILA-DA by observing consistent improvements in accuracy over popular domain adaptation approaches on a variety of benchmark datasets.
arXiv Detail & Related papers (2021-04-03T01:33:14Z) - Unsupervised and self-adaptative techniques for cross-domain person
re-identification [82.54691433502335]
Person Re-Identification (ReID) across non-overlapping cameras is a challenging task.
Unsupervised Domain Adaptation (UDA) is a promising alternative, as it performs feature-learning adaptation from a model trained on a source to a target domain without identity-label annotation.
In this paper, we propose a novel UDA-based ReID method that takes advantage of triplets of samples created by a new offline strategy.
arXiv Detail & Related papers (2021-03-21T23:58:39Z) - Self-Supervised Noisy Label Learning for Source-Free Unsupervised Domain
Adaptation [87.60688582088194]
We propose a novel Self-Supervised Noisy Label Learning method.
Our method can easily achieve state-of-the-art results and surpass other methods by a very large margin.
arXiv Detail & Related papers (2021-02-23T10:51:45Z) - Weak Adaptation Learning -- Addressing Cross-domain Data Insufficiency
with Weak Annotator [2.8672054847109134]
In some target problem domains, there are not many data samples available, which could hinder the learning process.
We propose a weak adaptation learning (WAL) approach that leverages unlabeled data from a similar source domain.
Our experiments demonstrate the effectiveness of our approach in learning an accurate classifier with limited labeled data in the target domain.
arXiv Detail & Related papers (2021-02-15T06:19:25Z) - Diversity-Based Generalization for Unsupervised Text Classification
under Domain Shift [16.522910268114504]
We propose a novel method for domain adaptation of single-task text classification problems based on a simple but effective idea of diversity-based generalization.
Our results demonstrate that machine learning architectures that ensure sufficient diversity can generalize better.
arXiv Detail & Related papers (2020-02-25T15:11:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.