Streaming Self-Training via Domain-Agnostic Unlabeled Images
- URL: http://arxiv.org/abs/2104.03309v1
- Date: Wed, 7 Apr 2021 17:58:39 GMT
- Title: Streaming Self-Training via Domain-Agnostic Unlabeled Images
- Authors: Zhiqiu Lin and Deva Ramanan and Aayush Bansal
- Abstract summary: We present streaming self-training (SST) that aims to democratize the process of learning visual recognition models.
Key to SST are two crucial observations: (1) domain-agnostic unlabeled images enable us to learn better models with a few labeled examples without any additional knowledge or supervision; and (2) learning is a continuous process and can be done by constructing a schedule of learning updates.
- Score: 62.57647373581592
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present streaming self-training (SST) that aims to democratize the process
of learning visual recognition models such that a non-expert user can define a
new task depending on their needs via a few labeled examples and minimal domain
knowledge. Key to SST are two crucial observations: (1) domain-agnostic
unlabeled images enable us to learn better models with a few labeled examples
without any additional knowledge or supervision; and (2) learning is a
continuous process and can be done by constructing a schedule of learning
updates that iterates between pre-training on novel segments of the streams of
unlabeled data, and fine-tuning on the small and fixed labeled dataset. This
allows SST to overcome the need for a large number of domain-specific labeled
and unlabeled examples, exorbitant computational resources, and
domain/task-specific knowledge. In this setting, classical semi-supervised
approaches require a large amount of domain-specific labeled and unlabeled
examples, immense resources to process data, and expert knowledge of a
particular task. Due to these reasons, semi-supervised learning has been
restricted to a few places that can house required computational and human
resources. In this work, we overcome these challenges and demonstrate our
findings for a wide range of visual recognition tasks including fine-grained
image classification, surface normal estimation, and semantic segmentation. We
also demonstrate our findings for diverse domains including medical, satellite,
and agricultural imagery, where there does not exist a large amount of labeled
or unlabeled data.
Related papers
- A Review of Pseudo-Labeling for Computer Vision [2.79239659248295]
Deep neural networks often require large datasets of labeled samples to generalize effectively.
An important area of active research is semi-supervised learning, which attempts to instead utilize large quantities of (easily acquired) unlabeled samples.
In this work we explore a broader interpretation of pseudo-labels within both self-supervised and unsupervised methods.
arXiv Detail & Related papers (2024-08-13T22:17:48Z) - Transfer learning with generative models for object detection on limited datasets [1.4999444543328293]
In some fields, such as marine biology, it is necessary to have correctly labeled bounding boxes around each object.
We propose a transfer learning framework that is valid for a generic scenario.
Our results pave the way for new generative AI-based protocols for machine learning applications in various domains.
arXiv Detail & Related papers (2024-02-09T21:17:31Z) - Open Set Domain Recognition via Attention-Based GCN and Semantic
Matching Optimization [8.831857715361624]
This work presents an end-to-end model based on attention-based GCN and semantic matching optimization.
Experimental results validate that the proposed model not only has superiority on recognizing the images of known and unknown classes, but also can adapt to various openness of the target domain.
arXiv Detail & Related papers (2021-05-11T12:05:36Z) - Adaptive Self-training for Few-shot Neural Sequence Labeling [55.43109437200101]
We develop techniques to address the label scarcity challenge for neural sequence labeling models.
Self-training serves as an effective mechanism to learn from large amounts of unlabeled data.
meta-learning helps in adaptive sample re-weighting to mitigate error propagation from noisy pseudo-labels.
arXiv Detail & Related papers (2020-10-07T22:29:05Z) - Knowledge-Guided Multi-Label Few-Shot Learning for General Image
Recognition [75.44233392355711]
KGGR framework exploits prior knowledge of statistical label correlations with deep neural networks.
It first builds a structured knowledge graph to correlate different labels based on statistical label co-occurrence.
Then, it introduces the label semantics to guide learning semantic-specific features.
It exploits a graph propagation network to explore graph node interactions.
arXiv Detail & Related papers (2020-09-20T15:05:29Z) - A Review of Single-Source Deep Unsupervised Visual Domain Adaptation [81.07994783143533]
Large-scale labeled training datasets have enabled deep neural networks to excel across a wide range of benchmark vision tasks.
In many applications, it is prohibitively expensive and time-consuming to obtain large quantities of labeled data.
To cope with limited labeled training data, many have attempted to directly apply models trained on a large-scale labeled source domain to another sparsely labeled or unlabeled target domain.
arXiv Detail & Related papers (2020-09-01T00:06:50Z) - Adversarial Knowledge Transfer from Unlabeled Data [62.97253639100014]
We present a novel Adversarial Knowledge Transfer framework for transferring knowledge from internet-scale unlabeled data to improve the performance of a classifier.
An important novel aspect of our method is that the unlabeled source data can be of different classes from those of the labeled target data, and there is no need to define a separate pretext task.
arXiv Detail & Related papers (2020-08-13T08:04:27Z) - Unsupervised Transfer Learning with Self-Supervised Remedy [60.315835711438936]
Generalising deep networks to novel domains without manual labels is challenging to deep learning.
Pre-learned knowledge does not transfer well without making strong assumptions about the learned and the novel domains.
In this work, we aim to learn a discriminative latent space of the unlabelled target data in a novel domain by knowledge transfer from labelled related domains.
arXiv Detail & Related papers (2020-06-08T16:42:17Z) - Learning Cross-domain Generalizable Features by Representation
Disentanglement [11.74643883335152]
Deep learning models exhibit limited generalizability across different domains.
We propose Mutual-Information-based Disentangled Neural Networks (MIDNet) to extract generalizable features that enable transferring knowledge to unseen categorical features in target domains.
We demonstrate our method on handwritten digits datasets and a fetal ultrasound dataset for image classification tasks.
arXiv Detail & Related papers (2020-02-29T17:53:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.