Webly Supervised Image Classification with Self-Contained Confidence
- URL: http://arxiv.org/abs/2008.11894v1
- Date: Thu, 27 Aug 2020 02:49:51 GMT
- Title: Webly Supervised Image Classification with Self-Contained Confidence
- Authors: Jingkang Yang, Litong Feng, Weirong Chen, Xiaopeng Yan, Huabin Zheng,
Ping Luo, Wayne Zhang
- Abstract summary: This paper focuses on webly supervised learning (WSL), where datasets are built by crawling samples from the Internet and directly using search queries as web labels.
We introduce Self-Contained Confidence ( SCC) by adapting model uncertainty for WSL setting, and use it to sample-wisely balance $mathcalL_s$ and $mathcalL_w$.
The proposed WSL framework has achieved the state-of-the-art results on two large-scale WSL datasets, WebVision-1000 and Food101-N. Code.
- Score: 36.87209906372911
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper focuses on webly supervised learning (WSL), where datasets are
built by crawling samples from the Internet and directly using search queries
as web labels. Although WSL benefits from fast and low-cost data collection,
noises in web labels hinder better performance of the image classification
model. To alleviate this problem, in recent works, self-label supervised loss
$\mathcal{L}_s$ is utilized together with webly supervised loss
$\mathcal{L}_w$. $\mathcal{L}_s$ relies on pseudo labels predicted by the model
itself. Since the correctness of the web label or pseudo label is usually on a
case-by-case basis for each web sample, it is desirable to adjust the balance
between $\mathcal{L}_s$ and $\mathcal{L}_w$ on sample level. Inspired by the
ability of Deep Neural Networks (DNNs) in confidence prediction, we introduce
Self-Contained Confidence (SCC) by adapting model uncertainty for WSL setting,
and use it to sample-wisely balance $\mathcal{L}_s$ and $\mathcal{L}_w$.
Therefore, a simple yet effective WSL framework is proposed. A series of
SCC-friendly regularization approaches are investigated, among which the
proposed graph-enhanced mixup is the most effective method to provide
high-quality confidence to enhance our framework. The proposed WSL framework
has achieved the state-of-the-art results on two large-scale WSL datasets,
WebVision-1000 and Food101-N. Code is available at
https://github.com/bigvideoresearch/SCC.
Related papers
- RankMatch: A Novel Approach to Semi-Supervised Label Distribution
Learning Leveraging Inter-label Correlations [52.549807652527306]
This paper introduces RankMatch, an innovative approach for Semi-Supervised Label Distribution Learning (SSLDL)
RankMatch effectively utilizes a small number of labeled examples in conjunction with a larger quantity of unlabeled data.
We establish a theoretical generalization bound for RankMatch, and through extensive experiments, demonstrate its superiority in performance against existing SSLDL methods.
arXiv Detail & Related papers (2023-12-11T12:47:29Z) - ProtoCon: Pseudo-label Refinement via Online Clustering and Prototypical
Consistency for Efficient Semi-supervised Learning [60.57998388590556]
ProtoCon is a novel method for confidence-based pseudo-labeling.
Online nature of ProtoCon allows it to utilise the label history of the entire dataset in one training cycle.
It delivers significant gains and faster convergence over state-of-the-art datasets.
arXiv Detail & Related papers (2023-03-22T23:51:54Z) - Data-Centric Debugging: mitigating model failures via targeted data
collection [4.599792546344752]
Deep neural networks can be unreliable in the real world when the training set does not adequately cover all the settings where they are deployed.
We propose a general methodology for model debug that can systemically improve model performance on $mathcalE$ while maintaining its performance on the original test set.
arXiv Detail & Related papers (2022-11-17T19:44:02Z) - Self Supervised Learning for Few Shot Hyperspectral Image Classification [57.2348804884321]
We propose to leverage Self Supervised Learning (SSL) for HSI classification.
We show that by pre-training an encoder on unlabeled pixels using Barlow-Twins, a state-of-the-art SSL algorithm, we can obtain accurate models with a handful of labels.
arXiv Detail & Related papers (2022-06-24T07:21:53Z) - Deep Low-Density Separation for Semi-Supervised Classification [0.0]
We introduce a novel hybrid method that applies low-density separation to the embedded features.
Our approach effectively classifies thousands of unlabeled users from a relatively small number of hand-classified examples.
arXiv Detail & Related papers (2022-05-22T11:00:55Z) - CLS: Cross Labeling Supervision for Semi-Supervised Learning [9.929229055862491]
Cross Labeling Supervision ( CLS) is a framework that generalizes the typical pseudo-labeling process.
CLS allows the creation of both pseudo and complementary labels to support both positive and negative learning.
arXiv Detail & Related papers (2022-02-17T08:09:40Z) - Matching Distributions via Optimal Transport for Semi-Supervised
Learning [31.533832244923843]
Semi-Supervised Learning (SSL) approaches have been an influential framework for the usage of unlabeled data.
We propose a new approach that adopts an Optimal Transport (OT) technique serving as a metric of similarity between discrete empirical probability measures.
We have evaluated our proposed method with state-of-the-art SSL algorithms on standard datasets to demonstrate the superiority and effectiveness of our SSL algorithm.
arXiv Detail & Related papers (2020-12-04T11:15:14Z) - OSLNet: Deep Small-Sample Classification with an Orthogonal Softmax
Layer [77.90012156266324]
This paper aims to find a subspace of neural networks that can facilitate a large decision margin.
We propose the Orthogonal Softmax Layer (OSL), which makes the weight vectors in the classification layer remain during both the training and test processes.
Experimental results demonstrate that the proposed OSL has better performance than the methods used for comparison on four small-sample benchmark datasets.
arXiv Detail & Related papers (2020-04-20T02:41:01Z) - FixMatch: Simplifying Semi-Supervised Learning with Consistency and
Confidence [93.91751021370638]
Semi-supervised learning (SSL) provides an effective means of leveraging unlabeled data to improve a model's performance.
In this paper, we demonstrate the power of a simple combination of two common SSL methods: consistency regularization and pseudo-labeling.
Our algorithm, FixMatch, first generates pseudo-labels using the model's predictions on weakly-augmented unlabeled images.
arXiv Detail & Related papers (2020-01-21T18:32:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.