Unsupervised Deep Metric Learning via Orthogonality based Probabilistic
Loss
- URL: http://arxiv.org/abs/2008.09880v1
- Date: Sat, 22 Aug 2020 17:13:33 GMT
- Title: Unsupervised Deep Metric Learning via Orthogonality based Probabilistic
Loss
- Authors: Ujjal Kr Dutta, Mehrtash Harandi and Chellu Chandra Sekhar
- Abstract summary: Existing state-of-the-art metric learning approaches require class labels to learn a metric.
We propose an unsupervised approach that learns a metric without making use of class labels.
The pseudo-labels are used to form triplets of examples, which guide the metric learning.
- Score: 27.955068939695042
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Metric learning is an important problem in machine learning. It aims to group
similar examples together. Existing state-of-the-art metric learning approaches
require class labels to learn a metric. As obtaining class labels in all
applications is not feasible, we propose an unsupervised approach that learns a
metric without making use of class labels. The lack of class labels is
compensated by obtaining pseudo-labels of data using a graph-based clustering
approach. The pseudo-labels are used to form triplets of examples, which guide
the metric learning. We propose a probabilistic loss that minimizes the chances
of each triplet violating an angular constraint. A weight function, and an
orthogonality constraint in the objective speeds up the convergence and avoids
a model collapse. We also provide a stochastic formulation of our method to
scale up to large-scale datasets. Our studies demonstrate the competitiveness
of our approach against state-of-the-art methods. We also thoroughly study the
effect of the different components of our method.
Related papers
- Dual-Decoupling Learning and Metric-Adaptive Thresholding for Semi-Supervised Multi-Label Learning [81.83013974171364]
Semi-supervised multi-label learning (SSMLL) is a powerful framework for leveraging unlabeled data to reduce the expensive cost of collecting precise multi-label annotations.
Unlike semi-supervised learning, one cannot select the most probable label as the pseudo-label in SSMLL due to multiple semantics contained in an instance.
We propose a dual-perspective method to generate high-quality pseudo-labels.
arXiv Detail & Related papers (2024-07-26T09:33:53Z) - Partial-Label Regression [54.74984751371617]
Partial-label learning is a weakly supervised learning setting that allows each training example to be annotated with a set of candidate labels.
Previous studies on partial-label learning only focused on the classification setting where candidate labels are all discrete.
In this paper, we provide the first attempt to investigate partial-label regression, where each training example is annotated with a set of real-valued candidate labels.
arXiv Detail & Related papers (2023-06-15T09:02:24Z) - All Points Matter: Entropy-Regularized Distribution Alignment for
Weakly-supervised 3D Segmentation [67.30502812804271]
Pseudo-labels are widely employed in weakly supervised 3D segmentation tasks where only sparse ground-truth labels are available for learning.
We propose a novel learning strategy to regularize the generated pseudo-labels and effectively narrow the gaps between pseudo-labels and model predictions.
arXiv Detail & Related papers (2023-05-25T08:19:31Z) - Are labels informative in semi-supervised learning? -- Estimating and
leveraging the missing-data mechanism [4.675583319625962]
Semi-supervised learning is a powerful technique for leveraging unlabeled data to improve machine learning models.
It can be affected by the presence of informative'' labels, which occur when some classes are more likely to be labeled than others.
We propose a novel approach to address this issue by estimating the missing-data mechanism and using inverse propensity weighting to debias any SSL algorithm.
arXiv Detail & Related papers (2023-02-15T09:18:46Z) - SLaM: Student-Label Mixing for Distillation with Unlabeled Examples [15.825078347452024]
We present a principled method for knowledge distillation with unlabeled examples that we call Student-Label Mixing (SLaM)
SLaM consistently improves over prior approaches by evaluating it on several standard benchmarks.
We give an algorithm improving the best-known sample complexity for learning halfspaces with margin under random classification noise.
arXiv Detail & Related papers (2023-02-08T00:14:44Z) - Learning from Multiple Unlabeled Datasets with Partial Risk
Regularization [80.54710259664698]
In this paper, we aim to learn an accurate classifier without any class labels.
We first derive an unbiased estimator of the classification risk that can be estimated from the given unlabeled sets.
We then find that the classifier obtained as such tends to cause overfitting as its empirical risks go negative during training.
Experiments demonstrate that our method effectively mitigates overfitting and outperforms state-of-the-art methods for learning from multiple unlabeled sets.
arXiv Detail & Related papers (2022-07-04T16:22:44Z) - Learning with Proper Partial Labels [87.65718705642819]
Partial-label learning is a kind of weakly-supervised learning with inexact labels.
We show that this proper partial-label learning framework includes many previous partial-label learning settings.
We then derive a unified unbiased estimator of the classification risk.
arXiv Detail & Related papers (2021-12-23T01:37:03Z) - It Takes Two to Tango: Mixup for Deep Metric Learning [16.60855728302127]
State-of-the-art methods focus mostly on sophisticated loss functions or mining strategies.
Mixup is a powerful data augmentation approach interpolating two or more examples and corresponding target labels at a time.
We show that mixing inputs, intermediate representations or embeddings along with target labels significantly improves representations and outperforms state-of-the-art metric learning methods on four benchmark datasets.
arXiv Detail & Related papers (2021-06-09T11:20:03Z) - Not All Unlabeled Data are Equal: Learning to Weight Data in
Semi-supervised Learning [135.89676456312247]
We show how to use a different weight for every unlabeled example.
We adjust those weights via an algorithm based on the influence function.
We demonstrate that this technique outperforms state-of-the-art methods on semi-supervised image and language classification tasks.
arXiv Detail & Related papers (2020-07-02T17:59:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.