Masked Proxy Loss For Text-Independent Speaker Verification
- URL: http://arxiv.org/abs/2011.04491v2
- Date: Fri, 25 Jun 2021 03:10:18 GMT
- Title: Masked Proxy Loss For Text-Independent Speaker Verification
- Authors: Jiachen Lian, Aiswarya Vinod Kumar, Hira Dhamyal, Bhiksha Raj, Rita
Singh
- Abstract summary: This paper proposes a Masked Proxy (MP) loss which directly incorporates both proxy-based relationships and pair-based relationships.
We further propose Multinomial Masked Proxy (MMP) loss to leverage the hardness of speaker pairs.
- Score: 27.417484680749784
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Open-set speaker recognition can be regarded as a metric learning problem,
which is to maximize inter-class variance and minimize intra-class variance.
Supervised metric learning can be categorized into entity-based learning and
proxy-based learning. Most of the existing metric learning objectives like
Contrastive, Triplet, Prototypical, GE2E, etc all belong to the former
division, the performance of which is either highly dependent on sample mining
strategy or restricted by insufficient label information in the mini-batch.
Proxy-based losses mitigate both shortcomings, however, fine-grained
connections among entities are either not or indirectly leveraged. This paper
proposes a Masked Proxy (MP) loss which directly incorporates both proxy-based
relationships and pair-based relationships. We further propose Multinomial
Masked Proxy (MMP) loss to leverage the hardness of speaker pairs. These
methods have been applied to evaluate on VoxCeleb test set and reach
state-of-the-art Equal Error Rate(EER).
Related papers
- Deep Metric Learning with Soft Orthogonal Proxies [1.823505080809275]
We propose a novel approach that introduces Soft Orthogonality (SO) constraint on proxies.
Our approach leverages Data-Efficient Image Transformer (DeiT) as an encoder to extract contextual features from images along with a DML objective.
Our evaluations demonstrate the superiority of our proposed approach over state-of-the-art methods by a significant margin.
arXiv Detail & Related papers (2023-06-22T17:22:15Z) - Robust Calibrate Proxy Loss for Deep Metric Learning [6.784952050036532]
We propose a Calibrate Proxy structure, which uses the real sample information to improve the similarity calculation in proxy-based loss.
We show that our approach can effectively improve the performance of commonly used proxy-based losses on both regular and noisy datasets.
arXiv Detail & Related papers (2023-04-06T02:43:10Z) - On Medians of (Randomized) Pairwise Means [8.497456090408084]
Tournament procedures, recently introduced in Lugosi & Mendelson, offer an appealing alternative to the principle of Empirical Risk Minimization in machine learning.
This paper extends this approach to address other learning problems, in particular for which the performance criterion takes the form of an expectation over pairs of observations.
arXiv Detail & Related papers (2022-11-01T17:18:15Z) - ADPS: Asymmetric Distillation Post-Segmentation for Image Anomaly
Detection [75.68023968735523]
Knowledge Distillation-based Anomaly Detection (KDAD) methods rely on the teacher-student paradigm to detect and segment anomalous regions.
We propose an innovative approach called Asymmetric Distillation Post-Segmentation (ADPS)
Our ADPS employs an asymmetric distillation paradigm that takes distinct forms of the same image as the input of the teacher-student networks.
We show that ADPS significantly improves Average Precision (AP) metric by 9% and 20% on the MVTec AD and KolektorSDD2 datasets.
arXiv Detail & Related papers (2022-10-19T12:04:47Z) - A Non-isotropic Probabilistic Take on Proxy-based Deep Metric Learning [49.999268109518255]
Proxy-based Deep Metric Learning learns by embedding images close to their class representatives (proxies)
In addition, proxy-based DML struggles to learn class-internal structures.
We introduce non-isotropic probabilistic proxy-based DML to address both issues.
arXiv Detail & Related papers (2022-07-08T09:34:57Z) - How Fine-Tuning Allows for Effective Meta-Learning [50.17896588738377]
We present a theoretical framework for analyzing representations derived from a MAML-like algorithm.
We provide risk bounds on the best predictor found by fine-tuning via gradient descent, demonstrating that the algorithm can provably leverage the shared structure.
This separation result underscores the benefit of fine-tuning-based methods, such as MAML, over methods with "frozen representation" objectives in few-shot learning.
arXiv Detail & Related papers (2021-05-05T17:56:00Z) - Proxy Synthesis: Learning with Synthetic Classes for Deep Metric
Learning [13.252164137961332]
We propose a simple regularizer called Proxy Synthesis that exploits synthetic classes for stronger generalization in deep metric learning.
The proposed method generates synthetic embeddings and proxies that work as synthetic classes, and they mimic unseen classes when computing proxy-based losses.
Our method is applicable to any proxy-based losses, including softmax and its variants.
arXiv Detail & Related papers (2021-03-29T09:39:07Z) - Fewer is More: A Deep Graph Metric Learning Perspective Using Fewer
Proxies [65.92826041406802]
We propose a Proxy-based deep Graph Metric Learning approach from the perspective of graph classification.
Multiple global proxies are leveraged to collectively approximate the original data points for each class.
We design a novel reverse label propagation algorithm, by which the neighbor relationships are adjusted according to ground-truth labels.
arXiv Detail & Related papers (2020-10-26T14:52:42Z) - Towards Certified Robustness of Distance Metric Learning [53.96113074344632]
We advocate imposing an adversarial margin in the input space so as to improve the generalization and robustness of metric learning algorithms.
We show that the enlarged margin is beneficial to the generalization ability by using the theoretical technique of algorithmic robustness.
arXiv Detail & Related papers (2020-06-10T16:51:53Z) - Proxy Anchor Loss for Deep Metric Learning [47.832107446521626]
We present a new proxy-based loss that takes advantages of both pair- and proxy-based methods and overcomes their limitations.
Thanks to the use of proxies, our loss boosts the speed of convergence and is robust against noisy labels and outliers.
Our method is evaluated on four public benchmarks, where a standard network trained with our loss achieves state-of-the-art performance and most quickly converges.
arXiv Detail & Related papers (2020-03-31T02:05:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.