A Comparison of Metric Learning Loss Functions for End-To-End Speaker
Verification
- URL: http://arxiv.org/abs/2003.14021v1
- Date: Tue, 31 Mar 2020 08:36:07 GMT
- Title: A Comparison of Metric Learning Loss Functions for End-To-End Speaker
Verification
- Authors: Juan M. Coria, Herv\'e Bredin, Sahar Ghannay, Sophie Rosset
- Abstract summary: We compare several metric learning loss functions in a systematic manner on the VoxCeleb dataset.
We show that the additive angular margin loss function outperforms all other loss functions in the study.
Based on a combination of SincNet trainable features and the x-vector architecture, the network used in this paper brings us a step closer to a really-end-to-end speaker verification system.
- Score: 4.617249742207066
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the growing popularity of metric learning approaches, very little
work has attempted to perform a fair comparison of these techniques for speaker
verification. We try to fill this gap and compare several metric learning loss
functions in a systematic manner on the VoxCeleb dataset. The first family of
loss functions is derived from the cross entropy loss (usually used for
supervised classification) and includes the congenerous cosine loss, the
additive angular margin loss, and the center loss. The second family of loss
functions focuses on the similarity between training samples and includes the
contrastive loss and the triplet loss. We show that the additive angular margin
loss function outperforms all other loss functions in the study, while learning
more robust representations. Based on a combination of SincNet trainable
features and the x-vector architecture, the network used in this paper brings
us a step closer to a really-end-to-end speaker verification system, when
combined with the additive angular margin loss, while still being competitive
with the x-vector baseline. In the spirit of reproducible research, we also
release open source Python code for reproducing our results, and share
pretrained PyTorch models on torch.hub that can be used either directly or
after fine-tuning.
Related papers
- Class Anchor Margin Loss for Content-Based Image Retrieval [97.81742911657497]
We propose a novel repeller-attractor loss that falls in the metric learning paradigm, yet directly optimize for the L2 metric without the need of generating pairs.
We evaluate the proposed objective in the context of few-shot and full-set training on the CBIR task, by using both convolutional and transformer architectures.
arXiv Detail & Related papers (2023-06-01T12:53:10Z) - SuSana Distancia is all you need: Enforcing class separability in metric
learning via two novel distance-based loss functions for few-shot image
classification [0.9236074230806579]
We propose two loss functions which consider the importance of the embedding vectors by looking at the intra-class and inter-class distance between the few data.
Our results show a significant improvement in accuracy in the miniImagenNet benchmark compared to other metric-based few-shot learning methods by a margin of 2%.
arXiv Detail & Related papers (2023-05-15T23:12:09Z) - Alternate Loss Functions for Classification and Robust Regression Can Improve the Accuracy of Artificial Neural Networks [6.452225158891343]
This paper shows that training speed and final accuracy of neural networks can significantly depend on the loss function used to train neural networks.
Two new classification loss functions that significantly improve performance on a wide variety of benchmark tasks are proposed.
arXiv Detail & Related papers (2023-03-17T12:52:06Z) - Xtreme Margin: A Tunable Loss Function for Binary Classification
Problems [0.0]
We provide an overview of a novel loss function, the Xtreme Margin loss function.
Unlike the binary cross-entropy and the hinge loss functions, this loss function provides researchers and practitioners flexibility with their training process.
arXiv Detail & Related papers (2022-10-31T22:39:32Z) - Do Lessons from Metric Learning Generalize to Image-Caption Retrieval? [67.45267657995748]
The triplet loss with semi-hard negatives has become the de facto choice for image-caption retrieval (ICR) methods that are optimized from scratch.
Recent progress in metric learning has given rise to new loss functions that outperform the triplet loss on tasks such as image retrieval and representation learning.
We ask whether these findings generalize to the setting of ICR by comparing three loss functions on two ICR methods.
arXiv Detail & Related papers (2022-02-14T15:18:00Z) - Incremental Class Learning using Variational Autoencoders with
Similarity Learning [0.0]
Catastrophic forgetting in neural networks during incremental learning remains a challenging problem.
Our research investigates catastrophic forgetting for four well-known metric-based loss functions during incremental class learning.
The angular loss was least affected, followed by contrastive, triplet loss, and centre loss with good mining techniques.
arXiv Detail & Related papers (2021-10-04T10:19:53Z) - Learning with Noisy Labels via Sparse Regularization [76.31104997491695]
Learning with noisy labels is an important task for training accurate deep neural networks.
Some commonly-used loss functions, such as Cross Entropy (CE), suffer from severe overfitting to noisy labels.
We introduce the sparse regularization strategy to approximate the one-hot constraint.
arXiv Detail & Related papers (2021-07-31T09:40:23Z) - Asymmetric Loss Functions for Learning with Noisy Labels [82.50250230688388]
We propose a new class of loss functions, namely textitasymmetric loss functions, which are robust to learning with noisy labels for various types of noise.
Experimental results on benchmark datasets demonstrate that asymmetric loss functions can outperform state-of-the-art methods.
arXiv Detail & Related papers (2021-06-06T12:52:48Z) - Searching for Robustness: Loss Learning for Noisy Classification Tasks [81.70914107917551]
We parameterize a flexible family of loss functions using Taylors and apply evolutionary strategies to search for noise-robust losses in this space.
The resulting white-box loss provides a simple and fast "plug-and-play" module that enables effective noise-robust learning in diverse downstream tasks.
arXiv Detail & Related papers (2021-02-27T15:27:22Z) - An Equivalence between Loss Functions and Non-Uniform Sampling in
Experience Replay [72.23433407017558]
We show that any loss function evaluated with non-uniformly sampled data can be transformed into another uniformly sampled loss function.
Surprisingly, we find in some environments PER can be replaced entirely by this new loss function without impact to empirical performance.
arXiv Detail & Related papers (2020-07-12T17:45:24Z) - All your loss are belong to Bayes [28.393499629583786]
Loss functions are a cornerstone of machine learning and the starting point of most algorithms.
We introduce a trick on squared Gaussian Processes to obtain a random process whose paths are compliant source functions.
Experimental results demonstrate substantial improvements over the state of the art.
arXiv Detail & Related papers (2020-06-08T14:31:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.