Spherical Feature Transform for Deep Metric Learning
- URL: http://arxiv.org/abs/2008.01469v1
- Date: Tue, 4 Aug 2020 11:32:23 GMT
- Title: Spherical Feature Transform for Deep Metric Learning
- Authors: Yuke Zhu, Yan Bai, Yichen Wei
- Abstract summary: This work proposes a novel spherical feature transform approach.
It relaxes the assumption of identical covariance between classes to an assumption of similar covariances of different classes on a hypersphere.
We provide a simple and effective training method, and in depth analysis on the relation between the two different transforms.
- Score: 58.35971328774927
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data augmentation in feature space is effective to increase data diversity.
Previous methods assume that different classes have the same covariance in
their feature distributions. Thus, feature transform between different classes
is performed via translation. However, this approach is no longer valid for
recent deep metric learning scenarios, where feature normalization is widely
adopted and all features lie on a hypersphere.
This work proposes a novel spherical feature transform approach. It relaxes
the assumption of identical covariance between classes to an assumption of
similar covariances of different classes on a hypersphere. Consequently, the
feature transform is performed by a rotation that respects the spherical data
distributions. We provide a simple and effective training method, and in depth
analysis on the relation between the two different transforms. Comprehensive
experiments on various deep metric learning benchmarks and different baselines
verify that our method achieves consistent performance improvement and
state-of-the-art results.
Related papers
- Equivariant Spatio-Temporal Self-Supervision for LiDAR Object Detection [37.142470149311904]
We propose atemporal equivariant learning framework by considering both spatial and temporal augmentations jointly.
We show our pre-training method for 3D object detection which outperforms existing equivariant and invariant approaches in many settings.
arXiv Detail & Related papers (2024-04-17T20:41:49Z) - Information Geometrically Generalized Covariate Shift Adaptation [5.990174495635326]
Many machine learning methods assume that the training and test data follow the same distribution.
We show that the well-known family of covariate shift adaptation methods is unified in the framework of information geometry.
Numerical experiments show that our generalization can achieve better performance than the existing methods it encompasses.
arXiv Detail & Related papers (2023-04-19T02:52:54Z) - Equivariance Discovery by Learned Parameter-Sharing [153.41877129746223]
We study how to discover interpretable equivariances from data.
Specifically, we formulate this discovery process as an optimization problem over a model's parameter-sharing schemes.
Also, we theoretically analyze the method for Gaussian data and provide a bound on the mean squared gap between the studied discovery scheme and the oracle scheme.
arXiv Detail & Related papers (2022-04-07T17:59:19Z) - Measuring dissimilarity with diffeomorphism invariance [94.02751799024684]
We introduce DID, a pairwise dissimilarity measure applicable to a wide range of data spaces.
We prove that DID enjoys properties which make it relevant for theoretical study and practical use.
arXiv Detail & Related papers (2022-02-11T13:51:30Z) - Revisiting Consistency Regularization for Semi-Supervised Learning [80.28461584135967]
We propose an improved consistency regularization framework by a simple yet effective technique, FeatDistLoss.
Experimental results show that our model defines a new state of the art for various datasets and settings.
arXiv Detail & Related papers (2021-12-10T20:46:13Z) - Towards Improved and Interpretable Deep Metric Learning via Attentive
Grouping [103.71992720794421]
Grouping has been commonly used in deep metric learning for computing diverse features.
We propose an improved and interpretable grouping method to be integrated flexibly with any metric learning framework.
arXiv Detail & Related papers (2020-11-17T19:08:24Z) - What causes the test error? Going beyond bias-variance via ANOVA [21.359033212191218]
Modern machine learning methods are often overparametrized, allowing adaptation to the data at a fine level.
Recent work aimed to understand in greater depth why overparametrization is helpful for generalization.
We propose using the analysis of variance (ANOVA) to decompose the variance in the test error in a symmetric way.
arXiv Detail & Related papers (2020-10-11T05:21:13Z) - Invariant Integration in Deep Convolutional Feature Space [77.99182201815763]
We show how to incorporate prior knowledge to a deep neural network architecture in a principled manner.
We report state-of-the-art performance on the Rotated-MNIST dataset.
arXiv Detail & Related papers (2020-04-20T09:45:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.