Learning Cross-view Geo-localization Embeddings via Dynamic Weighted
Decorrelation Regularization
- URL: http://arxiv.org/abs/2211.05296v1
- Date: Thu, 10 Nov 2022 02:13:10 GMT
- Title: Learning Cross-view Geo-localization Embeddings via Dynamic Weighted
Decorrelation Regularization
- Authors: Tingyu Wang, Zhedong Zheng, Zunjie Zhu, Yuhan Gao, Yi Yang and
Chenggang Yan
- Abstract summary: Cross-view geo-localization aims to spot images of the same location shot from two platforms, e.g., the drone platform and the satellite platform.
Existing methods usually focus on optimizing the distance between one embedding with others in the feature space.
In this paper, we argue that the low redundancy is also of importance, which motivates the model to mine more diverse patterns.
- Score: 52.493240055559916
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cross-view geo-localization aims to spot images of the same location shot
from two platforms, e.g., the drone platform and the satellite platform.
Existing methods usually focus on optimizing the distance between one embedding
with others in the feature space, while neglecting the redundancy of the
embedding itself. In this paper, we argue that the low redundancy is also of
importance, which motivates the model to mine more diverse patterns. To verify
this point, we introduce a simple yet effective regularization, i.e., Dynamic
Weighted Decorrelation Regularization (DWDR), to explicitly encourage networks
to learn independent embedding channels. As the name implies, DWDR regresses
the embedding correlation coefficient matrix to a sparse matrix, i.e., the
identity matrix, with dynamic weights. The dynamic weights are applied to focus
on still correlated channels during training. Besides, we propose a cross-view
symmetric sampling strategy, which keeps the example balance between different
platforms. Albeit simple, the proposed method has achieved competitive results
on three large-scale benchmarks, i.e., University-1652, CVUSA and CVACT.
Moreover, under the harsh circumstance, e.g., the extremely short feature of 64
dimensions, the proposed method surpasses the baseline model by a clear margin.
Related papers
- Exploiting Point-Wise Attention in 6D Object Pose Estimation Based on
Bidirectional Prediction [22.894810893732416]
The paper proposes a bidirectional correspondence prediction network with a point-wise attention-aware mechanism.
Our key insight is that the correlations between each model point and scene point provide essential information for learning point-pair matches.
Experimental results on the public datasets of LineMOD, YCB-Video, and Occ-LineMOD show that the proposed method achieves better performance than other state-of-the-art methods.
arXiv Detail & Related papers (2023-08-16T17:13:45Z) - Uncertainty Guided Adaptive Warping for Robust and Efficient Stereo
Matching [77.133400999703]
Correlation based stereo matching has achieved outstanding performance.
Current methods with a fixed model do not work uniformly well across various datasets.
This paper proposes a new perspective to dynamically calculate correlation for robust stereo matching.
arXiv Detail & Related papers (2023-07-26T09:47:37Z) - Phantom Embeddings: Using Embedding Space for Model Regularization in
Deep Neural Networks [12.293294756969477]
The strength of machine learning models stems from their ability to learn complex function approximations from data.
The complex models tend to memorize the training data, which results in poor regularization performance on test data.
We present a novel approach to regularize the models by leveraging the information-rich latent embeddings and their high intra-class correlation.
arXiv Detail & Related papers (2023-04-14T17:15:54Z) - Robust Representation via Dynamic Feature Aggregation [44.927408735490005]
Deep convolutional neural network (CNN) based models are vulnerable to adversarial attacks.
We propose a method, denoted as Dynamic Feature Aggregation, to compress the embedding space with a novel regularization.
An averaging accuracy of 56.91% is achieved by our method on CIFAR-10 against various attack methods.
arXiv Detail & Related papers (2022-05-16T06:22:15Z) - Distributionally Robust Models with Parametric Likelihood Ratios [123.05074253513935]
Three simple ideas allow us to train models with DRO using a broader class of parametric likelihood ratios.
We find that models trained with the resulting parametric adversaries are consistently more robust to subpopulation shifts when compared to other DRO approaches.
arXiv Detail & Related papers (2022-04-13T12:43:12Z) - Scaling Structured Inference with Randomization [64.18063627155128]
We propose a family of dynamic programming (RDP) randomized for scaling structured models to tens of thousands of latent states.
Our method is widely applicable to classical DP-based inference.
It is also compatible with automatic differentiation so can be integrated with neural networks seamlessly.
arXiv Detail & Related papers (2021-12-07T11:26:41Z) - Index $t$-SNE: Tracking Dynamics of High-Dimensional Datasets with
Coherent Embeddings [1.7188280334580195]
This paper presents a methodology to reuse an embedding to create a new one, where cluster positions are preserved.
The proposed algorithm has the same complexity as the original $t$-SNE to embed new items, and a lower one when considering the embedding of a dataset sliced into sub-pieces.
arXiv Detail & Related papers (2021-09-22T06:45:37Z) - GELATO: Geometrically Enriched Latent Model for Offline Reinforcement
Learning [54.291331971813364]
offline reinforcement learning approaches can be divided into proximal and uncertainty-aware methods.
In this work, we demonstrate the benefit of combining the two in a latent variational model.
Our proposed metrics measure both the quality of out of distribution samples as well as the discrepancy of examples in the data.
arXiv Detail & Related papers (2021-02-22T19:42:40Z) - Attentional-Biased Stochastic Gradient Descent [74.49926199036481]
We present a provable method (named ABSGD) for addressing the data imbalance or label noise problem in deep learning.
Our method is a simple modification to momentum SGD where we assign an individual importance weight to each sample in the mini-batch.
ABSGD is flexible enough to combine with other robust losses without any additional cost.
arXiv Detail & Related papers (2020-12-13T03:41:52Z) - Robust Locality-Aware Regression for Labeled Data Classification [5.432221650286726]
We propose a new discriminant feature extraction framework, namely Robust Locality-Aware Regression (RLAR)
In our model, we introduce a retargeted regression to perform the marginal representation learning adaptively instead of using the general average inter-class margin.
To alleviate the disturbance of outliers and prevent overfitting, we measure the regression term and locality-aware term together with the regularization term by the L2,1 norm.
arXiv Detail & Related papers (2020-06-15T11:36:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.