Low-Rank Robust Online Distance/Similarity Learning based on the
Rescaled Hinge Loss
- URL: http://arxiv.org/abs/2010.03268v2
- Date: Sat, 10 Oct 2020 09:30:36 GMT
- Title: Low-Rank Robust Online Distance/Similarity Learning based on the
Rescaled Hinge Loss
- Authors: Davood Zabihzadeh, Amar Tuama, Ali Karami-Mollaee
- Abstract summary: Existing online methods usually assume training triplets or pairwise constraints are exist in advance.
We formulate the online Distance-Similarity learning problem with the robust Rescaled hinge loss function.
The proposed model is rather general and can be applied to any PA-based online Distance-Similarity algorithm.
- Score: 0.34376560669160383
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An important challenge in metric learning is scalability to both size and
dimension of input data. Online metric learning algorithms are proposed to
address this challenge. Existing methods are commonly based on (Passive
Aggressive) PA approach. Hence, they can rapidly process large volumes of data
with an adaptive learning rate. However, these algorithms are based on the
Hinge loss and so are not robust against outliers and label noise. Also,
existing online methods usually assume training triplets or pairwise
constraints are exist in advance. However, many datasets in real-world
applications are in the form of input data and their associated labels. We
address these challenges by formulating the online Distance-Similarity learning
problem with the robust Rescaled hinge loss function. The proposed model is
rather general and can be applied to any PA-based online Distance-Similarity
algorithm. Also, we develop an efficient robust one-pass triplet construction
algorithm. Finally, to provide scalability in high dimensional DML
environments, the low-rank version of the proposed methods is presented that
not only reduces the computational cost significantly but also keeps the
predictive performance of the learned metrics. Also, it provides a
straightforward extension of our methods for deep Distance-Similarity learning.
We conduct several experiments on datasets from various applications. The
results confirm that the proposed methods significantly outperform
state-of-the-art online DML methods in the presence of label noise and outliers
by a large margin.
Related papers
- Stabilizing Linear Passive-Aggressive Online Learning with Weighted Reservoir Sampling [46.01254613933967]
Online learning methods are still highly effective for high-dimensional streaming data, out-of-core processing, and other throughput-sensitive applications.
Many such algorithms rely on fast adaptation to individual errors as a key to their convergence.
While such algorithms enjoy low theoretical regret, in real-world deployment they can be sensitive to individual outliers that cause the algorithm to over-correct.
arXiv Detail & Related papers (2024-10-31T03:35:48Z) - Simple Ingredients for Offline Reinforcement Learning [86.1988266277766]
offline reinforcement learning algorithms have proven effective on datasets highly connected to the target downstream task.
We show that existing methods struggle with diverse data: their performance considerably deteriorates as data collected for related but different tasks is simply added to the offline buffer.
We show that scale, more than algorithmic considerations, is the key factor influencing performance.
arXiv Detail & Related papers (2024-03-19T18:57:53Z) - Fighting Uncertainty with Gradients: Offline Reinforcement Learning via
Diffusion Score Matching [22.461036967440723]
We study smoothed distance to data as an uncertainty metric, and claim that it has two beneficial properties.
We show these gradients can be efficiently learned with score-matching techniques.
We propose Score-Guided Planning (SGP) to enable first-order planning in high-dimensional problems.
arXiv Detail & Related papers (2023-06-24T23:40:58Z) - A Survey of Learning on Small Data: Generalization, Optimization, and
Challenge [101.27154181792567]
Learning on small data that approximates the generalization ability of big data is one of the ultimate purposes of AI.
This survey follows the active sampling theory under a PAC framework to analyze the generalization error and label complexity of learning on small data.
Multiple data applications that may benefit from efficient small data representation are surveyed.
arXiv Detail & Related papers (2022-07-29T02:34:19Z) - One-Pass Learning via Bridging Orthogonal Gradient Descent and Recursive
Least-Squares [8.443742714362521]
We develop an algorithm for one-pass learning which seeks to perfectly fit every new datapoint while changing the parameters in a direction that causes the least change to the predictions on previous datapoints.
Our algorithm uses the memory efficiently by exploiting the structure of the streaming data via an incremental principal component analysis (IPCA)
Our experiments show the effectiveness of the proposed method compared to the baselines.
arXiv Detail & Related papers (2022-07-28T02:01:31Z) - Simple Stochastic and Online Gradient DescentAlgorithms for Pairwise
Learning [65.54757265434465]
Pairwise learning refers to learning tasks where the loss function depends on a pair instances.
Online descent (OGD) is a popular approach to handle streaming data in pairwise learning.
In this paper, we propose simple and online descent to methods for pairwise learning.
arXiv Detail & Related papers (2021-11-23T18:10:48Z) - Attentional-Biased Stochastic Gradient Descent [74.49926199036481]
We present a provable method (named ABSGD) for addressing the data imbalance or label noise problem in deep learning.
Our method is a simple modification to momentum SGD where we assign an individual importance weight to each sample in the mini-batch.
ABSGD is flexible enough to combine with other robust losses without any additional cost.
arXiv Detail & Related papers (2020-12-13T03:41:52Z) - An Online Method for A Class of Distributionally Robust Optimization
with Non-Convex Objectives [54.29001037565384]
We propose a practical online method for solving a class of online distributionally robust optimization (DRO) problems.
Our studies demonstrate important applications in machine learning for improving the robustness of networks.
arXiv Detail & Related papers (2020-06-17T20:19:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.