Understanding and Exploiting Dependent Variables with Deep Metric
Learning
- URL: http://arxiv.org/abs/2009.03820v1
- Date: Tue, 8 Sep 2020 15:30:45 GMT
- Title: Understanding and Exploiting Dependent Variables with Deep Metric
Learning
- Authors: Niall O' Mahony, Sean Campbell, Anderson Carvalho, Lenka Krpalkova,
Gustavo Velasco-Hernandez, Daniel Riordan, Joseph Walsh
- Abstract summary: Deep Metric Learning (DML) approaches learn to represent inputs to a lower-dimensional latent space.
This paper investigates how the mapping element of DML may be exploited in situations where the salient features in arbitrary classification problems vary over time.
- Score: 0.5025737475817937
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Deep Metric Learning (DML) approaches learn to represent inputs to a
lower-dimensional latent space such that the distance between representations
in this space corresponds with a predefined notion of similarity. This paper
investigates how the mapping element of DML may be exploited in situations
where the salient features in arbitrary classification problems vary over time
or due to changing underlying variables. Examples of such variable features
include seasonal and time-of-day variations in outdoor scenes in place
recognition tasks for autonomous navigation and age/gender variations in
human/animal subjects in classification tasks for medical/ethological studies.
Through the use of visualisation tools for observing the distribution of DML
representations per each query variable for which prior information is
available, the influence of each variable on the classification task may be
better understood. Based on these relationships, prior information on these
salient background variables may be exploited at the inference stage of the DML
approach by using a clustering algorithm to improve classification performance.
This research proposes such a methodology establishing the saliency of query
background variables and formulating clustering algorithms for better
separating latent-space representations at run-time. The paper also discusses
online management strategies to preserve the quality and diversity of data and
the representation of each class in the gallery of embeddings in the DML
approach. We also discuss latent works towards understanding the relevance of
underlying/multiple variables with DML.
Related papers
- Interpetable Target-Feature Aggregation for Multi-Task Learning based on Bias-Variance Analysis [53.38518232934096]
Multi-task learning (MTL) is a powerful machine learning paradigm designed to leverage shared knowledge across tasks to improve generalization and performance.
We propose an MTL approach at the intersection between task clustering and feature transformation based on a two-phase iterative aggregation of targets and features.
In both phases, a key aspect is to preserve the interpretability of the reduced targets and features through the aggregation with the mean, which is motivated by applications to Earth science.
arXiv Detail & Related papers (2024-06-12T08:30:16Z) - Deep Metric Learning for Computer Vision: A Brief Overview [4.980117530293724]
Objective functions that optimize deep neural networks play a vital role in creating an enhanced feature representation of the input data.
Deep Metric Learning seeks to develop methods that aim to measure the similarity between data samples.
We will provide an overview of recent progress in this area and discuss state-of-the-art Deep Metric Learning approaches.
arXiv Detail & Related papers (2023-12-01T21:53:36Z) - Learning Invariant Molecular Representation in Latent Discrete Space [52.13724532622099]
We propose a new framework for learning molecular representations that exhibit invariance and robustness against distribution shifts.
Our model achieves stronger generalization against state-of-the-art baselines in the presence of various distribution shifts.
arXiv Detail & Related papers (2023-10-22T04:06:44Z) - DCID: Deep Canonical Information Decomposition [84.59396326810085]
We consider the problem of identifying the signal shared between two one-dimensional target variables.
We propose ICM, an evaluation metric which can be used in the presence of ground-truth labels.
We also propose Deep Canonical Information Decomposition (DCID) - a simple, yet effective approach for learning the shared variables.
arXiv Detail & Related papers (2023-06-27T16:59:06Z) - Distinguishability Calibration to In-Context Learning [31.375797763897104]
We propose a method to map a PLM-encoded embedding into a new metric space to guarantee the distinguishability of the resulting embeddings.
We also take the advantage of hyperbolic embeddings to capture the hierarchical relations among fine-grained class-associated token embedding.
arXiv Detail & Related papers (2023-02-13T09:15:00Z) - Identifiable Latent Causal Content for Domain Adaptation under Latent Covariate Shift [82.14087963690561]
Multi-source domain adaptation (MSDA) addresses the challenge of learning a label prediction function for an unlabeled target domain.
We present an intricate causal generative model by introducing latent noises across domains, along with a latent content variable and a latent style variable.
The proposed approach showcases exceptional performance and efficacy on both simulated and real-world datasets.
arXiv Detail & Related papers (2022-08-30T11:25:15Z) - Regressing Relative Fine-Grained Change for Sub-Groups in Unreliable
Heterogeneous Data Through Deep Multi-Task Metric Learning [0.5999777817331317]
We investigate how techniques in multi-task metric learning can be applied for theregression of fine-grained change in real data.
The techniques investigated are specifically tailored for handling heterogeneous data sources.
arXiv Detail & Related papers (2022-08-11T12:57:11Z) - Consistency and Diversity induced Human Motion Segmentation [231.36289425663702]
We propose a novel Consistency and Diversity induced human Motion (CDMS) algorithm.
Our model factorizes the source and target data into distinct multi-layer feature spaces.
A multi-mutual learning strategy is carried out to reduce the domain gap between the source and target data.
arXiv Detail & Related papers (2022-02-10T06:23:56Z) - Learning Conditional Invariance through Cycle Consistency [60.85059977904014]
We propose a novel approach to identify meaningful and independent factors of variation in a dataset.
Our method involves two separate latent subspaces for the target property and the remaining input information.
We demonstrate on synthetic and molecular data that our approach identifies more meaningful factors which lead to sparser and more interpretable models.
arXiv Detail & Related papers (2021-11-25T17:33:12Z) - A Framework to Enhance Generalization of Deep Metric Learning methods
using General Discriminative Feature Learning and Class Adversarial Neural
Networks [1.5469452301122175]
Metric learning algorithms aim to learn a distance function that brings semantically similar data items together and keeps dissimilar ones at a distance.
Deep Metric Learning (DML) methods are proposed that automatically extract features from data and learn a non-linear transformation from input space to a semantically embedding space.
We propose a framework to enhance the generalization power of existing DML methods in a Zero-Shot Learning (ZSL) setting.
arXiv Detail & Related papers (2021-06-11T14:24:40Z) - Explainable Multivariate Time Series Classification: A Deep Neural
Network Which Learns To Attend To Important Variables As Well As Informative
Time Intervals [32.30627405832656]
Time series data is prevalent in a wide variety of real-world applications.
Key criterion to understand such predictive models involves elucidating and quantifying the contribution of time varying input variables to the classification.
We introduce a novel, modular, convolution-based feature extraction and attention mechanism that simultaneously identifies the variables as well as time intervals which determine the classification output.
arXiv Detail & Related papers (2020-11-23T19:16:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.