Set Representation Learning with Generalized Sliced-Wasserstein
Embeddings
- URL: http://arxiv.org/abs/2103.03892v1
- Date: Fri, 5 Mar 2021 19:00:34 GMT
- Title: Set Representation Learning with Generalized Sliced-Wasserstein
Embeddings
- Authors: Navid Naderializadeh, Soheil Kolouri, Joseph F. Comer, Reed W.
Andrews, Heiko Hoffmann
- Abstract summary: We propose a geometrically-interpretable framework for learning representations from set-structured data.
In particular, we treat elements of a set as samples from a probability measure and propose an exact Euclidean embedding for Generalized Sliced Wasserstein.
We evaluate our proposed framework on multiple supervised and unsupervised set learning tasks and demonstrate its superiority over state-of-the-art set representation learning approaches.
- Score: 22.845403993200932
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An increasing number of machine learning tasks deal with learning
representations from set-structured data. Solutions to these problems involve
the composition of permutation-equivariant modules (e.g., self-attention, or
individual processing via feed-forward neural networks) and
permutation-invariant modules (e.g., global average pooling, or pooling by
multi-head attention). In this paper, we propose a geometrically-interpretable
framework for learning representations from set-structured data, which is
rooted in the optimal mass transportation problem. In particular, we treat
elements of a set as samples from a probability measure and propose an exact
Euclidean embedding for Generalized Sliced Wasserstein (GSW) distances to learn
from set-structured data effectively. We evaluate our proposed framework on
multiple supervised and unsupervised set learning tasks and demonstrate its
superiority over state-of-the-art set representation learning approaches.
Related papers
- On Discriminative Probabilistic Modeling for Self-Supervised Representation Learning [85.75164588939185]
We study the discriminative probabilistic modeling problem on a continuous domain for (multimodal) self-supervised representation learning.
We conduct generalization error analysis to reveal the limitation of current InfoNCE-based contrastive loss for self-supervised representation learning.
arXiv Detail & Related papers (2024-10-11T18:02:46Z) - Enhancing Neural Subset Selection: Integrating Background Information into Set Representations [53.15923939406772]
We show that when the target value is conditioned on both the input set and subset, it is essential to incorporate an textitinvariant sufficient statistic of the superset into the subset of interest.
This ensures that the output value remains invariant to permutations of the subset and its corresponding superset, enabling identification of the specific superset from which the subset originated.
arXiv Detail & Related papers (2024-02-05T16:09:35Z) - Structurally Diverse Sampling Reduces Spurious Correlations in Semantic
Parsing Datasets [51.095144091781734]
We propose a novel algorithm for sampling a structurally diverse set of instances from a labeled instance pool with structured outputs.
We show that our algorithm performs competitively with or better than prior algorithms in not only compositional template splits but also traditional IID splits.
In general, we find that diverse train sets lead to better generalization than random training sets of the same size in 9 out of 10 dataset-split pairs.
arXiv Detail & Related papers (2022-03-16T07:41:27Z) - Leveraging Ensembles and Self-Supervised Learning for Fully-Unsupervised
Person Re-Identification and Text Authorship Attribution [77.85461690214551]
Learning from fully-unlabeled data is challenging in Multimedia Forensics problems, such as Person Re-Identification and Text Authorship Attribution.
Recent self-supervised learning methods have shown to be effective when dealing with fully-unlabeled data in cases where the underlying classes have significant semantic differences.
We propose a strategy to tackle Person Re-Identification and Text Authorship Attribution by enabling learning from unlabeled data even when samples from different classes are not prominently diverse.
arXiv Detail & Related papers (2022-02-07T13:08:11Z) - Learning to Generalize Compositionally by Transferring Across Semantic
Parsing Tasks [37.66114618645146]
We investigate learning representations that facilitate transfer learning from one compositional task to another.
We apply this method to semantic parsing, using three very different datasets.
Our method significantly improves compositional generalization over baselines on the test set of the target task.
arXiv Detail & Related papers (2021-11-09T09:10:21Z) - Learning Prototype-oriented Set Representations for Meta-Learning [85.19407183975802]
Learning from set-structured data is a fundamental problem that has recently attracted increasing attention.
This paper provides a novel optimal transport based way to improve existing summary networks.
We further instantiate it to the cases of few-shot classification and implicit meta generative modeling.
arXiv Detail & Related papers (2021-10-18T09:49:05Z) - Set-to-Sequence Methods in Machine Learning: a Review [0.0]
Machine learning on sets towards sequential output is an important and ubiquitous task, with applications ranging from language modelling and meta-learning to multi-agent strategy games and power grid optimization.
This paper provides a comprehensive introduction to the field as well as an overview of important machine learning methods tackling both of these key challenges.
arXiv Detail & Related papers (2021-03-17T13:52:33Z) - A Trainable Optimal Transport Embedding for Feature Aggregation and its
Relationship to Attention [96.77554122595578]
We introduce a parametrized representation of fixed size, which embeds and then aggregates elements from a given input set according to the optimal transport plan between the set and a trainable reference.
Our approach scales to large datasets and allows end-to-end training of the reference, while also providing a simple unsupervised learning mechanism with small computational cost.
arXiv Detail & Related papers (2020-06-22T08:35:58Z) - Improved Structural Discovery and Representation Learning of Multi-Agent
Data [5.40729975786985]
We present a dynamic alignment method which provides a robust ordering of structured multi-agent data.
We demonstrate the value of this approach using a large amount of soccer tracking data from a professional league.
arXiv Detail & Related papers (2019-12-30T22:49:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.