Learning to Represent and Predict Sets with Deep Neural Networks
- URL: http://arxiv.org/abs/2103.04957v1
- Date: Mon, 8 Mar 2021 18:27:08 GMT
- Title: Learning to Represent and Predict Sets with Deep Neural Networks
- Authors: Yan Zhang
- Abstract summary: We develop various techniques for working with sets in machine learning.
The first focus of this thesis is the learning of better set representations (sets as input)
The second focus of this thesis is the prediction of sets (sets as output)
- Score: 4.310814582717413
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this thesis, we develop various techniques for working with sets in
machine learning. Each input or output is not an image or a sequence, but a
set: an unordered collection of multiple objects, each object described by a
feature vector. Their unordered nature makes them suitable for modeling a wide
variety of data, ranging from objects in images to point clouds to graphs. Deep
learning has recently shown great success on other types of structured data, so
we aim to build the necessary structures for sets into deep neural networks.
The first focus of this thesis is the learning of better set representations
(sets as input). Existing approaches have bottlenecks that prevent them from
properly modeling relations between objects within the set. To address this
issue, we develop a variety of techniques for different scenarios and show that
alleviating the bottleneck leads to consistent improvements across many
experiments.
The second focus of this thesis is the prediction of sets (sets as output).
Current approaches do not take the unordered nature of sets into account
properly. We determine that this results in a problem that causes discontinuity
issues with many set prediction tasks and prevents them from learning some
extremely simple datasets. To avoid this problem, we develop two models that
properly take the structure of sets into account. Various experiments show that
our set prediction techniques can significantly benefit over existing
approaches.
Related papers
- A Closer Look at Deep Learning Methods on Tabular Datasets [78.61845513154502]
We present an extensive study on TALENT, a collection of 300+ datasets spanning broad ranges of size.<n>Our evaluation shows that ensembling benefits both tree-based and neural approaches.
arXiv Detail & Related papers (2024-07-01T04:24:07Z) - Pushing the Limits of Pre-training for Time Series Forecasting in the
CloudOps Domain [54.67888148566323]
We introduce three large-scale time series forecasting datasets from the cloud operations domain.
We show it is a strong zero-shot baseline and benefits from further scaling, both in model and dataset size.
Accompanying these datasets and results is a suite of comprehensive benchmark results comparing classical and deep learning baselines to our pre-trained method.
arXiv Detail & Related papers (2023-10-08T08:09:51Z) - On Inductive Biases for Machine Learning in Data Constrained Settings [0.0]
This thesis explores a different answer to the problem of learning expressive models in data constrained settings.
Instead of relying on big datasets to learn neural networks, we will replace some modules by known functions reflecting the structure of the data.
Our approach falls under the hood of "inductive biases", which can be defined as hypothesis on the data at hand restricting the space of models to explore.
arXiv Detail & Related papers (2023-02-21T14:22:01Z) - Semi-Structured Object Sequence Encoders [9.257633944317735]
We focus on the problem of developing a structure-aware input representation for semi-structured object sequences.
This type of data is often represented as a sequence of sets of key-value pairs over time.
We propose a two-part approach, which first considers each key independently and encodes a representation of its values over time.
arXiv Detail & Related papers (2023-01-03T09:19:41Z) - Mutual Exclusivity Training and Primitive Augmentation to Induce
Compositionality [84.94877848357896]
Recent datasets expose the lack of the systematic generalization ability in standard sequence-to-sequence models.
We analyze this behavior of seq2seq models and identify two contributing factors: a lack of mutual exclusivity bias and the tendency to memorize whole examples.
We show substantial empirical improvements using standard sequence-to-sequence models on two widely-used compositionality datasets.
arXiv Detail & Related papers (2022-11-28T17:36:41Z) - Multivariate Time Series Regression with Graph Neural Networks [0.6124773188525718]
Recent advances in adapting Deep Learning to graphs have shown promising potential in various graph-related tasks.
However, these methods have not been adapted for time series related tasks to a great extent.
In this work, we propose an architecture capable of processing these long sequences in a multivariate time series regression task.
arXiv Detail & Related papers (2022-01-03T16:11:46Z) - Self-Supervision by Prediction for Object Discovery in Videos [62.87145010885044]
In this paper, we use the prediction task as self-supervision and build a novel object-centric model for image sequence representation.
Our framework can be trained without the help of any manual annotation or pretrained network.
Initial experiments confirm that the proposed pipeline is a promising step towards object-centric video prediction.
arXiv Detail & Related papers (2021-03-09T19:14:33Z) - Contrastive Prototype Learning with Augmented Embeddings for Few-Shot
Learning [58.2091760793799]
We propose a novel contrastive prototype learning with augmented embeddings (CPLAE) model.
With a class prototype as an anchor, CPL aims to pull the query samples of the same class closer and those of different classes further away.
Extensive experiments on several benchmarks demonstrate that our proposed CPLAE achieves new state-of-the-art.
arXiv Detail & Related papers (2021-01-23T13:22:44Z) - Assisting Scene Graph Generation with Self-Supervision [21.89909688056478]
We propose a set of three novel yet simple self-supervision tasks and train them as auxiliary multi-tasks to the main model.
While comparing, we train the base-model from scratch with these self-supervision tasks, we achieve state-of-the-art results in all the metrics and recall settings.
arXiv Detail & Related papers (2020-08-08T16:38:03Z) - Predicting Temporal Sets with Deep Neural Networks [50.53727580527024]
We propose an integrated solution based on the deep neural networks for temporal sets prediction.
A unique perspective is to learn element relationship by constructing set-level co-occurrence graph.
We design an attention-based module to adaptively learn the temporal dependency of elements and sets.
arXiv Detail & Related papers (2020-06-20T03:29:02Z) - Learn to Predict Sets Using Feed-Forward Neural Networks [63.91494644881925]
This paper addresses the task of set prediction using deep feed-forward neural networks.
We present a novel approach for learning to predict sets with unknown permutation and cardinality.
We demonstrate the validity of our set formulations on relevant vision problems.
arXiv Detail & Related papers (2020-01-30T01:52:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.