Neural Processes with Stochastic Attention: Paying more attention to the
context dataset
- URL: http://arxiv.org/abs/2204.05449v1
- Date: Mon, 11 Apr 2022 23:57:19 GMT
- Title: Neural Processes with Stochastic Attention: Paying more attention to the
context dataset
- Authors: Mingyu Kim, Kyeongryeol Go, Se-Young Yun
- Abstract summary: Neural processes (NPs) aim to complete unseen data points based on a given context dataset.
We propose a attention mechanism for NPs to capture appropriate context information.
We empirically show that our approach substantially outperforms conventional NPs in various domains.
- Score: 11.301294319986477
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Neural processes (NPs) aim to stochastically complete unseen data points
based on a given context dataset. NPs essentially leverage a given dataset as a
context representation to derive a suitable identifier for a novel task. To
improve the prediction accuracy, many variants of NPs have investigated context
embedding approaches that generally design novel network architectures and
aggregation functions satisfying permutation invariant. In this work, we
propose a stochastic attention mechanism for NPs to capture appropriate context
information. From the perspective of information theory, we demonstrate that
the proposed method encourages context embedding to be differentiated from a
target dataset, allowing NPs to consider features in a target dataset and
context embedding independently. We observe that the proposed method can
appropriately capture context embedding even under noisy data sets and
restricted task distributions, where typical NPs suffer from a lack of context
embeddings. We empirically show that our approach substantially outperforms
conventional NPs in various domains through 1D regression, predator-prey model,
and image completion. Moreover, the proposed method is also validated by
MovieLens-10k dataset, a real-world problem.
Related papers
- In-Context In-Context Learning with Transformer Neural Processes [50.57807892496024]
We develop the in-context in-context learning pseudo-token TNP (ICICL-TNP)
The ICICL-TNP is capable of conditioning on both sets of datapoints and sets of datasets, enabling it to perform in-context in-context learning.
We demonstrate the importance of in-context in-context learning and the effectiveness of the ICICL-TNP in a number of experiments.
arXiv Detail & Related papers (2024-06-19T12:26:36Z) - Enhancing Neural Subset Selection: Integrating Background Information into Set Representations [53.15923939406772]
We show that when the target value is conditioned on both the input set and subset, it is essential to incorporate an textitinvariant sufficient statistic of the superset into the subset of interest.
This ensures that the output value remains invariant to permutations of the subset and its corresponding superset, enabling identification of the specific superset from which the subset originated.
arXiv Detail & Related papers (2024-02-05T16:09:35Z) - Latent Bottlenecked Attentive Neural Processes [71.18817592128207]
We present Latent Bottlenecked Attentive Neural Processes (LBANPs)
LBANPs have a querying computational complexity independent of the number of context datapoints.
We show LBANPs achieve results competitive with the state-of-the-art on meta-regression, image completion, and contextual multi-armed bandits.
arXiv Detail & Related papers (2022-11-15T19:21:41Z) - Message Passing Neural Processes [3.0969191504482247]
We introduce Message Passing Neural Processes (MPNPs), which explicitly makes use of relational structure within the model.
MPNPs thrive at lower sampling rates, on existing benchmarks and newly-proposed CA and Cora-Branched tasks.
We report strong generalisation over density-based CA rulesets and significant gains in challenging arbitrary-labelling and few-shot learning setups.
arXiv Detail & Related papers (2020-09-29T09:40:09Z) - Bootstrapping Neural Processes [114.97111530885093]
Neural Processes (NPs) implicitly define a broad class of processes with neural networks.
NPs still rely on an assumption that uncertainty in processes is modeled by a single latent variable.
We propose the Boostrapping Neural Process (BNP), a novel extension of the NP family using the bootstrap.
arXiv Detail & Related papers (2020-08-07T02:23:34Z) - Meta-Learning Stationary Stochastic Process Prediction with
Convolutional Neural Processes [32.02612871707347]
We propose ConvNP, which endows Neural Processes (NPs) with translation equivariance and extends convolutional conditional NPs to allow for dependencies in the predictive distribution.
We demonstrate the strong performance and generalization capabilities of ConvNPs on 1D, regression image completion, and various tasks with real-world-temporal data.
arXiv Detail & Related papers (2020-07-02T18:25:27Z) - A Revised Generative Evaluation of Visual Dialogue [80.17353102854405]
We propose a revised evaluation scheme for the VisDial dataset.
We measure consensus between answers generated by the model and a set of relevant answers.
We release these sets and code for the revised evaluation scheme as DenseVisDial.
arXiv Detail & Related papers (2020-04-20T13:26:45Z) - Supervised Learning for Non-Sequential Data: A Canonical Polyadic
Decomposition Approach [85.12934750565971]
Efficient modelling of feature interactions underpins supervised learning for non-sequential tasks.
To alleviate this issue, it has been proposed to implicitly represent the model parameters as a tensor.
For enhanced expressiveness, we generalize the framework to allow feature mapping to arbitrarily high-dimensional feature vectors.
arXiv Detail & Related papers (2020-01-27T22:38:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.