Contrastive Conditional Neural Processes
- URL: http://arxiv.org/abs/2203.03978v1
- Date: Tue, 8 Mar 2022 10:08:45 GMT
- Title: Contrastive Conditional Neural Processes
- Authors: Zesheng Ye, Lina Yao
- Abstract summary: Conditional Neural Processes(CNPs) bridge neural networks with probabilistic inference to approximate functions of Processes under meta-learning settings.
Two auxiliary contrastive branches are set up hierarchically, namely in-instantiation temporal contrastive learning(tt TCL) and cross-instantiation function contrastive learning(tt FCL)
We empirically show that tt TCL captures high-level abstraction of observations, whereas tt FCL helps identify underlying functions, which in turn provides more efficient representations.
- Score: 45.70735205041254
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conditional Neural Processes~(CNPs) bridge neural networks with probabilistic
inference to approximate functions of Stochastic Processes under meta-learning
settings. Given a batch of non-{\it i.i.d} function instantiations, CNPs are
jointly optimized for in-instantiation observation prediction and
cross-instantiation meta-representation adaptation within a generative
reconstruction pipeline. There can be a challenge in tying together such two
targets when the distribution of function observations scales to
high-dimensional and noisy spaces. Instead, noise contrastive estimation might
be able to provide more robust representations by learning distributional
matching objectives to combat such inherent limitation of generative models. In
light of this, we propose to equip CNPs by 1) aligning prediction with encoded
ground-truth observation, and 2) decoupling meta-representation adaptation from
generative reconstruction. Specifically, two auxiliary contrastive branches are
set up hierarchically, namely in-instantiation temporal contrastive
learning~({\tt TCL}) and cross-instantiation function contrastive
learning~({\tt FCL}), to facilitate local predictive alignment and global
function consistency, respectively. We empirically show that {\tt TCL} captures
high-level abstraction of observations, whereas {\tt FCL} helps identify
underlying functions, which in turn provides more efficient representations.
Our model outperforms other CNPs variants when evaluating function distribution
reconstruction and parameter identification across 1D, 2D and high-dimensional
time-series.
Related papers
- Spectral Convolutional Conditional Neural Processes [4.52069311861025]
Conditional Neural Processes (CNPs) constitute a family of probabilistic models that harness the flexibility of neural networks to parameterize processes.
We propose Spectral Convolutional Conditional Neural Processes (SConvCNPs), a new addition to the NPs family that allows for more efficient representation of functions in the frequency domain.
arXiv Detail & Related papers (2024-04-19T21:13:18Z) - Point-Based Value Iteration for POMDPs with Neural Perception Mechanisms [31.51588071503617]
We introduce neuro-symbolic partially observable Markov decision processes (NS-POMDPs)
We propose a novel piecewise linear and convex representation (P-PWLC) in terms of polyhedra covering the state space and value vectors.
We show the practical applicability of our approach on two case studies that employ (trained) ReLU neural networks as perception functions.
arXiv Detail & Related papers (2023-06-30T13:26:08Z) - Learning Unnormalized Statistical Models via Compositional Optimization [73.30514599338407]
Noise-contrastive estimation(NCE) has been proposed by formulating the objective as the logistic loss of the real data and the artificial noise.
In this paper, we study it a direct approach for optimizing the negative log-likelihood of unnormalized models.
arXiv Detail & Related papers (2023-06-13T01:18:16Z) - Intensity Profile Projection: A Framework for Continuous-Time
Representation Learning for Dynamic Networks [50.2033914945157]
We present a representation learning framework, Intensity Profile Projection, for continuous-time dynamic network data.
The framework consists of three stages: estimating pairwise intensity functions, learning a projection which minimises a notion of intensity reconstruction error.
Moreoever, we develop estimation theory providing tight control on the error of any estimated trajectory, indicating that the representations could even be used in quite noise-sensitive follow-on analyses.
arXiv Detail & Related papers (2023-06-09T15:38:25Z) - Versatile Neural Processes for Learning Implicit Neural Representations [57.090658265140384]
We propose Versatile Neural Processes (VNP), which largely increases the capability of approximating functions.
Specifically, we introduce a bottleneck encoder that produces fewer and informative context tokens, relieving the high computational cost.
We demonstrate the effectiveness of the proposed VNP on a variety of tasks involving 1D, 2D and 3D signals.
arXiv Detail & Related papers (2023-01-21T04:08:46Z) - Parameter Decoupling Strategy for Semi-supervised 3D Left Atrium
Segmentation [0.0]
We present a novel semi-supervised segmentation model based on parameter decoupling strategy to encourage consistent predictions from diverse views.
Our method has achieved a competitive result over the state-of-the-art semisupervised methods on the Atrial Challenge dataset.
arXiv Detail & Related papers (2021-09-20T14:51:42Z) - Mitigating Performance Saturation in Neural Marked Point Processes:
Architectures and Loss Functions [50.674773358075015]
We propose a simple graph-based network structure called GCHP, which utilizes only graph convolutional layers.
We show that GCHP can significantly reduce training time and the likelihood ratio loss with interarrival time probability assumptions can greatly improve the model performance.
arXiv Detail & Related papers (2021-07-07T16:59:14Z) - Meta-Learning Stationary Stochastic Process Prediction with
Convolutional Neural Processes [32.02612871707347]
We propose ConvNP, which endows Neural Processes (NPs) with translation equivariance and extends convolutional conditional NPs to allow for dependencies in the predictive distribution.
We demonstrate the strong performance and generalization capabilities of ConvNPs on 1D, regression image completion, and various tasks with real-world-temporal data.
arXiv Detail & Related papers (2020-07-02T18:25:27Z) - Supervised Learning for Non-Sequential Data: A Canonical Polyadic
Decomposition Approach [85.12934750565971]
Efficient modelling of feature interactions underpins supervised learning for non-sequential tasks.
To alleviate this issue, it has been proposed to implicitly represent the model parameters as a tensor.
For enhanced expressiveness, we generalize the framework to allow feature mapping to arbitrarily high-dimensional feature vectors.
arXiv Detail & Related papers (2020-01-27T22:38:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.