Autoregressive Conditional Neural Processes
- URL: http://arxiv.org/abs/2303.14468v1
- Date: Sat, 25 Mar 2023 13:34:12 GMT
- Title: Autoregressive Conditional Neural Processes
- Authors: Wessel P. Bruinsma, Stratis Markou, James Requiema, Andrew Y. K.
Foong, Tom R. Andersson, Anna Vaughan, Anthony Buonomo, J. Scott Hosking,
Richard E. Turner
- Abstract summary: Conditional neural processes (CNPs) are attractive meta-learning models.
They produce well-calibrated predictions and are trainable via a simple maximum likelihood procedure.
CNPs are unable to model dependencies in their predictions.
We propose to change how CNPs are deployed at test time, without any modifications to the model or training procedure.
- Score: 20.587835119831595
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Conditional neural processes (CNPs; Garnelo et al., 2018a) are attractive
meta-learning models which produce well-calibrated predictions and are
trainable via a simple maximum likelihood procedure. Although CNPs have many
advantages, they are unable to model dependencies in their predictions. Various
works propose solutions to this, but these come at the cost of either requiring
approximate inference or being limited to Gaussian predictions. In this work,
we instead propose to change how CNPs are deployed at test time, without any
modifications to the model or training procedure. Instead of making predictions
independently for every target point, we autoregressively define a joint
predictive distribution using the chain rule of probability, taking inspiration
from the neural autoregressive density estimator (NADE) literature. We show
that this simple procedure allows factorised Gaussian CNPs to model highly
dependent, non-Gaussian predictive distributions. Perhaps surprisingly, in an
extensive range of tasks with synthetic and real data, we show that CNPs in
autoregressive (AR) mode not only significantly outperform non-AR CNPs, but are
also competitive with more sophisticated models that are significantly more
computationally expensive and challenging to train. This performance is
remarkable given that AR CNPs are not trained to model joint dependencies. Our
work provides an example of how ideas from neural distribution estimation can
benefit neural processes, and motivates research into the AR deployment of
other neural process models.
Related papers
- Computation-Aware Gaussian Processes: Model Selection And Linear-Time Inference [55.150117654242706]
We show that model selection for computation-aware GPs trained on 1.8 million data points can be done within a few hours on a single GPU.
As a result of this work, Gaussian processes can be trained on large-scale datasets without significantly compromising their ability to quantify uncertainty.
arXiv Detail & Related papers (2024-11-01T21:11:48Z) - Convolutional Conditional Neural Processes [6.532867867011488]
This thesis advances neural processes in three ways.
ConvNPs improve data efficiency by building in a symmetry called translationvariance.
GNPs directly parametrise dependencies in the predictions of a neural process.
AR CNPs train a neural process without any modifications to the model or training procedure and, at test time, roll out the model in an autoregressive fashion.
arXiv Detail & Related papers (2024-08-18T19:53:38Z) - Human Trajectory Forecasting with Explainable Behavioral Uncertainty [63.62824628085961]
Human trajectory forecasting helps to understand and predict human behaviors, enabling applications from social robots to self-driving cars.
Model-free methods offer superior prediction accuracy but lack explainability, while model-based methods provide explainability but cannot predict well.
We show that BNSP-SFM achieves up to a 50% improvement in prediction accuracy, compared with 11 state-of-the-art methods.
arXiv Detail & Related papers (2023-07-04T16:45:21Z) - Conditional Neural Processes for Molecules [0.0]
Neural processes (NPs) are models for transfer learning with properties reminiscent of Gaussian Processes (GPs)
This paper applies the conditional neural process (CNP) to DOCKSTRING, a dataset of docking scores for benchmarking ML models.
CNPs show competitive performance in few-shot learning tasks relative to supervised learning baselines common in QSAR modelling, as well as an alternative model for transfer learning based on pre-training and refining neural network regressors.
arXiv Detail & Related papers (2022-10-17T16:10:12Z) - Correcting Model Bias with Sparse Implicit Processes [0.9187159782788579]
We show that Sparse Implicit Processes (SIP) is capable of correcting model bias when the data generating mechanism differs strongly from the one implied by the model.
We use synthetic datasets to show that SIP is capable of providing predictive distributions that reflect the data better than the exact predictions of the initial, but wrongly assumed model.
arXiv Detail & Related papers (2022-07-21T18:00:01Z) - Practical Conditional Neural Processes Via Tractable Dependent
Predictions [25.15531845287349]
Conditional Neural Processes (CNPs) are meta-learning models which leverage the flexibility of deep learning to produce well-calibrated predictions.
CNPs do not produce correlated predictions, making them inappropriate for many estimation and decision making tasks.
We present a new class of Neural Process models that make correlated predictions and support exact maximum likelihood training.
arXiv Detail & Related papers (2022-03-16T17:37:41Z) - Efficient Gaussian Neural Processes for Regression [7.149677544861951]
Conditional Neural Processes (CNPs) produce well-calibrated predictions, enable fast inference at test time, and are trainable via a simple maximum likelihood procedure.
A limitation of CNPs is their inability to model dependencies in the outputs.
We present an alternative way to model output dependencies which also lends itself maximum likelihood training.
arXiv Detail & Related papers (2021-08-22T09:31:50Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - When in Doubt: Neural Non-Parametric Uncertainty Quantification for
Epidemic Forecasting [70.54920804222031]
Most existing forecasting models disregard uncertainty quantification, resulting in mis-calibrated predictions.
Recent works in deep neural models for uncertainty-aware time-series forecasting also have several limitations.
We model the forecasting task as a probabilistic generative process and propose a functional neural process model called EPIFNP.
arXiv Detail & Related papers (2021-06-07T18:31:47Z) - Bootstrapping Neural Processes [114.97111530885093]
Neural Processes (NPs) implicitly define a broad class of processes with neural networks.
NPs still rely on an assumption that uncertainty in processes is modeled by a single latent variable.
We propose the Boostrapping Neural Process (BNP), a novel extension of the NP family using the bootstrap.
arXiv Detail & Related papers (2020-08-07T02:23:34Z) - Efficient Model-Based Reinforcement Learning through Optimistic Policy
Search and Planning [93.1435980666675]
We show how optimistic exploration can be easily combined with state-of-the-art reinforcement learning algorithms.
Our experiments demonstrate that optimistic exploration significantly speeds-up learning when there are penalties on actions.
arXiv Detail & Related papers (2020-06-15T18:37:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.