Autoencoding Conditional Neural Processes for Representation Learning
- URL: http://arxiv.org/abs/2305.18485v2
- Date: Sat, 17 Feb 2024 23:00:14 GMT
- Title: Autoencoding Conditional Neural Processes for Representation Learning
- Authors: Victor Prokhorov, Ivan Titov, N. Siddharth
- Abstract summary: Conditional neural processes (CNPs) are a flexible and efficient family of models that learn to learn from data.
We develop the Partial Pixel Space Variational Autoencoder (PPS-VAE), an amortised variational framework that casts CNP context as latent variables learnt simultaneously with the CNP.
- Score: 31.63717849083666
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Conditional neural processes (CNPs) are a flexible and efficient family of
models that learn to learn a stochastic process from data. They have seen
particular application in contextual image completion - observing pixel values
at some locations to predict a distribution over values at other unobserved
locations. However, the choice of pixels in learning CNPs is typically either
random or derived from a simple statistical measure (e.g. pixel variance).
Here, we turn the problem on its head and ask: which pixels would a CNP like to
observe - do they facilitate fitting better CNPs, and do such pixels tell us
something meaningful about the underlying image? To this end we develop the
Partial Pixel Space Variational Autoencoder (PPS-VAE), an amortised variational
framework that casts CNP context as latent variables learnt simultaneously with
the CNP. We evaluate PPS-VAE over a number of tasks across different visual
data, and find that not only can it facilitate better-fit CNPs, but also that
the spatial arrangement and values meaningfully characterise image information
- evaluated through the lens of classification on both within and out-of-data
distributions. Our model additionally allows for dynamic adaption of
context-set size and the ability to scale-up to larger images, providing a
promising avenue to explore learning meaningful and effective visual
representations.
Related papers
- Hierarchical Sparse Attention Framework for Computationally Efficient Classification of Biological Cells [0.0]
We present SparseAttnNet, a new hierarchical attention-driven framework for efficient image classification.<n>For biological cell images, we demonstrate that SparseAttnNet can process approximately 15% of the pixels instead of the full image.
arXiv Detail & Related papers (2025-05-12T15:29:08Z) - Learning Invariant Inter-pixel Correlations for Superpixel Generation [12.605604620139497]
Learnable features exhibit constrained discriminative capability, resulting in unsatisfactory pixel grouping performance.
We propose the Content Disentangle Superpixel algorithm to selectively separate the invariant inter-pixel correlations and statistical properties.
The experimental results on four benchmark datasets demonstrate the superiority of our approach to existing state-of-the-art methods.
arXiv Detail & Related papers (2024-02-28T09:46:56Z) - Differentiable Registration of Images and LiDAR Point Clouds with
VoxelPoint-to-Pixel Matching [58.10418136917358]
Cross-modality registration between 2D images from cameras and 3D point clouds from LiDARs is a crucial task in computer vision and robotic training.
Previous methods estimate 2D-3D correspondences by matching point and pixel patterns learned by neural networks.
We learn a structured cross-modality matching solver to represent 3D features via a different latent pixel space.
arXiv Detail & Related papers (2023-12-07T05:46:10Z) - NP-SemiSeg: When Neural Processes meet Semi-Supervised Semantic
Segmentation [87.50830107535533]
Semi-supervised semantic segmentation involves assigning pixel-wise labels to unlabeled images at training time.
Current approaches to semi-supervised semantic segmentation work by predicting pseudo-labels for each pixel from a class-wise probability distribution output by a model.
In this work, we move one step forward by adapting NPs to semi-supervised semantic segmentation, resulting in a new model called NP-SemiSeg.
arXiv Detail & Related papers (2023-08-05T12:42:15Z) - Versatile Neural Processes for Learning Implicit Neural Representations [57.090658265140384]
We propose Versatile Neural Processes (VNP), which largely increases the capability of approximating functions.
Specifically, we introduce a bottleneck encoder that produces fewer and informative context tokens, relieving the high computational cost.
We demonstrate the effectiveness of the proposed VNP on a variety of tasks involving 1D, 2D and 3D signals.
arXiv Detail & Related papers (2023-01-21T04:08:46Z) - Probabilistic Deep Metric Learning for Hyperspectral Image
Classification [91.5747859691553]
This paper proposes a probabilistic deep metric learning framework for hyperspectral image classification.
It aims to predict the category of each pixel for an image captured by hyperspectral sensors.
Our framework can be readily applied to existing hyperspectral image classification methods.
arXiv Detail & Related papers (2022-11-15T17:57:12Z) - Conditional Neural Processes for Molecules [0.0]
Neural processes (NPs) are models for transfer learning with properties reminiscent of Gaussian Processes (GPs)
This paper applies the conditional neural process (CNP) to DOCKSTRING, a dataset of docking scores for benchmarking ML models.
CNPs show competitive performance in few-shot learning tasks relative to supervised learning baselines common in QSAR modelling, as well as an alternative model for transfer learning based on pre-training and refining neural network regressors.
arXiv Detail & Related papers (2022-10-17T16:10:12Z) - Visual Recognition with Deep Nearest Centroids [57.35144702563746]
We devise deep nearest centroids (DNC), a conceptually elegant yet surprisingly effective network for large-scale visual recognition.
Compared with parametric counterparts, DNC performs better on image classification (CIFAR-10, ImageNet) and greatly boots pixel recognition (ADE20K, Cityscapes)
arXiv Detail & Related papers (2022-09-15T15:47:31Z) - Class Balanced PixelNet for Neurological Image Segmentation [20.56747443955369]
We propose an automatic brain tumor segmentation approach (e.g., PixelNet) using a pixel-level convolutional neural network (CNN)
The proposed model has achieved promising results in brain tumor and ischemic stroke segmentation datasets.
arXiv Detail & Related papers (2022-04-23T10:57:54Z) - Neural Knitworks: Patched Neural Implicit Representation Networks [1.0470286407954037]
We propose Knitwork, an architecture for neural implicit representation learning of natural images that achieves image synthesis.
To the best of our knowledge, this is the first implementation of a coordinate-based patch tailored for synthesis tasks such as image inpainting, super-resolution, and denoising.
The results show that modeling natural images using patches, rather than pixels, produces results of higher fidelity.
arXiv Detail & Related papers (2021-09-29T13:10:46Z) - Learning to Learn Parameterized Classification Networks for Scalable
Input Images [76.44375136492827]
Convolutional Neural Networks (CNNs) do not have a predictable recognition behavior with respect to the input resolution change.
We employ meta learners to generate convolutional weights of main networks for various input scales.
We further utilize knowledge distillation on the fly over model predictions based on different input resolutions.
arXiv Detail & Related papers (2020-07-13T04:27:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.