A new perspective on probabilistic image modeling
- URL: http://arxiv.org/abs/2203.11034v1
- Date: Mon, 21 Mar 2022 14:53:57 GMT
- Title: A new perspective on probabilistic image modeling
- Authors: Alexander Gepperth
- Abstract summary: We present a new probabilistic approach for image modeling capable of density estimation, sampling and tractable inference.
DCGMMs can be trained end-to-end by SGD from random initial conditions, much like CNNs.
We show that DCGMMs compare favorably to several recent PC and SPN models in terms of inference, classification and sampling.
- Score: 92.89846887298852
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present the Deep Convolutional Gaussian Mixture Model (DCGMM), a new
probabilistic approach for image modeling capable of density estimation,
sampling and tractable inference. DCGMM instances exhibit a CNN-like layered
structure, in which the principal building blocks are convolutional Gaussian
Mixture (cGMM) layers. A key innovation w.r.t. related models like sum-product
networks (SPNs) and probabilistic circuits (PCs) is that each cGMM layer
optimizes an independent loss function and therefore has an independent
probabilistic interpretation. This modular approach permits intervening
transformation layers to harness the full spectrum of (potentially
non-invertible) mappings available to CNNs, e.g., max-pooling or
half-convolutions. DCGMM sampling and inference are realized by a deep chain of
hierarchical priors, where a sample generated by a given cGMM layer defines the
parameters of sampling in the next-lower cGMM layer. For sampling through
non-invertible transformation layers, we introduce a new gradient-based
sharpening technique that exploits redundancy (overlap) in, e.g.,
half-convolutions. DCGMMs can be trained end-to-end by SGD from random initial
conditions, much like CNNs. We show that DCGMMs compare favorably to several
recent PC and SPN models in terms of inference, classification and sampling,
the latter particularly for challenging datasets such as SVHN. We provide a
public TF2 implementation.
Related papers
- Steering Masked Discrete Diffusion Models via Discrete Denoising Posterior Prediction [88.65168366064061]
We introduce Discrete Denoising Posterior Prediction (DDPP), a novel framework that casts the task of steering pre-trained MDMs as a problem of probabilistic inference.
Our framework leads to a family of three novel objectives that are all simulation-free, and thus scalable.
We substantiate our designs via wet-lab validation, where we observe transient expression of reward-optimized protein sequences.
arXiv Detail & Related papers (2024-10-10T17:18:30Z) - Deep Gaussian mixture model for unsupervised image segmentation [1.3654846342364308]
In many tasks sufficient pixel-level labels are very difficult to obtain.
We propose a method which combines a Gaussian mixture model (GMM) with unsupervised deep learning techniques.
We demonstrate the advantages of our method in various experiments on the example of infarct segmentation on multi-sequence MRI images.
arXiv Detail & Related papers (2024-04-18T15:20:59Z) - Consistency Trajectory Models: Learning Probability Flow ODE Trajectory of Diffusion [56.38386580040991]
Consistency Trajectory Model (CTM) is a generalization of Consistency Models (CM)
CTM enables the efficient combination of adversarial training and denoising score matching loss to enhance performance.
Unlike CM, CTM's access to the score function can streamline the adoption of established controllable/conditional generation methods.
arXiv Detail & Related papers (2023-10-01T05:07:17Z) - A differentiable Gaussian Prototype Layer for explainable Segmentation [3.258592531141818]
We introduce a gradient-based prototype layer for gradient-based prototype learning.
We show it can be used as a novel building block for explainable neural networks.
arXiv Detail & Related papers (2023-06-25T22:33:21Z) - Image Modeling with Deep Convolutional Gaussian Mixture Models [79.0660895390689]
We present a new formulation of deep hierarchical Gaussian Mixture Models (GMMs) that is suitable for describing and generating images.
DCGMMs avoid this by a stacked architecture of multiple GMM layers, linked by convolution and pooling operations.
For generating sharp images with DCGMMs, we introduce a new gradient-based technique for sampling through non-invertible operations like convolution and pooling.
Based on the MNIST and FashionMNIST datasets, we validate the DCGMMs model by demonstrating its superiority over flat GMMs for clustering, sampling and outlier detection.
arXiv Detail & Related papers (2021-04-19T12:08:53Z) - DS-UI: Dual-Supervised Mixture of Gaussian Mixture Models for
Uncertainty Inference [52.899219617256655]
We propose a dual-supervised uncertainty inference (DS-UI) framework for improving Bayesian estimation-based uncertainty inference (UI) in deep neural network (DNN)-based image recognition.
In the DS-UI, we combine the last fully-connected (FC) layer with a mixture of Gaussian mixture models (MoGMM) to obtain an MoGMM-FC layer.
Experimental results show the DS-UI outperforms the state-of-the-art UI methods in misclassification detection.
arXiv Detail & Related papers (2020-11-17T12:35:02Z) - RG-Flow: A hierarchical and explainable flow model based on
renormalization group and sparse prior [2.274915755738124]
Flow-based generative models have become an important class of unsupervised learning approaches.
In this work, we incorporate the key ideas of renormalization group (RG) and sparse prior distribution to design a hierarchical flow-based generative model, RG-Flow.
Our proposed method has $O(log L)$ complexity for inpainting of an image with edge length $L$, compared to previous generative models with $O(L2)$ complexity.
arXiv Detail & Related papers (2020-09-30T18:04:04Z) - Semi-Supervised Learning with Normalizing Flows [54.376602201489995]
FlowGMM is an end-to-end approach to generative semi supervised learning with normalizing flows.
We show promising results on a wide range of applications, including AG-News and Yahoo Answers text data.
arXiv Detail & Related papers (2019-12-30T17:36:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.