Neural Estimators for Conditional Mutual Information Using Nearest
Neighbors Sampling
- URL: http://arxiv.org/abs/2006.07225v3
- Date: Wed, 30 Dec 2020 10:15:17 GMT
- Title: Neural Estimators for Conditional Mutual Information Using Nearest
Neighbors Sampling
- Authors: Sina Molavipour, Germ\'an Bassi, Mikael Skoglund
- Abstract summary: estimation of mutual information (MI) or conditional mutual information (CMI) from a set of samples is a long-standing problem.
Recent work has leveraged the approximation power of artificial neural networks and has shown improvements over conventional methods.
We introduce a new technique, based on k nearest neighbors (k-NN), to perform the resampling and derive high-confidence concentration bounds for the sample average.
- Score: 36.35382677479192
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The estimation of mutual information (MI) or conditional mutual information
(CMI) from a set of samples is a long-standing problem. A recent line of work
in this area has leveraged the approximation power of artificial neural
networks and has shown improvements over conventional methods. One important
challenge in this new approach is the need to obtain, given the original
dataset, a different set where the samples are distributed according to a
specific product density function. This is particularly challenging when
estimating CMI.
In this paper, we introduce a new technique, based on k nearest neighbors
(k-NN), to perform the resampling and derive high-confidence concentration
bounds for the sample average. Then the technique is employed to train a neural
network classifier and the CMI is estimated accordingly. We propose three
estimators using this technique and prove their consistency, make a comparison
between them and similar approaches in the literature, and experimentally show
improvements in estimating the CMI in terms of accuracy and variance of the
estimators.
Related papers
- Nearest Neighbour Score Estimators for Diffusion Generative Models [16.189734871742743]
We introduce a novel nearest neighbour score function estimator which utilizes multiple samples from the training set to dramatically decrease estimator variance.
In diffusion models, we show that our estimator can replace a learned network for probability-flow ODE integration, opening promising new avenues of future research.
arXiv Detail & Related papers (2024-02-12T19:27:30Z) - Implicit Variational Inference for High-Dimensional Posteriors [7.924706533725115]
In variational inference, the benefits of Bayesian models rely on accurately capturing the true posterior distribution.
We propose using neural samplers that specify implicit distributions, which are well-suited for approximating complex multimodal and correlated posteriors.
Our approach introduces novel bounds for approximate inference using implicit distributions by locally linearising the neural sampler.
arXiv Detail & Related papers (2023-10-10T14:06:56Z) - Compound Batch Normalization for Long-tailed Image Classification [77.42829178064807]
We propose a compound batch normalization method based on a Gaussian mixture.
It can model the feature space more comprehensively and reduce the dominance of head classes.
The proposed method outperforms existing methods on long-tailed image classification.
arXiv Detail & Related papers (2022-12-02T07:31:39Z) - ScoreMix: A Scalable Augmentation Strategy for Training GANs with
Limited Data [93.06336507035486]
Generative Adversarial Networks (GANs) typically suffer from overfitting when limited training data is available.
We present ScoreMix, a novel and scalable data augmentation approach for various image synthesis tasks.
arXiv Detail & Related papers (2022-10-27T02:55:15Z) - Fake It Till You Make It: Near-Distribution Novelty Detection by
Score-Based Generative Models [54.182955830194445]
existing models either fail or face a dramatic drop under the so-called near-distribution" setting.
We propose to exploit a score-based generative model to produce synthetic near-distribution anomalous data.
Our method improves the near-distribution novelty detection by 6% and passes the state-of-the-art by 1% to 5% across nine novelty detection benchmarks.
arXiv Detail & Related papers (2022-05-28T02:02:53Z) - Mixtures of Laplace Approximations for Improved Post-Hoc Uncertainty in
Deep Learning [24.3370326359959]
We propose to predict with a Gaussian mixture model posterior that consists of a weighted sum of Laplace approximations of independently trained deep neural networks.
We theoretically validate that our approach mitigates overconfidence "far away" from the training data and empirically compare against state-of-the-art baselines on standard uncertainty quantification benchmarks.
arXiv Detail & Related papers (2021-11-05T15:52:48Z) - A Variational Bayesian Approach to Learning Latent Variables for
Acoustic Knowledge Transfer [55.20627066525205]
We propose a variational Bayesian (VB) approach to learning distributions of latent variables in deep neural network (DNN) models.
Our proposed VB approach can obtain good improvements on target devices, and consistently outperforms 13 state-of-the-art knowledge transfer algorithms.
arXiv Detail & Related papers (2021-10-16T15:54:01Z) - Neural Methods for Point-wise Dependency Estimation [129.93860669802046]
We focus on estimating point-wise dependency (PD), which quantitatively measures how likely two outcomes co-occur.
We demonstrate the effectiveness of our approaches in 1) MI estimation, 2) self-supervised representation learning, and 3) cross-modal retrieval task.
arXiv Detail & Related papers (2020-06-09T23:26:15Z) - C-MI-GAN : Estimation of Conditional Mutual Information using MinMax
formulation [20.57104064155529]
We study conditional mutual information (CMI) estimation by utilizing its formulation as a minmax optimization problem.
We find that our proposed estimator provides better estimates than the existing approaches on a variety of simulated data sets.
As an application of CMI estimation, we deploy our estimator for conditional independence (CI) testing on real data and obtain better results than state-of-the-art CI testers.
arXiv Detail & Related papers (2020-05-17T11:22:12Z) - DCMD: Distance-based Classification Using Mixture Distributions on
Microbiome Data [10.171660468645603]
We present an innovative approach for distance-based classification using mixture distributions (DCMD)
This approach models the inherent uncertainty in sparse counts by estimating a mixture distribution for the sample data.
Results are compared against a number of existing machine learning and distance-based approaches.
arXiv Detail & Related papers (2020-03-29T23:30:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.