Profile Entropy: A Fundamental Measure for the Learnability and
Compressibility of Discrete Distributions
- URL: http://arxiv.org/abs/2002.11665v1
- Date: Wed, 26 Feb 2020 17:49:04 GMT
- Title: Profile Entropy: A Fundamental Measure for the Learnability and
Compressibility of Discrete Distributions
- Authors: Yi Hao, Alon Orlitsky
- Abstract summary: We show that for samples of discrete distributions, profile entropy is a fundamental measure unifying the concepts of estimation, inference, and compression.
Specifically, profile entropy a) determines the speed of estimating the distribution relative to the best natural estimator; b) characterizes the rate of inferring all symmetric properties compared with the best estimator over any label-invariant distribution collection; c) serves as the limit of profile compression.
- Score: 63.60499266361255
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The profile of a sample is the multiset of its symbol frequencies. We show
that for samples of discrete distributions, profile entropy is a fundamental
measure unifying the concepts of estimation, inference, and compression.
Specifically, profile entropy a) determines the speed of estimating the
distribution relative to the best natural estimator; b) characterizes the rate
of inferring all symmetric properties compared with the best estimator over any
label-invariant distribution collection; c) serves as the limit of profile
compression, for which we derive optimal near-linear-time block and sequential
algorithms. To further our understanding of profile entropy, we investigate its
attributes, provide algorithms for approximating its value, and determine its
magnitude for numerous structural distribution families.
Related papers
- Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis [56.442307356162864]
We study the theoretical aspects of score-based discrete diffusion models under the Continuous Time Markov Chain (CTMC) framework.
We introduce a discrete-time sampling algorithm in the general state space $[S]d$ that utilizes score estimators at predefined time points.
Our convergence analysis employs a Girsanov-based method and establishes key properties of the discrete score function.
arXiv Detail & Related papers (2024-10-03T09:07:13Z) - Adaptive Online Bayesian Estimation of Frequency Distributions with Local Differential Privacy [0.4604003661048266]
We propose a novel approach for the adaptive and online estimation of the frequency distribution of a finite number of categories under the local differential privacy (LDP) framework.
The proposed algorithm performs Bayesian parameter estimation via posterior sampling and adapts the randomization mechanism for LDP based on the obtained posterior samples.
We provide a theoretical analysis showing that (i) the posterior distribution targeted by the algorithm converges to the true parameter even for approximate posterior sampling, and (ii) the algorithm selects the optimal subset with high probability if posterior sampling is performed exactly.
arXiv Detail & Related papers (2024-05-11T13:59:52Z) - MESSY Estimation: Maximum-Entropy based Stochastic and Symbolic densitY
Estimation [4.014524824655106]
MESSY estimation is a Maximum-Entropy based Gradient and Symbolic densitY estimation method.
We construct a gradient-based drift-diffusion process that connects samples of the unknown distribution function to a guess symbolic expression.
We find that the addition of a symbolic search for basis functions improves the accuracy of the estimation at a reasonable additional computational cost.
arXiv Detail & Related papers (2023-06-07T03:28:47Z) - Mean-Square Analysis of Discretized It\^o Diffusions for Heavy-tailed
Sampling [17.415391025051434]
We analyze the complexity of sampling from a class of heavy-tailed distributions by discretizing a natural class of Ito diffusions associated with weighted Poincar'e inequalities.
Based on a mean-square analysis, we establish the iteration complexity for obtaining a sample whose distribution is $epsilon$ close to the target distribution in the Wasserstein-2 metric.
arXiv Detail & Related papers (2023-03-01T15:16:03Z) - Statistical Efficiency of Score Matching: The View from Isoperimetry [96.65637602827942]
We show a tight connection between statistical efficiency of score matching and the isoperimetric properties of the distribution being estimated.
We formalize these results both in the sample regime and in the finite regime.
arXiv Detail & Related papers (2022-10-03T06:09:01Z) - Learning Structured Gaussians to Approximate Deep Ensembles [10.055143995729415]
This paper proposes using a sparse-structured multivariate Gaussian to provide a closed-form approxorimator for dense image prediction tasks.
We capture the uncertainty and structured correlations in the predictions explicitly in a formal distribution, rather than implicitly through sampling alone.
We demonstrate the merits of our approach on monocular depth estimation and show that the advantages of our approach are obtained with comparable quantitative performance.
arXiv Detail & Related papers (2022-03-29T12:34:43Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z) - Instance-Optimal Compressed Sensing via Posterior Sampling [101.43899352984774]
We show for Gaussian measurements and emphany prior distribution on the signal, that the posterior sampling estimator achieves near-optimal recovery guarantees.
We implement the posterior sampling estimator for deep generative priors using Langevin dynamics, and empirically find that it produces accurate estimates with more diversity than MAP.
arXiv Detail & Related papers (2021-06-21T22:51:56Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z) - On the Estimation of Information Measures of Continuous Distributions [25.395010130602287]
estimation of information measures of continuous distributions based on samples is a fundamental problem in statistics and machine learning.
We provide confidence bounds for simple histogram based estimation of differential entropy from a fixed number of samples.
Our focus is on differential entropy, but we provide examples that show that similar results hold for mutual information and relative entropy as well.
arXiv Detail & Related papers (2020-02-07T15:36:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.