A Unifying Framework for Some Directed Distances in Statistics
- URL: http://arxiv.org/abs/2203.00863v1
- Date: Wed, 2 Mar 2022 04:24:13 GMT
- Title: A Unifying Framework for Some Directed Distances in Statistics
- Authors: Michel Broniatowski and Wolfgang Stummer
- Abstract summary: Density-based directed distances -- particularly known as divergences -- are widely used in statistics.
We provide a general framework which covers in particular both the density-based and distribution-function-based divergence approaches.
We deduce new concepts of dependence between random variables, as alternatives to the celebrated mutual information.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Density-based directed distances -- particularly known as divergences --
between probability distributions are widely used in statistics as well as in
the adjacent research fields of information theory, artificial intelligence and
machine learning. Prominent examples are the Kullback-Leibler information
distance (relative entropy) which e.g. is closely connected to the omnipresent
maximum likelihood estimation method, and Pearson's chisquare-distance which
e.g. is used for the celebrated chisquare goodness-of-fit test. Another line of
statistical inference is built upon distribution-function-based divergences
such as e.g. the prominent (weighted versions of) Cramer-von Mises test
statistics respectively Anderson-Darling test statistics which are frequently
applied for goodness-of-fit investigations; some more recent methods deal with
(other kinds of) cumulative paired divergences and closely related concepts. In
this paper, we provide a general framework which covers in particular both the
above-mentioned density-based and distribution-function-based divergence
approaches; the dissimilarity of quantiles respectively of other statistical
functionals will be included as well. From this framework, we structurally
extract numerous classical and also state-of-the-art (including new)
procedures. Furthermore, we deduce new concepts of dependence between random
variables, as alternatives to the celebrated mutual information. Some
variational representations are discussed, too.
Related papers
- Synthetic Tabular Data Validation: A Divergence-Based Approach [8.062368743143388]
Divergences quantify discrepancies between data distributions.
Traditional approaches calculate divergences independently for each feature.
We propose a novel approach that uses divergence estimation to overcome the limitations of marginal comparisons.
arXiv Detail & Related papers (2024-05-13T15:07:52Z) - Estimation of mutual information via quantum kernel method [0.0]
Estimating mutual information (MI) plays a critical role to investigate the relationship among multiple random variables with a nonlinear correlation.
We propose a method for estimating mutual information using the quantum kernel.
arXiv Detail & Related papers (2023-10-19T00:53:16Z) - MINDE: Mutual Information Neural Diffusion Estimation [7.399561232927219]
We present a new method for the estimation of Mutual Information (MI) between random variables.
We use score-based diffusion models to estimate the Kullback Leibler divergence between two densities as a difference between their score functions.
As a by-product, our method also enables the estimation of the entropy of random variables.
arXiv Detail & Related papers (2023-10-13T11:47:41Z) - Conformal inference for regression on Riemannian Manifolds [49.7719149179179]
We investigate prediction sets for regression scenarios when the response variable, denoted by $Y$, resides in a manifold, and the covariable, denoted by X, lies in Euclidean space.
We prove the almost sure convergence of the empirical version of these regions on the manifold to their population counterparts.
arXiv Detail & Related papers (2023-10-12T10:56:25Z) - Beyond Normal: On the Evaluation of Mutual Information Estimators [52.85079110699378]
We show how to construct a diverse family of distributions with known ground-truth mutual information.
We provide guidelines for practitioners on how to select appropriate estimator adapted to the difficulty of problem considered.
arXiv Detail & Related papers (2023-06-19T17:26:34Z) - Nonparametric Conditional Local Independence Testing [69.31200003384122]
Conditional local independence is an independence relation among continuous time processes.
No nonparametric test of conditional local independence has been available.
We propose such a nonparametric test based on double machine learning.
arXiv Detail & Related papers (2022-03-25T10:31:02Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z) - Bayesian Sparse Factor Analysis with Kernelized Observations [67.60224656603823]
Multi-view problems can be faced with latent variable models.
High-dimensionality and non-linear issues are traditionally handled by kernel methods.
We propose merging both approaches into single model.
arXiv Detail & Related papers (2020-06-01T14:25:38Z) - copent: Estimating Copula Entropy and Transfer Entropy in R [2.3980064191633232]
Copula Entropy (CE) has been applied to solve several related statistical or machine learning problems.
This paper introduces copent, the R package which implements proposed methods for estimating copula entropy and transfer entropy.
arXiv Detail & Related papers (2020-05-27T10:01:12Z) - Marginal likelihood computation for model selection and hypothesis
testing: an extensive review [66.37504201165159]
This article provides a comprehensive study of the state-of-the-art of the topic.
We highlight limitations, benefits, connections and differences among the different techniques.
Problems and possible solutions with the use of improper priors are also described.
arXiv Detail & Related papers (2020-05-17T18:31:58Z) - Measuring the Discrepancy between Conditional Distributions: Methods,
Properties and Applications [18.293397644865458]
We propose a simple yet powerful test statistic to quantify the discrepancy between two conditional distributions.
The new statistic avoids the explicit estimation of the underlying distributions in highdimensional space.
It inherits the merits of the correntropy function to explicitly incorporate high-order statistics in the data.
arXiv Detail & Related papers (2020-05-05T14:03:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.