Measuring Statistical Dependencies via Maximum Norm and Characteristic
Functions
- URL: http://arxiv.org/abs/2208.07934v1
- Date: Tue, 16 Aug 2022 20:24:31 GMT
- Title: Measuring Statistical Dependencies via Maximum Norm and Characteristic
Functions
- Authors: Povilas Daniu\v{s}is, Shubham Juneja, Lukas Kuzma, Virginijus
Marcinkevi\v{c}ius
- Abstract summary: We propose a statistical dependence measure based on the maximum-norm of the difference between joint and product-marginal characteristic functions.
The proposed measure can detect arbitrary statistical dependence between two random vectors of possibly different dimensions.
We conduct experiments both with simulated and real data.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we focus on the problem of statistical dependence estimation
using characteristic functions. We propose a statistical dependence measure,
based on the maximum-norm of the difference between joint and product-marginal
characteristic functions. The proposed measure can detect arbitrary statistical
dependence between two random vectors of possibly different dimensions, is
differentiable, and easily integrable into modern machine learning and deep
learning pipelines. We also conduct experiments both with simulated and real
data. Our simulations show, that the proposed method can measure statistical
dependencies in high-dimensional, non-linear data, and is less affected by the
curse of dimensionality, compared to the previous work in this line of
research. The experiments with real data demonstrate the potential
applicability of our statistical measure for two different empirical inference
scenarios, showing statistically significant improvement in the performance
characteristics when applied for supervised feature extraction and deep neural
network regularization. In addition, we provide a link to the accompanying
open-source repository https://bit.ly/3d4ch5I.
Related papers
- Meta-Statistical Learning: Supervised Learning of Statistical Inference [59.463430294611626]
This work demonstrates that the tools and principles driving the success of large language models (LLMs) can be repurposed to tackle distribution-level tasks.
We propose meta-statistical learning, a framework inspired by multi-instance learning that reformulates statistical inference tasks as supervised learning problems.
arXiv Detail & Related papers (2025-02-17T18:04:39Z) - Model-free Methods for Event History Analysis and Efficient Adjustment (PhD Thesis) [55.2480439325792]
This thesis is a series of independent contributions to statistics unified by a model-free perspective.
The first chapter elaborates on how a model-free perspective can be used to formulate flexible methods that leverage prediction techniques from machine learning.
The second chapter studies the concept of local independence, which describes whether the evolution of one process is directly influenced by another.
arXiv Detail & Related papers (2025-02-11T19:24:09Z) - Statistical Inference for Temporal Difference Learning with Linear Function Approximation [62.69448336714418]
We study the consistency properties of TD learning with Polyak-Ruppert averaging and linear function approximation.
First, we derive a novel high-dimensional probability convergence guarantee that depends explicitly on the variance and holds under weak conditions.
We further establish refined high-dimensional Berry-Esseen bounds over the class of convex sets that guarantee faster rates than those in the literature.
arXiv Detail & Related papers (2024-10-21T15:34:44Z) - Training Normalizing Flows from Dependent Data [31.42053454078623]
We propose a likelihood objective of normalizing flows incorporating dependencies between the data points.
We show that respecting dependencies between observations can improve empirical results on both synthetic and real-world data.
arXiv Detail & Related papers (2022-09-29T16:50:34Z) - Data-Driven Influence Functions for Optimization-Based Causal Inference [105.5385525290466]
We study a constructive algorithm that approximates Gateaux derivatives for statistical functionals by finite differencing.
We study the case where probability distributions are not known a priori but need to be estimated from data.
arXiv Detail & Related papers (2022-08-29T16:16:22Z) - Statistical inference of travelers' route choice preferences with
system-level data [4.120057972557892]
We develop a methodology to estimate travelers' utility functions with multiple attributes using system-level data.
Experiments on synthetic data show that the coefficients are consistently recovered and that hypothesis tests are a reliable statistic to identify which attributes are determinants of travelers' route choices.
The methodology is also deployed at a large scale using real Fresnoworld multisource data collected during the COVID outbreak.
arXiv Detail & Related papers (2022-04-23T00:38:32Z) - Uncertainty Modeling for Out-of-Distribution Generalization [56.957731893992495]
We argue that the feature statistics can be properly manipulated to improve the generalization ability of deep learning models.
Common methods often consider the feature statistics as deterministic values measured from the learned features.
We improve the network generalization ability by modeling the uncertainty of domain shifts with synthesized feature statistics during training.
arXiv Detail & Related papers (2022-02-08T16:09:12Z) - Differential privacy and robust statistics in high dimensions [49.50869296871643]
High-dimensional Propose-Test-Release (HPTR) builds upon three crucial components: the exponential mechanism, robust statistics, and the Propose-Test-Release mechanism.
We show that HPTR nearly achieves the optimal sample complexity under several scenarios studied in the literature.
arXiv Detail & Related papers (2021-11-12T06:36:40Z) - Efficient Multidimensional Functional Data Analysis Using Marginal
Product Basis Systems [2.4554686192257424]
We propose a framework for learning continuous representations from a sample of multidimensional functional data.
We show that the resulting estimation problem can be solved efficiently by the tensor decomposition.
We conclude with a real data application in neuroimaging.
arXiv Detail & Related papers (2021-07-30T16:02:15Z) - Measuring Dependence with Matrix-based Entropy Functional [21.713076360132195]
Measuring the dependence of data plays a central role in statistics and machine learning.
We propose two measures, namely the matrix-based normalized total correlation ($T_alpha*$) and the matrix-based normalized dual total correlation ($D_alpha*$)
We show that our measures are differentiable and statistically more powerful than prevalent ones.
arXiv Detail & Related papers (2021-01-25T15:18:16Z) - Measuring the Discrepancy between Conditional Distributions: Methods,
Properties and Applications [18.293397644865458]
We propose a simple yet powerful test statistic to quantify the discrepancy between two conditional distributions.
The new statistic avoids the explicit estimation of the underlying distributions in highdimensional space.
It inherits the merits of the correntropy function to explicitly incorporate high-order statistics in the data.
arXiv Detail & Related papers (2020-05-05T14:03:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.