The Normalized Cross Density Functional: A Framework to Quantify
Statistical Dependence for Random Processes
- URL: http://arxiv.org/abs/2212.04631v3
- Date: Tue, 20 Feb 2024 23:47:44 GMT
- Title: The Normalized Cross Density Functional: A Framework to Quantify
Statistical Dependence for Random Processes
- Authors: Bo Hu and Jose C. Principe
- Abstract summary: We present a novel approach to measuring statistical dependence between two random processes (r.p.) using a positive-definite function called the Normalized Cross Density (NCD)
NCD is derived directly from the probability density functions of two r.p. and constructs a data-dependent Hilbert space, the Normalized Cross-Density Hilbert Space (NCD-HS)
We mathematically prove that FMCA learns the dominant eigenvalues and eigenfunctions of NCD directly from realizations.
- Score: 6.625320950808605
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a novel approach to measuring statistical dependence
between two random processes (r.p.) using a positive-definite function called
the Normalized Cross Density (NCD). NCD is derived directly from the
probability density functions of two r.p. and constructs a data-dependent
Hilbert space, the Normalized Cross-Density Hilbert Space (NCD-HS). By Mercer's
Theorem, the NCD norm can be decomposed into its eigenspectrum, which we name
the Multivariate Statistical Dependence (MSD) measure, and their sum, the Total
Dependence Measure (TSD). Hence, the NCD-HS eigenfunctions serve as a novel
embedded feature space, suitable for quantifying r.p. statistical dependence.
In order to apply NCD directly to r.p. realizations, we introduce an
architecture with two multiple-output neural networks, a cost function, and an
algorithm named the Functional Maximal Correlation Algorithm (FMCA). With FMCA,
the two networks learn concurrently by approximating each other's outputs,
extending the Alternating Conditional Expectation (ACE) for multivariate
functions. We mathematically prove that FMCA learns the dominant eigenvalues
and eigenfunctions of NCD directly from realizations. Preliminary results with
synthetic data and medium-sized image datasets corroborate the theory.
Different strategies for applying NCD are proposed and discussed, demonstrating
the method's versatility and stability beyond supervised learning.
Specifically, when the two r.p. are high-dimensional real-world images and a
white uniform noise process, FMCA learns factorial codes, i.e., the occurrence
of a code guarantees that a specific training set image was present, which is
important for feature learning.
Related papers
- Statistical Inference for Temporal Difference Learning with Linear Function Approximation [62.69448336714418]
Temporal Difference (TD) learning, arguably the most widely used for policy evaluation, serves as a natural framework for this purpose.
In this paper, we study the consistency properties of TD learning with Polyak-Ruppert averaging and linear function approximation, and obtain three significant improvements over existing results.
arXiv Detail & Related papers (2024-10-21T15:34:44Z) - MEP: Multiple Kernel Learning Enhancing Relative Positional Encoding Length Extrapolation [5.298814565953444]
Relative position encoding methods address the length extrapolation challenge exclusively through the implementation of a single kernel function.
This study proposes a novel relative positional encoding method, called MEP, which employs a weighted average to combine distinct kernel functions.
We present two distinct versions of our method: a parameter-free variant that requires no new learnable parameters, and a parameterized variant capable of integrating state-of-the-art techniques.
arXiv Detail & Related papers (2024-03-26T13:38:06Z) - Latent Semantic Consensus For Deterministic Geometric Model Fitting [109.44565542031384]
We propose an effective method called Latent Semantic Consensus (LSC)
LSC formulates the model fitting problem into two latent semantic spaces based on data points and model hypotheses.
LSC is able to provide consistent and reliable solutions within only a few milliseconds for general multi-structural model fitting.
arXiv Detail & Related papers (2024-03-11T05:35:38Z) - Max-Sliced Mutual Information [17.667315953598788]
Quantifying the dependence between high-dimensional random variables is central to statistical learning and inference.
Two classical methods are canonical correlation analysis (CCA), which identifies maximally correlated projected versions of the original variables, and Shannon's mutual information, which is a universal dependence measure.
This work proposes a middle ground in the form of a scalable information-theoretic generalization of CCA, termed max-sliced mutual information (mSMI)
arXiv Detail & Related papers (2023-09-28T06:49:25Z) - The Normal Distributions Indistinguishability Spectrum and its Application to Privacy-Preserving Machine Learning [7.61316504036937]
Differential Privacy is the most common method for machine learning (ML) on privacy-sensitive data.
In big data analytics, one often uses randomized sketching/aggregation algorithms to make processing high-dimensional data tractable.
arXiv Detail & Related papers (2023-09-03T19:07:31Z) - Validation Diagnostics for SBI algorithms based on Normalizing Flows [55.41644538483948]
This work proposes easy to interpret validation diagnostics for multi-dimensional conditional (posterior) density estimators based on NF.
It also offers theoretical guarantees based on results of local consistency.
This work should help the design of better specified models or drive the development of novel SBI-algorithms.
arXiv Detail & Related papers (2022-11-17T15:48:06Z) - Post-Regularization Confidence Bands for Ordinary Differential Equations [6.3582148777824115]
We construct confidence band for individual regulatory function in ODE with unknown functionals and noisy data observations.
We show that the constructed confidence band has the desired kernel coverage probability, and the recovered regulatory network approaches the truth with probability tending to one.
arXiv Detail & Related papers (2021-10-24T19:21:10Z) - A Note on Optimizing Distributions using Kernel Mean Embeddings [94.96262888797257]
Kernel mean embeddings represent probability measures by their infinite-dimensional mean embeddings in a reproducing kernel Hilbert space.
We show that when the kernel is characteristic, distributions with a kernel sum-of-squares density are dense.
We provide algorithms to optimize such distributions in the finite-sample setting.
arXiv Detail & Related papers (2021-06-18T08:33:45Z) - A Functional Perspective on Learning Symmetric Functions with Neural
Networks [48.80300074254758]
We study the learning and representation of neural networks defined on measures.
We establish approximation and generalization bounds under different choices of regularization.
The resulting models can be learned efficiently and enjoy generalization guarantees that extend across input sizes.
arXiv Detail & Related papers (2020-08-16T16:34:33Z) - Reciprocal Adversarial Learning via Characteristic Functions [12.961770002117142]
Generative adversarial nets (GANs) have become a preferred tool for tasks involving complicated distributions.
We show how to use the characteristic function (CF) to compare the distributions rather than their moments.
We then prove an equivalence between the embedded and data domains when a reciprocal exists, where we naturally develop the GAN in an auto-encoder structure.
This efficient structure uses only two modules, together with a simple training strategy, to achieve bi-directionally generating clear images.
arXiv Detail & Related papers (2020-06-15T14:04:55Z) - Learning Likelihoods with Conditional Normalizing Flows [54.60456010771409]
Conditional normalizing flows (CNFs) are efficient in sampling and inference.
We present a study of CNFs where the base density to output space mapping is conditioned on an input x, to model conditional densities p(y|x)
arXiv Detail & Related papers (2019-11-29T19:17:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.