A Manifold Two-Sample Test Study: Integral Probability Metric with
Neural Networks
- URL: http://arxiv.org/abs/2205.02043v2
- Date: Wed, 19 Apr 2023 23:28:10 GMT
- Title: A Manifold Two-Sample Test Study: Integral Probability Metric with
Neural Networks
- Authors: Jie Wang, Minshuo Chen, Tuo Zhao, Wenjing Liao, Yao Xie
- Abstract summary: Two-sample tests are important areas aiming to determine whether two collections of observations follow the same distribution or not.
We propose two-sample tests based on integral probability metric (IPM) for high-dimensional samples supported on a low-dimensional manifold.
Our proposed tests are adaptive to low-dimensional geometric structure because their performance crucially depends on the intrinsic dimension instead of the data dimension.
- Score: 46.62713126719579
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Two-sample tests are important areas aiming to determine whether two
collections of observations follow the same distribution or not. We propose
two-sample tests based on integral probability metric (IPM) for
high-dimensional samples supported on a low-dimensional manifold. We
characterize the properties of proposed tests with respect to the number of
samples $n$ and the structure of the manifold with intrinsic dimension $d$.
When an atlas is given, we propose two-step test to identify the difference
between general distributions, which achieves the type-II risk in the order of
$n^{-1/\max\{d,2\}}$. When an atlas is not given, we propose H\"older IPM test
that applies for data distributions with $(s,\beta)$-H\"older densities, which
achieves the type-II risk in the order of $n^{-(s+\beta)/d}$. To mitigate the
heavy computation burden of evaluating the H\"older IPM, we approximate the
H\"older function class using neural networks. Based on the approximation
theory of neural networks, we show that the neural network IPM test has the
type-II risk in the order of $n^{-(s+\beta)/d}$, which is in the same order of
the type-II risk as the H\"older IPM test. Our proposed tests are adaptive to
low-dimensional geometric structure because their performance crucially depends
on the intrinsic dimension instead of the data dimension.
Related papers
- Learning with Norm Constrained, Over-parameterized, Two-layer Neural Networks [54.177130905659155]
Recent studies show that a reproducing kernel Hilbert space (RKHS) is not a suitable space to model functions by neural networks.
In this paper, we study a suitable function space for over- parameterized two-layer neural networks with bounded norms.
arXiv Detail & Related papers (2024-04-29T15:04:07Z) - Computational-Statistical Gaps in Gaussian Single-Index Models [77.1473134227844]
Single-Index Models are high-dimensional regression problems with planted structure.
We show that computationally efficient algorithms, both within the Statistical Query (SQ) and the Low-Degree Polynomial (LDP) framework, necessarily require $Omega(dkstar/2)$ samples.
arXiv Detail & Related papers (2024-03-08T18:50:19Z) - Collaborative non-parametric two-sample testing [55.98760097296213]
The goal is to identify nodes where the null hypothesis $p_v = q_v$ should be rejected.
We propose the non-parametric collaborative two-sample testing (CTST) framework that efficiently leverages the graph structure.
Our methodology integrates elements from f-divergence estimation, Kernel Methods, and Multitask Learning.
arXiv Detail & Related papers (2024-02-08T14:43:56Z) - Maximum Mean Discrepancy Meets Neural Networks: The
Radon-Kolmogorov-Smirnov Test [5.255750357176021]
We study the function of $mathcalF$ to be the unit ball in the RBV space of a given smoothness order $k geq 0$.
This test can be viewed as a generalization of the well-known and classical Kolmogorov-Smirnov (KS) test to multiple dimensions and higher orders of smoothness.
We leverage the power of modern deep learning toolkits to optimize the criterion that underlies the RKS test.
arXiv Detail & Related papers (2023-09-05T17:51:00Z) - Kernel-Based Tests for Likelihood-Free Hypothesis Testing [21.143798051525646]
Given $n$ observations from two balanced classes, consider the task of labeling an additional $m$ inputs that are known to all belong to emphone of the two classes.
Special cases of this problem are well-known; when $m=1$ it corresponds to binary classification; and when $mapprox n$ it is equivalent to two-sample testing.
In recent work it was discovered that there is a fundamental trade-off between $m$ and $n$: increasing the data sample $m$ reduces the amount $n$ of training/simulation data needed.
arXiv Detail & Related papers (2023-08-17T15:24:03Z) - High-Dimensional Smoothed Entropy Estimation via Dimensionality
Reduction [14.53979700025531]
We consider the estimation of the differential entropy $h(X+Z)$ via $n$ independently and identically distributed samples of $X$.
Under the absolute-error loss, the above problem has a parametric estimation rate of $fraccDsqrtn$.
We overcome this exponential sample complexity by projecting $X$ to a low-dimensional space via principal component analysis (PCA) before the entropy estimation.
arXiv Detail & Related papers (2023-05-08T13:51:48Z) - On the Identifiability and Estimation of Causal Location-Scale Noise
Models [122.65417012597754]
We study the class of location-scale or heteroscedastic noise models (LSNMs)
We show the causal direction is identifiable up to some pathological cases.
We propose two estimators for LSNMs: an estimator based on (non-linear) feature maps, and one based on neural networks.
arXiv Detail & Related papers (2022-10-13T17:18:59Z) - Fundamental tradeoffs between memorization and robustness in random
features and neural tangent regimes [15.76663241036412]
We prove for a large class of activation functions that, if the model memorizes even a fraction of the training, then its Sobolev-seminorm is lower-bounded.
Experiments reveal for the first time, (iv) a multiple-descent phenomenon in the robustness of the min-norm interpolator.
arXiv Detail & Related papers (2021-06-04T17:52:50Z) - Learning Deep Kernels for Non-Parametric Two-Sample Tests [50.92621794426821]
We propose a class of kernel-based two-sample tests, which aim to determine whether two sets of samples are drawn from the same distribution.
Our tests are constructed from kernels parameterized by deep neural nets, trained to maximize test power.
arXiv Detail & Related papers (2020-02-21T03:54:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.