Low coordinate degree algorithms I: Universality of computational
thresholds for hypothesis testing
- URL: http://arxiv.org/abs/2403.07862v1
- Date: Tue, 12 Mar 2024 17:52:35 GMT
- Title: Low coordinate degree algorithms I: Universality of computational
thresholds for hypothesis testing
- Authors: Dmitriy Kunisky
- Abstract summary: Low coordinate degree functions (LCDF) can hypothesis test between high-dimensional probability measures.
We show that LCDF can be tested for the presence of sufficiently "dilute" random signals through noisy channels.
These results are the first computational lower bounds against any large class of algorithms for all of these models.
- Score: 2.4889993472438383
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study when low coordinate degree functions (LCDF) -- linear combinations
of functions depending on small subsets of entries of a vector -- can
hypothesis test between high-dimensional probability measures. These functions
are a generalization, proposed in Hopkins' 2018 thesis but seldom studied
since, of low degree polynomials (LDP), a class widely used in recent
literature as a proxy for all efficient algorithms for tasks in statistics and
optimization. Instead of the orthogonal polynomial decompositions used in LDP
calculations, our analysis of LCDF is based on the Efron-Stein or ANOVA
decomposition, making it much more broadly applicable. By way of illustration,
we prove channel universality for the success of LCDF in testing for the
presence of sufficiently "dilute" random signals through noisy channels: the
efficacy of LCDF depends on the channel only through the scalar Fisher
information for a class of channels including nearly arbitrary additive i.i.d.
noise and nearly arbitrary exponential families. As applications, we extend
lower bounds against LDP for spiked matrix and tensor models under additive
Gaussian noise to lower bounds against LCDF under general noisy channels. We
also give a simple and unified treatment of the effect of censoring models by
erasing observations at random and of quantizing models by taking the sign of
the observations. These results are the first computational lower bounds
against any large class of algorithms for all of these models when the channel
is not one of a few special cases, and thereby give the first substantial
evidence for the universality of several statistical-to-computational gaps.
Related papers
- Low coordinate degree algorithms II: Categorical signals and generalized stochastic block models [2.4889993472438383]
We study when low coordinate degree functions can test for the presence of categorical structure in high-dimensional data.
This complements the first paper of this series, which studied the power of LCDF in testing for continuous structure.
arXiv Detail & Related papers (2024-12-30T18:34:36Z) - Unveiling the Statistical Foundations of Chain-of-Thought Prompting Methods [59.779795063072655]
Chain-of-Thought (CoT) prompting and its variants have gained popularity as effective methods for solving multi-step reasoning problems.
We analyze CoT prompting from a statistical estimation perspective, providing a comprehensive characterization of its sample complexity.
arXiv Detail & Related papers (2024-08-25T04:07:18Z) - Proximal Interacting Particle Langevin Algorithms [0.0]
We introduce Proximal Interacting Particle Langevin Algorithms (PIPLA) for inference and learning in latent variable models.
We propose several variants within the novel proximal IPLA family, tailored to the problem of estimating parameters in a non-differentiable statistical model.
Our theory and experiments together show that PIPLA family can be the de facto choice for parameter estimation problems in latent variable models for non-differentiable models.
arXiv Detail & Related papers (2024-06-20T13:16:41Z) - Computational-Statistical Gaps in Gaussian Single-Index Models [77.1473134227844]
Single-Index Models are high-dimensional regression problems with planted structure.
We show that computationally efficient algorithms, both within the Statistical Query (SQ) and the Low-Degree Polynomial (LDP) framework, necessarily require $Omega(dkstar/2)$ samples.
arXiv Detail & Related papers (2024-03-08T18:50:19Z) - Max-affine regression via first-order methods [7.12511675782289]
The max-affine model ubiquitously arises in applications in signal processing and statistics.
We present a non-asymptotic convergence analysis of gradient descent (GD) and mini-batch gradient descent (SGD) for max-affine regression.
arXiv Detail & Related papers (2023-08-15T23:46:44Z) - A Framework for Analyzing Cross-correlators using Price's Theorem and
Piecewise-Linear Decomposition [5.094549132183797]
We present a general mathematical framework that allows us to analyze cross-correlators constructed using a mixture of piece-wise linear functions.
We show that some of the most promising cross-correlators are based on Huber's loss functions, margin-propagation (MP) functions, and the log-sum-exp (LSE) functions.
arXiv Detail & Related papers (2023-04-18T19:03:27Z) - On High dimensional Poisson models with measurement error: hypothesis
testing for nonlinear nonconvex optimization [13.369004892264146]
We estimation and testing regression model with high dimensionals, which has wide applications in analyzing data.
We propose to estimate regression parameter through minimizing penalized consistency.
The proposed method is applied to the Alzheimer's Disease Initiative.
arXiv Detail & Related papers (2022-12-31T06:58:42Z) - Asymptotically Unbiased Instance-wise Regularized Partial AUC
Optimization: Theory and Algorithm [101.44676036551537]
One-way Partial AUC (OPAUC) and Two-way Partial AUC (TPAUC) measures the average performance of a binary classifier.
Most of the existing methods could only optimize PAUC approximately, leading to inevitable biases that are not controllable.
We present a simpler reformulation of the PAUC problem via distributional robust optimization AUC.
arXiv Detail & Related papers (2022-10-08T08:26:22Z) - Faster One-Sample Stochastic Conditional Gradient Method for Composite
Convex Minimization [61.26619639722804]
We propose a conditional gradient method (CGM) for minimizing convex finite-sum objectives formed as a sum of smooth and non-smooth terms.
The proposed method, equipped with an average gradient (SAG) estimator, requires only one sample per iteration. Nevertheless, it guarantees fast convergence rates on par with more sophisticated variance reduction techniques.
arXiv Detail & Related papers (2022-02-26T19:10:48Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z) - Probabilistic Circuits for Variational Inference in Discrete Graphical
Models [101.28528515775842]
Inference in discrete graphical models with variational methods is difficult.
Many sampling-based methods have been proposed for estimating Evidence Lower Bound (ELBO)
We propose a new approach that leverages the tractability of probabilistic circuit models, such as Sum Product Networks (SPN)
We show that selective-SPNs are suitable as an expressive variational distribution, and prove that when the log-density of the target model is aweighted the corresponding ELBO can be computed analytically.
arXiv Detail & Related papers (2020-10-22T05:04:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.