Sparse network asymptotics for logistic regression
- URL: http://arxiv.org/abs/2010.04703v1
- Date: Fri, 9 Oct 2020 17:46:29 GMT
- Title: Sparse network asymptotics for logistic regression
- Authors: Bryan S. Graham
- Abstract summary: Asymptotic normality of the logistic regression is shown using a martingale central limit theorem (CLT) for triangular arrays.
Sparse networks may lead to better inference since they suggest variances which incorporate additional sources of sampling variation and (ii) are valid under varying degrees of dyadic dependence.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Consider a bipartite network where $N$ consumers choose to buy or not to buy
$M$ different products. This paper considers the properties of the logistic
regression of the $N\times M$ array of i-buys-j purchase decisions,
$\left[Y_{ij}\right]_{1\leq i\leq N,1\leq j\leq M}$, onto known functions of
consumer and product attributes under asymptotic sequences where (i) both $N$
and $M$ grow large and (ii) the average number of products purchased per
consumer is finite in the limit. This latter assumption implies that the
network of purchases is sparse: only a (very) small fraction of all possible
purchases are actually made (concordant with many real-world settings). Under
sparse network asymptotics, the first and last terms in an extended
Hoeffding-type variance decomposition of the score of the logit composite
log-likelihood are of equal order. In contrast, under dense network
asymptotics, the last term is asymptotically negligible. Asymptotic normality
of the logistic regression coefficients is shown using a martingale central
limit theorem (CLT) for triangular arrays. Unlike in the dense case, the
normality result derived here also holds under degeneracy of the network
graphon. Relatedly, when there happens to be no dyadic dependence in the
dataset in hand, it specializes to recently derived results on the behavior of
logistic regression with rare events and iid data. Sparse network asymptotics
may lead to better inference in practice since they suggest variance estimators
which (i) incorporate additional sources of sampling variation and (ii) are
valid under varying degrees of dyadic dependence.
Related papers
- Nonasymptotic Analysis of Stochastic Gradient Descent with the Richardson-Romberg Extrapolation [22.652143194356864]
We address the problem of solving strongly convex and smooth problems using a descent gradient (SGD) algorithm with a constant step size.
We provide an expansion of the mean-squared error of the resulting estimator with respect to the number iterations of $n$.
We establish that this chain is geometrically ergodic with respect to a defined weighted Wasserstein semimetric.
arXiv Detail & Related papers (2024-10-07T15:02:48Z) - The Curse of Memory in Stochastic Approximation: Extended Version [1.534667887016089]
Theory and application of approximation (SA) has grown within the control systems community since the earliest days of adaptive control.
Recent results establish remarkable performance of SA with (sufficiently small) constant step-size $alpha>0$.
arXiv Detail & Related papers (2023-09-06T12:22:32Z) - Towards Faster Non-Asymptotic Convergence for Diffusion-Based Generative
Models [49.81937966106691]
We develop a suite of non-asymptotic theory towards understanding the data generation process of diffusion models.
In contrast to prior works, our theory is developed based on an elementary yet versatile non-asymptotic approach.
arXiv Detail & Related papers (2023-06-15T16:30:08Z) - Degeneracy is OK: Logarithmic Regret for Network Revenue Management with
Indiscrete Distributions [14.959412298776199]
We study the classical Network Revenue Management (NRM) problem with accept/reject decisions and $T$ IID arrivals.
We develop an online algorithm that achieves $O(log2 T)$ regret under this model.
We derive a second result that achieves $O(log T)$ regret under an additional assumption of second-order growth.
arXiv Detail & Related papers (2022-10-14T17:52:19Z) - On the Effective Number of Linear Regions in Shallow Univariate ReLU
Networks: Convergence Guarantees and Implicit Bias [50.84569563188485]
We show that gradient flow converges in direction when labels are determined by the sign of a target network with $r$ neurons.
Our result may already hold for mild over- parameterization, where the width is $tildemathcalO(r)$ and independent of the sample size.
arXiv Detail & Related papers (2022-05-18T16:57:10Z) - An efficient projection neural network for $\ell_1$-regularized logistic
regression [10.517079029721257]
This paper presents a simple projection neural network for $ell_$-regularized logistics regression.
The proposed neural network does not require any extra auxiliary variable nor any smooth approximation.
We also investigate the convergence of the proposed neural network by using the Lyapunov theory and show that it converges to a solution of the problem with any arbitrary initial value.
arXiv Detail & Related papers (2021-05-12T06:13:44Z) - Convergence Rates of Stochastic Gradient Descent under Infinite Noise
Variance [14.06947898164194]
Heavy tails emerge in gradient descent (SGD) in various scenarios.
We provide convergence guarantees for SGD under a state-dependent and heavy-tailed noise with a potentially infinite variance.
Our results indicate that even under heavy-tailed noise with infinite variance, SGD can converge to the global optimum.
arXiv Detail & Related papers (2021-02-20T13:45:11Z) - Stochastic Approximation for Online Tensorial Independent Component
Analysis [98.34292831923335]
Independent component analysis (ICA) has been a popular dimension reduction tool in statistical machine learning and signal processing.
In this paper, we present a by-product online tensorial algorithm that estimates for each independent component.
arXiv Detail & Related papers (2020-12-28T18:52:37Z) - ROOT-SGD: Sharp Nonasymptotics and Near-Optimal Asymptotics in a Single Algorithm [71.13558000599839]
We study the problem of solving strongly convex and smooth unconstrained optimization problems using first-order algorithms.
We devise a novel, referred to as Recursive One-Over-T SGD, based on an easily implementable, averaging of past gradients.
We prove that it simultaneously achieves state-of-the-art performance in both a finite-sample, nonasymptotic sense and an sense.
arXiv Detail & Related papers (2020-08-28T14:46:56Z) - A Random Matrix Analysis of Random Fourier Features: Beyond the Gaussian
Kernel, a Precise Phase Transition, and the Corresponding Double Descent [85.77233010209368]
This article characterizes the exacts of random Fourier feature (RFF) regression, in the realistic setting where the number of data samples $n$ is all large and comparable.
This analysis also provides accurate estimates of training and test regression errors for large $n,p,N$.
arXiv Detail & Related papers (2020-06-09T02:05:40Z) - On Linear Stochastic Approximation: Fine-grained Polyak-Ruppert and
Non-Asymptotic Concentration [115.1954841020189]
We study the inequality and non-asymptotic properties of approximation procedures with Polyak-Ruppert averaging.
We prove a central limit theorem (CLT) for the averaged iterates with fixed step size and number of iterations going to infinity.
arXiv Detail & Related papers (2020-04-09T17:54:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.