Balancing central and marginal rejection when combining independent
significance tests
- URL: http://arxiv.org/abs/2310.16600v2
- Date: Mon, 13 Nov 2023 20:22:03 GMT
- Title: Balancing central and marginal rejection when combining independent
significance tests
- Authors: Chris Salahub and Wayne Oldford
- Abstract summary: A common approach to evaluating the significance of a collection of $p$-values combines them with a pooling function.
A series of alternative hypotheses are introduced that communicate the strength and prevalence of non-null evidence in the $p$-values.
It is proven that central rejection is always greater than or equal to marginal rejection, motivating a quotient to measure the balance between the two.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A common approach to evaluating the significance of a collection of
$p$-values combines them with a pooling function, in particular when the
original data are not available. These pooled $p$-values convert a sample of
$p$-values into a single number which behaves like a univariate $p$-value. To
clarify discussion of these functions, a telescoping series of alternative
hypotheses are introduced that communicate the strength and prevalence of
non-null evidence in the $p$-values before general pooling formulae are
discussed. A pattern noticed in the UMP pooled $p$-value for a particular
alternative motivates the definition and discussion of central and marginal
rejection levels at $\alpha$. It is proven that central rejection is always
greater than or equal to marginal rejection, motivating a quotient to measure
the balance between the two for pooled $p$-values. A combining function based
on the $\chi^2_{\kappa}$ quantile transformation is proposed to control this
quotient and shown to be robust to mis-specified parameters relative to the
UMP. Different powers for different parameter settings motivate a map of
plausible alternatives based on where this pooled $p$-value is minimized.
Related papers
- Data subsampling for Poisson regression with pth-root-link [53.63838219437508]
We develop and analyze data subsampling techniques for Poisson regression.
In particular, we consider the Poisson generalized linear model with ID- and square root-link functions.
arXiv Detail & Related papers (2024-10-30T10:09:05Z) - A Statistical Analysis of Deep Federated Learning for Intrinsically Low-dimensional Data [32.98264375121064]
Federated Learning (FL) has emerged as a groundbreaking paradigm in collaborative machine learning.
This paper investigates the generalization properties of deep federated regression within a two-stage sampling model.
arXiv Detail & Related papers (2024-10-28T01:36:25Z) - Mind the Gap: A Causal Perspective on Bias Amplification in Prediction & Decision-Making [58.06306331390586]
We introduce the notion of a margin complement, which measures how much a prediction score $S$ changes due to a thresholding operation.
We show that under suitable causal assumptions, the influences of $X$ on the prediction score $S$ are equal to the influences of $X$ on the true outcome $Y$.
arXiv Detail & Related papers (2024-05-24T11:22:19Z) - Kernelized Normalizing Constant Estimation: Bridging Bayesian Quadrature
and Bayesian Optimization [51.533164528799084]
We show that to estimate the normalizing constant within a small relative error, the level of difficulty depends on the value of $lambda$.
We find that this pattern holds true even when the function evaluations are noisy.
arXiv Detail & Related papers (2024-01-11T07:45:09Z) - Group Pose: A Simple Baseline for End-to-End Multi-person Pose
Estimation [102.02917299051757]
We present a simple yet effective transformer approach, named Group Pose.
We replace single self-attention over all the $Ntimes(K+1)$ queries with two subsequent group self-attentions.
Experimental results on MS COCO and CrowdPose show that our approach without human box supervision is superior to previous methods.
arXiv Detail & Related papers (2023-08-14T17:58:04Z) - A Scale-Invariant Sorting Criterion to Find a Causal Order in Additive
Noise Models [49.038420266408586]
We show that sorting variables by increasing variance often yields an ordering close to a causal order.
We propose an efficient baseline algorithm termed $R2$-SortnRegress that exploits high $R2$-sortability.
Our findings reveal high $R2$-sortability as an assumption about the data generating process relevant to causal discovery.
arXiv Detail & Related papers (2023-03-31T17:05:46Z) - The Projected Covariance Measure for assumption-lean variable significance testing [3.8936058127056357]
A simple but common approach is to specify a linear model, and then test whether the regression coefficient for $X$ is non-zero.
We study the problem of testing the model-free null of conditional mean independence, i.e. that the conditional mean of $Y$ given $X$ and $Z$ does not depend on $X$.
We propose a simple and general framework that can leverage flexible nonparametric or machine learning methods, such as additive models or random forests.
arXiv Detail & Related papers (2022-11-03T17:55:50Z) - Observable adjustments in single-index models for regularized
M-estimators [3.5353632767823506]
In regime where sample size $n$ and dimension $p$ are both increasing, the behavior of the empirical distribution of $hatbeta$ and the predicted values $Xhatbeta$ has been previously characterized.
This paper develops a different theory to describe the empirical distribution of $hatbeta$ and $Xhatbeta$.
arXiv Detail & Related papers (2022-04-14T14:32:02Z) - Distributionally Robust Prescriptive Analytics with Wasserstein Distance [10.475438374386886]
This paper proposes a new distributionally robust approach under Wasserstein ambiguity sets.
We show that the nominal distribution converges to the actual conditional distribution under the Wasserstein distance.
arXiv Detail & Related papers (2021-06-10T13:08:17Z) - Sample Complexity of Asynchronous Q-Learning: Sharper Analysis and
Variance Reduction [63.41789556777387]
Asynchronous Q-learning aims to learn the optimal action-value function (or Q-function) of a Markov decision process (MDP)
We show that the number of samples needed to yield an entrywise $varepsilon$-accurate estimate of the Q-function is at most on the order of $frac1mu_min (1-gamma)5varepsilon2+ fract_mixmu_min (1-gamma)$ up to some logarithmic factor.
arXiv Detail & Related papers (2020-06-04T17:51:00Z) - Optimal rates for independence testing via $U$-statistic permutation
tests [7.090165638014331]
We study the problem of independence testing given independent and identically distributed pairs taking values in a $sigma$-finite, separable measure space.
We first show that there is no valid test of independence that is uniformly consistent against alternatives of the form $f: D(f) geq rho2 $.
arXiv Detail & Related papers (2020-01-15T19:04:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.