Learning False Discovery Rate Control via Model-Based Neural Networks
- URL: http://arxiv.org/abs/2602.05798v1
- Date: Thu, 05 Feb 2026 15:53:11 GMT
- Title: Learning False Discovery Rate Control via Model-Based Neural Networks
- Authors: Arnau Vilella, Jasin Machkour, Michael Muma, Daniel P. Palomar,
- Abstract summary: We introduce a learning-augmented enhancement of the T-Rex Selector framework that narrows the gap between the realized false discovery proportion (FDP) and the target false discovery rate (FDR)<n>Our approach replaces the analytical FDP estimator with a neural network trained solely on diverse synthetic datasets, enabling a substantially tighter and more accurate approximation of the FDP.
- Score: 14.45679797184966
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Controlling the false discovery rate (FDR) in high-dimensional variable selection requires balancing rigorous error control with statistical power. Existing methods with provable guarantees are often overly conservative, creating a persistent gap between the realized false discovery proportion (FDP) and the target FDR level. We introduce a learning-augmented enhancement of the T-Rex Selector framework that narrows this gap. Our approach replaces the analytical FDP estimator with a neural network trained solely on diverse synthetic datasets, enabling a substantially tighter and more accurate approximation of the FDP. This refinement allows the procedure to operate much closer to the desired FDR level, thereby increasing discovery power while maintaining effective approximate control. Through extensive simulations and a challenging synthetic genome-wide association study (GWAS), we demonstrate that our method achieves superior detection of true variables compared to existing approaches.
Related papers
- GRIP2: A Robust and Powerful Deep Knockoff Method for Feature Selection [9.889589777434283]
Group Regularization Importance Persistence in 2 Dimensions (GRIP2), integrates first-layer feature activity over a two-dimensional regularization surface.<n>In experiments on synthetic and semi-real data, GRIP2 demonstrates improved robustness to feature correlation and noise level.<n>On real-world HIV drug resistance data, GRIP2 recovers known resistance-associated mutations with power better than established linear baselines.
arXiv Detail & Related papers (2026-01-30T16:30:49Z) - LEC: Linear Expectation Constraints for False-Discovery Control in Selective Prediction and Routing Systems [95.35293543918762]
Large language models (LLMs) often generate unreliable answers, while uncertainty methods fail to fully distinguish correct from incorrect predictions.<n>We address this issue through the lens of false discovery rate (FDR) control, ensuring that among all accepted predictions, the proportion of errors does not exceed a target risk level.<n>We propose LEC, which reinterprets selective prediction as a constrained decision problem by enforcing a Linear Expectation Constraint.
arXiv Detail & Related papers (2025-12-01T11:27:09Z) - A False Discovery Rate Control Method Using a Fully Connected Hidden Markov Random Field for Neuroimaging Data [1.7582682214679273]
We propose fcHMRF-LIS, a powerful, stable, and scalable spatial FDR control method for voxel-wise multiple testing.<n>fcHMRF-LIS achieves accurate FDR control, lower FNR, reduced variability in FDP and FNP, and a higher number of true positives compared to existing methods.
arXiv Detail & Related papers (2025-05-27T03:59:02Z) - Adding Additional Control to One-Step Diffusion with Joint Distribution Matching [58.37264951734603]
JDM is a novel approach that minimizes the reverse KL divergence between image-condition joint distributions.<n>By deriving a tractable upper bound, JDM decouples fidelity learning from condition learning.<n>This asymmetric distillation scheme enables our one-step student to handle controls unknown to the teacher model.
arXiv Detail & Related papers (2025-03-09T15:06:50Z) - High-Dimensional False Discovery Rate Control for Dependent Variables [10.86851797584794]
We propose a dependency-aware T-Rex selector that harnesses the dependency structure among variables.
We prove that our variable penalization mechanism ensures FDR control.
We formulate a fully integrated optimal calibration algorithm that concurrently determines the parameters of the graphical model and the T-Rex framework.
arXiv Detail & Related papers (2024-01-28T22:56:16Z) - Provable Guarantees for Generative Behavior Cloning: Bridging Low-Level
Stability and High-Level Behavior [51.60683890503293]
We propose a theoretical framework for studying behavior cloning of complex expert demonstrations using generative modeling.
We show that pure supervised cloning can generate trajectories matching the per-time step distribution of arbitrary expert trajectories.
arXiv Detail & Related papers (2023-07-27T04:27:26Z) - Probabilistic Model Incorporating Auxiliary Covariates to Control FDR [6.270317798744481]
Controlling False Discovery Rate (FDR) while leveraging the side information of multiple hypothesis testing is an emerging research topic in modern data science.
We propose a deep Black-Box framework controlling FDR (named as NeurT-FDR) which boosts statistical power and controls FDR for multiple-hypothesis testing.
We show that NeurT-FDR makes substantially more discoveries in three real datasets compared to competitive baselines.
arXiv Detail & Related papers (2022-10-06T19:35:53Z) - Error-based Knockoffs Inference for Controlled Feature Selection [49.99321384855201]
We propose an error-based knockoff inference method by integrating the knockoff features, the error-based feature importance statistics, and the stepdown procedure together.
The proposed inference procedure does not require specifying a regression model and can handle feature selection with theoretical guarantees.
arXiv Detail & Related papers (2022-03-09T01:55:59Z) - NeurT-FDR: Controlling FDR by Incorporating Feature Hierarchy [7.496622386458525]
We propose NeurT-FDR which boosts statistical power and controls FDR for multiple hypothesis testing.
We show that NeurT-FDR has strong FDR guarantees and makes substantially more discoveries in synthetic and real datasets.
arXiv Detail & Related papers (2021-01-24T21:55:10Z) - Uncertainty Estimation Using a Single Deep Deterministic Neural Network [66.26231423824089]
We propose a method for training a deterministic deep model that can find and reject out of distribution data points at test time with a single forward pass.
We scale training in these with a novel loss function and centroid updating scheme and match the accuracy of softmax models.
arXiv Detail & Related papers (2020-03-04T12:27:36Z) - Stochastic-Sign SGD for Federated Learning with Theoretical Guarantees [49.91477656517431]
Quantization-based solvers have been widely adopted in Federated Learning (FL)
No existing methods enjoy all the aforementioned properties.
We propose an intuitively-simple yet theoretically-simple method based on SIGNSGD to bridge the gap.
arXiv Detail & Related papers (2020-02-25T15:12:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.