Quasiprobabilistic Density Ratio Estimation with a Reverse Engineered Classification Loss Function
- URL: http://arxiv.org/abs/2512.19913v1
- Date: Mon, 22 Dec 2025 22:37:19 GMT
- Title: Quasiprobabilistic Density Ratio Estimation with a Reverse Engineered Classification Loss Function
- Authors: Matthew Drnevich, Stephen Jiggins, Kyle Cranmer,
- Abstract summary: We introduce a convex loss function that is well-suited for both probabilistic and quasiprobabilistic density ratio estimation.<n>We demonstrate our approach on a real-world example from particle physics, of di-Higgs production in association with jets via gluon-gluon fusion.
- Score: 0.7769607568805288
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider a generalization of the classifier-based density-ratio estimation task to a quasiprobabilistic setting where probability densities can be negative. The problem with most loss functions used for this task is that they implicitly define a relationship between the optimal classifier and the target quasiprobabilistic density ratio which is discontinuous or not surjective. We address these problems by introducing a convex loss function that is well-suited for both probabilistic and quasiprobabilistic density ratio estimation. To quantify performance, an extended version of the Sliced-Wasserstein distance is introduced which is compatible with quasiprobability distributions. We demonstrate our approach on a real-world example from particle physics, of di-Higgs production in association with jets via gluon-gluon fusion, and achieve state-of-the-art results.
Related papers
- Flow-Based Density Ratio Estimation for Intractable Distributions with Applications in Genomics [80.05951561886123]
We leverage condition-aware flow matching to derive a single dynamical formulation for tracking density ratios along generative trajectories.<n>We demonstrate competitive performance on simulated benchmarks for closed-form ratio estimation, and show that our method supports versatile tasks in single-cell genomics data analysis.
arXiv Detail & Related papers (2026-02-27T17:27:55Z) - Coarsening Bias from Variable Discretization in Causal Functionals [0.8880611506199766]
A class of causal effect functionals requires integration over conditional densities of continuous variables.<n>Discretization alters the population-level functional and can induce non-negligible approximation bias.<n>We propose a simple bias-reduced functional that evaluates the outcome regression at within-bin conditional means.
arXiv Detail & Related papers (2026-02-25T16:32:04Z) - Density estimation with atoms, and functional estimation for mixed discrete-continuous data [29.43493007296859]
In density-functional estimation, it is standard to assume that the underlying distribution has a density with respect to the Lebesgue measure.<n>When the data distribution is a mixture of continuous and discrete components, the resulting methods are inconsistent in theory and perform poorly in practice.<n>We modify existing estimators for a broad class of functionals of the continuous component of the mixture.
arXiv Detail & Related papers (2025-08-03T10:22:35Z) - Estimating Unbounded Density Ratios: Applications in Error Control under Covariate Shift [17.924340241624204]
We study density ratio estimators using loss functions based on at least squares and logistic regression.<n>We establish upper bounds on estimation errors with standard minimax optimal rates, up to logarithmic factors.<n>Our results accommodate density ratio functions with unbounded domains and ranges.
arXiv Detail & Related papers (2025-03-29T11:35:39Z) - Regulating Model Reliance on Non-Robust Features by Smoothing Input Marginal Density [93.32594873253534]
Trustworthy machine learning requires meticulous regulation of model reliance on non-robust features.
We propose a framework to delineate and regulate such features by attributing model predictions to the input.
arXiv Detail & Related papers (2024-07-05T09:16:56Z) - Binary Losses for Density Ratio Estimation [2.512309434783062]
Estimating the ratio of two probability densities from a finite number of observations is a central machine learning problem.<n>In this work, we characterize all loss functions that result in density ratio estimators with small error.<n>We obtain a simple recipe for constructing loss functions with certain properties, such as those that prioritize an accurate estimation of large density ratio values.
arXiv Detail & Related papers (2024-07-01T15:24:34Z) - Sobolev Space Regularised Pre Density Models [51.558848491038916]
We propose a new approach to non-parametric density estimation that is based on regularizing a Sobolev norm of the density.
This method is statistically consistent, and makes the inductive validation model clear and consistent.
arXiv Detail & Related papers (2023-07-25T18:47:53Z) - Learning Unnormalized Statistical Models via Compositional Optimization [73.30514599338407]
Noise-contrastive estimation(NCE) has been proposed by formulating the objective as the logistic loss of the real data and the artificial noise.
In this paper, we study it a direct approach for optimizing the negative log-likelihood of unnormalized models.
arXiv Detail & Related papers (2023-06-13T01:18:16Z) - Machine-Learned Exclusion Limits without Binning [0.0]
We extend the Machine-Learned Likelihoods (MLL) method by including Kernel Density Estimators (KDE) to extract one-dimensional signal and background probability density functions.
We apply the method to two cases of interest at the LHC: a search for exotic Higgs bosons, and a $Z'$ boson decaying into lepton pairs.
arXiv Detail & Related papers (2022-11-09T11:04:50Z) - Nonparametric Probabilistic Regression with Coarse Learners [1.8275108630751844]
We show that we can compute precise conditional densities with minimal assumptions on the shape or form of the density.
We demonstrate this approach on a variety of datasets and show competitive performance, particularly on larger datasets.
arXiv Detail & Related papers (2022-10-28T16:25:26Z) - Statistical Efficiency of Score Matching: The View from Isoperimetry [96.65637602827942]
We show a tight connection between statistical efficiency of score matching and the isoperimetric properties of the distribution being estimated.
We formalize these results both in the sample regime and in the finite regime.
arXiv Detail & Related papers (2022-10-03T06:09:01Z) - Data-Driven Influence Functions for Optimization-Based Causal Inference [105.5385525290466]
We study a constructive algorithm that approximates Gateaux derivatives for statistical functionals by finite differencing.
We study the case where probability distributions are not known a priori but need to be estimated from data.
arXiv Detail & Related papers (2022-08-29T16:16:22Z) - Robust Estimation for Nonparametric Families via Generative Adversarial
Networks [92.64483100338724]
We provide a framework for designing Generative Adversarial Networks (GANs) to solve high dimensional robust statistics problems.
Our work extend these to robust mean estimation, second moment estimation, and robust linear regression.
In terms of techniques, our proposed GAN losses can be viewed as a smoothed and generalized Kolmogorov-Smirnov distance.
arXiv Detail & Related papers (2022-02-02T20:11:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.