C-SURE: Shrinkage Estimator and Prototype Classifier for Complex-Valued
Deep Learning
- URL: http://arxiv.org/abs/2006.12590v1
- Date: Mon, 22 Jun 2020 20:02:20 GMT
- Title: C-SURE: Shrinkage Estimator and Prototype Classifier for Complex-Valued
Deep Learning
- Authors: Yifei Xing, Rudrasis Chakraborty, Minxuan Duan, Stella Yu
- Abstract summary: We propose C-SURE, a novel Stein's unbiased risk estimate (SURE) of the JS estimator on the manifold of complex-valued data.
C-SURE is more accurate and robust than SurReal, and the shrinkage estimator is always better than MLE for the same prototype classifier.
- Score: 15.906530504220179
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The James-Stein (JS) shrinkage estimator is a biased estimator that captures
the mean of Gaussian random vectors.While it has a desirable statistical
property of dominance over the maximum likelihood estimator (MLE) in terms of
mean squared error (MSE), not much progress has been made on extending the
estimator onto manifold-valued data.
We propose C-SURE, a novel Stein's unbiased risk estimate (SURE) of the JS
estimator on the manifold of complex-valued data with a theoretically proven
optimum over MLE. Adapting the architecture of the complex-valued SurReal
classifier, we further incorporate C-SURE into a prototype convolutional neural
network (CNN) classifier. We compare C-SURE with SurReal and a real-valued
baseline on complex-valued MSTAR and RadioML datasets.
C-SURE is more accurate and robust than SurReal, and the shrinkage estimator
is always better than MLE for the same prototype classifier. Like SurReal,
C-SURE is much smaller, outperforming the real-valued baseline on MSTAR
(RadioML) with less than 1 percent (3 percent) of the baseline size
Related papers
- Precise Model Benchmarking with Only a Few Observations [6.092112060364272]
We propose an empirical Bayes (EB) estimator that balances direct and regression estimates for each subgroup separately.
EB consistently provides more precise estimates of the LLM performance compared to the direct and regression approaches.
arXiv Detail & Related papers (2024-10-07T17:26:31Z) - How Much is Unseen Depends Chiefly on Information About the Seen [2.169081345816618]
We find that the proportion of data points in an unknown population-that belong to classes that do not appear in the training data-is almost entirely determined by the number $f_k$ of classes that do appear in the training data the same number of times.
We develop a genetic algorithm that, given only the sample, searches for an estimator with minimal mean-squared error (MSE)
arXiv Detail & Related papers (2024-02-08T17:12:49Z) - Sparse PCA with Oracle Property [115.72363972222622]
We propose a family of estimators based on the semidefinite relaxation of sparse PCA with novel regularizations.
We prove that, another estimator within the family achieves a sharper statistical rate of convergence than the standard semidefinite relaxation of sparse PCA.
arXiv Detail & Related papers (2023-12-28T02:52:54Z) - Provably Efficient CVaR RL in Low-rank MDPs [58.58570425202862]
We study risk-sensitive Reinforcement Learning (RL)
We propose a novel Upper Confidence Bound (UCB) bonus-driven algorithm to balance interplay between exploration, exploitation, and representation learning in CVaR RL.
We prove that our algorithm achieves a sample complexity of $epsilon$-optimal CVaR, where $H$ is the length of each episode, $A$ is the capacity of action space, and $d$ is the dimension of representations.
arXiv Detail & Related papers (2023-11-20T17:44:40Z) - Error Reduction from Stacked Regressions [12.657895453939298]
Stacking regressions is an ensemble technique that forms linear combinations of different regression estimators to enhance predictive accuracy.
In this paper, we learn these weights analogously by minimizing a regularized version of the empirical risk subject to a nonnegativity constraint.
Thanks to an adaptive shrinkage effect, the resulting stacked estimator has strictly smaller population risk than best single estimator among them.
arXiv Detail & Related papers (2023-09-18T15:42:12Z) - Consensus-Adaptive RANSAC [104.87576373187426]
We propose a new RANSAC framework that learns to explore the parameter space by considering the residuals seen so far via a novel attention layer.
The attention mechanism operates on a batch of point-to-model residuals, and updates a per-point estimation state to take into account the consensus found through a lightweight one-step transformer.
arXiv Detail & Related papers (2023-07-26T08:25:46Z) - A Model-Based Method for Minimizing CVaR and Beyond [7.751691910877239]
We develop a variant of the prox-linear method for minimizing the Conditional Value-at-Risk (CVaR) objective.
CVaR is a risk measure focused on minimizing worst-case performance, defined as the average of the top quantile of the losses.
In machine learning, such a risk measure is useful to train more robust models.
arXiv Detail & Related papers (2023-05-27T15:38:53Z) - Learning to Estimate Without Bias [57.82628598276623]
Gauss theorem states that the weighted least squares estimator is a linear minimum variance unbiased estimation (MVUE) in linear models.
In this paper, we take a first step towards extending this result to non linear settings via deep learning with bias constraints.
A second motivation to BCE is in applications where multiple estimates of the same unknown are averaged for improved performance.
arXiv Detail & Related papers (2021-10-24T10:23:51Z) - SLOE: A Faster Method for Statistical Inference in High-Dimensional
Logistic Regression [68.66245730450915]
We develop an improved method for debiasing predictions and estimating frequentist uncertainty for practical datasets.
Our main contribution is SLOE, an estimator of the signal strength with convergence guarantees that reduces the computation time of estimation and inference by orders of magnitude.
arXiv Detail & Related papers (2021-03-23T17:48:56Z) - Breaking the Sample Size Barrier in Model-Based Reinforcement Learning
with a Generative Model [50.38446482252857]
This paper is concerned with the sample efficiency of reinforcement learning, assuming access to a generative model (or simulator)
We first consider $gamma$-discounted infinite-horizon Markov decision processes (MDPs) with state space $mathcalS$ and action space $mathcalA$.
We prove that a plain model-based planning algorithm suffices to achieve minimax-optimal sample complexity given any target accuracy level.
arXiv Detail & Related papers (2020-05-26T17:53:18Z) - Robust subgaussian estimation with VC-dimension [0.0]
This work proposes a new general way to bound the excess risk for MOM estimators.
The core technique is the use of VC-dimension (instead of Rademacher complexity) to measure the statistical complexity.
arXiv Detail & Related papers (2020-04-24T13:21:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.