Density-Softmax: Efficient Test-time Model for Uncertainty Estimation and Robustness under Distribution Shifts
- URL: http://arxiv.org/abs/2302.06495v3
- Date: Tue, 28 May 2024 02:27:54 GMT
- Title: Density-Softmax: Efficient Test-time Model for Uncertainty Estimation and Robustness under Distribution Shifts
- Authors: Ha Manh Bui, Anqi Liu,
- Abstract summary: Density-Softmax is a sampling-free deterministic framework for uncertainty estimation.
We show that our model is the solution of minimax uncertainty risk.
Our method enjoys competitive results with state-of-the-art techniques in terms of uncertainty and robustness.
- Score: 8.431465371266391
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sampling-based methods, e.g., Deep Ensembles and Bayesian Neural Nets have become promising approaches to improve the quality of uncertainty estimation and robust generalization. However, they suffer from a large model size and high latency at test-time, which limits the scalability needed for low-resource devices and real-time applications. To resolve these computational issues, we propose Density-Softmax, a sampling-free deterministic framework via combining a density function built on a Lipschitz-constrained feature extractor with the softmax layer. Theoretically, we show that our model is the solution of minimax uncertainty risk and is distance-aware on feature space, thus reducing the over-confidence of the standard softmax under distribution shifts. Empirically, our method enjoys competitive results with state-of-the-art techniques in terms of uncertainty and robustness, while having a lower number of model parameters and a lower latency at test-time.
Related papers
- Regulating Model Reliance on Non-Robust Features by Smoothing Input Marginal Density [93.32594873253534]
Trustworthy machine learning requires meticulous regulation of model reliance on non-robust features.
We propose a framework to delineate and regulate such features by attributing model predictions to the input.
arXiv Detail & Related papers (2024-07-05T09:16:56Z) - Density-Regression: Efficient and Distance-Aware Deep Regressor for
Uncertainty Estimation under Distribution Shifts [11.048463491646993]
Density-Regression is a method that leverages the density function in uncertainty estimation and achieves fast inference by a single forward pass.
We show that Density-Regression has competitive uncertainty estimation performance under distribution shifts with modern deep regressors.
arXiv Detail & Related papers (2024-03-07T23:20:34Z) - Tight Certified Robustness via Min-Max Representations of ReLU Neural
Networks [9.771011198361865]
The reliable deployment of neural networks in control systems requires rigorous robustness guarantees.
In this paper, we obtain tight robustness certificates over convex representations of ReLU neural networks.
arXiv Detail & Related papers (2023-10-07T21:07:45Z) - Softmax Bias Correction for Quantized Generative Models [8.953308552614438]
Post-training quantization (PTQ) is the go-to compression technique for large generative models, such as stable diffusion or large language models.
PTQ methods commonly keep the softmax activation in higher precision as it has been shown to be very sensitive to quantization noise.
This can lead to a significant runtime and power overhead during inference on resource-constraint edge devices.
We propose an offline bias correction technique that improves the quantizability of softmax without additional compute during deployment.
arXiv Detail & Related papers (2023-09-04T17:29:31Z) - Distribution-restrained Softmax Loss for the Model Robustness [11.166004203932351]
We identify a significant factor that affects the robustness of models.
We propose a loss function to suppress the distribution diversity of softmax.
A large number of experiments have shown that our method can improve robustness without significant time consumption.
arXiv Detail & Related papers (2023-03-22T07:56:59Z) - On Efficient Uncertainty Estimation for Resource-Constrained Mobile
Applications [0.0]
Predictive uncertainty supplements model predictions and enables improved functionality of downstream tasks.
We tackle this problem by building upon Monte Carlo Dropout (MCDO) models using the Axolotl framework.
We conduct experiments on (1) a multi-class classification task using the CIFAR10 dataset, and (2) a more complex human body segmentation task.
arXiv Detail & Related papers (2021-11-11T22:24:15Z) - Robust Implicit Networks via Non-Euclidean Contractions [63.91638306025768]
Implicit neural networks show improved accuracy and significant reduction in memory consumption.
They can suffer from ill-posedness and convergence instability.
This paper provides a new framework to design well-posed and robust implicit neural networks.
arXiv Detail & Related papers (2021-06-06T18:05:02Z) - Fast Distributionally Robust Learning with Variance Reduced Min-Max
Optimization [85.84019017587477]
Distributionally robust supervised learning is emerging as a key paradigm for building reliable machine learning systems for real-world applications.
Existing algorithms for solving Wasserstein DRSL involve solving complex subproblems or fail to make use of gradients.
We revisit Wasserstein DRSL through the lens of min-max optimization and derive scalable and efficiently implementable extra-gradient algorithms.
arXiv Detail & Related papers (2021-04-27T16:56:09Z) - Amortized Conditional Normalized Maximum Likelihood: Reliable Out of
Distribution Uncertainty Estimation [99.92568326314667]
We propose the amortized conditional normalized maximum likelihood (ACNML) method as a scalable general-purpose approach for uncertainty estimation.
Our algorithm builds on the conditional normalized maximum likelihood (CNML) coding scheme, which has minimax optimal properties according to the minimum description length principle.
We demonstrate that ACNML compares favorably to a number of prior techniques for uncertainty estimation in terms of calibration on out-of-distribution inputs.
arXiv Detail & Related papers (2020-11-05T08:04:34Z) - Fast Objective & Duality Gap Convergence for Non-Convex Strongly-Concave
Min-Max Problems with PL Condition [52.08417569774822]
This paper focuses on methods for solving smooth non-concave min-max problems, which have received increasing attention due to deep learning (e.g., deep AUC)
arXiv Detail & Related papers (2020-06-12T00:32:21Z) - Uncertainty Estimation Using a Single Deep Deterministic Neural Network [66.26231423824089]
We propose a method for training a deterministic deep model that can find and reject out of distribution data points at test time with a single forward pass.
We scale training in these with a novel loss function and centroid updating scheme and match the accuracy of softmax models.
arXiv Detail & Related papers (2020-03-04T12:27:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.