Out of Distribution Detection, Generalization, and Robustness Triangle
with Maximum Probability Theorem
- URL: http://arxiv.org/abs/2203.12145v1
- Date: Wed, 23 Mar 2022 02:42:08 GMT
- Title: Out of Distribution Detection, Generalization, and Robustness Triangle
with Maximum Probability Theorem
- Authors: Amir Emad Marvasti, Ehsan Emad Marvasti, Ulas Bagci
- Abstract summary: MPT uses the probability distribution that the models assume on random variables to provide an upper bound on probability of the model.
We apply MPT to challenging out-of-distribution (OOD) detection problems in computer vision by incorporating MPT as a regularization scheme in training of CNNs and their energy based variants.
- Score: 2.0654955576087084
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Maximum Probability Framework, powered by Maximum Probability Theorem, is a
recent theoretical development, aiming to formally define probabilistic models,
guiding development of objective functions, and regularization of probabilistic
models. MPT uses the probability distribution that the models assume on random
variables to provide an upper bound on probability of the model. We apply MPT
to challenging out-of-distribution (OOD) detection problems in computer vision
by incorporating MPT as a regularization scheme in training of CNNs and their
energy based variants. We demonstrate the effectiveness of the proposed method
on 1080 trained models, with varying hyperparameters, and conclude that MPT
based regularization strategy both stabilizes and improves the generalization
and robustness of base models in addition to improved OOD performance on
CIFAR10, CIFAR100 and MNIST datasets.
Related papers
- MITA: Bridging the Gap between Model and Data for Test-time Adaptation [68.62509948690698]
Test-Time Adaptation (TTA) has emerged as a promising paradigm for enhancing the generalizability of models.
We propose Meet-In-The-Middle based MITA, which introduces energy-based optimization to encourage mutual adaptation of the model and data from opposing directions.
arXiv Detail & Related papers (2024-10-12T07:02:33Z) - When Rigidity Hurts: Soft Consistency Regularization for Probabilistic
Hierarchical Time Series Forecasting [69.30930115236228]
Probabilistic hierarchical time-series forecasting is an important variant of time-series forecasting.
Most methods focus on point predictions and do not provide well-calibrated probabilistic forecasts distributions.
We propose PROFHiT, a fully probabilistic hierarchical forecasting model that jointly models forecast distribution of entire hierarchy.
arXiv Detail & Related papers (2023-10-17T20:30:16Z) - pTSE: A Multi-model Ensemble Method for Probabilistic Time Series
Forecasting [10.441994923253596]
pTSE is a multi-model distribution ensemble method for probabilistic forecasting based on Hidden Markov Model (HMM)
We provide a complete theoretical analysis of pTSE to prove that the empirical distribution of time series subject to an HMM will converge to the stationary distribution almost surely.
Experiments on benchmarks show the superiority of pTSE overall member models and competitive ensemble methods.
arXiv Detail & Related papers (2023-05-16T07:00:57Z) - Meta-Uncertainty in Bayesian Model Comparison [0.0]
We propose a fully probabilistic framework for quantifying meta-uncertainty.
We demonstrate the utility of the proposed method in the context of conjugate Bayesian regression, likelihood-based inference with Markov chain Monte Carlo, and simulation-based inference with neural networks.
arXiv Detail & Related papers (2022-10-13T18:10:29Z) - When Rigidity Hurts: Soft Consistency Regularization for Probabilistic
Hierarchical Time Series Forecasting [69.30930115236228]
Probabilistic hierarchical time-series forecasting is an important variant of time-series forecasting.
Most methods focus on point predictions and do not provide well-calibrated probabilistic forecasts distributions.
We propose PROFHiT, a fully probabilistic hierarchical forecasting model that jointly models forecast distribution of entire hierarchy.
arXiv Detail & Related papers (2022-06-16T06:13:53Z) - PSD Representations for Effective Probability Models [117.35298398434628]
We show that a recently proposed class of positive semi-definite (PSD) models for non-negative functions is particularly suited to this end.
We characterize both approximation and generalization capabilities of PSD models, showing that they enjoy strong theoretical guarantees.
Our results open the way to applications of PSD models to density estimation, decision theory and inference.
arXiv Detail & Related papers (2021-06-30T15:13:39Z) - Probabilistic Gradient Boosting Machines for Large-Scale Probabilistic
Regression [51.770998056563094]
Probabilistic Gradient Boosting Machines (PGBM) is a method to create probabilistic predictions with a single ensemble of decision trees.
We empirically demonstrate the advantages of PGBM compared to existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-03T08:32:13Z) - Deep neural network enabled corrective source term approach to hybrid
analysis and modeling [0.0]
Hybrid Analysis and Modeling (HAM) is an emerging modeling paradigm which aims to combine physics-based modeling and data-driven modeling.
We introduce, justify and demonstrate a novel approach to HAM -- the Corrective Source Term Approach (CoSTA)
arXiv Detail & Related papers (2021-05-24T20:17:13Z) - Community Detection in the Stochastic Block Model by Mixed Integer
Programming [3.8073142980733]
Degree-Corrected Block Model (DCSBM) is a popular model to generate random graphs with community structure given an expected degree sequence.
Standard approach of community detection based on the DCSBM is to search for the model parameters that are the most likely to have produced the observed network data through maximum likelihood estimation (MLE)
We present mathematical programming formulations and exact solution methods that can provably find the model parameters and community assignments of maximum likelihood given an observed graph.
arXiv Detail & Related papers (2021-01-26T22:04:40Z) - Probabilistic electric load forecasting through Bayesian Mixture Density
Networks [70.50488907591463]
Probabilistic load forecasting (PLF) is a key component in the extended tool-chain required for efficient management of smart energy grids.
We propose a novel PLF approach, framed on Bayesian Mixture Density Networks.
To achieve reliable and computationally scalable estimators of the posterior distributions, both Mean Field variational inference and deep ensembles are integrated.
arXiv Detail & Related papers (2020-12-23T16:21:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.