Improving importance estimation in covariate shift for providing
accurate prediction error
- URL: http://arxiv.org/abs/2402.01450v1
- Date: Fri, 2 Feb 2024 14:39:39 GMT
- Title: Improving importance estimation in covariate shift for providing
accurate prediction error
- Authors: Laura Fdez-D\'iaz, Sara Gonz\'alez Tomillo, Elena Monta\~n\'es, Jos\'e
Ram\'on Quevedo
- Abstract summary: The Kullback-Leibler Importance Estimation Procedure (KLIEP) is capable of estimating importance in a promising way.
This paper explores the potential performance improvement if target information is considered in the computation of the importance.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In traditional Machine Learning, the algorithms predictions are based on the
assumption that the data follows the same distribution in both the training and
the test datasets. However, in real world data this condition does not hold
and, for instance, the distribution of the covariates changes whereas the
conditional distribution of the targets remains unchanged. This situation is
called covariate shift problem where standard error estimation may be no longer
accurate. In this context, the importance is a measure commonly used to
alleviate the influence of covariate shift on error estimations. The main
drawback is that it is not easy to compute. The Kullback-Leibler Importance
Estimation Procedure (KLIEP) is capable of estimating importance in a promising
way. Despite its good performance, it fails to ignore target information, since
it only includes the covariates information for computing the importance. In
this direction, this paper explores the potential performance improvement if
target information is considered in the computation of the importance. Then, a
redefinition of the importance arises in order to be generalized in this way.
Besides the potential improvement in performance, including target information
make possible the application to a real application about plankton
classification that motivates this research and characterized by its great
dimensionality, since considering targets rather than covariates reduces the
computation and the noise in the covariates. The impact of taking target
information is also explored when Logistic Regression (LR), Kernel Mean
Matching (KMM), Ensemble Kernel Mean Matching (EKMM) and the naive predecessor
of KLIEP called Kernel Density Estimation (KDE) methods estimate the
importance. The experimental results lead to a more accurate error estimation
using target information, especially in case of the more promising method
KLIEP.
Related papers
- Generative Principal Component Regression via Variational Inference [2.4415762506639944]
One approach to designing appropriate manipulations is to target key features of predictive models.
We develop a novel objective based on supervised variational autoencoders (SVAEs) that enforces such information is represented in the latent space.
We show in simulations that gPCR dramatically improves target selection in manipulation as compared to standard PCR and SVAEs.
arXiv Detail & Related papers (2024-09-03T22:38:55Z) - Uncertainty Estimation by Fisher Information-based Evidential Deep
Learning [61.94125052118442]
Uncertainty estimation is a key factor that makes deep learning reliable in practical applications.
We propose a novel method, Fisher Information-based Evidential Deep Learning ($mathcalI$-EDL)
In particular, we introduce Fisher Information Matrix (FIM) to measure the informativeness of evidence carried by each sample, according to which we can dynamically reweight the objective loss terms to make the network more focused on the representation learning of uncertain classes.
arXiv Detail & Related papers (2023-03-03T16:12:59Z) - Model Optimization in Imbalanced Regression [2.580765958706854]
Imbalanced domain learning aims to produce accurate models in predicting instances that, though underrepresented, are of utmost importance for the domain.
One of the main reasons for this is the lack of loss functions capable of focusing on minimizing the errors of extreme (rare) values.
Recently, an evaluation metric was introduced: Squared Error Relevance Area (SERA)
This metric posits a bigger emphasis on the errors committed at extreme values while also accounting for the performance in the overall target variable domain.
arXiv Detail & Related papers (2022-06-20T20:23:56Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Robust Importance Sampling for Error Estimation in the Context of
Optimal Bayesian Transfer Learning [13.760785726194591]
We introduce a novel class of Bayesian minimum mean-square error (MMSE) estimators for optimal Bayesian transfer learning (OBTL)
We employ the proposed estimator to evaluate the classification accuracy of a broad family of classifiers that span diverse learning capabilities.
Experimental results based on both synthetic data as well as real-world RNA sequencing (RNA-seq) data show that our proposed OBTL error estimation scheme clearly outperforms standard error estimators.
arXiv Detail & Related papers (2021-09-05T19:11:33Z) - KL Guided Domain Adaptation [88.19298405363452]
Domain adaptation is an important problem and often needed for real-world applications.
A common approach in the domain adaptation literature is to learn a representation of the input that has the same distributions over the source and the target domain.
We show that with a probabilistic representation network, the KL term can be estimated efficiently via minibatch samples.
arXiv Detail & Related papers (2021-06-14T22:24:23Z) - A new framework for experimental design using Bayesian Evidential
Learning: the case of wellhead protection area [0.0]
We predict the wellhead protection area (WHPA), the shape and extent of which is influenced by the distribution of hydraulic conductivity (K), from a small number of tracing experiments (predictors)
Our first objective is to make predictions of the WHPA within the Bayesian Evidential Learning framework, which aims to find a direct relationship between predictor and target using machine learning.
Our second objective is to extend BEL to identify the optimal design of data source locations that minimizes the posterior uncertainty of the WHPA.
arXiv Detail & Related papers (2021-05-12T09:40:28Z) - Scalable Marginal Likelihood Estimation for Model Selection in Deep
Learning [78.83598532168256]
Marginal-likelihood based model-selection is rarely used in deep learning due to estimation difficulties.
Our work shows that marginal likelihoods can improve generalization and be useful when validation data is unavailable.
arXiv Detail & Related papers (2021-04-11T09:50:24Z) - Estimating and Evaluating Regression Predictive Uncertainty in Deep
Object Detectors [9.273998041238224]
We show that training variance networks with negative log likelihood (NLL) can lead to high entropy predictive distributions.
We propose to use the energy score as a non-local proper scoring rule and find that when used for training, the energy score leads to better calibrated and lower entropy predictive distributions.
arXiv Detail & Related papers (2021-01-13T12:53:54Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z) - Evaluating Prediction-Time Batch Normalization for Robustness under
Covariate Shift [81.74795324629712]
We call prediction-time batch normalization, which significantly improves model accuracy and calibration under covariate shift.
We show that prediction-time batch normalization provides complementary benefits to existing state-of-the-art approaches for improving robustness.
The method has mixed results when used alongside pre-training, and does not seem to perform as well under more natural types of dataset shift.
arXiv Detail & Related papers (2020-06-19T05:08:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.