Automation for Interpretable Machine Learning Through a Comparison of
Loss Functions to Regularisers
- URL: http://arxiv.org/abs/2106.03428v1
- Date: Mon, 7 Jun 2021 08:50:56 GMT
- Title: Automation for Interpretable Machine Learning Through a Comparison of
Loss Functions to Regularisers
- Authors: A. I. Parkes, J. Camilleri, D. A. Hudson and A. J. Sobey
- Abstract summary: This paper explores the use of the Fit to Median Error measure in machine learning regression automation.
It improves interpretability by regularising learnt input-output relationships to the conditional median.
Networks optimised for their Fit to Median Error are shown to approximate the ground truth more consistently.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To increase the ubiquity of machine learning it needs to be automated.
Automation is cost-effective as it allows experts to spend less time tuning the
approach, which leads to shorter development times. However, while this
automation produces highly accurate architectures, they can be uninterpretable,
acting as `black-boxes' which produce low conventional errors but fail to model
the underlying input-output relationships -- the ground truth. This paper
explores the use of the Fit to Median Error measure in machine learning
regression automation, using evolutionary computation in order to improve the
approximation of the ground truth. When used alongside conventional error
measures it improves interpretability by regularising learnt input-output
relationships to the conditional median. It is compared to traditional
regularisers to illustrate that the use of the Fit to Median Error produces
regression neural networks which model more consistent input-output
relationships. The problem considered is ship power prediction using a
fuel-saving air lubrication system, which is highly stochastic in nature. The
networks optimised for their Fit to Median Error are shown to approximate the
ground truth more consistently, without sacrificing conventional Minkowski-r
error values.
Related papers
- Automatic debiasing of neural networks via moment-constrained learning [0.0]
Naively learning the regression function and taking a sample mean of the target functional results in biased estimators.
We propose moment-constrained learning as a new RR learning approach that addresses some shortcomings in automatic debiasing.
arXiv Detail & Related papers (2024-09-29T20:56:54Z) - Root Causing Prediction Anomalies Using Explainable AI [3.970146574042422]
We present a novel application of explainable AI (XAI) for root-causing performance degradation in machine learning models.
A single feature corruption can cause cascading feature, label and concept drifts.
We have successfully applied this technique to improve the reliability of models used in personalized advertising.
arXiv Detail & Related papers (2024-03-04T19:38:50Z) - Accelerating Generalized Linear Models by Trading off Computation for
Uncertainty [29.877181350448193]
Generalized Linear Models (GLMs) define a flexible probabilistic framework to model categorical, ordinal and continuous data.
The resulting approximation error adversely impacts the reliability of the model and is not accounted for in the uncertainty of the prediction.
We introduce a family of iterative methods that explicitly model this error.
arXiv Detail & Related papers (2023-10-31T08:58:16Z) - Improving Adaptive Conformal Prediction Using Self-Supervised Learning [72.2614468437919]
We train an auxiliary model with a self-supervised pretext task on top of an existing predictive model and use the self-supervised error as an additional feature to estimate nonconformity scores.
We empirically demonstrate the benefit of the additional information using both synthetic and real data on the efficiency (width), deficit, and excess of conformal prediction intervals.
arXiv Detail & Related papers (2023-02-23T18:57:14Z) - Self-supervised debiasing using low rank regularization [59.84695042540525]
Spurious correlations can cause strong biases in deep neural networks, impairing generalization ability.
We propose a self-supervised debiasing framework potentially compatible with unlabeled samples.
Remarkably, the proposed debiasing framework significantly improves the generalization performance of self-supervised learning baselines.
arXiv Detail & Related papers (2022-10-11T08:26:19Z) - Real-to-Sim: Predicting Residual Errors of Robotic Systems with Sparse
Data using a Learning-based Unscented Kalman Filter [65.93205328894608]
We learn the residual errors between a dynamic and/or simulator model and the real robot.
We show that with the learned residual errors, we can further close the reality gap between dynamic models, simulations, and actual hardware.
arXiv Detail & Related papers (2022-09-07T15:15:12Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Distribution Mismatch Correction for Improved Robustness in Deep Neural
Networks [86.42889611784855]
normalization methods increase the vulnerability with respect to noise and input corruptions.
We propose an unsupervised non-parametric distribution correction method that adapts the activation distribution of each layer.
In our experiments, we empirically show that the proposed method effectively reduces the impact of intense image corruptions.
arXiv Detail & Related papers (2021-10-05T11:36:25Z) - Autocalibration and Tweedie-dominance for Insurance Pricing with Machine
Learning [0.0]
It is shown that minimizing deviance involves a trade-off between the integral of weighted differences of lower partial moments and the bias measured on a specific scale.
This new method to correct for bias adds extra local GLM step to the analysis.
The convex order appears to be the natural tool to compare competing models.
arXiv Detail & Related papers (2021-03-05T12:40:30Z) - Error Autocorrelation Objective Function for Improved System Modeling [1.2760453906939444]
We introduce a "whitening" cost function, the Ljung-Box statistic, which not only minimizes the error but also minimizes the correlations between errors.
The results show significant improvement in generalization for recurrent neural networks (RNNs) and image autoencoders (2d)
arXiv Detail & Related papers (2020-08-08T19:20:32Z) - Evaluating Prediction-Time Batch Normalization for Robustness under
Covariate Shift [81.74795324629712]
We call prediction-time batch normalization, which significantly improves model accuracy and calibration under covariate shift.
We show that prediction-time batch normalization provides complementary benefits to existing state-of-the-art approaches for improving robustness.
The method has mixed results when used alongside pre-training, and does not seem to perform as well under more natural types of dataset shift.
arXiv Detail & Related papers (2020-06-19T05:08:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.