Incorporating Experts' Judgment into Machine Learning Models
- URL: http://arxiv.org/abs/2304.11870v2
- Date: Sat, 29 Apr 2023 20:13:42 GMT
- Title: Incorporating Experts' Judgment into Machine Learning Models
- Authors: Hogun Park and Aly Megahed and Peifeng Yin and Yuya Ong and Pravar
Mahajan and Pei Guo
- Abstract summary: In some cases, domain experts might have a judgment about the expected outcome that might conflict with the prediction of machine learning models.
We present a novel framework that aims at leveraging experts' judgment to mitigate the conflict.
- Score: 2.5363839239628843
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine learning (ML) models have been quite successful in predicting
outcomes in many applications. However, in some cases, domain experts might
have a judgment about the expected outcome that might conflict with the
prediction of ML models. One main reason for this is that the training data
might not be totally representative of the population. In this paper, we
present a novel framework that aims at leveraging experts' judgment to mitigate
the conflict. The underlying idea behind our framework is that we first
determine, using a generative adversarial network, the degree of representation
of an unlabeled data point in the training data. Then, based on such degree, we
correct the \textcolor{black}{machine learning} model's prediction by
incorporating the experts' judgment into it, where the higher that
aforementioned degree of representation, the less the weight we put on the
expert intuition that we add to our corrected output, and vice-versa. We
perform multiple numerical experiments on synthetic data as well as two
real-world case studies (one from the IT services industry and the other from
the financial industry). All results show the effectiveness of our framework;
it yields much higher closeness to the experts' judgment with minimal sacrifice
in the prediction accuracy, when compared to multiple baseline methods. We also
develop a new evaluation metric that combines prediction accuracy with the
closeness to experts' judgment. Our framework yields statistically significant
results when evaluated on that metric.
Related papers
- AMEND: A Mixture of Experts Framework for Long-tailed Trajectory Prediction [6.724750970258851]
We propose a modular model-agnostic framework for trajectory prediction.
Each expert is trained with a specialized skill with respect to a particular part of the data.
To produce predictions, we utilise a router network that selects the best expert by generating relative confidence scores.
arXiv Detail & Related papers (2024-02-13T02:43:41Z) - On Least Square Estimation in Softmax Gating Mixture of Experts [78.3687645289918]
We investigate the performance of the least squares estimators (LSE) under a deterministic MoE model.
We establish a condition called strong identifiability to characterize the convergence behavior of various types of expert functions.
Our findings have important practical implications for expert selection.
arXiv Detail & Related papers (2024-02-05T12:31:18Z) - Designing Decision Support Systems Using Counterfactual Prediction Sets [15.121082690769525]
Decision support systems for classification tasks are predominantly designed to predict the value of the ground truth labels.
This paper revisits the design of this type of systems from the perspective of online learning.
We develop a methodology that does not require, nor assumes, an expert model.
arXiv Detail & Related papers (2023-06-06T18:00:09Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - Learning to Predict Trustworthiness with Steep Slope Loss [69.40817968905495]
We study the problem of predicting trustworthiness on real-world large-scale datasets.
We observe that the trustworthiness predictors trained with prior-art loss functions are prone to view both correct predictions and incorrect predictions to be trustworthy.
We propose a novel steep slope loss to separate the features w.r.t. correct predictions from the ones w.r.t. incorrect predictions by two slide-like curves that oppose each other.
arXiv Detail & Related papers (2021-09-30T19:19:09Z) - Hessian-based toolbox for reliable and interpretable machine learning in
physics [58.720142291102135]
We present a toolbox for interpretability and reliability, extrapolation of the model architecture.
It provides a notion of the influence of the input data on the prediction at a given test point, an estimation of the uncertainty of the model predictions, and an agnostic score for the model predictions.
Our work opens the road to the systematic use of interpretability and reliability methods in ML applied to physics and, more generally, science.
arXiv Detail & Related papers (2021-08-04T16:32:59Z) - Test-time Collective Prediction [73.74982509510961]
Multiple parties in machine learning want to jointly make predictions on future test points.
Agents wish to benefit from the collective expertise of the full set of agents, but may not be willing to release their data or model parameters.
We explore a decentralized mechanism to make collective predictions at test time, leveraging each agent's pre-trained model.
arXiv Detail & Related papers (2021-06-22T18:29:58Z) - Towards Unbiased and Accurate Deferral to Multiple Experts [19.24068936057053]
We propose a framework that simultaneously learns a classifier and a deferral system, with the deferral system choosing to defer to one or more human experts.
We test our framework on a synthetic dataset and a content moderation dataset with biased synthetic experts, and show that it significantly improves the accuracy and fairness of the final predictions.
arXiv Detail & Related papers (2021-02-25T17:08:39Z) - Financial Data Analysis Using Expert Bayesian Framework For Bankruptcy
Prediction [0.0]
We propose another route of generative modeling using Expert Bayesian framework.
The biggest advantage of the proposed framework is an explicit inclusion of expert judgment in the modeling process.
The proposed approach is well suited for highly regulated or safety critical applications such as in finance or in medical diagnosis.
arXiv Detail & Related papers (2020-10-19T19:09:02Z) - Competing AI: How does competition feedback affect machine learning? [14.350250426090893]
We show that competition causes predictors to specialize for specific sub-populations at the cost of worse performance over the general population.
We show that having too few or too many competing predictors in a market can hurt the overall prediction quality.
arXiv Detail & Related papers (2020-09-15T00:13:32Z) - Ambiguity in Sequential Data: Predicting Uncertain Futures with
Recurrent Models [110.82452096672182]
We propose an extension of the Multiple Hypothesis Prediction (MHP) model to handle ambiguous predictions with sequential data.
We also introduce a novel metric for ambiguous problems, which is better suited to account for uncertainties.
arXiv Detail & Related papers (2020-03-10T09:15:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.