Machine Learning in Sports: A Case Study on Using Explainable Models for
Predicting Outcomes of Volleyball Matches
- URL: http://arxiv.org/abs/2206.09258v1
- Date: Sat, 18 Jun 2022 18:09:15 GMT
- Title: Machine Learning in Sports: A Case Study on Using Explainable Models for
Predicting Outcomes of Volleyball Matches
- Authors: Abhinav Lalwani, Aman Saraiya, Apoorv Singh, Aditya Jain, Tirtharaj
Dash
- Abstract summary: This paper explores a two-phased Explainable Artificial Intelligence(XAI) approach to predict outcomes of matches in the Brazilian volleyball League (SuperLiga)
In the first phase, we directly use the interpretable rule-based ML models that provide a global understanding of the model's behaviors.
In the second phase, we construct non-linear models such as Support Vector Machine (SVM) and Deep Neural Network (DNN) to obtain predictive performance on the volleyball matches' outcomes.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Machine Learning has become an integral part of engineering design and
decision making in several domains, including sports. Deep Neural Networks
(DNNs) have been the state-of-the-art methods for predicting outcomes of
professional sports events. However, apart from getting highly accurate
predictions on these sports events outcomes, it is necessary to answer
questions such as "Why did the model predict that Team A would win Match X
against Team B?" DNNs are inherently black-box in nature. Therefore, it is
required to provide high-quality interpretable, and understandable explanations
for a model's prediction in sports. This paper explores a two-phased
Explainable Artificial Intelligence(XAI) approach to predict outcomes of
matches in the Brazilian volleyball League (SuperLiga). In the first phase, we
directly use the interpretable rule-based ML models that provide a global
understanding of the model's behaviors based on Boolean Rule Column Generation
(BRCG; extracts simple AND-OR classification rules) and Logistic Regression
(LogReg; allows to estimate the feature importance scores). In the second
phase, we construct non-linear models such as Support Vector Machine (SVM) and
Deep Neural Network (DNN) to obtain predictive performance on the volleyball
matches' outcomes. We construct the "post-hoc" explanations for each data
instance using ProtoDash, a method that finds prototypes in the training
dataset that are most similar to the test instance, and SHAP, a method that
estimates the contribution of each feature on the model's prediction. We
evaluate the SHAP explanations using the faithfulness metric. Our results
demonstrate the effectiveness of the explanations for the model's predictions.
Related papers
- Sparse Prototype Network for Explainable Pedestrian Behavior Prediction [60.80524827122901]
We present Sparse Prototype Network (SPN), an explainable method designed to simultaneously predict a pedestrian's future action, trajectory, and pose.
Regularized by mono-semanticity and clustering constraints, the prototypes learn consistent and human-understandable features.
arXiv Detail & Related papers (2024-10-16T03:33:40Z) - Machine Learning for Soccer Match Result Prediction [0.9002260638342727]
This chapter discusses available datasets, the types of models and features, and ways of evaluating model performance.
The aim of this chapter is to give a broad overview of the current state and potential future developments in machine learning for soccer match results prediction.
arXiv Detail & Related papers (2024-03-12T14:00:50Z) - ShuttleSHAP: A Turn-Based Feature Attribution Approach for Analyzing
Forecasting Models in Badminton [52.21869064818728]
Deep learning approaches for player tactic forecasting in badminton show promising performance partially attributed to effective reasoning about rally-player interactions.
We propose a turn-based feature attribution approach, ShuttleSHAP, for analyzing forecasting models in badminton based on variants of Shapley values.
arXiv Detail & Related papers (2023-12-18T05:37:51Z) - Supervised Learning for Table Tennis Match Prediction [2.7835697868135902]
This paper proposes the use of machine learning to predict the outcome of table tennis single matches.
We use player and match statistics as features and evaluate their relative importance in an ablation study.
The results can serve as a baseline for future table tennis prediction models, and can feed back to prediction research in similar ball sports.
arXiv Detail & Related papers (2023-03-28T17:42:13Z) - Rationalizing Predictions by Adversarial Information Calibration [65.19407304154177]
We train two models jointly: one is a typical neural model that solves the task at hand in an accurate but black-box manner, and the other is a selector-predictor model that additionally produces a rationale for its prediction.
We use an adversarial technique to calibrate the information extracted by the two models such that the difference between them is an indicator of the missed or over-selected features.
arXiv Detail & Related papers (2023-01-15T03:13:09Z) - Boosted Dynamic Neural Networks [53.559833501288146]
A typical EDNN has multiple prediction heads at different layers of the network backbone.
To optimize the model, these prediction heads together with the network backbone are trained on every batch of training data.
Treating training and testing inputs differently at the two phases will cause the mismatch between training and testing data distributions.
We formulate an EDNN as an additive model inspired by gradient boosting, and propose multiple training techniques to optimize the model effectively.
arXiv Detail & Related papers (2022-11-30T04:23:12Z) - Evaluating the Explainers: Black-Box Explainable Machine Learning for
Student Success Prediction in MOOCs [5.241055914181294]
We implement five state-of-the-art methodologies for explaining black-box machine learning models.
We examine the strengths of each approach on the downstream task of student performance prediction.
Our results come to the concerning conclusion that the choice of explainer is an important decision.
arXiv Detail & Related papers (2022-07-01T17:09:17Z) - Explain, Edit, and Understand: Rethinking User Study Design for
Evaluating Model Explanations [97.91630330328815]
We conduct a crowdsourcing study, where participants interact with deception detection models that have been trained to distinguish between genuine and fake hotel reviews.
We observe that for a linear bag-of-words model, participants with access to the feature coefficients during training are able to cause a larger reduction in model confidence in the testing phase when compared to the no-explanation control.
arXiv Detail & Related papers (2021-12-17T18:29:56Z) - Hessian-based toolbox for reliable and interpretable machine learning in
physics [58.720142291102135]
We present a toolbox for interpretability and reliability, extrapolation of the model architecture.
It provides a notion of the influence of the input data on the prediction at a given test point, an estimation of the uncertainty of the model predictions, and an agnostic score for the model predictions.
Our work opens the road to the systematic use of interpretability and reliability methods in ML applied to physics and, more generally, science.
arXiv Detail & Related papers (2021-08-04T16:32:59Z) - Learning from the Best: Rationalizing Prediction by Adversarial
Information Calibration [39.685626118667074]
We train two models jointly: one is a typical neural model that solves the task at hand in an accurate but black-box manner, and the other is a selector-predictor model that additionally produces a rationale for its prediction.
We use an adversarial-based technique to calibrate the information extracted by the two models.
For natural language tasks, we propose to use a language-model-based regularizer to encourage the extraction of fluent rationales.
arXiv Detail & Related papers (2020-12-16T11:54:15Z) - Explainable Artificial Intelligence: How Subsets of the Training Data
Affect a Prediction [2.3204178451683264]
We propose a novel methodology which we call Shapley values for training data subset importance.
We show how the proposed explanations can be used to reveal biasedness in models and erroneous training data.
We argue that the explanations enable us to perceive more of the inner workings of the algorithms, and illustrate how models producing similar predictions can be based on very different parts of the training data.
arXiv Detail & Related papers (2020-12-07T12:15:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.