Utilizing Explainable AI for improving the Performance of Neural
Networks
- URL: http://arxiv.org/abs/2210.04686v1
- Date: Fri, 7 Oct 2022 09:39:20 GMT
- Title: Utilizing Explainable AI for improving the Performance of Neural
Networks
- Authors: Huawei Sun, Lorenzo Servadei, Hao Feng, Michael Stephan, Robert Wille,
Avik Santra
- Abstract summary: We propose a retraining pipeline that consistently improves the model predictions starting from XAI.
In order to benchmark our method, we evaluate it on both real-life and public datasets.
Experiments using the SHAP-based retraining approach achieve a 4% more accuracy w.r.t. the standard equal weight retraining for people counting tasks.
- Score: 6.670483888835783
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Nowadays, deep neural networks are widely used in a variety of fields that
have a direct impact on society. Although those models typically show
outstanding performance, they have been used for a long time as black boxes. To
address this, Explainable Artificial Intelligence (XAI) has been developing as
a field that aims to improve the transparency of the model and increase their
trustworthiness. We propose a retraining pipeline that consistently improves
the model predictions starting from XAI and utilizing state-of-the-art
techniques. To do that, we use the XAI results, namely SHapley Additive
exPlanations (SHAP) values, to give specific training weights to the data
samples. This leads to an improved training of the model and, consequently,
better performance. In order to benchmark our method, we evaluate it on both
real-life and public datasets. First, we perform the method on a radar-based
people counting scenario. Afterward, we test it on the CIFAR-10, a public
Computer Vision dataset. Experiments using the SHAP-based retraining approach
achieve a 4% more accuracy w.r.t. the standard equal weight retraining for
people counting tasks. Moreover, on the CIFAR-10, our SHAP-based weighting
strategy ends up with a 3% accuracy rate than the training procedure with equal
weighted samples.
Related papers
- Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback [110.16220825629749]
Learning from preference feedback has emerged as an essential step for improving the generation quality and performance of modern language models.
In this work, we identify four core aspects of preference-based learning: preference data, learning algorithm, reward model, and policy training prompts.
Our findings indicate that all aspects are important for performance, with better preference data leading to the largest improvements.
arXiv Detail & Related papers (2024-06-13T16:17:21Z) - It's all about PR -- Smart Benchmarking AI Accelerators using Performance Representatives [40.197673152937256]
Training of statistical performance models often requires vast amounts of data, leading to a significant time investment and can be difficult in case of limited hardware availability.
We propose a novel performance modeling methodology that significantly reduces the number of training samples while maintaining good accuracy.
We achieve a Mean Absolute Percentage Error (MAPE) of as low as 0.02% for single-layer estimations and 0.68% for whole estimations with less than 10000 training samples.
arXiv Detail & Related papers (2024-06-12T15:34:28Z) - Getting More Juice Out of the SFT Data: Reward Learning from Human Demonstration Improves SFT for LLM Alignment [65.15914284008973]
We propose to leverage an Inverse Reinforcement Learning (IRL) technique to simultaneously build an reward model and a policy model.
We show that the proposed algorithms converge to the stationary solutions of the IRL problem.
Our results indicate that it is beneficial to leverage reward learning throughout the entire alignment process.
arXiv Detail & Related papers (2024-05-28T07:11:05Z) - Towards An Online Incremental Approach to Predict Students Performance [0.8287206589886879]
We propose a memory-based online incremental learning approach for updating an online classifier.
Our approach achieves a notable improvement in model accuracy, with an enhancement of nearly 10% compared to the current state-of-the-art.
arXiv Detail & Related papers (2024-05-03T17:13:26Z) - Always-Sparse Training by Growing Connections with Guided Stochastic
Exploration [46.4179239171213]
We propose an efficient always-sparse training algorithm with excellent scaling to larger and sparser models.
We evaluate our method on CIFAR-10/100 and ImageNet using VGG, and ViT models, and compare it against a range of sparsification methods.
arXiv Detail & Related papers (2024-01-12T21:32:04Z) - SHAPNN: Shapley Value Regularized Tabular Neural Network [4.587122314291091]
We present SHAPNN, a novel deep data modeling architecture designed for supervised learning.
Our neural network is trained using standard backward propagation optimization methods, and is regularized with realtime estimated Shapley values.
We evaluate our method on various publicly available datasets and compare it with state-of-the-art deep neural network models.
arXiv Detail & Related papers (2023-09-15T22:45:05Z) - PILOT: A Pre-Trained Model-Based Continual Learning Toolbox [71.63186089279218]
This paper introduces a pre-trained model-based continual learning toolbox known as PILOT.
On the one hand, PILOT implements some state-of-the-art class-incremental learning algorithms based on pre-trained models, such as L2P, DualPrompt, and CODA-Prompt.
On the other hand, PILOT fits typical class-incremental learning algorithms within the context of pre-trained models to evaluate their effectiveness.
arXiv Detail & Related papers (2023-09-13T17:55:11Z) - Value-Consistent Representation Learning for Data-Efficient
Reinforcement Learning [105.70602423944148]
We propose a novel method, called value-consistent representation learning (VCR), to learn representations that are directly related to decision-making.
Instead of aligning this imagined state with a real state returned by the environment, VCR applies a $Q$-value head on both states and obtains two distributions of action values.
It has been demonstrated that our methods achieve new state-of-the-art performance for search-free RL algorithms.
arXiv Detail & Related papers (2022-06-25T03:02:25Z) - Training Sparse Neural Networks using Compressed Sensing [13.84396596420605]
We develop and test a novel method based on compressed sensing which combines the pruning and training into a single step.
Specifically, we utilize an adaptively weighted $ell1$ penalty on the weights during training, which we combine with a generalization of the regularized dual averaging (RDA) algorithm in order to train sparse neural networks.
arXiv Detail & Related papers (2020-08-21T19:35:54Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z) - Adversarial Robustness: From Self-Supervised Pre-Training to Fine-Tuning [134.15174177472807]
We introduce adversarial training into self-supervision, to provide general-purpose robust pre-trained models for the first time.
We conduct extensive experiments to demonstrate that the proposed framework achieves large performance margins.
arXiv Detail & Related papers (2020-03-28T18:28:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.