Deep Explainable Learning with Graph Based Data Assessing and Rule
Reasoning
- URL: http://arxiv.org/abs/2211.04693v2
- Date: Thu, 10 Nov 2022 11:41:56 GMT
- Title: Deep Explainable Learning with Graph Based Data Assessing and Rule
Reasoning
- Authors: Yuanlong Li, Gaopan Huang, Min Zhou, Chuan Fu, Honglin Qiao, Yan He
- Abstract summary: We propose an end-to-end deep explainable learning approach that combines the advantage of deep model in noise handling and expert rule-based interpretability.
The proposed method is tested in an industry production system, showing comparable prediction accuracy, much higher generalization stability and better interpretability.
- Score: 4.369058206183195
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Learning an explainable classifier often results in low accuracy model or
ends up with a huge rule set, while learning a deep model is usually more
capable of handling noisy data at scale, but with the cost of hard to explain
the result and weak at generalization. To mitigate this gap, we propose an
end-to-end deep explainable learning approach that combines the advantage of
deep model in noise handling and expert rule-based interpretability.
Specifically, we propose to learn a deep data assessing model which models the
data as a graph to represent the correlations among different observations,
whose output will be used to extract key data features. The key features are
then fed into a rule network constructed following predefined noisy expert
rules with trainable parameters. As these models are correlated, we propose an
end-to-end training framework, utilizing the rule classification loss to
optimize the rule learning model and data assessing model at the same time. As
the rule-based computation is none-differentiable, we propose a gradient
linking search module to carry the gradient information from the rule learning
model to the data assessing model. The proposed method is tested in an industry
production system, showing comparable prediction accuracy, much higher
generalization stability and better interpretability when compared with a
decent deep ensemble baseline, and shows much better fitting power than pure
rule-based approach.
Related papers
- Surprisal Driven $k$-NN for Robust and Interpretable Nonparametric
Learning [1.4293924404819704]
We shed new light on the traditional nearest neighbors algorithm from the perspective of information theory.
We propose a robust and interpretable framework for tasks such as classification, regression, density estimation, and anomaly detection using a single model.
Our work showcases the architecture's versatility by achieving state-of-the-art results in classification and anomaly detection.
arXiv Detail & Related papers (2023-11-17T00:35:38Z) - Pedagogical Rule Extraction for Learning Interpretable Models [0.0]
We propose a framework dubbed PRELIM to learn better rules from small data.
It augments data using statistical models and employs it to learn a rulebased model.
In our experiments, we identified PRELIM configurations that outperform state-of-the-art.
arXiv Detail & Related papers (2021-12-25T20:54:53Z) - General Greedy De-bias Learning [163.65789778416172]
We propose a General Greedy De-bias learning framework (GGD), which greedily trains the biased models and the base model like gradient descent in functional space.
GGD can learn a more robust base model under the settings of both task-specific biased models with prior knowledge and self-ensemble biased model without prior knowledge.
arXiv Detail & Related papers (2021-12-20T14:47:32Z) - On-Policy Model Errors in Reinforcement Learning [9.507323314334572]
We present a novel method that combines real world data and a learned model in order to get the best of both worlds.
The core idea is to exploit the real world data for on-policy predictions and use the learned model only to generalize to different actions.
We show that our method can drastically improve existing model-based approaches without introducing additional tuning parameters.
arXiv Detail & Related papers (2021-10-15T10:15:53Z) - Evaluating State-of-the-Art Classification Models Against Bayes
Optimality [106.50867011164584]
We show that we can compute the exact Bayes error of generative models learned using normalizing flows.
We use our approach to conduct a thorough investigation of state-of-the-art classification models.
arXiv Detail & Related papers (2021-06-07T06:21:20Z) - Churn Reduction via Distillation [54.5952282395487]
We show an equivalence between training with distillation using the base model as the teacher and training with an explicit constraint on the predictive churn.
We then show that distillation performs strongly for low churn training against a number of recent baselines.
arXiv Detail & Related papers (2021-06-04T18:03:31Z) - Distilling Interpretable Models into Human-Readable Code [71.11328360614479]
Human-readability is an important and desirable standard for machine-learned model interpretability.
We propose to train interpretable models using conventional methods, and then distill them into concise, human-readable code.
We describe a piecewise-linear curve-fitting algorithm that produces high-quality results efficiently and reliably across a broad range of use cases.
arXiv Detail & Related papers (2021-01-21T01:46:36Z) - S^3-Rec: Self-Supervised Learning for Sequential Recommendation with
Mutual Information Maximization [104.87483578308526]
We propose the model S3-Rec, which stands for Self-Supervised learning for Sequential Recommendation.
For our task, we devise four auxiliary self-supervised objectives to learn the correlations among attribute, item, subsequence, and sequence.
Extensive experiments conducted on six real-world datasets demonstrate the superiority of our proposed method over existing state-of-the-art methods.
arXiv Detail & Related papers (2020-08-18T11:44:10Z) - Model-Augmented Actor-Critic: Backpropagating through Paths [81.86992776864729]
Current model-based reinforcement learning approaches use the model simply as a learned black-box simulator.
We show how to make more effective use of the model by exploiting its differentiability.
arXiv Detail & Related papers (2020-05-16T19:18:10Z) - Metafeatures-based Rule-Extraction for Classifiers on Behavioral and
Textual Data [0.0]
Rule-extraction techniques have been proposed to combine the desired predictive accuracy of complex "black-box" models with global explainability.
We develop and test a rule-extraction methodology based on higher-level, less-sparse metafeatures.
A key finding of our analysis is that metafeatures-based explanations are better at mimicking the behavior of the black-box prediction model.
arXiv Detail & Related papers (2020-03-10T15:08:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.