Explaining Inference Queries with Bayesian Optimization
- URL: http://arxiv.org/abs/2102.05308v1
- Date: Wed, 10 Feb 2021 08:08:32 GMT
- Title: Explaining Inference Queries with Bayesian Optimization
- Authors: Brandon Lockhart, Jinglin Peng, Weiyuan Wu, Jiannan Wang, Eugene Wu
- Abstract summary: Inference query explanation seeks to explain unexpected aggregate query results on inference data.
An explanation may need to be derived from the source, training, or inference data in an ML pipeline.
We propose BOExplain, a novel framework for explaining inference queries using Bayesian optimization (BO)
- Score: 16.448164301763168
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Obtaining an explanation for an SQL query result can enrich the analysis
experience, reveal data errors, and provide deeper insight into the data.
Inference query explanation seeks to explain unexpected aggregate query results
on inference data; such queries are challenging to explain because an
explanation may need to be derived from the source, training, or inference data
in an ML pipeline. In this paper, we model an objective function as a black-box
function and propose BOExplain, a novel framework for explaining inference
queries using Bayesian optimization (BO). An explanation is a predicate
defining the input tuples that should be removed so that the query result of
interest is significantly affected. BO - a technique for finding the global
optimum of a black-box function - is used to find the best predicate. We
develop two new techniques (individual contribution encoding and warm start) to
handle categorical variables. We perform experiments showing that the
predicates found by BOExplain have a higher degree of explanation compared to
those found by the state-of-the-art query explanation engines. We also show
that BOExplain is effective at deriving explanations for inference queries from
source and training data on three real-world datasets.
Related papers
- CoTEVer: Chain of Thought Prompting Annotation Toolkit for Explanation
Verification [1.658938566492109]
Chain-of-thought (CoT) prompting enables large language models (LLMs) to solve complex reasoning tasks by generating an explanation before the final prediction.
Despite it's promising ability, a critical downside of CoT prompting is that the performance is greatly affected by the factuality of the generated explanation.
To improve the correctness of the explanations, fine-tuning language models with explanation data is needed.
CoTEVer is a tool-kit for annotating the factual correctness of generated explanations and collecting revision data of wrong explanations.
arXiv Detail & Related papers (2023-03-07T03:23:14Z) - Explanation Selection Using Unlabeled Data for Chain-of-Thought
Prompting [80.9896041501715]
Explanations that have not been "tuned" for a task, such as off-the-shelf explanations written by nonexperts, may lead to mediocre performance.
This paper tackles the problem of how to optimize explanation-infused prompts in a blackbox fashion.
arXiv Detail & Related papers (2023-02-09T18:02:34Z) - ExaRanker: Explanation-Augmented Neural Ranker [67.4894325619275]
In this work, we show that neural rankers also benefit from explanations.
We use LLMs such as GPT-3.5 to augment retrieval datasets with explanations.
Our model, dubbed ExaRanker, finetuned on a few thousand examples with synthetic explanations performs on par with models finetuned on 3x more examples without explanations.
arXiv Detail & Related papers (2023-01-25T11:03:04Z) - Complementary Explanations for Effective In-Context Learning [77.83124315634386]
Large language models (LLMs) have exhibited remarkable capabilities in learning from explanations in prompts.
This work aims to better understand the mechanisms by which explanations are used for in-context learning.
arXiv Detail & Related papers (2022-11-25T04:40:47Z) - Interpretable by Design: Learning Predictors by Composing Interpretable
Queries [8.054701719767293]
We argue that machine learning algorithms should be interpretable by design.
We minimize the expected number of queries needed for accurate prediction.
Experiments on vision and NLP tasks demonstrate the efficacy of our approach.
arXiv Detail & Related papers (2022-07-03T02:40:34Z) - Graph Enhanced BERT for Query Understanding [55.90334539898102]
query understanding plays a key role in exploring users' search intents and facilitating users to locate their most desired information.
In recent years, pre-trained language models (PLMs) have advanced various natural language processing tasks.
We propose a novel graph-enhanced pre-training framework, GE-BERT, which can leverage both query content and the query graph.
arXiv Detail & Related papers (2022-04-03T16:50:30Z) - Are Training Resources Insufficient? Predict First Then Explain! [54.184609286094044]
We argue that the predict-then-explain (PtE) architecture is a more efficient approach in terms of the modelling perspective.
We show that the PtE structure is the most data-efficient approach when explanation data are lacking.
arXiv Detail & Related papers (2021-08-29T07:04:50Z) - Search Methods for Sufficient, Socially-Aligned Feature Importance
Explanations with In-Distribution Counterfactuals [72.00815192668193]
Feature importance (FI) estimates are a popular form of explanation, and they are commonly created and evaluated by computing the change in model confidence caused by removing certain input features at test time.
We study several under-explored dimensions of FI-based explanations, providing conceptual and empirical improvements for this form of explanation.
arXiv Detail & Related papers (2021-06-01T20:36:48Z) - ExplanationLP: Abductive Reasoning for Explainable Science Question
Answering [4.726777092009554]
This paper frames question answering as an abductive reasoning problem.
We construct plausible explanations for each choice and then selecting the candidate with the best explanation as the final answer.
Our system, ExplanationLP, elicits explanations by constructing a weighted graph of relevant facts for each candidate answer.
arXiv Detail & Related papers (2020-10-25T14:49:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.