Rank-LIME: Local Model-Agnostic Feature Attribution for Learning to Rank
- URL: http://arxiv.org/abs/2212.12722v1
- Date: Sat, 24 Dec 2022 12:14:32 GMT
- Title: Rank-LIME: Local Model-Agnostic Feature Attribution for Learning to Rank
- Authors: Tanya Chowdhury, Razieh Rahimi, James Allan
- Abstract summary: Rank-LIME is a model-agnostic, local, post-hoc linear feature attribution method for the task of learning to rank.
We employ novel correlation-based perturbations, differentiable ranking loss functions and introduce new metrics to evaluate ranking based additive feature attribution models.
- Score: 16.780058676633914
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Understanding why a model makes certain predictions is crucial when adapting
it for real world decision making. LIME is a popular model-agnostic feature
attribution method for the tasks of classification and regression. However, the
task of learning to rank in information retrieval is more complex in comparison
with either classification or regression. In this work, we extend LIME to
propose Rank-LIME, a model-agnostic, local, post-hoc linear feature attribution
method for the task of learning to rank that generates explanations for ranked
lists.
We employ novel correlation-based perturbations, differentiable ranking loss
functions and introduce new metrics to evaluate ranking based additive feature
attribution models. We compare Rank-LIME with a variety of competing systems,
with models trained on the MS MARCO datasets and observe that Rank-LIME
outperforms existing explanation algorithms in terms of Model Fidelity and
Explain-NDCG. With this we propose one of the first algorithms to generate
additive feature attributions for explaining ranked lists.
Related papers
- Ranking Unraveled: Recipes for LLM Rankings in Head-to-Head AI Combat [7.8905223445925055]
Pairwise ranking has emerged as a new method for evaluating human preferences for large language models (LLM)
We explore the effectiveness of ranking systems for head-to-head comparisons of LLMs.
Our analysis uncovers key insights into the factors that affect ranking accuracy and efficiency.
arXiv Detail & Related papers (2024-11-19T20:16:26Z) - RankUp: Boosting Semi-Supervised Regression with an Auxiliary Ranking Classifier [21.37308028303897]
State-of-the-art (SOTA) semi-supervised learning techniques, such as FixMatch and it's variants, have demonstrated impressive performance in classification tasks.
We present RankUp, a simple yet effective approach that adapts existing semi-supervised classification techniques to enhance the performance of regression tasks.
arXiv Detail & Related papers (2024-10-29T15:25:21Z) - RankSHAP: Shapley Value Based Feature Attributions for Learning to Rank [28.438428292619577]
We adopt an axiomatic game-theoretic approach, popular in the feature attribution community, to identify a set of fundamental axioms that every ranking-based feature attribution method should satisfy.
We then introduce Rank-SHAP, extending classical Shapley values to ranking.
We also perform an axiomatic analysis of existing rank attribution algorithms to determine their compliance with our proposed axioms.
arXiv Detail & Related papers (2024-05-03T04:43:24Z) - Evaluating Generative Language Models in Information Extraction as Subjective Question Correction [49.729908337372436]
We propose a new evaluation method, SQC-Score.
Inspired by the principles in subjective question correction, we propose a new evaluation method, SQC-Score.
Results on three information extraction tasks show that SQC-Score is more preferred by human annotators than the baseline metrics.
arXiv Detail & Related papers (2024-04-04T15:36:53Z) - Learning List-Level Domain-Invariant Representations for Ranking [59.3544317373004]
We propose list-level alignment -- learning domain-invariant representations at the higher level of lists.
The benefits are twofold: it leads to the first domain adaptation generalization bound for ranking, in turn providing theoretical support for the proposed method.
arXiv Detail & Related papers (2022-12-21T04:49:55Z) - Self-Supervised Class Incremental Learning [51.62542103481908]
Existing Class Incremental Learning (CIL) methods are based on a supervised classification framework sensitive to data labels.
When updating them based on the new class data, they suffer from catastrophic forgetting: the model cannot discern old class data clearly from the new.
In this paper, we explore the performance of Self-Supervised representation learning in Class Incremental Learning (SSCIL) for the first time.
arXiv Detail & Related papers (2021-11-18T06:58:19Z) - Search Methods for Sufficient, Socially-Aligned Feature Importance
Explanations with In-Distribution Counterfactuals [72.00815192668193]
Feature importance (FI) estimates are a popular form of explanation, and they are commonly created and evaluated by computing the change in model confidence caused by removing certain input features at test time.
We study several under-explored dimensions of FI-based explanations, providing conceptual and empirical improvements for this form of explanation.
arXiv Detail & Related papers (2021-06-01T20:36:48Z) - PiRank: Learning To Rank via Differentiable Sorting [85.28916333414145]
We propose PiRank, a new class of differentiable surrogates for ranking.
We show that PiRank exactly recovers the desired metrics in the limit of zero temperature.
arXiv Detail & Related papers (2020-12-12T05:07:36Z) - Interpretable Learning-to-Rank with Generalized Additive Models [78.42800966500374]
Interpretability of learning-to-rank models is a crucial yet relatively under-examined research area.
Recent progress on interpretable ranking models largely focuses on generating post-hoc explanations for existing black-box ranking models.
We lay the groundwork for intrinsically interpretable learning-to-rank by introducing generalized additive models (GAMs) into ranking tasks.
arXiv Detail & Related papers (2020-05-06T01:51:30Z) - Valid Explanations for Learning to Rank Models [5.320400771224103]
We propose a model agnostic local explanation method that seeks to identify a small subset of input features as explanation to a ranking decision.
We introduce new notions of validity and completeness of explanations specifically for rankings, based on the presence or absence of selected features.
arXiv Detail & Related papers (2020-04-29T06:21:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.