Interpretable Learning-to-Rank with Generalized Additive Models
- URL: http://arxiv.org/abs/2005.02553v2
- Date: Thu, 14 May 2020 18:44:23 GMT
- Title: Interpretable Learning-to-Rank with Generalized Additive Models
- Authors: Honglei Zhuang, Xuanhui Wang, Michael Bendersky, Alexander Grushetsky,
Yonghui Wu, Petr Mitrichev, Ethan Sterling, Nathan Bell, Walker Ravina, Hai
Qian
- Abstract summary: Interpretability of learning-to-rank models is a crucial yet relatively under-examined research area.
Recent progress on interpretable ranking models largely focuses on generating post-hoc explanations for existing black-box ranking models.
We lay the groundwork for intrinsically interpretable learning-to-rank by introducing generalized additive models (GAMs) into ranking tasks.
- Score: 78.42800966500374
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Interpretability of learning-to-rank models is a crucial yet relatively
under-examined research area. Recent progress on interpretable ranking models
largely focuses on generating post-hoc explanations for existing black-box
ranking models, whereas the alternative option of building an intrinsically
interpretable ranking model with transparent and self-explainable structure
remains unexplored. Developing fully-understandable ranking models is necessary
in some scenarios (e.g., due to legal or policy constraints) where post-hoc
methods cannot provide sufficiently accurate explanations. In this paper, we
lay the groundwork for intrinsically interpretable learning-to-rank by
introducing generalized additive models (GAMs) into ranking tasks. Generalized
additive models (GAMs) are intrinsically interpretable machine learning models
and have been extensively studied on regression and classification tasks. We
study how to extend GAMs into ranking models which can handle both item-level
and list-level features and propose a novel formulation of ranking GAMs. To
instantiate ranking GAMs, we employ neural networks instead of traditional
splines or regression trees. We also show that our neural ranking GAMs can be
distilled into a set of simple and compact piece-wise linear functions that are
much more efficient to evaluate with little accuracy loss. We conduct
experiments on three data sets and show that our proposed neural ranking GAMs
can achieve significantly better performance than other traditional GAM
baselines while maintaining similar interpretability.
Related papers
- GAMformer: In-Context Learning for Generalized Additive Models [53.08263343627232]
We introduce GAMformer, the first method to leverage in-context learning to estimate shape functions of a GAM in a single forward pass.
Our experiments show that GAMformer performs on par with other leading GAMs across various classification benchmarks.
arXiv Detail & Related papers (2024-10-06T17:28:20Z) - Towards interpretable-by-design deep learning algorithms [11.154826546951414]
A proposed framework named I recasts the standard supervised classification problem into a function of similarity to a set of prototypes derived from the training data.
We show that one can turn such DL models into conceptually simpler, explainable-through-prototypes ones.
arXiv Detail & Related papers (2023-11-19T18:40:49Z) - Universal Domain Adaptation from Foundation Models: A Baseline Study [58.51162198585434]
We make empirical studies of state-of-the-art UniDA methods using foundation models.
We introduce textitCLIP distillation, a parameter-free method specifically designed to distill target knowledge from CLIP models.
Although simple, our method outperforms previous approaches in most benchmark tasks.
arXiv Detail & Related papers (2023-05-18T16:28:29Z) - Rank-LIME: Local Model-Agnostic Feature Attribution for Learning to Rank [16.780058676633914]
Rank-LIME is a model-agnostic, local, post-hoc linear feature attribution method for the task of learning to rank.
We employ novel correlation-based perturbations, differentiable ranking loss functions and introduce new metrics to evaluate ranking based additive feature attribution models.
arXiv Detail & Related papers (2022-12-24T12:14:32Z) - Rethinking Log Odds: Linear Probability Modelling and Expert Advice in
Interpretable Machine Learning [8.831954614241234]
We introduce a family of interpretable machine learning models, with two broad additions: Linearised Additive Models (LAMs) and SubscaleHedge.
LAMs replace the ubiquitous logistic link function in General Additive Models (GAMs); and SubscaleHedge is an expert advice algorithm for combining base models trained on subsets of features called subscales.
arXiv Detail & Related papers (2022-11-11T17:21:57Z) - Self-Supervised Class Incremental Learning [51.62542103481908]
Existing Class Incremental Learning (CIL) methods are based on a supervised classification framework sensitive to data labels.
When updating them based on the new class data, they suffer from catastrophic forgetting: the model cannot discern old class data clearly from the new.
In this paper, we explore the performance of Self-Supervised representation learning in Class Incremental Learning (SSCIL) for the first time.
arXiv Detail & Related papers (2021-11-18T06:58:19Z) - Scalable Rule-Based Representation Learning for Interpretable
Classification [12.736847587988853]
Rule-based Learner Representation (RRL) learns interpretable non-fuzzy rules for data representation and classification.
RRL can be easily adjusted to obtain a trade-off between classification accuracy and model complexity for different scenarios.
arXiv Detail & Related papers (2021-09-30T13:07:42Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - Document Ranking with a Pretrained Sequence-to-Sequence Model [56.44269917346376]
We show how a sequence-to-sequence model can be trained to generate relevance labels as "target words"
Our approach significantly outperforms an encoder-only model in a data-poor regime.
arXiv Detail & Related papers (2020-03-14T22:29:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.