FOLD-TR: A Scalable and Efficient Inductive Learning Algorithm for
Learning To Rank
- URL: http://arxiv.org/abs/2206.07295v1
- Date: Wed, 15 Jun 2022 04:46:49 GMT
- Title: FOLD-TR: A Scalable and Efficient Inductive Learning Algorithm for
Learning To Rank
- Authors: Huaduo Wang and Gopal Gupta
- Abstract summary: FOLD-R++ is a new inductive learning algorithm for binary classification tasks.
We present a customized FOLD-R++ algorithm with the ranking framework, called FOLD-TR.
- Score: 3.1981440103815717
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: FOLD-R++ is a new inductive learning algorithm for binary classification
tasks. It generates an (explainable) normal logic program for mixed type
(numerical and categorical) data. We present a customized FOLD-R++ algorithm
with the ranking framework, called FOLD-TR, that aims to rank new items
following the ranking pattern in the training data. Like FOLD-R++, the FOLD-TR
algorithm is able to handle mixed-type data directly and provide native
justification to explain the comparison between a pair of items.
Related papers
- Contextual Dual Learning Algorithm with Listwise Distillation for Unbiased Learning to Rank [26.69630281310365]
Unbiased Learning to Rank (ULTR) aims to leverage biased implicit user feedback (e.g., click) to optimize an unbiased ranking model.
We propose a Contextual Dual Learning Algorithm with Listwise Distillation (CDLA-LD) to address both position bias and contextual bias.
arXiv Detail & Related papers (2024-08-19T09:13:52Z) - RankDNN: Learning to Rank for Few-shot Learning [70.49494297554537]
This paper introduces a new few-shot learning pipeline that casts relevance ranking for image retrieval as binary ranking relation classification.
It provides a new perspective on few-shot learning and is complementary to state-of-the-art methods.
arXiv Detail & Related papers (2022-11-28T13:59:31Z) - FOLD-SE: Scalable Explainable AI [3.1981440103815717]
We present an improvement over the FOLD-R++ algorithm, termed FOLD-SE, that provides scalable explainability (SE)
The number of learned rules and learned literals stay small and, hence, understandable by human beings, while maintaining good performance in classification.
arXiv Detail & Related papers (2022-08-16T19:15:11Z) - FOLD-RM: A Scalable and Efficient Inductive Learning Algorithm for
Multi-Category Classification of Mixed Data [3.1981440103815717]
FOLD-RM is an automated inductive learning algorithm for learning default rules for mixed (numerical and categorical) data.
It generates an (explainable) answer set programming (ASP) rule set for multi-category classification tasks.
arXiv Detail & Related papers (2022-02-14T18:07:54Z) - GNNRank: Learning Global Rankings from Pairwise Comparisons via Directed
Graph Neural Networks [68.61934077627085]
We introduce GNNRank, a modeling framework compatible with any GNN capable of learning digraph embeddings.
We show that our methods attain competitive and often superior performance compared with existing approaches.
arXiv Detail & Related papers (2022-02-01T04:19:50Z) - Reinforcement Learning Based Query Vertex Ordering Model for Subgraph
Matching [58.39970828272366]
Subgraph matching algorithms enumerate all is embeddings of a query graph in a data graph G.
matching order plays a critical role in time efficiency of these backtracking based subgraph matching algorithms.
In this paper, for the first time we apply the Reinforcement Learning (RL) and Graph Neural Networks (GNNs) techniques to generate the high-quality matching order for subgraph matching algorithms.
arXiv Detail & Related papers (2022-01-25T00:10:03Z) - FOLD-R++: A Toolset for Automated Inductive Learning of Default Theories
from Mixed Data [2.741266294612776]
FOLD-R is an automated inductive learning algorithm for learning default rules with exceptions for mixed (numerical and categorical) data.
We present an improved FOLD-R algorithm, called FOLD-R++, that significantly increases the efficiency and scalability of FOLD-R.
arXiv Detail & Related papers (2021-10-15T03:55:13Z) - A Clustering and Demotion Based Algorithm for Inductive Learning of
Default Theories [4.640835690336653]
We present a clustering- and demotion-based algorithm called Kmeans-FOLD to induce nonmonotonic logic programs from positive and negative examples.
Our algorithm generates a more concise logic program compared to the FOLD algorithm.
Experiments on the UCI dataset show that a combination of K-Means clustering and our demotion strategy produces significant improvement for datasets with more than one cluster of positive examples.
arXiv Detail & Related papers (2021-09-26T14:50:18Z) - Rank-R FNN: A Tensor-Based Learning Model for High-Order Data
Classification [69.26747803963907]
Rank-R Feedforward Neural Network (FNN) is a tensor-based nonlinear learning model that imposes Canonical/Polyadic decomposition on its parameters.
First, it handles inputs as multilinear arrays, bypassing the need for vectorization, and can thus fully exploit the structural information along every data dimension.
We establish the universal approximation and learnability properties of Rank-R FNN, and we validate its performance on real-world hyperspectral datasets.
arXiv Detail & Related papers (2021-04-11T16:37:32Z) - Train your classifier first: Cascade Neural Networks Training from upper
layers to lower layers [54.47911829539919]
We develop a novel top-down training method which can be viewed as an algorithm for searching for high-quality classifiers.
We tested this method on automatic speech recognition (ASR) tasks and language modelling tasks.
The proposed method consistently improves recurrent neural network ASR models on Wall Street Journal, self-attention ASR models on Switchboard, and AWD-LSTM language models on WikiText-2.
arXiv Detail & Related papers (2021-02-09T08:19:49Z) - Interpretable Learning-to-Rank with Generalized Additive Models [78.42800966500374]
Interpretability of learning-to-rank models is a crucial yet relatively under-examined research area.
Recent progress on interpretable ranking models largely focuses on generating post-hoc explanations for existing black-box ranking models.
We lay the groundwork for intrinsically interpretable learning-to-rank by introducing generalized additive models (GAMs) into ranking tasks.
arXiv Detail & Related papers (2020-05-06T01:51:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.