piRank: A Probabilistic Intent Based Ranking Framework for Facebook
Search
- URL: http://arxiv.org/abs/2203.14363v1
- Date: Sun, 27 Mar 2022 18:12:56 GMT
- Title: piRank: A Probabilistic Intent Based Ranking Framework for Facebook
Search
- Authors: Zhen Liao
- Abstract summary: We propose a probabilistic intent based ranking framework (short for piRank) to address various ranking issues for different query intents.
We conducted extensive experiments and studies on top of Facebook search engine system and validated the effectiveness of this new ranking architecture.
- Score: 0.07614628596146598
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While numerous studies have been conducted in the literature exploring
different types of machine learning approaches for search ranking, most of them
are focused on specific pre-defined problems but only a few of them have
studied the ranking framework which can be applied in a commercial search
engine in a scalable way. In the meantime, existing ranking models are often
optimized for normalized discounted cumulative gains (NDCG) or online
click-through rate (CTR), and both types of machine learning models are built
based on the assumption that high-quality training data can be easily obtained
and well applied to unseen cases. In practice at Facebook search, we observed
that our training data for ML models have certain issues. First, tail query
intents are hardly covered in our human rating dataset. Second, search click
logs are often noisy and hard to clean up due to various reasons. To address
the above issues, in this paper, we propose a probabilistic intent based
ranking framework (short for piRank), which can: 1) provide a scalable
framework to address various ranking issues for different query intents in a
divide-and-conquer way; 2) improve system development agility including
iteration speed and system debuggability; 3) combine both machine learning and
empirical-based algorithmic methods in a systematic way. We conducted extensive
experiments and studies on top of Facebook search engine system and validated
the effectiveness of this new ranking architecture.
Related papers
- Meta Learning to Rank for Sparsely Supervised Queries [10.422527051110526]
In many real-world search and retrieval scenarios, supervisory signals may not be readily available or could be costly to obtain for some queries.
We propose a novel meta learning to rank framework which leverages fast learning and adaption capability of meta-learning.
The proposed method would yield significant advantages especially when new queries are of different characteristics with the training queries.
arXiv Detail & Related papers (2024-09-29T04:24:38Z) - Zero-shot Retrieval: Augmenting Pre-trained Models with Search Engines [83.65380507372483]
Large pre-trained models can dramatically reduce the amount of task-specific data required to solve a problem, but they often fail to capture domain-specific nuances out of the box.
This paper shows how to leverage recent advances in NLP and multi-modal learning to augment a pre-trained model with search engine retrieval.
arXiv Detail & Related papers (2023-11-29T05:33:28Z) - A novel evaluation methodology for supervised Feature Ranking algorithms [0.0]
This paper proposes a new evaluation methodology for Feature Rankers.
By making use of synthetic datasets, feature importance scores can be known beforehand, allowing more systematic evaluation.
To facilitate large-scale experimentation using the new methodology, a benchmarking framework was built in Python, called fseval.
arXiv Detail & Related papers (2022-07-09T12:00:36Z) - Meta Navigator: Search for a Good Adaptation Policy for Few-shot
Learning [113.05118113697111]
Few-shot learning aims to adapt knowledge learned from previous tasks to novel tasks with only a limited amount of labeled data.
Research literature on few-shot learning exhibits great diversity, while different algorithms often excel at different few-shot learning scenarios.
We present Meta Navigator, a framework that attempts to solve the limitation in few-shot learning by seeking a higher-level strategy.
arXiv Detail & Related papers (2021-09-13T07:20:01Z) - FEAR: A Simple Lightweight Method to Rank Architectures [14.017656480004955]
We propose a simple but powerful method which we call FEAR, for ranking architectures in any search space.
FEAR can cut down the search time by approximately 2.4X without losing accuracy.
We additionally empirically study very recently proposed zero-cost measures for ranking and find that they breakdown in ranking performance as training proceeds.
arXiv Detail & Related papers (2021-06-07T23:38:21Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - PiRank: Learning To Rank via Differentiable Sorting [85.28916333414145]
We propose PiRank, a new class of differentiable surrogates for ranking.
We show that PiRank exactly recovers the desired metrics in the limit of zero temperature.
arXiv Detail & Related papers (2020-12-12T05:07:36Z) - Analysis of Multivariate Scoring Functions for Automatic Unbiased
Learning to Rank [14.827143632277274]
AutoULTR algorithms that jointly learn user bias models (i.e., propensity models) with unbiased rankers have received a lot of attention due to their superior performance and low deployment cost in practice.
Recent advances in context-aware learning-to-rank models have shown that multivariate scoring functions, which read multiple documents together and predict their ranking scores jointly, are more powerful than uni-variate ranking functions in ranking tasks with human-annotated relevance labels.
Our experiments with synthetic clicks on two large-scale benchmark datasets show that AutoULTR models with permutation-invariant multivariate scoring functions significantly outperform
arXiv Detail & Related papers (2020-08-20T16:31:59Z) - Towards Model-Agnostic Post-Hoc Adjustment for Balancing Ranking
Fairness and Algorithm Utility [54.179859639868646]
Bipartite ranking aims to learn a scoring function that ranks positive individuals higher than negative ones from labeled data.
There have been rising concerns on whether the learned scoring function can cause systematic disparity across different protected groups.
We propose a model post-processing framework for balancing them in the bipartite ranking scenario.
arXiv Detail & Related papers (2020-06-15T10:08:39Z) - Deep-n-Cheap: An Automated Search Framework for Low Complexity Deep
Learning [3.479254848034425]
We present Deep-n-Cheap -- an open-source AutoML framework to search for deep learning models.
Our framework is targeted for deployment on both benchmark and custom datasets.
Deep-n-Cheap includes a user-customizable complexity penalty which trades off performance with training time or number of parameters.
arXiv Detail & Related papers (2020-03-27T13:00:21Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.