Multi-Objective Personalized Product Retrieval in Taobao Search
- URL: http://arxiv.org/abs/2210.04170v1
- Date: Sun, 9 Oct 2022 05:18:42 GMT
- Title: Multi-Objective Personalized Product Retrieval in Taobao Search
- Authors: Yukun Zheng, Jiang Bian, Guanghao Meng, Chao Zhang, Honggang Wang,
Zhixuan Zhang, Sen Li, Tao Zhuang, Qingwen Liu, and Xiaoyi Zeng
- Abstract summary: We propose a novel Multi-Objective Personalized Product Retrieval (MOPPR) model with four hierarchical optimization objectives: relevance, exposure, click and purchase.
MOPPR achieves 0.96% transaction and 1.29% GMV improvements in a 28-day online A/B test.
Since the Double-11 shopping festival of 2021, MOPPR has been fully deployed in mobile Taobao search, replacing the previous MGDSPR.
- Score: 27.994166796745496
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In large-scale e-commerce platforms like Taobao, it is a big challenge to
retrieve products that satisfy users from billions of candidates. This has been
a common concern of academia and industry. Recently, plenty of works in this
domain have achieved significant improvements by enhancing embedding-based
retrieval (EBR) methods, including the Multi-Grained Deep Semantic Product
Retrieval (MGDSPR) model [16] in Taobao search engine. However, we find that
MGDSPR still has problems of poor relevance and weak personalization compared
to other retrieval methods in our online system, such as lexical matching and
collaborative filtering. These problems promote us to further strengthen the
capabilities of our EBR model in both relevance estimation and personalized
retrieval. In this paper, we propose a novel Multi-Objective Personalized
Product Retrieval (MOPPR) model with four hierarchical optimization objectives:
relevance, exposure, click and purchase. We construct entire-space
multi-positive samples to train MOPPR, rather than the single-positive samples
for existing EBR models.We adopt a modified softmax loss for optimizing
multiple objectives. Results of extensive offline and online experiments show
that MOPPR outperforms the baseline MGDSPR on evaluation metrics of relevance
estimation and personalized retrieval. MOPPR achieves 0.96% transaction and
1.29% GMV improvements in a 28-day online A/B test. Since the Double-11
shopping festival of 2021, MOPPR has been fully deployed in mobile Taobao
search, replacing the previous MGDSPR. Finally, we discuss several advanced
topics of our deeper explorations on multi-objective retrieval and ranking to
contribute to the community.
Related papers
- Learning to Rank for Multiple Retrieval-Augmented Models through Iterative Utility Maximization [21.115495457454365]
This paper investigates the design of a unified search engine to serve multiple retrieval-augmented generation (RAG) agents.
We introduce an iterative approach where the search engine generates retrieval results for these RAG agents and gathers feedback on the quality of the retrieved documents during an offline phase.
We adapt this approach to an online setting, allowing the search engine to refine its behavior based on real-time individual agents feedback.
arXiv Detail & Related papers (2024-10-13T17:53:50Z) - MRSE: An Efficient Multi-modality Retrieval System for Large Scale E-commerce [42.3177388371158]
Current Embedding-based Retrieval Systems embed queries and items into a shared low-dimensional space.
We propose MRSE, a Multi-modality Retrieval System that integrates text, item images, and user preferences.
MRSE achieves an 18.9% improvement in offline relevance and a 3.7% gain in online core metrics compared to Shopee's state-of-the-art uni-modality system.
arXiv Detail & Related papers (2024-08-27T11:21:19Z) - MIND: Multimodal Shopping Intention Distillation from Large Vision-language Models for E-commerce Purchase Understanding [67.26334044239161]
MIND is a framework that infers purchase intentions from multimodal product metadata and prioritizes human-centric ones.
Using Amazon Review data, we create a multimodal intention knowledge base, which contains 1,264,441 million intentions.
Our obtained intentions significantly enhance large language models in two intention comprehension tasks.
arXiv Detail & Related papers (2024-06-15T17:56:09Z) - Large Language Models for Relevance Judgment in Product Search [48.56992980315751]
High relevance of retrieved and re-ranked items to the search query is the cornerstone of successful product search.
We present an array of techniques for leveraging Large Language Models (LLMs) for automating the relevance judgment of query-item pairs (QIPs) at scale.
Our findings have immediate implications for the growing field of relevance judgment automation in product search.
arXiv Detail & Related papers (2024-06-01T00:52:41Z) - Monte Carlo Tree Search Boosts Reasoning via Iterative Preference Learning [55.96599486604344]
We introduce an approach aimed at enhancing the reasoning capabilities of Large Language Models (LLMs) through an iterative preference learning process.
We use Monte Carlo Tree Search (MCTS) to iteratively collect preference data, utilizing its look-ahead ability to break down instance-level rewards into more granular step-level signals.
The proposed algorithm employs Direct Preference Optimization (DPO) to update the LLM policy using this newly generated step-level preference data.
arXiv Detail & Related papers (2024-05-01T11:10:24Z) - Que2Engage: Embedding-based Retrieval for Relevant and Engaging Products
at Facebook Marketplace [15.054431410052851]
We present Que2Engage, a search EBR system built towards bridging the gap between retrieval and ranking for end-to-end optimizations.
We show the effectiveness of our approach via a multitask evaluation framework and thorough baseline comparisons and ablation studies.
arXiv Detail & Related papers (2023-02-21T23:10:16Z) - Multi-Task Fusion via Reinforcement Learning for Long-Term User
Satisfaction in Recommender Systems [3.4394890850129007]
We propose a Batch Reinforcement Learning based Multi-Task Fusion framework (BatchRL-MTF)
We learn an optimal recommendation policy from the fixed batch data offline for long-term user satisfaction.
With a comprehensive investigation on user behaviors, we model the user satisfaction reward with subtles from two aspects of user stickiness and user activeness.
arXiv Detail & Related papers (2022-08-09T06:35:05Z) - Entity-Graph Enhanced Cross-Modal Pretraining for Instance-level Product
Retrieval [152.3504607706575]
This research aims to conduct weakly-supervised multi-modal instance-level product retrieval for fine-grained product categories.
We first contribute the Product1M datasets, and define two real practical instance-level retrieval tasks.
We exploit to train a more effective cross-modal model which is adaptively capable of incorporating key concept information from the multi-modal data.
arXiv Detail & Related papers (2022-06-17T15:40:45Z) - ItemSage: Learning Product Embeddings for Shopping Recommendations at
Pinterest [60.841761065439414]
At Pinterest, we build a single set of product embeddings called ItemSage to provide relevant recommendations in all shopping use cases.
This approach has led to significant improvements in engagement and conversion metrics, while reducing both infrastructure and maintenance cost.
arXiv Detail & Related papers (2022-05-24T02:28:58Z) - Product1M: Towards Weakly Supervised Instance-Level Product Retrieval
via Cross-modal Pretraining [108.86502855439774]
We investigate a more realistic setting that aims to perform weakly-supervised multi-modal instance-level product retrieval.
We contribute Product1M, one of the largest multi-modal cosmetic datasets for real-world instance-level retrieval.
We propose a novel model named Cross-modal contrAstive Product Transformer for instance-level prodUct REtrieval (CAPTURE)
arXiv Detail & Related papers (2021-07-30T12:11:24Z) - Generator and Critic: A Deep Reinforcement Learning Approach for Slate
Re-ranking in E-commerce [17.712394984304336]
We present a novel Generator and Critic slate re-ranking approach, where the Critic evaluates the slate and the Generator ranks the items by the reinforcement learning approach.
For the Generator, to tackle the problem of large action space, we propose a new exploration reinforcement learning algorithm, called PPO-Exploration.
arXiv Detail & Related papers (2020-05-25T16:24:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.