GUIM -- General User and Item Embedding with Mixture of Representation
in E-commerce
- URL: http://arxiv.org/abs/2207.00750v1
- Date: Sat, 2 Jul 2022 06:27:54 GMT
- Title: GUIM -- General User and Item Embedding with Mixture of Representation
in E-commerce
- Authors: Chao Yang, Ru He, Fangquan Lin, Suoyuan Song, Jingqiao Zhang, Cheng
Yang
- Abstract summary: Our goal is to build general representation (embedding) for each user and each product item across Alibaba's businesses.
Inspired by the BERT model in natural language processing (NLP) domain, we propose a GUIM (General User Item embedding with Mixture of representation) model.
We utilize mixture of representation (MoR) as a novel representation form to model the diverse interests of each user.
- Score: 13.142842265419262
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Our goal is to build general representation (embedding) for each user and
each product item across Alibaba's businesses, including Taobao and Tmall which
are among the world's biggest e-commerce websites. The representation of users
and items has been playing a critical role in various downstream applications,
including recommendation system, search, marketing, demand forecasting and so
on. Inspired from the BERT model in natural language processing (NLP) domain,
we propose a GUIM (General User Item embedding with Mixture of representation)
model to achieve the goal with massive, structured, multi-modal data including
the interactions among hundreds of millions of users and items. We utilize
mixture of representation (MoR) as a novel representation form to model the
diverse interests of each user. In addition, we use the InfoNCE from
contrastive learning to avoid intractable computational costs due to the
numerous size of item (token) vocabulary. Finally, we propose a set of
representative downstream tasks to serve as a standard benchmark to evaluate
the quality of the learned user and/or item embeddings, analogous to the GLUE
benchmark in NLP domain. Our experimental results in these downstream tasks
clearly show the comparative value of embeddings learned from our GUIM model.
Related papers
- Adaptive Learning on User Segmentation: Universal to Specific Representation via Bipartite Neural Interaction [15.302921887305283]
We propose a novel learning framework that can first learn general universal user representation through information bottleneck.
Then, merge and learn a segmentation-specific or a task-specific representation through neural interaction.
Our proposed method is evaluated in two open-source benchmarks, two offline business datasets, and deployed on two online marketing applications to predict users' CVR.
arXiv Detail & Related papers (2024-09-23T12:02:23Z) - EmbSum: Leveraging the Summarization Capabilities of Large Language Models for Content-Based Recommendations [38.44534579040017]
We introduce EmbSum, a framework that enables offline pre-computations of users and candidate items.
The model's ability to generate summaries of user interests serves as a valuable by-product, enhancing its usefulness for personalized content recommendations.
arXiv Detail & Related papers (2024-05-19T04:31:54Z) - Scaling User Modeling: Large-scale Online User Representations for Ads Personalization in Meta [25.100342052990793]
Scaling User Modeling (SUM) is a framework widely deployed in Meta's ads ranking system.
To date, SUM has been launched to hundreds of ads ranking models in Meta, processing hundreds of billions of user requests daily.
arXiv Detail & Related papers (2023-11-16T03:47:48Z) - MM-BigBench: Evaluating Multimodal Models on Multimodal Content
Comprehension Tasks [56.60050181186531]
We introduce MM-BigBench, which incorporates a diverse range of metrics to offer an extensive evaluation of the performance of various models and instructions.
Our paper evaluates a total of 20 language models (14 MLLMs) on 14 multimodal datasets spanning 6 tasks, with 10 instructions for each task, and derives novel insights.
arXiv Detail & Related papers (2023-10-13T11:57:04Z) - Pre-trained Neural Recommenders: A Transferable Zero-Shot Framework for
Recommendation Systems [5.597511654202054]
We show how to learn universal (i.e., supporting zero-shot adaptation without user or item auxiliary information) representations for nodes and edges from the bipartite user-item interaction graph.
arXiv Detail & Related papers (2023-09-03T14:18:31Z) - Recommender AI Agent: Integrating Large Language Models for Interactive
Recommendations [53.76682562935373]
We introduce an efficient framework called textbfInteRecAgent, which employs LLMs as the brain and recommender models as tools.
InteRecAgent achieves satisfying performance as a conversational recommender system, outperforming general-purpose LLMs.
arXiv Detail & Related papers (2023-08-31T07:36:44Z) - The Minority Matters: A Diversity-Promoting Collaborative Metric
Learning Algorithm [154.47590401735323]
Collaborative Metric Learning (CML) has recently emerged as a popular method in recommendation systems.
This paper focuses on a challenging scenario where a user has multiple categories of interests.
We propose a novel method called textitDiversity-Promoting Collaborative Metric Learning (DPCML)
arXiv Detail & Related papers (2022-09-30T08:02:18Z) - e-CLIP: Large-Scale Vision-Language Representation Learning in
E-commerce [9.46186546774799]
We propose a contrastive learning framework that aligns language and visual models using unlabeled raw product text and images.
We present techniques we used to train large-scale representation learning models and share solutions that address domain-specific challenges.
arXiv Detail & Related papers (2022-07-01T05:16:47Z) - Entity-Graph Enhanced Cross-Modal Pretraining for Instance-level Product
Retrieval [152.3504607706575]
This research aims to conduct weakly-supervised multi-modal instance-level product retrieval for fine-grained product categories.
We first contribute the Product1M datasets, and define two real practical instance-level retrieval tasks.
We exploit to train a more effective cross-modal model which is adaptively capable of incorporating key concept information from the multi-modal data.
arXiv Detail & Related papers (2022-06-17T15:40:45Z) - ItemSage: Learning Product Embeddings for Shopping Recommendations at
Pinterest [60.841761065439414]
At Pinterest, we build a single set of product embeddings called ItemSage to provide relevant recommendations in all shopping use cases.
This approach has led to significant improvements in engagement and conversion metrics, while reducing both infrastructure and maintenance cost.
arXiv Detail & Related papers (2022-05-24T02:28:58Z) - Pre-training Graph Transformer with Multimodal Side Information for
Recommendation [82.4194024706817]
We propose a pre-training strategy to learn item representations by considering both item side information and their relationships.
We develop a novel sampling algorithm named MCNSampling to select contextual neighbors for each item.
The proposed Pre-trained Multimodal Graph Transformer (PMGT) learns item representations with two objectives: 1) graph structure reconstruction, and 2) masked node feature reconstruction.
arXiv Detail & Related papers (2020-10-23T10:30:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.