GUIM -- General User and Item Embedding with Mixture of Representation
in E-commerce
- URL: http://arxiv.org/abs/2207.00750v1
- Date: Sat, 2 Jul 2022 06:27:54 GMT
- Title: GUIM -- General User and Item Embedding with Mixture of Representation
in E-commerce
- Authors: Chao Yang, Ru He, Fangquan Lin, Suoyuan Song, Jingqiao Zhang, Cheng
Yang
- Abstract summary: Our goal is to build general representation (embedding) for each user and each product item across Alibaba's businesses.
Inspired by the BERT model in natural language processing (NLP) domain, we propose a GUIM (General User Item embedding with Mixture of representation) model.
We utilize mixture of representation (MoR) as a novel representation form to model the diverse interests of each user.
- Score: 13.142842265419262
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Our goal is to build general representation (embedding) for each user and
each product item across Alibaba's businesses, including Taobao and Tmall which
are among the world's biggest e-commerce websites. The representation of users
and items has been playing a critical role in various downstream applications,
including recommendation system, search, marketing, demand forecasting and so
on. Inspired from the BERT model in natural language processing (NLP) domain,
we propose a GUIM (General User Item embedding with Mixture of representation)
model to achieve the goal with massive, structured, multi-modal data including
the interactions among hundreds of millions of users and items. We utilize
mixture of representation (MoR) as a novel representation form to model the
diverse interests of each user. In addition, we use the InfoNCE from
contrastive learning to avoid intractable computational costs due to the
numerous size of item (token) vocabulary. Finally, we propose a set of
representative downstream tasks to serve as a standard benchmark to evaluate
the quality of the learned user and/or item embeddings, analogous to the GLUE
benchmark in NLP domain. Our experimental results in these downstream tasks
clearly show the comparative value of embeddings learned from our GUIM model.
Related papers
- Adaptive Learning on User Segmentation: Universal to Specific Representation via Bipartite Neural Interaction [15.302921887305283]
We propose a novel learning framework that can first learn general universal user representation through information bottleneck.
Then, merge and learn a segmentation-specific or a task-specific representation through neural interaction.
Our proposed method is evaluated in two open-source benchmarks, two offline business datasets, and deployed on two online marketing applications to predict users' CVR.
arXiv Detail & Related papers (2024-09-23T12:02:23Z) - EmbSum: Leveraging the Summarization Capabilities of Large Language Models for Content-Based Recommendations [38.44534579040017]
We introduce EmbSum, a framework that enables offline pre-computations of users and candidate items.
The model's ability to generate summaries of user interests serves as a valuable by-product, enhancing its usefulness for personalized content recommendations.
arXiv Detail & Related papers (2024-05-19T04:31:54Z) - MMGRec: Multimodal Generative Recommendation with Transformer Model [81.61896141495144]
MMGRec aims to introduce a generative paradigm into multimodal recommendation.
We first devise a hierarchical quantization method Graph CF-RQVAE to assign Rec-ID for each item from its multimodal information.
We then train a Transformer-based recommender to generate the Rec-IDs of user-preferred items based on historical interaction sequences.
arXiv Detail & Related papers (2024-04-25T12:11:27Z) - Scaling User Modeling: Large-scale Online User Representations for Ads Personalization in Meta [25.100342052990793]
Scaling User Modeling (SUM) is a framework widely deployed in Meta's ads ranking system.
To date, SUM has been launched to hundreds of ads ranking models in Meta, processing hundreds of billions of user requests daily.
arXiv Detail & Related papers (2023-11-16T03:47:48Z) - Pre-trained Neural Recommenders: A Transferable Zero-Shot Framework for
Recommendation Systems [5.597511654202054]
We show how to learn universal (i.e., supporting zero-shot adaptation without user or item auxiliary information) representations for nodes and edges from the bipartite user-item interaction graph.
arXiv Detail & Related papers (2023-09-03T14:18:31Z) - Recommender AI Agent: Integrating Large Language Models for Interactive
Recommendations [53.76682562935373]
We introduce an efficient framework called textbfInteRecAgent, which employs LLMs as the brain and recommender models as tools.
InteRecAgent achieves satisfying performance as a conversational recommender system, outperforming general-purpose LLMs.
arXiv Detail & Related papers (2023-08-31T07:36:44Z) - The Minority Matters: A Diversity-Promoting Collaborative Metric
Learning Algorithm [154.47590401735323]
Collaborative Metric Learning (CML) has recently emerged as a popular method in recommendation systems.
This paper focuses on a challenging scenario where a user has multiple categories of interests.
We propose a novel method called textitDiversity-Promoting Collaborative Metric Learning (DPCML)
arXiv Detail & Related papers (2022-09-30T08:02:18Z) - Entity-Graph Enhanced Cross-Modal Pretraining for Instance-level Product
Retrieval [152.3504607706575]
This research aims to conduct weakly-supervised multi-modal instance-level product retrieval for fine-grained product categories.
We first contribute the Product1M datasets, and define two real practical instance-level retrieval tasks.
We exploit to train a more effective cross-modal model which is adaptively capable of incorporating key concept information from the multi-modal data.
arXiv Detail & Related papers (2022-06-17T15:40:45Z) - ItemSage: Learning Product Embeddings for Shopping Recommendations at
Pinterest [60.841761065439414]
At Pinterest, we build a single set of product embeddings called ItemSage to provide relevant recommendations in all shopping use cases.
This approach has led to significant improvements in engagement and conversion metrics, while reducing both infrastructure and maintenance cost.
arXiv Detail & Related papers (2022-05-24T02:28:58Z) - Pre-training Graph Transformer with Multimodal Side Information for
Recommendation [82.4194024706817]
We propose a pre-training strategy to learn item representations by considering both item side information and their relationships.
We develop a novel sampling algorithm named MCNSampling to select contextual neighbors for each item.
The proposed Pre-trained Multimodal Graph Transformer (PMGT) learns item representations with two objectives: 1) graph structure reconstruction, and 2) masked node feature reconstruction.
arXiv Detail & Related papers (2020-10-23T10:30:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.