Sparse Attentive Memory Network for Click-through Rate Prediction with
Long Sequences
- URL: http://arxiv.org/abs/2208.04022v1
- Date: Mon, 8 Aug 2022 10:11:46 GMT
- Title: Sparse Attentive Memory Network for Click-through Rate Prediction with
Long Sequences
- Authors: Qianying Lin, Wen-Ji Zhou, Yanshi Wang, Qing Da, Qing-Guo Chen, Bing
Wang
- Abstract summary: We propose a Sparse Attentive Memory network for long sequential user behavior modeling.
SAM supports efficient training and real-time inference for user behavior sequences with lengths on the scale of thousands.
SAM is successfully deployed on one of the largest international E-commerce platforms.
- Score: 10.233015715433602
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sequential recommendation predicts users' next behaviors with their
historical interactions. Recommending with longer sequences improves
recommendation accuracy and increases the degree of personalization. As
sequences get longer, existing works have not yet addressed the following two
main challenges. Firstly, modeling long-range intra-sequence dependency is
difficult with increasing sequence lengths. Secondly, it requires efficient
memory and computational speeds. In this paper, we propose a Sparse Attentive
Memory (SAM) network for long sequential user behavior modeling. SAM supports
efficient training and real-time inference for user behavior sequences with
lengths on the scale of thousands. In SAM, we model the target item as the
query and the long sequence as the knowledge database, where the former
continuously elicits relevant information from the latter. SAM simultaneously
models target-sequence dependencies and long-range intra-sequence dependencies
with O(L) complexity and O(1) number of sequential updates, which can only be
achieved by the self-attention mechanism with O(L^2) complexity. Extensive
empirical results demonstrate that our proposed solution is effective not only
in long user behavior modeling but also on short sequences modeling.
Implemented on sequences of length 1000, SAM is successfully deployed on one of
the largest international E-commerce platforms. This inference time is within
30ms, with a substantial 7.30% click-through rate improvement for the online
A/B test. To the best of our knowledge, it is the first end-to-end long user
sequence modeling framework that models intra-sequence and target-sequence
dependencies with the aforementioned degree of efficiency and successfully
deployed on a large-scale real-time industrial recommender system.
Related papers
- Multi-granularity Interest Retrieval and Refinement Network for Long-Term User Behavior Modeling in CTR Prediction [68.90783662117936]
Click-through Rate (CTR) prediction is crucial for online personalization platforms.
Recent advancements have shown that modeling rich user behaviors can significantly improve the performance of CTR prediction.
We propose Multi-granularity Interest Retrieval and Refinement Network (MIRRN)
arXiv Detail & Related papers (2024-11-22T15:29:05Z) - Long-Sequence Recommendation Models Need Decoupled Embeddings [49.410906935283585]
We identify and characterize a neglected deficiency in existing long-sequence recommendation models.
A single set of embeddings struggles with learning both attention and representation, leading to interference between these two processes.
We propose the Decoupled Attention and Representation Embeddings (DARE) model, where two distinct embedding tables are learned separately to fully decouple attention and representation.
arXiv Detail & Related papers (2024-10-03T15:45:15Z) - TWIN V2: Scaling Ultra-Long User Behavior Sequence Modeling for Enhanced CTR Prediction at Kuaishou [28.809014888174932]
We introduce TWIN-V2, an enhancement of SIM, where a divide-and-conquer approach is applied to compress life-cycle behaviors and uncover more accurate and diverse user interests.
Under an efficient deployment framework, TWIN-V2 has been successfully deployed to the primary traffic that serves hundreds of millions of daily active users at Kuaishou.
arXiv Detail & Related papers (2024-07-23T10:00:45Z) - LongVQ: Long Sequence Modeling with Vector Quantization on Structured Memory [63.41820940103348]
Self-attention mechanism's computational cost limits its practicality for long sequences.
We propose a new method called LongVQ to compress the global abstraction as a length-fixed codebook.
LongVQ effectively maintains dynamic global and local patterns, which helps to complement the lack of long-range dependency issues.
arXiv Detail & Related papers (2024-04-17T08:26:34Z) - Learning Sequence Representations by Non-local Recurrent Neural Memory [61.65105481899744]
We propose a Non-local Recurrent Neural Memory (NRNM) for supervised sequence representation learning.
Our model is able to capture long-range dependencies and latent high-level features can be distilled by our model.
Our model compares favorably against other state-of-the-art methods specifically designed for each of these sequence applications.
arXiv Detail & Related papers (2022-07-20T07:26:15Z) - Sequential Search with Off-Policy Reinforcement Learning [48.88165680363482]
We propose a highly scalable hybrid learning model that consists of an RNN learning framework and an attention model.
As a novel optimization step, we fit multiple short user sequences in a single RNN pass within a training batch, by solving a greedy knapsack problem on the fly.
We also explore the use of off-policy reinforcement learning in multi-session personalized search ranking.
arXiv Detail & Related papers (2022-02-01T06:52:40Z) - Sequence Adaptation via Reinforcement Learning in Recommender Systems [8.909115457491522]
We propose the SAR model, which learns the sequential patterns and adjusts the sequence length of user-item interactions in a personalized manner.
In addition, we optimize a joint loss function to align the accuracy of the sequential recommendations with the expected cumulative rewards of the critic network.
Our experimental evaluation on four real-world datasets demonstrates the superiority of our proposed model over several baseline approaches.
arXiv Detail & Related papers (2021-07-31T13:56:46Z) - Dynamic Memory based Attention Network for Sequential Recommendation [79.5901228623551]
We propose a novel long sequential recommendation model called Dynamic Memory-based Attention Network (DMAN)
It segments the overall long behavior sequence into a series of sub-sequences, then trains the model and maintains a set of memory blocks to preserve long-term interests of users.
Based on the dynamic memory, the user's short-term and long-term interests can be explicitly extracted and combined for efficient joint recommendation.
arXiv Detail & Related papers (2021-02-18T11:08:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.