InfoBehavior: Self-supervised Representation Learning for Ultra-long
Behavior Sequence via Hierarchical Grouping
- URL: http://arxiv.org/abs/2106.06905v1
- Date: Sun, 13 Jun 2021 03:45:45 GMT
- Title: InfoBehavior: Self-supervised Representation Learning for Ultra-long
Behavior Sequence via Hierarchical Grouping
- Authors: Runshi Liu, Pengda Qin, Yuhong Li, Weigao Wen, Dong Li, Kefeng Deng,
Qiang Wu
- Abstract summary: E-commerce companies have to face abnormal sellers who sell potentially-risky products.
Traditional feature extraction techniques heavily depend on domain experts and adapt poorly to new tasks.
We propose a self-supervised method InfoBehavior to automatically extract meaningful representations from ultra-long raw behavior sequences.
- Score: 14.80873165144865
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: E-commerce companies have to face abnormal sellers who sell potentially-risky
products. Typically, the risk can be identified by jointly considering product
content (e.g., title and image) and seller behavior. This work focuses on
behavior feature extraction as behavior sequences can provide valuable clues
for the risk discovery by reflecting the sellers' operation habits. Traditional
feature extraction techniques heavily depend on domain experts and adapt poorly
to new tasks. In this paper, we propose a self-supervised method InfoBehavior
to automatically extract meaningful representations from ultra-long raw
behavior sequences instead of the costly feature selection procedure.
InfoBehavior utilizes Bidirectional Transformer as feature encoder due to its
excellent capability in modeling long-term dependency. However, it is
intractable for commodity GPUs because the time and memory required by
Transformer grow quadratically with the increase of sequence length. Thus, we
propose a hierarchical grouping strategy to aggregate ultra-long raw behavior
sequences to length-processable high-level embedding sequences. Moreover, we
introduce two types of pretext tasks. Sequence-related pretext task defines a
contrastive-based training objective to correctly select the masked-out
coarse-grained/fine-grained behavior sequences against other "distractor"
behavior sequences; Domain-related pretext task designs a classification
training objective to correctly predict the domain-specific statistical results
of anomalous behavior. We show that behavior representations from the
pre-trained InfoBehavior can be directly used or integrated with features from
other side information to support a wide range of downstream tasks.
Experimental results demonstrate that InfoBehavior significantly improves the
performance of Product Risk Management and Intellectual Property Protection.
Related papers
- Uniting contrastive and generative learning for event sequences models [51.547576949425604]
This study investigates the integration of two self-supervised learning techniques - instance-wise contrastive learning and a generative approach based on restoring masked events in latent space.
Experiments conducted on several public datasets, focusing on sequence classification and next-event type prediction, show that the integrated method achieves superior performance compared to individual approaches.
arXiv Detail & Related papers (2024-08-19T13:47:17Z) - Unsupervised Generative Feature Transformation via Graph Contrastive Pre-training and Multi-objective Fine-tuning [28.673952870674146]
We develop a measurement-pretrain-finetune paradigm for Unsupervised Feature Transformation Learning.
For unsupervised feature set utility measurement, we propose a feature value consistency preservation perspective.
For generative transformation finetuning, we regard a feature set as a feature cross sequence and feature transformation as sequential generation.
arXiv Detail & Related papers (2024-05-27T06:50:00Z) - Personalized Behavior-Aware Transformer for Multi-Behavior Sequential
Recommendation [25.400756652696895]
We propose a Personalized Behavior-Aware Transformer framework (PBAT) for Multi-Behavior Sequential Recommendation (MBSR) problem.
PBAT develops a personalized behavior pattern generator in the representation layer, which extracts dynamic and discriminative behavior patterns for sequential learning.
We conduct experiments on three benchmark datasets and the results demonstrate the effectiveness and interpretability of our framework.
arXiv Detail & Related papers (2024-02-22T12:03:21Z) - Deep Group Interest Modeling of Full Lifelong User Behaviors for CTR Prediction [32.35797628133921]
Deep Group Interest Network (DGIN) is an end-to-end method to model the user's entire behavior history.
DGIN grouping reduces the behavior length significantly, from O(104) to O(102)
We identify a subset of behaviors that share the same item_id with the candidate item from the lifelong behavior sequence.
arXiv Detail & Related papers (2023-11-15T06:36:11Z) - Let Offline RL Flow: Training Conservative Agents in the Latent Space of
Normalizing Flows [58.762959061522736]
offline reinforcement learning aims to train a policy on a pre-recorded and fixed dataset without any additional environment interactions.
We build upon recent works on learning policies in latent action spaces and use a special form of Normalizing Flows for constructing a generative model.
We evaluate our method on various locomotion and navigation tasks, demonstrating that our approach outperforms recently proposed algorithms.
arXiv Detail & Related papers (2022-11-20T21:57:10Z) - Recommender Transformers with Behavior Pathways [50.842316273120744]
We build the Recommender Transformer (RETR) with a novel Pathway Attention mechanism.
We empirically verify the effectiveness of RETR on seven real-world datasets.
arXiv Detail & Related papers (2022-06-13T08:58:37Z) - TraSeTR: Track-to-Segment Transformer with Contrastive Query for
Instance-level Instrument Segmentation in Robotic Surgery [60.439434751619736]
We propose TraSeTR, a Track-to-Segment Transformer that exploits tracking cues to assist surgical instrument segmentation.
TraSeTR jointly reasons about the instrument type, location, and identity with instance-level predictions.
The effectiveness of our method is demonstrated with state-of-the-art instrument type segmentation results on three public datasets.
arXiv Detail & Related papers (2022-02-17T05:52:18Z) - ASFormer: Transformer for Action Segmentation [9.509416095106493]
We present an efficient Transformer-based model for action segmentation task, named ASFormer.
It constrains the hypothesis space within a reliable scope, and is beneficial for the action segmentation task to learn a proper target function with small training sets.
We apply a pre-defined hierarchical representation pattern that efficiently handles long input sequences.
arXiv Detail & Related papers (2021-10-16T13:07:20Z) - Explainable Deep Behavioral Sequence Clustering for Transaction Fraud
Detection [3.9505606841402607]
We propose a Deep learning based behavior data representation method for Clustering (FinDeepBehaviorCluster) to detect fraudulent transactions.
To utilize the behavior sequence data, we treat click stream data as event sequence, use time attention based Bi-LSTM to learn the sequence embedding in an unsupervised fashion, and combine them with intuitive features generated by risk experts to form a hybrid feature representation.
Our experimental results show that the proposed FinDeepBehaviorCluster framework is able to catch missed fraudulent transactions with considerable business values.
arXiv Detail & Related papers (2021-01-12T04:12:18Z) - A Trainable Optimal Transport Embedding for Feature Aggregation and its
Relationship to Attention [96.77554122595578]
We introduce a parametrized representation of fixed size, which embeds and then aggregates elements from a given input set according to the optimal transport plan between the set and a trainable reference.
Our approach scales to large datasets and allows end-to-end training of the reference, while also providing a simple unsupervised learning mechanism with small computational cost.
arXiv Detail & Related papers (2020-06-22T08:35:58Z) - Self-Attention Attribution: Interpreting Information Interactions Inside
Transformer [89.21584915290319]
We propose a self-attention attribution method to interpret the information interactions inside Transformer.
We show that the attribution results can be used as adversarial patterns to implement non-targeted attacks towards BERT.
arXiv Detail & Related papers (2020-04-23T14:58:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.