SimCURL: Simple Contrastive User Representation Learning from Command
Sequences
- URL: http://arxiv.org/abs/2207.14760v1
- Date: Fri, 29 Jul 2022 16:06:03 GMT
- Title: SimCURL: Simple Contrastive User Representation Learning from Command
Sequences
- Authors: Hang Chu, Amir Hosein Khasahmadi, Karl D.D. Willis, Fraser Anderson,
Yaoli Mao, Linh Tran, Justin Matejka, Jo Vermeulen
- Abstract summary: We propose SimCURL, a contrastive self-supervised deep learning framework that learns user representation from unlabeled command sequences.
We train and evaluate our method on a real-world command sequence dataset of more than half a billion commands.
- Score: 22.92215383896495
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: User modeling is crucial to understanding user behavior and essential for
improving user experience and personalized recommendations. When users interact
with software, vast amounts of command sequences are generated through logging
and analytics systems. These command sequences contain clues to the users'
goals and intents. However, these data modalities are highly unstructured and
unlabeled, making it difficult for standard predictive systems to learn from.
We propose SimCURL, a simple yet effective contrastive self-supervised deep
learning framework that learns user representation from unlabeled command
sequences. Our method introduces a user-session network architecture, as well
as session dropout as a novel way of data augmentation. We train and evaluate
our method on a real-world command sequence dataset of more than half a billion
commands. Our method shows significant improvement over existing methods when
the learned representation is transferred to downstream tasks such as
experience and expertise classification.
Related papers
- Leave No One Behind: Online Self-Supervised Self-Distillation for Sequential Recommendation [20.52842524024608]
Sequential recommendation methods play a pivotal role in modern recommendation systems.
Recent methods leverage contrastive learning to derive self-supervision signals.
We introduce a novel learning paradigm, named Online Self-Supervised Self-distillation for Sequential Recommendation.
arXiv Detail & Related papers (2024-03-22T12:27:21Z) - One-Shot Learning as Instruction Data Prospector for Large Language Models [108.81681547472138]
textscNuggets uses one-shot learning to select high-quality instruction data from extensive datasets.
We show that instruction tuning with the top 1% of examples curated by textscNuggets substantially outperforms conventional methods employing the entire dataset.
arXiv Detail & Related papers (2023-12-16T03:33:12Z) - Bootstrapping Adaptive Human-Machine Interfaces with Offline
Reinforcement Learning [82.91837418721182]
Adaptive interfaces can help users perform sequential decision-making tasks.
Recent advances in human-in-the-loop machine learning enable such systems to improve by interacting with users.
We propose a reinforcement learning algorithm to train an interface to map raw command signals to actions.
arXiv Detail & Related papers (2023-09-07T16:52:27Z) - First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual
Information Maximization [112.40598205054994]
We formalize this idea as a completely unsupervised objective for optimizing interfaces.
We conduct an observational study on 540K examples of users operating various keyboard and eye gaze interfaces for typing, controlling simulated robots, and playing video games.
The results show that our mutual information scores are predictive of the ground-truth task completion metrics in a variety of domains.
arXiv Detail & Related papers (2022-05-24T21:57:18Z) - UserBERT: Modeling Long- and Short-Term User Preferences via
Self-Supervision [6.8904125699168075]
This paper extends the BERT model to e-commerce user data for pre-training representations in a self-supervised manner.
By viewing user actions in sequences as analogous to words in sentences, we extend the existing BERT model to user behavior data.
We propose methods for the tokenization of different types of user behavior sequences, the generation of input representation, and a novel pretext task to enable the pre-trained model to learn from its own input.
arXiv Detail & Related papers (2022-02-14T08:31:36Z) - Learning to Learn a Cold-start Sequential Recommender [70.5692886883067]
The cold-start recommendation is an urgent problem in contemporary online applications.
We propose a meta-learning based cold-start sequential recommendation framework called metaCSR.
metaCSR holds the ability to learn the common patterns from regular users' behaviors.
arXiv Detail & Related papers (2021-10-18T08:11:24Z) - Hyper Meta-Path Contrastive Learning for Multi-Behavior Recommendation [61.114580368455236]
User purchasing prediction with multi-behavior information remains a challenging problem for current recommendation systems.
We propose the concept of hyper meta-path to construct hyper meta-paths or hyper meta-graphs to explicitly illustrate the dependencies among different behaviors of a user.
Thanks to the recent success of graph contrastive learning, we leverage it to learn embeddings of user behavior patterns adaptively instead of assigning a fixed scheme to understand the dependencies among different behaviors.
arXiv Detail & Related papers (2021-09-07T04:28:09Z) - Improved Representation Learning for Session-based Recommendation [0.0]
Session-based recommendation systems suggest relevant items to users by modeling user behavior and preferences using short-term anonymous sessions.
Existing methods leverage Graph Neural Networks (GNNs) that propagate and aggregate information from neighboring nodes.
We propose using a Transformer in combination with a target attentive GNN, which allows richer Representation Learning.
arXiv Detail & Related papers (2021-07-04T00:57:28Z) - Pre-training for low resource speech-to-intent applications [26.093156590824076]
We discuss a user-taught speech-to-intent (S2I) system in this paper.
The user-taught system learns from scratch from the users' spoken input with action demonstration.
In this paper we combine the encoder of an end-to-end ASR system with the prior NMF/capsule network-based user-taught decoder.
arXiv Detail & Related papers (2021-03-30T20:44:29Z) - Multi-Agent Task-Oriented Dialog Policy Learning with Role-Aware Reward
Decomposition [64.06167416127386]
We propose Multi-Agent Dialog Policy Learning, which regards both the system and the user as the dialog agents.
Two agents interact with each other and are jointly learned simultaneously.
Results show that our method can successfully build a system policy and a user policy simultaneously.
arXiv Detail & Related papers (2020-04-08T04:51:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.