Dichotomic Pattern Mining with Applications to Intent Prediction from
Semi-Structured Clickstream Datasets
- URL: http://arxiv.org/abs/2201.09178v1
- Date: Sun, 23 Jan 2022 05:00:50 GMT
- Title: Dichotomic Pattern Mining with Applications to Intent Prediction from
Semi-Structured Clickstream Datasets
- Authors: Xin Wang, Serdar Kadioglu
- Abstract summary: We introduce a pattern mining framework that operates on semi-structured datasets.
We show that pattern embeddings play an integrator role between semi-structured data and machine learning models.
- Score: 10.76469643992931
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We introduce a pattern mining framework that operates on semi-structured
datasets and exploits the dichotomy between outcomes. Our approach takes
advantage of constraint reasoning to find sequential patterns that occur
frequently and exhibit desired properties. This allows the creation of novel
pattern embeddings that are useful for knowledge extraction and predictive
modeling. Finally, we present an application on customer intent prediction from
digital clickstream data. Overall, we show that pattern embeddings play an
integrator role between semi-structured data and machine learning models,
improve the performance of the downstream task and retain interpretability.
Related papers
- Hierarchical Blockmodelling for Knowledge Graphs [0.5530212768657544]
We use blockmodels for the purpose of hierarchical entity clustering on knowledge graphs.
The integration of the Nested Chinese Restaurant Process and the Stick Breaking Process into the generative model allows for the induction of hierarchical clusterings.
We evaluate our model on synthetic and real-world datasets and quantitatively compare against benchmark models.
arXiv Detail & Related papers (2024-08-28T09:04:15Z) - Decomposing and Editing Predictions by Modeling Model Computation [75.37535202884463]
We introduce a task called component modeling.
The goal of component modeling is to decompose an ML model's prediction in terms of its components.
We present COAR, a scalable algorithm for estimating component attributions.
arXiv Detail & Related papers (2024-04-17T16:28:08Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Ensemble Modeling for Multimodal Visual Action Recognition [50.38638300332429]
We propose an ensemble modeling approach for multimodal action recognition.
We independently train individual modality models using a variant of focal loss tailored to handle the long-tailed distribution of the MECCANO [21] dataset.
arXiv Detail & Related papers (2023-08-10T08:43:20Z) - Efficient Model Selection for Predictive Pattern Mining Model by Safe
Pattern Pruning [14.892080531048956]
We propose the Safe Pattern Pruning (SPP) method to address the explosion of pattern numbers in predictive pattern mining.
To demonstrate the effectiveness of the proposed method, we conduct numerical experiments on regression and classification problems involving sets, graphs, and sequences.
arXiv Detail & Related papers (2023-06-23T15:34:20Z) - EAMDrift: An interpretable self retrain model for time series [0.0]
We present EAMDrift, a novel method that combines forecasts from multiple individual predictors by weighting each prediction according to a performance metric.
EAMDrift is designed to automatically adapt to out-of-distribution patterns in data and identify the most appropriate models to use at each moment.
Our study on real-world datasets shows that EAMDrift outperforms individual baseline models by 20% and achieves comparable accuracy results to non-interpretable ensemble models.
arXiv Detail & Related papers (2023-05-31T13:25:26Z) - Online learning techniques for prediction of temporal tabular datasets
with regime changes [0.0]
We propose a modular machine learning pipeline for ranking predictions on temporal panel datasets.
The modularity of the pipeline allows the use of different models, including Gradient Boosting Decision Trees (GBDTs) and Neural Networks.
Online learning techniques, which require no retraining of models, can be used post-prediction to enhance the results.
arXiv Detail & Related papers (2022-12-30T17:19:00Z) - Conditioned Human Trajectory Prediction using Iterative Attention Blocks [70.36888514074022]
We present a simple yet effective pedestrian trajectory prediction model aimed at pedestrians positions prediction in urban-like environments.
Our model is a neural-based architecture that can run several layers of attention blocks and transformers in an iterative sequential fashion.
We show that without explicit introduction of social masks, dynamical models, social pooling layers, or complicated graph-like structures, it is possible to produce on par results with SoTA models.
arXiv Detail & Related papers (2022-06-29T07:49:48Z) - A Graph-Enhanced Click Model for Web Search [67.27218481132185]
We propose a novel graph-enhanced click model (GraphCM) for web search.
We exploit both intra-session and inter-session information for the sparsity and cold-start problems.
arXiv Detail & Related papers (2022-06-17T08:32:43Z) - On Learning Prediction-Focused Mixtures [30.338543175315507]
We introduce prediction-focused modeling for mixtures, which automatically selects the dimensions relevant to the prediction task.
Our approach identifies relevant signal from the input, outperforms models that are not prediction-focused, and is easy to optimize.
arXiv Detail & Related papers (2021-10-25T19:14:36Z) - Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach [80.8446673089281]
We propose a new learning paradigm with graph representation and learning.
Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data.
arXiv Detail & Related papers (2021-10-09T09:02:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.