Field-Embedded Factorization Machines for Click-through rate prediction
- URL: http://arxiv.org/abs/2009.09931v2
- Date: Mon, 14 Jun 2021 18:45:02 GMT
- Title: Field-Embedded Factorization Machines for Click-through rate prediction
- Authors: Harshit Pande
- Abstract summary: Click-through rate (CTR) prediction models are common in many online applications such as digital advertising and recommender systems.
We propose a novel shallow Field-Embedded Factorization Machine (FEFM) and its deep counterpart Deep Field-Embedded Factorization Machine (DeepFEFM)
FEFM has significantly lower model complexity than FFM and roughly the same complexity as FwFM.
- Score: 2.942829992746068
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Click-through rate (CTR) prediction models are common in many online
applications such as digital advertising and recommender systems. Field-Aware
Factorization Machine (FFM) and Field-weighted Factorization Machine (FwFM) are
state-of-the-art among the shallow models for CTR prediction. Recently, many
deep learning-based models have also been proposed. Among deeper models,
DeepFM, xDeepFM, AutoInt+, and FiBiNet are state-of-the-art models. The deeper
models combine a core architectural component, which learns explicit feature
interactions, with a deep neural network (DNN) component. We propose a novel
shallow Field-Embedded Factorization Machine (FEFM) and its deep counterpart
Deep Field-Embedded Factorization Machine (DeepFEFM). FEFM learns symmetric
matrix embeddings for each field pair along with the usual single vector
embeddings for each feature. FEFM has significantly lower model complexity than
FFM and roughly the same complexity as FwFM. FEFM also has insightful
mathematical properties about important fields and field interactions. DeepFEFM
combines the FEFM interaction vectors learned by the FEFM component with a DNN
and is thus able to learn higher order interactions. We conducted comprehensive
experiments over a wide range of hyperparameters on two large publicly
available real-world datasets. When comparing test AUC and log loss, the
results show that FEFM and DeepFEFM outperform the existing state-of-the-art
shallow and deep models for CTR prediction tasks. We have made the code of FEFM
and DeepFEFM available in the DeepCTR library
(https://github.com/shenweichen/DeepCTR).
Related papers
- Towards Neural Scaling Laws for Time Series Foundation Models [63.5211738245487]
We examine two common TSFM architectures, encoder-only and decoder-only Transformers, and investigate their scaling behavior on both ID and OOD data.
Our experiments reveal that the log-likelihood loss of TSFMs exhibits similar scaling behavior in both OOD and ID settings.
We provide practical guidelines for designing and scaling larger TSFMs with enhanced model capabilities.
arXiv Detail & Related papers (2024-10-16T08:23:39Z) - EMOFM: Ensemble MLP mOdel with Feature-based Mixers for Click-Through
Rate Prediction [5.983194751474721]
A dataset contains millions of records and each field-wise feature in a record consists of hashed integers for privacy.
For this task, the keys of network-based methods might be type-wise feature extraction and information fusion across different fields.
We propose plug-in mixers for field/type-wise feature fusion, thus construct an field&type-wise ensemble model, namely EMOFM.
arXiv Detail & Related papers (2023-10-06T12:32:23Z) - Directed Acyclic Graph Factorization Machines for CTR Prediction via
Knowledge Distillation [65.62538699160085]
We propose a Directed Acyclic Graph Factorization Machine (KD-DAGFM) to learn the high-order feature interactions from existing complex interaction models for CTR prediction via Knowledge Distillation.
KD-DAGFM achieves the best performance with less than 21.5% FLOPs of the state-of-the-art method on both online and offline experiments.
arXiv Detail & Related papers (2022-11-21T03:09:42Z) - Boosting Factorization Machines via Saliency-Guided Mixup [125.15872106335692]
We present MixFM, inspired by Mixup, to generate auxiliary training data to boost Factorization machines (FMs)
We also put forward a novel Factorization Machine powered by Saliency-guided Mixup (denoted as SMFM)
arXiv Detail & Related papers (2022-06-17T09:49:00Z) - Leaf-FM: A Learnable Feature Generation Factorization Machine for
Click-Through Rate Prediction [2.412497918389292]
We propose LeafFM model based on FM to generate new features from the original feature embedding by learning the transformation functions automatically.
Experiments are conducted on three real-world datasets and the results show Leaf-FM model outperforms standard FMs by a large margin.
arXiv Detail & Related papers (2021-07-26T08:29:18Z) - Global Filter Networks for Image Classification [90.81352483076323]
We present a conceptually simple yet computationally efficient architecture that learns long-term spatial dependencies in the frequency domain with log-linear complexity.
Our results demonstrate that GFNet can be a very competitive alternative to transformer-style models and CNNs in efficiency, generalization ability and robustness.
arXiv Detail & Related papers (2021-07-01T17:58:16Z) - Rank-R FNN: A Tensor-Based Learning Model for High-Order Data
Classification [69.26747803963907]
Rank-R Feedforward Neural Network (FNN) is a tensor-based nonlinear learning model that imposes Canonical/Polyadic decomposition on its parameters.
First, it handles inputs as multilinear arrays, bypassing the need for vectorization, and can thus fully exploit the structural information along every data dimension.
We establish the universal approximation and learnability properties of Rank-R FNN, and we validate its performance on real-world hyperspectral datasets.
arXiv Detail & Related papers (2021-04-11T16:37:32Z) - Quaternion Factorization Machines: A Lightweight Solution to Intricate
Feature Interaction Modelling [76.89779231460193]
factorization machine (FM) is capable of automatically learning high-order interactions among features to make predictions without the need for manual feature engineering.
We propose the quaternion factorization machine (QFM) and quaternion neural factorization machine (QNFM) for sparse predictive analytics.
arXiv Detail & Related papers (2021-04-05T00:02:36Z) - $FM^2$: Field-matrixed Factorization Machines for Recommender Systems [9.461169933697379]
We propose a novel approach to model the field information effectively and efficiently.
The proposed approach is a direct improvement of FwFM, and is named as Field-matrixed Factorization Machines (FmFM)
arXiv Detail & Related papers (2021-02-20T00:03:37Z) - AdnFM: An Attentive DenseNet based Factorization Machine for CTR
Prediction [11.958336595818267]
We propose a novel model called Attentive DenseNet based Factorization Machines (AdnFM)
AdnFM can extract more comprehensive deep features by using all the hidden layers from a feed-forward neural network as implicit high-order features.
Experiments on two real-world datasets show that the proposed model can effectively improve the performance of Click-Through-Rate prediction.
arXiv Detail & Related papers (2020-12-20T01:00:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.