AdnFM: An Attentive DenseNet based Factorization Machine for CTR
Prediction
- URL: http://arxiv.org/abs/2012.10820v1
- Date: Sun, 20 Dec 2020 01:00:39 GMT
- Title: AdnFM: An Attentive DenseNet based Factorization Machine for CTR
Prediction
- Authors: Kai Wang, Chunxu Shen, Wenye Ma
- Abstract summary: We propose a novel model called Attentive DenseNet based Factorization Machines (AdnFM)
AdnFM can extract more comprehensive deep features by using all the hidden layers from a feed-forward neural network as implicit high-order features.
Experiments on two real-world datasets show that the proposed model can effectively improve the performance of Click-Through-Rate prediction.
- Score: 11.958336595818267
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we consider the Click-Through-Rate (CTR) prediction problem.
Factorization Machines and their variants consider pair-wise feature
interactions, but normally we won't do high-order feature interactions using FM
due to high time complexity. Given the success of deep neural networks (DNNs)
in many fields, researchers have proposed several DNN-based models to learn
high-order feature interactions. Multi-layer perceptrons (MLP) have been widely
employed to learn reliable mappings from feature embeddings to final logits. In
this paper, we aim to explore more about these high-order features
interactions. However, high-order feature interaction deserves more attention
and further development. Inspired by the great achievements of Densely
Connected Convolutional Networks (DenseNet) in computer vision, we propose a
novel model called Attentive DenseNet based Factorization Machines (AdnFM).
AdnFM can extract more comprehensive deep features by using all the hidden
layers from a feed-forward neural network as implicit high-order features, then
selects dominant features via an attention mechanism. Also, high-order
interactions in the implicit way using DNNs are more cost-efficient than in the
explicit way, for example in FM. Extensive experiments on two real-world
datasets show that the proposed model can effectively improve the performance
of CTR prediction.
Related papers
- Multilinear Operator Networks [60.7432588386185]
Polynomial Networks is a class of models that does not require activation functions.
We propose MONet, which relies solely on multilinear operators.
arXiv Detail & Related papers (2024-01-31T16:52:19Z) - Directed Acyclic Graph Factorization Machines for CTR Prediction via
Knowledge Distillation [65.62538699160085]
We propose a Directed Acyclic Graph Factorization Machine (KD-DAGFM) to learn the high-order feature interactions from existing complex interaction models for CTR prediction via Knowledge Distillation.
KD-DAGFM achieves the best performance with less than 21.5% FLOPs of the state-of-the-art method on both online and offline experiments.
arXiv Detail & Related papers (2022-11-21T03:09:42Z) - Deep Multi-Representation Model for Click-Through Rate Prediction [6.155158115218501]
Click-Through Rate prediction (CTR) is a crucial task in recommender systems.
We propose the Deep Multi-Representation model (DeepMR) that jointly trains a mixture of two powerful feature representation learning components.
Experiments on three real-world datasets show that the proposed model significantly outperforms all state-of-the-art models in the task of click-through rate prediction.
arXiv Detail & Related papers (2022-10-18T09:37:11Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - Global Filter Networks for Image Classification [90.81352483076323]
We present a conceptually simple yet computationally efficient architecture that learns long-term spatial dependencies in the frequency domain with log-linear complexity.
Our results demonstrate that GFNet can be a very competitive alternative to transformer-style models and CNNs in efficiency, generalization ability and robustness.
arXiv Detail & Related papers (2021-07-01T17:58:16Z) - DCAP: Deep Cross Attentional Product Network for User Response
Prediction [20.17934000984361]
We propose a novel architecture Deep Cross Attentional Product Network (DCAP)
DCAP keeps cross network's benefits in modeling high-order feature interactions explicitly at the vector-wise level.
Our proposed model can be easily implemented and train in parallel.
arXiv Detail & Related papers (2021-05-18T16:27:20Z) - Rank-R FNN: A Tensor-Based Learning Model for High-Order Data
Classification [69.26747803963907]
Rank-R Feedforward Neural Network (FNN) is a tensor-based nonlinear learning model that imposes Canonical/Polyadic decomposition on its parameters.
First, it handles inputs as multilinear arrays, bypassing the need for vectorization, and can thus fully exploit the structural information along every data dimension.
We establish the universal approximation and learnability properties of Rank-R FNN, and we validate its performance on real-world hyperspectral datasets.
arXiv Detail & Related papers (2021-04-11T16:37:32Z) - Field-Embedded Factorization Machines for Click-through rate prediction [2.942829992746068]
Click-through rate (CTR) prediction models are common in many online applications such as digital advertising and recommender systems.
We propose a novel shallow Field-Embedded Factorization Machine (FEFM) and its deep counterpart Deep Field-Embedded Factorization Machine (DeepFEFM)
FEFM has significantly lower model complexity than FFM and roughly the same complexity as FwFM.
arXiv Detail & Related papers (2020-09-13T15:32:42Z) - GateNet: Gating-Enhanced Deep Network for Click-Through Rate Prediction [3.201333208812837]
In recent years, many neural network based CTR models have been proposed and achieved success.
We propose a novel model named GateNet which introduces either the feature embedding gate or the hidden gate to the embedding layer or hidden CTR models.
arXiv Detail & Related papers (2020-07-06T12:45:46Z) - Feature Interaction based Neural Network for Click-Through Rate
Prediction [5.095988654970358]
We propose a Feature Interaction based Neural Network (FINN) which is able to model feature interaction via a 3-dimention relation tensor.
We show that our deep FINN model outperforms other state-of-the-art deep models such as PNN and DeepFM.
It also indicates that our models can effectively learn the feature interactions, and achieve better performances in real-world datasets.
arXiv Detail & Related papers (2020-06-07T03:53:24Z) - Neural Additive Models: Interpretable Machine Learning with Neural Nets [77.66871378302774]
Deep neural networks (DNNs) are powerful black-box predictors that have achieved impressive performance on a wide variety of tasks.
We propose Neural Additive Models (NAMs) which combine some of the expressivity of DNNs with the inherent intelligibility of generalized additive models.
NAMs learn a linear combination of neural networks that each attend to a single input feature.
arXiv Detail & Related papers (2020-04-29T01:28:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.