AutoFIS: Automatic Feature Interaction Selection in Factorization Models
for Click-Through Rate Prediction
- URL: http://arxiv.org/abs/2003.11235v3
- Date: Fri, 3 Jul 2020 14:19:47 GMT
- Title: AutoFIS: Automatic Feature Interaction Selection in Factorization Models
for Click-Through Rate Prediction
- Authors: Bin Liu, Chenxu Zhu, Guilin Li, Weinan Zhang, Jincai Lai, Ruiming
Tang, Xiuqiang He, Zhenguo Li, Yong Yu
- Abstract summary: We propose a two-stage algorithm called Automatic Feature Interaction Selection (AutoFIS)
AutoFIS can automatically identify important feature interactions for factorization models with computational cost just equivalent to training the target model to convergence.
AutoFIS has been deployed onto the training platform of Huawei App Store recommendation service.
- Score: 75.16836697734995
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning feature interactions is crucial for click-through rate (CTR)
prediction in recommender systems. In most existing deep learning models,
feature interactions are either manually designed or simply enumerated.
However, enumerating all feature interactions brings large memory and
computation cost. Even worse, useless interactions may introduce noise and
complicate the training process. In this work, we propose a two-stage algorithm
called Automatic Feature Interaction Selection (AutoFIS). AutoFIS can
automatically identify important feature interactions for factorization models
with computational cost just equivalent to training the target model to
convergence. In the \emph{search stage}, instead of searching over a discrete
set of candidate feature interactions, we relax the choices to be continuous by
introducing the architecture parameters. By implementing a regularized
optimizer over the architecture parameters, the model can automatically
identify and remove the redundant feature interactions during the training
process of the model. In the \emph{re-train stage}, we keep the architecture
parameters serving as an attention unit to further boost the performance.
Offline experiments on three large-scale datasets (two public benchmarks, one
private) demonstrate that AutoFIS can significantly improve various FM based
models. AutoFIS has been deployed onto the training platform of Huawei App
Store recommendation service, where a 10-day online A/B test demonstrated that
AutoFIS improved the DeepFM model by 20.3\% and 20.1\% in terms of CTR and CVR
respectively.
Related papers
- Wolf2Pack: The AutoFusion Framework for Dynamic Parameter Fusion [4.164728134421114]
We introduce AutoFusion, a framework that fuses distinct model parameters for multi-task learning without pre-trained checkpoints.
We validate AutoFusion's effectiveness through experiments on commonly used benchmark datasets.
Our framework offers a scalable and flexible solution for model integration, positioning it as a powerful tool for future research and practical applications.
arXiv Detail & Related papers (2024-10-08T07:21:24Z) - ToolACE: Winning the Points of LLM Function Calling [139.07157814653638]
ToolACE is an automatic agentic pipeline designed to generate accurate, complex, and diverse tool-learning data.
We demonstrate that models trained on our synthesized data, even with only 8B parameters, achieve state-of-the-art performance on the Berkeley Function-Calling Leaderboard.
arXiv Detail & Related papers (2024-09-02T03:19:56Z) - AutoXPCR: Automated Multi-Objective Model Selection for Time Series
Forecasting [1.0515439489916734]
We propose AutoXPCR - a novel method for automated and explainable multi-objective model selection.
Our approach leverages meta-learning to estimate any model's performance along PCR criteria, which encompass (P)redictive error, (C)omplexity, and (R)esource demand.
Our method clearly outperforms other model selection approaches - on average, it only requires 20% of computation costs for recommending models with 90% of the best-possible quality.
arXiv Detail & Related papers (2023-12-20T14:04:57Z) - Trajeglish: Traffic Modeling as Next-Token Prediction [67.28197954427638]
A longstanding challenge for self-driving development is simulating dynamic driving scenarios seeded from recorded driving logs.
We apply tools from discrete sequence modeling to model how vehicles, pedestrians and cyclists interact in driving scenarios.
Our model tops the Sim Agents Benchmark, surpassing prior work along the realism meta metric by 3.3% and along the interaction metric by 9.9%.
arXiv Detail & Related papers (2023-12-07T18:53:27Z) - Memorize, Factorize, or be Na\"ive: Learning Optimal Feature Interaction
Methods for CTR Prediction [29.343267933348372]
We propose a framework called OptInter which finds the most suitable modelling method for each feature interaction.
Our experiments show that OptInter improves the best performed state-of-the-art baseline deep CTR models by up to 2.21%.
arXiv Detail & Related papers (2021-08-03T03:03:34Z) - AutoDis: Automatic Discretization for Embedding Numerical Features in
CTR Prediction [45.69943728028556]
Learning sophisticated feature interactions is crucial for Click-Through Rate (CTR) prediction in recommender systems.
Various deep CTR models follow an Embedding & Feature Interaction paradigm.
We propose AutoDis, a framework that discretizes features in numerical fields automatically and is optimized with CTR models in an end-to-end manner.
arXiv Detail & Related papers (2020-12-16T14:31:31Z) - Towards Automated Neural Interaction Discovery for Click-Through Rate
Prediction [64.03526633651218]
Click-Through Rate (CTR) prediction is one of the most important machine learning tasks in recommender systems.
We propose an automated interaction architecture discovering framework for CTR prediction named AutoCTR.
arXiv Detail & Related papers (2020-06-29T04:33:01Z) - Auto-PyTorch Tabular: Multi-Fidelity MetaLearning for Efficient and
Robust AutoDL [53.40030379661183]
Auto-PyTorch is a framework to enable fully automated deep learning (AutoDL)
It combines multi-fidelity optimization with portfolio construction for warmstarting and ensembling of deep neural networks (DNNs)
We show that Auto-PyTorch performs better than several state-of-the-art competitors on average.
arXiv Detail & Related papers (2020-06-24T15:15:17Z) - ARDA: Automatic Relational Data Augmentation for Machine Learning [23.570173866941612]
We present system, an end-to-end system that takes as input a dataset and a data repository, and outputs an augmented data set.
Our system has two distinct components: (1) a framework to search and join data with the input data, based on various attributes of the input, and (2) an efficient feature selection algorithm that prunes out noisy or irrelevant features from the resulting join.
arXiv Detail & Related papers (2020-03-21T21:55:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.