i-Razor: A Differentiable Neural Input Razor for Feature Selection and
Dimension Search in DNN-Based Recommender Systems
- URL: http://arxiv.org/abs/2204.00281v2
- Date: Sun, 12 Nov 2023 02:05:16 GMT
- Title: i-Razor: A Differentiable Neural Input Razor for Feature Selection and
Dimension Search in DNN-Based Recommender Systems
- Authors: Yao Yao, Bin Liu, Haoxun He, Dakui Sheng, Ke Wang, Li Xiao, and
Huanhuan Cao
- Abstract summary: Noisy features and inappropriate embedding dimension assignments can deteriorate the performance of recommender systems.
We propose a differentiable neural input razor (i-Razor) that enables joint optimization of feature selection and dimension search.
- Score: 8.992480061695138
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Input features play a crucial role in DNN-based recommender systems with
thousands of categorical and continuous fields from users, items, contexts, and
interactions. Noisy features and inappropriate embedding dimension assignments
can deteriorate the performance of recommender systems and introduce
unnecessary complexity in model training and online serving. Optimizing the
input configuration of DNN models, including feature selection and embedding
dimension assignment, has become one of the essential topics in feature
engineering. However, in existing industrial practices, feature selection and
dimension search are optimized sequentially, i.e., feature selection is
performed first, followed by dimension search to determine the optimal
dimension size for each selected feature. Such a sequential optimization
mechanism increases training costs and risks generating suboptimal input
configurations. To address this problem, we propose a differentiable neural
input razor (i-Razor) that enables joint optimization of feature selection and
dimension search. Concretely, we introduce an end-to-end differentiable model
to learn the relative importance of different embedding regions of each
feature. Furthermore, a flexible pruning algorithm is proposed to achieve
feature filtering and dimension derivation simultaneously. Extensive
experiments on two large-scale public datasets in the Click-Through-Rate (CTR)
prediction task demonstrate the efficacy and superiority of i-Razor in
balancing model complexity and performance.
Related papers
- Unveiling the Power of Sparse Neural Networks for Feature Selection [60.50319755984697]
Sparse Neural Networks (SNNs) have emerged as powerful tools for efficient feature selection.
We show that SNNs trained with dynamic sparse training (DST) algorithms can achieve, on average, more than $50%$ memory and $55%$ FLOPs reduction.
Our findings show that feature selection with SNNs trained with DST algorithms can achieve, on average, more than $50%$ memory and $55%$ FLOPs reduction.
arXiv Detail & Related papers (2024-08-08T16:48:33Z) - Adaptive Preference Scaling for Reinforcement Learning with Human Feedback [103.36048042664768]
Reinforcement learning from human feedback (RLHF) is a prevalent approach to align AI systems with human values.
We propose a novel adaptive preference loss, underpinned by distributionally robust optimization (DRO)
Our method is versatile and can be readily adapted to various preference optimization frameworks.
arXiv Detail & Related papers (2024-06-04T20:33:22Z) - Feature Selection as Deep Sequential Generative Learning [50.00973409680637]
We develop a deep variational transformer model over a joint of sequential reconstruction, variational, and performance evaluator losses.
Our model can distill feature selection knowledge and learn a continuous embedding space to map feature selection decision sequences into embedding vectors associated with utility scores.
arXiv Detail & Related papers (2024-03-06T16:31:56Z) - Beyond Discrete Selection: Continuous Embedding Space Optimization for
Generative Feature Selection [34.32619834917906]
We reformulate the feature selection problem as a deep differentiable optimization task.
We propose a new principled research perspective: conceptualizing discrete feature subsetting as continuous embedding space.
Specifically, we utilize reinforcement feature selection learning to generate diverse and high-quality training data.
arXiv Detail & Related papers (2023-02-26T03:18:45Z) - Compactness Score: A Fast Filter Method for Unsupervised Feature
Selection [66.84571085643928]
We propose a fast unsupervised feature selection method, named as, Compactness Score (CSUFS) to select desired features.
Our proposed algorithm seems to be more accurate and efficient compared with existing algorithms.
arXiv Detail & Related papers (2022-01-31T13:01:37Z) - Sparse Centroid-Encoder: A Nonlinear Model for Feature Selection [1.2487990897680423]
We develop a sparse implementation of the centroid-encoder for nonlinear data reduction and visualization called Centro Sparseid-Encoder.
We also provide a feature selection framework that first ranks each feature by its occurrence, and the optimal number of features is chosen using a validation set.
The algorithm is applied to a wide variety of data sets including, single-cell biological data, high dimensional infectious disease data, hyperspectral data, image data, and speech data.
arXiv Detail & Related papers (2022-01-30T20:46:24Z) - RoMA: Robust Model Adaptation for Offline Model-based Optimization [115.02677045518692]
We consider the problem of searching an input maximizing a black-box objective function given a static dataset of input-output queries.
A popular approach to solving this problem is maintaining a proxy model that approximates the true objective function.
Here, the main challenge is how to avoid adversarially optimized inputs during the search.
arXiv Detail & Related papers (2021-10-27T05:37:12Z) - Joint Adaptive Graph and Structured Sparsity Regularization for
Unsupervised Feature Selection [6.41804410246642]
We propose a joint adaptive graph and structured sparsity regularization unsupervised feature selection (JASFS) method.
A subset of optimal features will be selected in group, and the number of selected features will be determined automatically.
Experimental results on eight benchmarks demonstrate the effectiveness and efficiency of the proposed method.
arXiv Detail & Related papers (2020-10-09T08:17:04Z) - Automated Human Activity Recognition by Colliding Bodies
Optimization-based Optimal Feature Selection with Recurrent Neural Network [0.0]
Human Activity Recognition (HAR) is considered to be an efficient model in pervasive computation from sensor readings.
This paper tempts to implement the HAR system using deep learning with the data collected from smart sensors that are publicly available in the UC Irvine Machine Learning Repository (UCI)
arXiv Detail & Related papers (2020-10-07T10:58:46Z) - Differentiable Neural Input Search for Recommender Systems [26.88124270897381]
Differentiable Neural Input Search (DNIS) is a method that searches for mixed feature embedding dimensions in a more flexible space.
DNIS is model-agnostic and can be seamlessly incorporated with existing latent factor models for recommendation.
arXiv Detail & Related papers (2020-06-08T10:43:59Z) - Self-Directed Online Machine Learning for Topology Optimization [58.920693413667216]
Self-directed Online Learning Optimization integrates Deep Neural Network (DNN) with Finite Element Method (FEM) calculations.
Our algorithm was tested by four types of problems including compliance minimization, fluid-structure optimization, heat transfer enhancement and truss optimization.
It reduced the computational time by 2 5 orders of magnitude compared with directly using methods, and outperformed all state-of-the-art algorithms tested in our experiments.
arXiv Detail & Related papers (2020-02-04T20:00:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.