Efficient and Joint Hyperparameter and Architecture Search for
Collaborative Filtering
- URL: http://arxiv.org/abs/2307.11004v1
- Date: Wed, 12 Jul 2023 10:56:25 GMT
- Title: Efficient and Joint Hyperparameter and Architecture Search for
Collaborative Filtering
- Authors: Yan Wen, Chen Gao, Lingling Yi, Liwei Qiu, Yaqing Wang, Yong Li
- Abstract summary: We propose a two-stage search algorithm for Collaborative Filtering models.
In the first stage, we leverage knowledge from subsampled datasets to reduce evaluation costs.
In the second stage, we efficiently fine-tune top candidate models on the whole dataset.
- Score: 31.25094171513831
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automated Machine Learning (AutoML) techniques have recently been introduced
to design Collaborative Filtering (CF) models in a data-specific manner.
However, existing works either search architectures or hyperparameters while
ignoring the fact they are intrinsically related and should be considered
together. This motivates us to consider a joint hyperparameter and architecture
search method to design CF models. However, this is not easy because of the
large search space and high evaluation cost. To solve these challenges, we
reduce the space by screening out usefulness yperparameter choices through a
comprehensive understanding of individual hyperparameters. Next, we propose a
two-stage search algorithm to find proper configurations from the reduced
space. In the first stage, we leverage knowledge from subsampled datasets to
reduce evaluation costs; in the second stage, we efficiently fine-tune top
candidate models on the whole dataset. Extensive experiments on real-world
datasets show better performance can be achieved compared with both
hand-designed and previous searched models. Besides, ablation and case studies
demonstrate the effectiveness of our search framework.
Related papers
- Implicitly Guided Design with PropEn: Match your Data to Follow the Gradient [52.2669490431145]
PropEn is inspired by'matching', which enables implicit guidance without training a discriminator.
We show that training with a matched dataset approximates the gradient of the property of interest while remaining within the data distribution.
arXiv Detail & Related papers (2024-05-28T11:30:19Z) - Efficient Architecture Search via Bi-level Data Pruning [70.29970746807882]
This work pioneers an exploration into the critical role of dataset characteristics for DARTS bi-level optimization.
We introduce a new progressive data pruning strategy that utilizes supernet prediction dynamics as the metric.
Comprehensive evaluations on the NAS-Bench-201 search space, DARTS search space, and MobileNet-like search space validate that BDP reduces search costs by over 50%.
arXiv Detail & Related papers (2023-12-21T02:48:44Z) - MGAS: Multi-Granularity Architecture Search for Trade-Off Between Model
Effectiveness and Efficiency [10.641875933652647]
We introduce multi-granularity architecture search (MGAS) to discover both effective and efficient neural networks.
We learn discretization functions specific to each granularity level to adaptively determine the unit remaining ratio according to the evolving architecture.
Extensive experiments on CIFAR-10, CIFAR-100 and ImageNet demonstrate that MGAS outperforms other state-of-the-art methods in achieving a better trade-off between model performance and model size.
arXiv Detail & Related papers (2023-10-23T16:32:18Z) - Fairer and More Accurate Tabular Models Through NAS [14.147928131445852]
We propose using multi-objective Neural Architecture Search (NAS) and Hyperparameter Optimization (HPO) in the first application to the very challenging domain of tabular data.
We show that models optimized solely for accuracy with NAS often fail to inherently address fairness concerns.
We produce architectures that consistently dominate state-of-the-art bias mitigation methods either in fairness, accuracy or both.
arXiv Detail & Related papers (2023-10-18T17:56:24Z) - Improved Distribution Matching for Dataset Condensation [91.55972945798531]
We propose a novel dataset condensation method based on distribution matching.
Our simple yet effective method outperforms most previous optimization-oriented methods with much fewer computational resources.
arXiv Detail & Related papers (2023-07-19T04:07:33Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Efficient Data-specific Model Search for Collaborative Filtering [56.60519991956558]
Collaborative filtering (CF) is a fundamental approach for recommender systems.
In this paper, motivated by the recent advances in automated machine learning (AutoML), we propose to design a data-specific CF model.
Key here is a new framework that unifies state-of-the-art (SOTA) CF methods and splits them into disjoint stages of input encoding, embedding function, interaction and prediction function.
arXiv Detail & Related papers (2021-06-14T14:30:32Z) - Generalization Guarantees for Neural Architecture Search with
Train-Validation Split [48.265305046655996]
This paper explores the statistical aspects of such problems with train-validation splits.
We show that refined properties of the validation loss such as risk and hyper-gradients are indicative of those of the true test loss.
We also highlight rigorous connections between NAS, multiple kernel learning, and low-rank matrix learning.
arXiv Detail & Related papers (2021-04-29T06:11:00Z) - Deep-n-Cheap: An Automated Search Framework for Low Complexity Deep
Learning [3.479254848034425]
We present Deep-n-Cheap -- an open-source AutoML framework to search for deep learning models.
Our framework is targeted for deployment on both benchmark and custom datasets.
Deep-n-Cheap includes a user-customizable complexity penalty which trades off performance with training time or number of parameters.
arXiv Detail & Related papers (2020-03-27T13:00:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.