Efficient Data-specific Model Search for Collaborative Filtering
- URL: http://arxiv.org/abs/2106.07453v1
- Date: Mon, 14 Jun 2021 14:30:32 GMT
- Title: Efficient Data-specific Model Search for Collaborative Filtering
- Authors: Chen Gao and Quanming Yao and Depeng Jin and Yong Li
- Abstract summary: Collaborative filtering (CF) is a fundamental approach for recommender systems.
In this paper, motivated by the recent advances in automated machine learning (AutoML), we propose to design a data-specific CF model.
Key here is a new framework that unifies state-of-the-art (SOTA) CF methods and splits them into disjoint stages of input encoding, embedding function, interaction and prediction function.
- Score: 56.60519991956558
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Collaborative filtering (CF), as a fundamental approach for recommender
systems, is usually built on the latent factor model with learnable parameters
to predict users' preferences towards items. However, designing a proper CF
model for a given data is not easy, since the properties of datasets are highly
diverse. In this paper, motivated by the recent advances in automated machine
learning (AutoML), we propose to design a data-specific CF model by AutoML
techniques. The key here is a new framework that unifies state-of-the-art
(SOTA) CF methods and splits them into disjoint stages of input encoding,
embedding function, interaction function, and prediction function. We further
develop an easy-to-use, robust, and efficient search strategy, which utilizes
random search and a performance predictor for efficient searching within the
above framework. In this way, we can combinatorially generalize data-specific
CF models, which have not been visited in the literature, from SOTA ones.
Extensive experiments on five real-world datasets demonstrate that our method
can consistently outperform SOTA ones for various CF tasks. Further experiments
verify the rationality of the proposed framework and the efficiency of the
search strategy. The searched CF models can also provide insights for exploring
more effective methods in the future
Related papers
- Wolf2Pack: The AutoFusion Framework for Dynamic Parameter Fusion [4.164728134421114]
We introduce AutoFusion, a framework that fuses distinct model parameters for multi-task learning without pre-trained checkpoints.
We validate AutoFusion's effectiveness through experiments on commonly used benchmark datasets.
Our framework offers a scalable and flexible solution for model integration, positioning it as a powerful tool for future research and practical applications.
arXiv Detail & Related papers (2024-10-08T07:21:24Z) - LLM-Select: Feature Selection with Large Language Models [64.5099482021597]
Large language models (LLMs) are capable of selecting the most predictive features, with performance rivaling the standard tools of data science.
Our findings suggest that LLMs may be useful not only for selecting the best features for training but also for deciding which features to collect in the first place.
arXiv Detail & Related papers (2024-07-02T22:23:40Z) - Implicitly Guided Design with PropEn: Match your Data to Follow the Gradient [52.2669490431145]
PropEn is inspired by'matching', which enables implicit guidance without training a discriminator.
We show that training with a matched dataset approximates the gradient of the property of interest while remaining within the data distribution.
arXiv Detail & Related papers (2024-05-28T11:30:19Z) - A PSO Based Method to Generate Actionable Counterfactuals for High
Dimensional Data [3.0320603363468845]
We describe an efficient and an actionable counterfactual (CF) generation method based on particle swarm optimization (PSO)
An algorithm is proposed that incorporates these features and it enables greater control over the proximity and sparsity properties over the generated CFs.
arXiv Detail & Related papers (2023-09-30T18:08:00Z) - Efficient and Joint Hyperparameter and Architecture Search for
Collaborative Filtering [31.25094171513831]
We propose a two-stage search algorithm for Collaborative Filtering models.
In the first stage, we leverage knowledge from subsampled datasets to reduce evaluation costs.
In the second stage, we efficiently fine-tune top candidate models on the whole dataset.
arXiv Detail & Related papers (2023-07-12T10:56:25Z) - Multidimensional Item Response Theory in the Style of Collaborative
Filtering [0.8057006406834467]
This paper presents a machine learning approach to multidimensional item response theory (MIRT)
Inspired by collaborative filtering, we define a general class of models that includes many MIRT models.
We discuss the use of penalized joint maximum likelihood (JML) to estimate individual models and cross-validation to select the best performing model.
arXiv Detail & Related papers (2023-01-03T00:56:27Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - DualCF: Efficient Model Extraction Attack from Counterfactual
Explanations [57.46134660974256]
Cloud service providers have launched Machine-Learning-as-a-Service platforms to allow users to access large-scale cloudbased models via APIs.
Such extra information inevitably causes the cloud models to be more vulnerable to extraction attacks.
We propose a novel simple yet efficient querying strategy to greatly enhance the querying efficiency to steal a classification model.
arXiv Detail & Related papers (2022-05-13T08:24:43Z) - Conservative Objective Models for Effective Offline Model-Based
Optimization [78.19085445065845]
Computational design problems arise in a number of settings, from synthetic biology to computer architectures.
We propose a method that learns a model of the objective function that lower bounds the actual value of the ground-truth objective on out-of-distribution inputs.
COMs are simple to implement and outperform a number of existing methods on a wide range of MBO problems.
arXiv Detail & Related papers (2021-07-14T17:55:28Z) - NASE: Learning Knowledge Graph Embedding for Link Prediction via Neural
Architecture Search [9.634626241415916]
Link prediction is the task of predicting missing connections between entities in the knowledge graph (KG)
Previous work has tried to use Automated Machine Learning (AutoML) to search for the best model for a given dataset.
We propose a novel Neural Architecture Search (NAS) framework for the link prediction task.
arXiv Detail & Related papers (2020-08-18T03:34:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.