Graph-incorporated Latent Factor Analysis for High-dimensional and
Sparse Matrices
- URL: http://arxiv.org/abs/2204.07818v1
- Date: Sat, 16 Apr 2022 15:04:34 GMT
- Title: Graph-incorporated Latent Factor Analysis for High-dimensional and
Sparse Matrices
- Authors: Di Wu, Yi He, Xin Luo
- Abstract summary: A High-dimensional and sparse (HiDS) matrix is frequently encountered in a big data-related application like an e-commerce system or a social network services system.
This paper proposes a graph-incorporated latent factor analysis (GLFA) model to perform representation learning on HiDS matrix.
Experimental results on three real-world datasets demonstrate that GLFA outperforms six state-of-the-art models in predicting the missing data of an HiDS matrix.
- Score: 9.51012204233452
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A High-dimensional and sparse (HiDS) matrix is frequently encountered in a
big data-related application like an e-commerce system or a social network
services system. To perform highly accurate representation learning on it is of
great significance owing to the great desire of extracting latent knowledge and
patterns from it. Latent factor analysis (LFA), which represents an HiDS matrix
by learning the low-rank embeddings based on its observed entries only, is one
of the most effective and efficient approaches to this issue. However, most
existing LFA-based models perform such embeddings on a HiDS matrix directly
without exploiting its hidden graph structures, thereby resulting in accuracy
loss. To address this issue, this paper proposes a graph-incorporated latent
factor analysis (GLFA) model. It adopts two-fold ideas: 1) a graph is
constructed for identifying the hidden high-order interaction (HOI) among nodes
described by an HiDS matrix, and 2) a recurrent LFA structure is carefully
designed with the incorporation of HOI, thereby improving the representa-tion
learning ability of a resultant model. Experimental results on three real-world
datasets demonstrate that GLFA outperforms six state-of-the-art models in
predicting the missing data of an HiDS matrix, which evidently supports its
strong representation learning ability to HiDS data.
Related papers
- Web-Scale Visual Entity Recognition: An LLM-Driven Data Approach [56.55633052479446]
Web-scale visual entity recognition presents significant challenges due to the lack of clean, large-scale training data.
We propose a novel methodology to curate such a dataset, leveraging a multimodal large language model (LLM) for label verification, metadata generation, and rationale explanation.
Experiments demonstrate that models trained on this automatically curated data achieve state-of-the-art performance on web-scale visual entity recognition tasks.
arXiv Detail & Related papers (2024-10-31T06:55:24Z) - Disentangled Generative Graph Representation Learning [51.59824683232925]
This paper introduces DiGGR (Disentangled Generative Graph Representation Learning), a self-supervised learning framework.
It aims to learn latent disentangled factors and utilize them to guide graph mask modeling.
Experiments on 11 public datasets for two different graph learning tasks demonstrate that DiGGR consistently outperforms many previous self-supervised methods.
arXiv Detail & Related papers (2024-08-24T05:13:02Z) - Mini-Hes: A Parallelizable Second-order Latent Factor Analysis Model [8.06111903129142]
This paper proposes a miniblock diagonal hessian-free (Mini-Hes) optimization for building an LFA model.
Experiment results indicate that, with Mini-Hes, the LFA model outperforms several state-of-the-art models in addressing missing data estimation task.
arXiv Detail & Related papers (2024-02-19T08:43:00Z) - A Dynamic Linear Bias Incorporation Scheme for Nonnegative Latent Factor
Analysis [5.029743143286546]
HDI data is commonly encountered in big data-related applications like social network services systems.
Nonnegative Latent Factor Analysis (NLFA) models have proven to possess the superiority to address this issue.
This paper innovatively presents the dynamic linear bias incorporation scheme.
arXiv Detail & Related papers (2023-09-19T13:48:26Z) - Challenging the Myth of Graph Collaborative Filtering: a Reasoned and Reproducibility-driven Analysis [50.972595036856035]
We present a code that successfully replicates results from six popular and recent graph recommendation models.
We compare these graph models with traditional collaborative filtering models that historically performed well in offline evaluations.
By investigating the information flow from users' neighborhoods, we aim to identify which models are influenced by intrinsic features in the dataset structure.
arXiv Detail & Related papers (2023-08-01T09:31:44Z) - Fast Latent Factor Analysis via a Fuzzy PID-Incorporated Stochastic
Gradient Descent Algorithm [1.984879854062214]
A gradient descent (SGD)-based latent factor analysis model is remarkably effective in extracting valuable information from an HDI matrix.
A standard SGD algorithm learns a latent factor relying on the gradient of current instance error only without considering past update information.
This paper proposes a Fuzzy PID-incorporated SGD algorithm with two-fold ideas: 1) rebuilding the instance error by considering the past update information in an efficient way following the principle of PID, and 2) implementing hyper-learnings and gain adaptation following the fuzzy rules.
arXiv Detail & Related papers (2023-03-07T14:51:09Z) - Data-Free Adversarial Knowledge Distillation for Graph Neural Networks [62.71646916191515]
We propose the first end-to-end framework for data-free adversarial knowledge distillation on graph structured data (DFAD-GNN)
To be specific, our DFAD-GNN employs a generative adversarial network, which mainly consists of three components: a pre-trained teacher model and a student model are regarded as two discriminators, and a generator is utilized for deriving training graphs to distill knowledge from the teacher model into the student model.
Our DFAD-GNN significantly surpasses state-of-the-art data-free baselines in the graph classification task.
arXiv Detail & Related papers (2022-05-08T08:19:40Z) - A Multi-Metric Latent Factor Model for Analyzing High-Dimensional and
Sparse data [11.800988109180285]
High-dimensional and sparse (HiDS) matrices are omnipresent in a variety of big data-related applications.
Current LFA-based models mainly focus on a single-metric representation, where the representation strategy designed for the approximation Loss function is fixed and exclusive.
We in this paper propose a multi-metric latent factor (MMLF) model.
Our proposed MMLF enjoys the merits originated from a set of disparate metric spaces all at once, achieving the comprehensive and unbiased representation of HiDS matrices.
arXiv Detail & Related papers (2022-04-16T15:08:00Z) - A Differential Evolution-Enhanced Latent Factor Analysis Model for
High-dimensional and Sparse Data [11.164847043777703]
This paper proposes a Sequential-Group-Differential- Evolution (SGDE) algorithm to refine the latent factors optimized by a PLFA model.
As demonstrated by the experiments on four HiDS matrices, a SGDE-PLFA model outperforms the state-of-the-art models.
arXiv Detail & Related papers (2022-04-02T13:41:19Z) - Learning Bijective Feature Maps for Linear ICA [73.85904548374575]
We show that existing probabilistic deep generative models (DGMs) which are tailor-made for image data, underperform on non-linear ICA tasks.
To address this, we propose a DGM which combines bijective feature maps with a linear ICA model to learn interpretable latent structures for high-dimensional data.
We create models that converge quickly, are easy to train, and achieve better unsupervised latent factor discovery than flow-based models, linear ICA, and Variational Autoencoders on images.
arXiv Detail & Related papers (2020-02-18T17:58:07Z) - Revisiting Graph based Collaborative Filtering: A Linear Residual Graph
Convolutional Network Approach [55.44107800525776]
Graph Convolutional Networks (GCNs) are state-of-the-art graph based representation learning models.
In this paper, we revisit GCN based Collaborative Filtering (CF) based Recommender Systems (RS)
We show that removing non-linearities would enhance recommendation performance, consistent with the theories in simple graph convolutional networks.
We propose a residual network structure that is specifically designed for CF with user-item interaction modeling.
arXiv Detail & Related papers (2020-01-28T04:41:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.