A Nonlinear PID-Enhanced Adaptive Latent Factor Analysis Model
- URL: http://arxiv.org/abs/2208.02513v1
- Date: Thu, 4 Aug 2022 07:48:19 GMT
- Title: A Nonlinear PID-Enhanced Adaptive Latent Factor Analysis Model
- Authors: Jinli Li, Ye Yuan
- Abstract summary: High-dimensional and incomplete (HDI) data holds tremendous interactive information in various industrial applications.
A latent factor (LF) model is remarkably effective in extracting valuable information from HDI data with decent gradient (SGD) algorithm.
An SGD-based LFA model suffers from slow convergence since it only considers the current learning error.
- Score: 6.2303427193075755
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: High-dimensional and incomplete (HDI) data holds tremendous interactive
information in various industrial applications. A latent factor (LF) model is
remarkably effective in extracting valuable information from HDI data with
stochastic gradient decent (SGD) algorithm. However, an SGD-based LFA model
suffers from slow convergence since it only considers the current learning
error. To address this critical issue, this paper proposes a Nonlinear
PID-enhanced Adaptive Latent Factor (NPALF) model with two-fold ideas: 1)
rebuilding the learning error via considering the past learning errors
following the principle of a nonlinear PID controller; b) implementing all
parameters adaptation effectively following the principle of a particle swarm
optimization (PSO) algorithm. Experience results on four representative HDI
datasets indicate that compared with five state-of-the-art LFA models, the
NPALF model achieves better convergence rate and prediction accuracy for
missing data of an HDI data.
Related papers
- PSLF: A PID Controller-incorporated Second-order Latent Factor Analysis Model for Recommender System [11.650076383080526]
A second-order-based HDI model (SLF) analysis demonstrates superior performance in graph learning, particularly for high- and incomplete factor data rates.
arXiv Detail & Related papers (2024-08-31T13:01:58Z) - An ADRC-Incorporated Stochastic Gradient Descent Algorithm for Latent
Factor Analysis [6.843073158719234]
A gradient descent (SGD)-based latent factor analysis (LFA) model is remarkably effective in extracting valuable information from an HDI matrix.
A standard SGD algorithm only considers the current learning error to compute the gradient without considering the historical and future state of the learning error.
This paper innovatively proposes an ADRC-incorporated SGD (ADS) algorithm by refining the instance learning error by considering the historical and future state.
arXiv Detail & Related papers (2024-01-13T08:38:54Z) - A PAC-Bayesian Perspective on the Interpolating Information Criterion [54.548058449535155]
We show how a PAC-Bayes bound is obtained for a general class of models, characterizing factors which influence performance in the interpolating regime.
We quantify how the test error for overparameterized models achieving effectively zero training error depends on the quality of the implicit regularization imposed by e.g. the combination of model, parameter-initialization scheme.
arXiv Detail & Related papers (2023-11-13T01:48:08Z) - Robust Learning with Progressive Data Expansion Against Spurious
Correlation [65.83104529677234]
We study the learning process of a two-layer nonlinear convolutional neural network in the presence of spurious features.
Our analysis suggests that imbalanced data groups and easily learnable spurious features can lead to the dominance of spurious features during the learning process.
We propose a new training algorithm called PDE that efficiently enhances the model's robustness for a better worst-group performance.
arXiv Detail & Related papers (2023-06-08T05:44:06Z) - Fast Latent Factor Analysis via a Fuzzy PID-Incorporated Stochastic
Gradient Descent Algorithm [1.984879854062214]
A gradient descent (SGD)-based latent factor analysis model is remarkably effective in extracting valuable information from an HDI matrix.
A standard SGD algorithm learns a latent factor relying on the gradient of current instance error only without considering past update information.
This paper proposes a Fuzzy PID-incorporated SGD algorithm with two-fold ideas: 1) rebuilding the instance error by considering the past update information in an efficient way following the principle of PID, and 2) implementing hyper-learnings and gain adaptation following the fuzzy rules.
arXiv Detail & Related papers (2023-03-07T14:51:09Z) - PI-NLF: A Proportional-Integral Approach for Non-negative Latent Factor
Analysis [9.087387628717952]
A non-negative latent factor (NLF) model performs efficient representation learning to an HDI matrix.
A PI-NLF model outperforms the state-of-the-art models in both computational efficiency and estimation accuracy for missing data of an HDI matrix.
arXiv Detail & Related papers (2022-05-05T12:04:52Z) - Adaptive Divergence-based Non-negative Latent Factor Analysis [6.265179945530255]
This study presents an Adaptive Divergence-based Non-negative Latent Factor (ADNLF) model with three-fold ideas.
An ADNLF model achieves significantly higher estimation accuracy for missing data of an HDI dataset with high computational efficiency.
arXiv Detail & Related papers (2022-03-30T11:28:36Z) - Improving Generalization via Uncertainty Driven Perturbations [107.45752065285821]
We consider uncertainty-driven perturbations of the training data points.
Unlike loss-driven perturbations, uncertainty-guided perturbations do not cross the decision boundary.
We show that UDP is guaranteed to achieve the robustness margin decision on linear models.
arXiv Detail & Related papers (2022-02-11T16:22:08Z) - Adaptive Anomaly Detection for Internet of Things in Hierarchical Edge
Computing: A Contextual-Bandit Approach [81.5261621619557]
We propose an adaptive anomaly detection scheme with hierarchical edge computing (HEC)
We first construct multiple anomaly detection DNN models with increasing complexity, and associate each of them to a corresponding HEC layer.
Then, we design an adaptive model selection scheme that is formulated as a contextual-bandit problem and solved by using a reinforcement learning policy network.
arXiv Detail & Related papers (2021-08-09T08:45:47Z) - Understanding Overparameterization in Generative Adversarial Networks [56.57403335510056]
Generative Adversarial Networks (GANs) are used to train non- concave mini-max optimization problems.
A theory has shown the importance of the gradient descent (GD) to globally optimal solutions.
We show that in an overized GAN with a $1$-layer neural network generator and a linear discriminator, the GDA converges to a global saddle point of the underlying non- concave min-max problem.
arXiv Detail & Related papers (2021-04-12T16:23:37Z) - Decomposed Adversarial Learned Inference [118.27187231452852]
We propose a novel approach, Decomposed Adversarial Learned Inference (DALI)
DALI explicitly matches prior and conditional distributions in both data and code spaces.
We validate the effectiveness of DALI on the MNIST, CIFAR-10, and CelebA datasets.
arXiv Detail & Related papers (2020-04-21T20:00:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.