An Adam-enhanced Particle Swarm Optimizer for Latent Factor Analysis
- URL: http://arxiv.org/abs/2302.11956v1
- Date: Thu, 23 Feb 2023 12:10:59 GMT
- Title: An Adam-enhanced Particle Swarm Optimizer for Latent Factor Analysis
- Authors: Jia Chen, Renyu Zhang, Yuanyi Liu
- Abstract summary: We propose an Adam-enhanced Hierarchical PSO-LFA model, which refines the latent factors with a sequential PSO algorithm.
The experimental results on four real datasets demonstrate that our proposed model achieves higher prediction accuracy with its peers.
- Score: 6.960453648000231
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Digging out the latent information from large-scale incomplete matrices is a
key issue with challenges. The Latent Factor Analysis (LFA) model has been
investigated in depth to an alyze the latent information. Recently, Swarm
Intelligence-related LFA models have been proposed and adopted widely to
improve the optimization process of LFA with high efficiency, i.e., the
Particle Swarm Optimization (PSO)-LFA model. However, the hyper-parameters of
the PSO-LFA model have to tune manually, which is inconvenient for widely
adoption and limits the learning rate as a fixed value. To address this issue,
we propose an Adam-enhanced Hierarchical PSO-LFA model, which refines the
latent factors with a sequential Adam-adjusting hyper-parameters PSO algorithm.
First, we design the Adam incremental vector for a particle and construct the
Adam-enhanced evolution process for particles. Second, we refine all the latent
factors of the target matrix sequentially with our proposed Adam-enhanced PSO's
process. The experimental results on four real datasets demonstrate that our
proposed model achieves higher prediction accuracy with its peers.
Related papers
- SMILE: Zero-Shot Sparse Mixture of Low-Rank Experts Construction From Pre-Trained Foundation Models [85.67096251281191]
We present an innovative approach to model fusion called zero-shot Sparse MIxture of Low-rank Experts (SMILE) construction.
SMILE allows for the upscaling of source models into an MoE model without extra data or further training.
We conduct extensive experiments across diverse scenarios, such as image classification and text generation tasks, using full fine-tuning and LoRA fine-tuning.
arXiv Detail & Related papers (2024-08-19T17:32:15Z) - The Hitchhiker's Guide to Human Alignment with *PO [43.4130314879284]
We focus on identifying the algorithm that, while being performant, is simultaneously more robust to varying hyper parameters.
Our analysis reveals that the widely adopted DPO method consistently produces lengthy responses of inferior quality.
Motivated by these findings, we propose an embarrassingly simple extension to the DPO algorithm, LN-DPO, resulting in more concise responses without sacrificing quality.
arXiv Detail & Related papers (2024-07-21T17:35:20Z) - FAdam: Adam is a natural gradient optimizer using diagonal empirical Fisher information [5.010523239708004]
We provide an accessible and detailed analysis of the diagonal empirical Fisher information matrix (FIM) in Adam.
Our analysis uncovers flaws in the original Adam algorithm, leading to proposed corrections.
Our modified algorithm, Fisher Adam (FAdam), demonstrates superior performance across diverse domains.
arXiv Detail & Related papers (2024-05-21T13:58:17Z) - Entropy-Regularized Token-Level Policy Optimization for Language Agent Reinforcement [67.1393112206885]
Large Language Models (LLMs) have shown promise as intelligent agents in interactive decision-making tasks.
We introduce Entropy-Regularized Token-level Policy Optimization (ETPO), an entropy-augmented RL method tailored for optimizing LLMs at the token level.
We assess the effectiveness of ETPO within a simulated environment that models data science code generation as a series of multi-step interactive tasks.
arXiv Detail & Related papers (2024-02-09T07:45:26Z) - Secrets of RLHF in Large Language Models Part I: PPO [81.01936993929127]
Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence.
reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit.
In this report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training.
arXiv Detail & Related papers (2023-07-11T01:55:24Z) - To Repeat or Not To Repeat: Insights from Scaling LLM under Token-Crisis [50.31589712761807]
Large language models (LLMs) are notoriously token-hungry during pre-training, and high-quality text data on the web is approaching its scaling limit for LLMs.
We investigate the consequences of repeating pre-training data, revealing that the model is susceptible to overfitting.
Second, we examine the key factors contributing to multi-epoch degradation, finding that significant factors include dataset size, model parameters, and training objectives.
arXiv Detail & Related papers (2023-05-22T17:02:15Z) - Prediction-Oriented Bayesian Active Learning [51.426960808684655]
Expected predictive information gain (EPIG) is an acquisition function that measures information gain in the space of predictions rather than parameters.
EPIG leads to stronger predictive performance compared with BALD across a range of datasets and models.
arXiv Detail & Related papers (2023-04-17T10:59:57Z) - A Dynamic-Neighbor Particle Swarm Optimizer for Accurate Latent Factor
Analysis [8.451827165005993]
The performance of an LFA model heavily rely on its optimization process.
Some prior studies employ the Particle Swarm Optimization to enhance an LFA model's optimization process.
This paper proposes a Dynamic-neighbor-cooperated Hierarchical PSO-enhanced LFA model with two-fold main ideas.
arXiv Detail & Related papers (2023-02-23T12:03:59Z) - An Adam-adjusting-antennae BAS Algorithm for Refining Latent Factors [10.129747457073389]
We propose a sequential Adam-adjusting-antennae BAS (A2BAS) optimization algorithm, which refines the latent factors obtained by the PSO-incorporated LFA model.
With experimental results on two real high-dimensional matrices, we demonstrate that our algorithm can effectively solve the premature convergence issue.
arXiv Detail & Related papers (2022-08-13T09:05:05Z) - Adaptive Latent Factor Analysis via Generalized Momentum-Incorporated
Particle Swarm Optimization [6.2303427193075755]
A gradient descent (SGD) algorithm is an effective learning strategy to build a latent factor analysis (LFA) model on a high-dimensional and incomplete (HDI) matrix.
A particle swarm optimization (PSO) algorithm is commonly adopted to make an SGD-based LFA model's hyper- parameters, i.e., learning rate and regularization coefficient, self-adaptation.
This paper incorporates more historical information into each particle's evolutionary process for avoiding premature convergence.
arXiv Detail & Related papers (2022-08-04T03:15:07Z) - Understanding Overparameterization in Generative Adversarial Networks [56.57403335510056]
Generative Adversarial Networks (GANs) are used to train non- concave mini-max optimization problems.
A theory has shown the importance of the gradient descent (GD) to globally optimal solutions.
We show that in an overized GAN with a $1$-layer neural network generator and a linear discriminator, the GDA converges to a global saddle point of the underlying non- concave min-max problem.
arXiv Detail & Related papers (2021-04-12T16:23:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.