Hierarchical Optimization-Derived Learning
- URL: http://arxiv.org/abs/2302.05587v2
- Date: Tue, 12 Sep 2023 13:52:55 GMT
- Title: Hierarchical Optimization-Derived Learning
- Authors: Risheng Liu, Xuan Liu, Shangzhi Zeng, Jin Zhang, and Yixuan Zhang
- Abstract summary: We establish a new framework, named Hierarchical ODL (HODL), to simultaneously investigate the intrinsic behaviors of optimization-derived model construction and its corresponding learning process.
This is the first theoretical guarantee for these two coupled ODL components: optimization and learning.
- Score: 58.69200830655009
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In recent years, by utilizing optimization techniques to formulate the
propagation of deep model, a variety of so-called Optimization-Derived Learning
(ODL) approaches have been proposed to address diverse learning and vision
tasks. Although having achieved relatively satisfying practical performance,
there still exist fundamental issues in existing ODL methods. In particular,
current ODL methods tend to consider model construction and learning as two
separate phases, and thus fail to formulate their underlying coupling and
depending relationship. In this work, we first establish a new framework, named
Hierarchical ODL (HODL), to simultaneously investigate the intrinsic behaviors
of optimization-derived model construction and its corresponding learning
process. Then we rigorously prove the joint convergence of these two sub-tasks,
from the perspectives of both approximation quality and stationary analysis. To
our best knowledge, this is the first theoretical guarantee for these two
coupled ODL components: optimization and learning. We further demonstrate the
flexibility of our framework by applying HODL to challenging learning tasks,
which have not been properly addressed by existing ODL methods. Finally, we
conduct extensive experiments on both synthetic data and real applications in
vision and other learning tasks to verify the theoretical properties and
practical performance of HODL in various application scenarios.
Related papers
- Aligned Multi Objective Optimization [14.320569438197271]
In machine learning practice, there are many scenarios where such conflict does not take place.
Recent findings from multi-task learning, reinforcement learning, and LLMs training show that diverse related tasks can enhance performance across objectives simultaneously.
We introduce the Aligned Multi-Objective Optimization framework, propose new algorithms for this setting, and provide theoretical guarantees of their superior performance.
arXiv Detail & Related papers (2025-02-19T20:50:03Z) - Learning Provably Improves the Convergence of Gradient Descent [9.82454981262489]
We study the convergence of Learning to Optimize (L2O) problems by training-based solvers.
An algorithm's tangent significantly enhances L2O's convergence.
Our findings indicate 50% outperformance over the GD methods.
arXiv Detail & Related papers (2025-01-30T02:03:30Z) - A Survey of Optimization Methods for Training DL Models: Theoretical Perspective on Convergence and Generalization [11.072619355813496]
We provide an extensive summary of theoretical foundations of optimization methods in deep learning (DL)
This paper includes theoretical analysis of popular gradient-based first-order second-order generalization methods.
We also discuss the analysis of the generic convex loss and explicitly encourage the discovery of well-generalizing optimal points.
arXiv Detail & Related papers (2025-01-24T12:42:38Z) - DaRec: A Disentangled Alignment Framework for Large Language Model and Recommender System [83.34921966305804]
Large language models (LLMs) have demonstrated remarkable performance in recommender systems.
We propose a novel plug-and-play alignment framework for LLMs and collaborative models.
Our method is superior to existing state-of-the-art algorithms.
arXiv Detail & Related papers (2024-08-15T15:56:23Z) - Faithful Explanations of Black-box NLP Models Using LLM-generated
Counterfactuals [67.64770842323966]
Causal explanations of predictions of NLP systems are essential to ensure safety and establish trust.
Existing methods often fall short of explaining model predictions effectively or efficiently.
We propose two approaches for counterfactual (CF) approximation.
arXiv Detail & Related papers (2023-10-01T07:31:04Z) - Latent Variable Representation for Reinforcement Learning [131.03944557979725]
It remains unclear theoretically and empirically how latent variable models may facilitate learning, planning, and exploration to improve the sample efficiency of model-based reinforcement learning.
We provide a representation view of the latent variable models for state-action value functions, which allows both tractable variational learning algorithm and effective implementation of the optimism/pessimism principle.
In particular, we propose a computationally efficient planning algorithm with UCB exploration by incorporating kernel embeddings of latent variable models.
arXiv Detail & Related papers (2022-12-17T00:26:31Z) - Scalable PAC-Bayesian Meta-Learning via the PAC-Optimal Hyper-Posterior:
From Theory to Practice [54.03076395748459]
A central question in the meta-learning literature is how to regularize to ensure generalization to unseen tasks.
We present a generalization bound for meta-learning, which was first derived by Rothfuss et al.
We provide a theoretical analysis and empirical case study under which conditions and to what extent these guarantees for meta-learning improve upon PAC-Bayesian per-task learning bounds.
arXiv Detail & Related papers (2022-11-14T08:51:04Z) - Optimization-Derived Learning with Essential Convergence Analysis of
Training and Hyper-training [52.39882976848064]
We design a Generalized Krasnoselskii-Mann (GKM) scheme based on fixed-point iterations as our fundamental ODL module.
Under the GKM scheme, a Bilevel Meta Optimization (BMO) algorithmic framework is constructed to solve the optimal training and hyper-training variables together.
arXiv Detail & Related papers (2022-06-16T01:50:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.