Hierarchical Optimization-Derived Learning
- URL: http://arxiv.org/abs/2302.05587v2
- Date: Tue, 12 Sep 2023 13:52:55 GMT
- Title: Hierarchical Optimization-Derived Learning
- Authors: Risheng Liu, Xuan Liu, Shangzhi Zeng, Jin Zhang, and Yixuan Zhang
- Abstract summary: We establish a new framework, named Hierarchical ODL (HODL), to simultaneously investigate the intrinsic behaviors of optimization-derived model construction and its corresponding learning process.
This is the first theoretical guarantee for these two coupled ODL components: optimization and learning.
- Score: 58.69200830655009
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In recent years, by utilizing optimization techniques to formulate the
propagation of deep model, a variety of so-called Optimization-Derived Learning
(ODL) approaches have been proposed to address diverse learning and vision
tasks. Although having achieved relatively satisfying practical performance,
there still exist fundamental issues in existing ODL methods. In particular,
current ODL methods tend to consider model construction and learning as two
separate phases, and thus fail to formulate their underlying coupling and
depending relationship. In this work, we first establish a new framework, named
Hierarchical ODL (HODL), to simultaneously investigate the intrinsic behaviors
of optimization-derived model construction and its corresponding learning
process. Then we rigorously prove the joint convergence of these two sub-tasks,
from the perspectives of both approximation quality and stationary analysis. To
our best knowledge, this is the first theoretical guarantee for these two
coupled ODL components: optimization and learning. We further demonstrate the
flexibility of our framework by applying HODL to challenging learning tasks,
which have not been properly addressed by existing ODL methods. Finally, we
conduct extensive experiments on both synthetic data and real applications in
vision and other learning tasks to verify the theoretical properties and
practical performance of HODL in various application scenarios.
Related papers
- A Unifying Framework for Action-Conditional Self-Predictive Reinforcement Learning [48.59516337905877]
Learning a good representation is a crucial challenge for Reinforcement Learning (RL) agents.
Recent work has developed theoretical insights into these algorithms.
We take a step towards bridging the gap between theory and practice by analyzing an action-conditional self-predictive objective.
arXiv Detail & Related papers (2024-06-04T07:22:12Z) - Faithful Explanations of Black-box NLP Models Using LLM-generated
Counterfactuals [67.64770842323966]
Causal explanations of predictions of NLP systems are essential to ensure safety and establish trust.
Existing methods often fall short of explaining model predictions effectively or efficiently.
We propose two approaches for counterfactual (CF) approximation.
arXiv Detail & Related papers (2023-10-01T07:31:04Z) - Latent Variable Representation for Reinforcement Learning [131.03944557979725]
It remains unclear theoretically and empirically how latent variable models may facilitate learning, planning, and exploration to improve the sample efficiency of model-based reinforcement learning.
We provide a representation view of the latent variable models for state-action value functions, which allows both tractable variational learning algorithm and effective implementation of the optimism/pessimism principle.
In particular, we propose a computationally efficient planning algorithm with UCB exploration by incorporating kernel embeddings of latent variable models.
arXiv Detail & Related papers (2022-12-17T00:26:31Z) - Scalable PAC-Bayesian Meta-Learning via the PAC-Optimal Hyper-Posterior:
From Theory to Practice [54.03076395748459]
A central question in the meta-learning literature is how to regularize to ensure generalization to unseen tasks.
We present a generalization bound for meta-learning, which was first derived by Rothfuss et al.
We provide a theoretical analysis and empirical case study under which conditions and to what extent these guarantees for meta-learning improve upon PAC-Bayesian per-task learning bounds.
arXiv Detail & Related papers (2022-11-14T08:51:04Z) - Task-Free Continual Learning via Online Discrepancy Distance Learning [11.540150938141034]
This paper develops a new theoretical analysis framework which provides generalization bounds based on the discrepancy distance between the visited samples and the entire information made available for training the model.
Inspired by this theoretical model, we propose a new approach enabled by the dynamic component expansion mechanism for a mixture model, namely the Online Discrepancy Distance Learning (ODDL)
arXiv Detail & Related papers (2022-10-12T20:44:09Z) - Optimization-Derived Learning with Essential Convergence Analysis of
Training and Hyper-training [52.39882976848064]
We design a Generalized Krasnoselskii-Mann (GKM) scheme based on fixed-point iterations as our fundamental ODL module.
Under the GKM scheme, a Bilevel Meta Optimization (BMO) algorithmic framework is constructed to solve the optimal training and hyper-training variables together.
arXiv Detail & Related papers (2022-06-16T01:50:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.