Pareto Navigation Gradient Descent: a First-Order Algorithm for
Optimization in Pareto Set
- URL: http://arxiv.org/abs/2110.08713v1
- Date: Sun, 17 Oct 2021 04:07:04 GMT
- Title: Pareto Navigation Gradient Descent: a First-Order Algorithm for
Optimization in Pareto Set
- Authors: Mao Ye, Qiang Liu
- Abstract summary: Modern machine learning applications, such as multi-task learning, require finding optimal model parameters to trade-off multiple objective functions.
We propose a first-order algorithm that approximately solves OPT-in-Pareto using only gradient information.
- Score: 17.617944390196286
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many modern machine learning applications, such as multi-task learning,
require finding optimal model parameters to trade-off multiple objective
functions that may conflict with each other. The notion of the Pareto set
allows us to focus on the set of (often infinite number of) models that cannot
be strictly improved. But it does not provide an actionable procedure for
picking one or a few special models to return to practical users. In this
paper, we consider \emph{optimization in Pareto set (OPT-in-Pareto)}, the
problem of finding Pareto models that optimize an extra reference criterion
function within the Pareto set. This function can either encode a specific
preference from the users, or represent a generic diversity measure for
obtaining a set of diversified Pareto models that are representative of the
whole Pareto set. Unfortunately, despite being a highly useful framework,
efficient algorithms for OPT-in-Pareto have been largely missing, especially
for large-scale, non-convex, and non-linear objectives in deep learning. A
naive approach is to apply Riemannian manifold gradient descent on the Pareto
set, which yields a high computational cost due to the need for
eigen-calculation of Hessian matrices. We propose a first-order algorithm that
approximately solves OPT-in-Pareto using only gradient information, with both
high practical efficiency and theoretically guaranteed convergence property.
Empirically, we demonstrate that our method works efficiently for a variety of
challenging multi-task-related problems.
Related papers
- Preference-Optimized Pareto Set Learning for Blackbox Optimization [1.9628841617148691]
No single solution exists that can optimize all the objectives simultaneously.
In a typical MOO problem, the goal is to find a set of optimum solutions (Pareto set) that trades off the preferences among objectives.
Our formulation leads to a bilevel optimization problem that can be solved by e.g. differentiable cross-entropy methods.
arXiv Detail & Related papers (2024-08-19T13:23:07Z) - Efficient Pareto Manifold Learning with Low-Rank Structure [31.082432589391953]
Multi-task learning is inherently a multi-objective optimization problem.
We propose a novel approach that integrates a main network with several low-rank matrices.
It significantly reduces the number of parameters and facilitates the extraction of shared features.
arXiv Detail & Related papers (2024-07-30T11:09:27Z) - Towards Efficient Pareto Set Approximation via Mixture of Experts Based Model Fusion [53.33473557562837]
Solving multi-objective optimization problems for large deep neural networks is a challenging task due to the complexity of the loss landscape and the expensive computational cost.
We propose a practical and scalable approach to solve this problem via mixture of experts (MoE) based model fusion.
By ensembling the weights of specialized single-task models, the MoE module can effectively capture the trade-offs between multiple objectives.
arXiv Detail & Related papers (2024-06-14T07:16:18Z) - MAP: Low-compute Model Merging with Amortized Pareto Fronts via Quadratic Approximation [80.47072100963017]
We introduce a novel and low-compute algorithm, Model Merging with Amortized Pareto Front (MAP)
MAP efficiently identifies a set of scaling coefficients for merging multiple models, reflecting the trade-offs involved.
We also introduce Bayesian MAP for scenarios with a relatively low number of tasks and Nested MAP for situations with a high number of tasks, further reducing the computational cost of evaluation.
arXiv Detail & Related papers (2024-06-11T17:55:25Z) - UCB-driven Utility Function Search for Multi-objective Reinforcement Learning [75.11267478778295]
In Multi-objective Reinforcement Learning (MORL) agents are tasked with optimising decision-making behaviours.
We focus on the case of linear utility functions parameterised by weight vectors w.
We introduce a method based on Upper Confidence Bound to efficiently search for the most promising weight vectors during different stages of the learning process.
arXiv Detail & Related papers (2024-05-01T09:34:42Z) - MultiZenoTravel: a Tunable Benchmark for Multi-Objective Planning with
Known Pareto Front [71.19090689055054]
Multi-objective AI planning suffers from a lack of benchmarks exhibiting known Pareto Fronts.
We propose a tunable benchmark generator, together with a dedicated solver that provably computes the true Pareto front of the resulting instances.
We show how to characterize the optimal plans for a constrained version of the problem, and then show how to reduce the general problem to the constrained one.
arXiv Detail & Related papers (2023-04-28T07:09:23Z) - Pareto Manifold Learning: Tackling multiple tasks via ensembles of
single-task models [50.33956216274694]
In Multi-Task Learning (MTL), tasks may compete and limit the performance achieved on each other, rather than guiding the optimization to a solution.
We propose textitPareto Manifold Learning, an ensembling method in weight space.
arXiv Detail & Related papers (2022-10-18T11:20:54Z) - Pareto Set Learning for Expensive Multi-Objective Optimization [5.419608513284392]
Expensive multi-objective optimization problems can be found in many real-world applications.
This paper develops a novel learning-based method to approximate the whole Pareto set for MOBO.
arXiv Detail & Related papers (2022-10-16T09:41:54Z) - A Hybrid 2-stage Neural Optimization for Pareto Front Extraction [3.918940900258555]
A major obstacle to optimal trade-off solutions is that they don't always converge to each other.
We propose a two-stage approach that is accurate and cost-effective.
arXiv Detail & Related papers (2021-01-27T20:56:19Z) - Pareto Multi-Task Learning [53.90732663046125]
Multi-task learning is a powerful method for solving multiple correlated tasks simultaneously.
It is often impossible to find one single solution to optimize all the tasks, since different tasks might conflict with each other.
Recently, a novel method is proposed to find one single Pareto optimal solution with good trade-off among different tasks by casting multi-task learning as multiobjective optimization.
arXiv Detail & Related papers (2019-12-30T08:58:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.