Pareto Manifold Learning: Tackling multiple tasks via ensembles of
single-task models
- URL: http://arxiv.org/abs/2210.09759v2
- Date: Wed, 14 Jun 2023 11:45:45 GMT
- Title: Pareto Manifold Learning: Tackling multiple tasks via ensembles of
single-task models
- Authors: Nikolaos Dimitriadis, Pascal Frossard, Fran\c{c}ois Fleuret
- Abstract summary: In Multi-Task Learning (MTL), tasks may compete and limit the performance achieved on each other, rather than guiding the optimization to a solution.
We propose textitPareto Manifold Learning, an ensembling method in weight space.
- Score: 50.33956216274694
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In Multi-Task Learning (MTL), tasks may compete and limit the performance
achieved on each other, rather than guiding the optimization to a solution,
superior to all its single-task trained counterparts. Since there is often not
a unique solution optimal for all tasks, practitioners have to balance
tradeoffs between tasks' performance, and resort to optimality in the Pareto
sense. Most MTL methodologies either completely neglect this aspect, and
instead of aiming at learning a Pareto Front, produce one solution predefined
by their optimization schemes, or produce diverse but discrete solutions.
Recent approaches parameterize the Pareto Front via neural networks, leading to
complex mappings from tradeoff to objective space. In this paper, we conjecture
that the Pareto Front admits a linear parameterization in parameter space,
which leads us to propose \textit{Pareto Manifold Learning}, an ensembling
method in weight space. Our approach produces a continuous Pareto Front in a
single training run, that allows to modulate the performance on each task
during inference. Experiments on multi-task learning benchmarks, ranging from
image classification to tabular datasets and scene understanding, show that
\textit{Pareto Manifold Learning} outperforms state-of-the-art single-point
algorithms, while learning a better Pareto parameterization than multi-point
baselines.
Related papers
- Efficient Pareto Manifold Learning with Low-Rank Structure [31.082432589391953]
Multi-task learning is inherently a multi-objective optimization problem.
We propose a novel approach that integrates a main network with several low-rank matrices.
It significantly reduces the number of parameters and facilitates the extraction of shared features.
arXiv Detail & Related papers (2024-07-30T11:09:27Z) - Pareto Low-Rank Adapters: Efficient Multi-Task Learning with Preferences [49.14535254003683]
PaLoRA is a novel parameter-efficient method that augments the original model with task-specific low-rank adapters.
Our experimental results show that PaLoRA outperforms MTL and PFL baselines across various datasets.
arXiv Detail & Related papers (2024-07-10T21:25:51Z) - Towards Efficient Pareto Set Approximation via Mixture of Experts Based Model Fusion [53.33473557562837]
Solving multi-objective optimization problems for large deep neural networks is a challenging task due to the complexity of the loss landscape and the expensive computational cost.
We propose a practical and scalable approach to solve this problem via mixture of experts (MoE) based model fusion.
By ensembling the weights of specialized single-task models, the MoE module can effectively capture the trade-offs between multiple objectives.
arXiv Detail & Related papers (2024-06-14T07:16:18Z) - MAP: Low-compute Model Merging with Amortized Pareto Fronts via Quadratic Approximation [80.47072100963017]
We introduce a novel and low-compute algorithm, Model Merging with Amortized Pareto Front (MAP)
MAP efficiently identifies a set of scaling coefficients for merging multiple models, reflecting the trade-offs involved.
We also introduce Bayesian MAP for scenarios with a relatively low number of tasks and Nested MAP for situations with a high number of tasks, further reducing the computational cost of evaluation.
arXiv Detail & Related papers (2024-06-11T17:55:25Z) - Multi-Task Learning with Multi-Task Optimization [31.518330903602095]
We show that a set of optimized yet well-distributed models embody different trade-offs in one algorithmic pass.
We investigate the proposed multi-task learning with multi-task optimization for solving various problem settings.
arXiv Detail & Related papers (2024-03-24T14:04:40Z) - Pareto Navigation Gradient Descent: a First-Order Algorithm for
Optimization in Pareto Set [17.617944390196286]
Modern machine learning applications, such as multi-task learning, require finding optimal model parameters to trade-off multiple objective functions.
We propose a first-order algorithm that approximately solves OPT-in-Pareto using only gradient information.
arXiv Detail & Related papers (2021-10-17T04:07:04Z) - Small Towers Make Big Differences [59.243296878666285]
Multi-task learning aims at solving multiple machine learning tasks at the same time.
A good solution to a multi-task learning problem should be generalizable in addition to being Pareto optimal.
We propose a method of under- parameterized self-auxiliaries for multi-task models to achieve the best of both worlds.
arXiv Detail & Related papers (2020-08-13T10:45:31Z) - Efficient Continuous Pareto Exploration in Multi-Task Learning [34.41682709915956]
We present a novel, efficient method for continuous analysis of optimal solutions in machine learning problems.
We scale up theoretical results in multi-objective optimization to modern machine learning problems by proposing a sample-based sparse linear system.
arXiv Detail & Related papers (2020-06-29T23:36:20Z) - Pareto Multi-Task Learning [53.90732663046125]
Multi-task learning is a powerful method for solving multiple correlated tasks simultaneously.
It is often impossible to find one single solution to optimize all the tasks, since different tasks might conflict with each other.
Recently, a novel method is proposed to find one single Pareto optimal solution with good trade-off among different tasks by casting multi-task learning as multiobjective optimization.
arXiv Detail & Related papers (2019-12-30T08:58:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.