Self-Evolutionary Optimization for Pareto Front Learning
- URL: http://arxiv.org/abs/2110.03461v1
- Date: Thu, 7 Oct 2021 13:38:57 GMT
- Title: Self-Evolutionary Optimization for Pareto Front Learning
- Authors: Simyung Chang, KiYoon Yoo, Jiho Jang and Nojun Kwak
- Abstract summary: Multi-objective optimization (MOO) approaches have been proposed for multitasking problems.
Recent MOO methods approximate multiple optimal solutions (Pareto front) with a single unified model.
We show that PFL can be re-formulated into another MOO problem with multiple objectives, each of which corresponds to different preference weights for the tasks.
- Score: 34.17125297176668
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multi-task learning (MTL), which aims to improve performance by learning
multiple tasks simultaneously, inherently presents an optimization challenge
due to multiple objectives. Hence, multi-objective optimization (MOO)
approaches have been proposed for multitasking problems. Recent MOO methods
approximate multiple optimal solutions (Pareto front) with a single unified
model, which is collectively referred to as Pareto front learning (PFL). In
this paper, we show that PFL can be re-formulated into another MOO problem with
multiple objectives, each of which corresponds to different preference weights
for the tasks. We leverage an evolutionary algorithm (EA) to propose a method
for PFL called self-evolutionary optimization (SEO) by directly maximizing the
hypervolume. By using SEO, the neural network learns to approximate the Pareto
front conditioned on multiple hyper-parameters that drastically affect the
hypervolume. Then, by generating a population of approximations simply by
inferencing the network, the hyper-parameters of the network can be optimized
by EA. Utilizing SEO for PFL, we also introduce self-evolutionary Pareto
networks (SEPNet), enabling the unified model to approximate the entire Pareto
front set that maximizes the hypervolume. Extensive experimental results
confirm that SEPNet can find a better Pareto front than the current
state-of-the-art methods while minimizing the increase in model size and
training cost.
Related papers
- Unlearning as multi-task optimization: A normalized gradient difference approach with an adaptive learning rate [105.86576388991713]
We introduce a normalized gradient difference (NGDiff) algorithm, enabling us to have better control over the trade-off between the objectives.
We provide a theoretical analysis and empirically demonstrate the superior performance of NGDiff among state-of-the-art unlearning methods on the TOFU and MUSE datasets.
arXiv Detail & Related papers (2024-10-29T14:41:44Z) - Pareto Low-Rank Adapters: Efficient Multi-Task Learning with Preferences [49.14535254003683]
PaLoRA is a novel parameter-efficient method that augments the original model with task-specific low-rank adapters.
Our experimental results show that PaLoRA outperforms MTL and PFL baselines across various datasets.
arXiv Detail & Related papers (2024-07-10T21:25:51Z) - Towards Efficient Pareto Set Approximation via Mixture of Experts Based Model Fusion [53.33473557562837]
Solving multi-objective optimization problems for large deep neural networks is a challenging task due to the complexity of the loss landscape and the expensive computational cost.
We propose a practical and scalable approach to solve this problem via mixture of experts (MoE) based model fusion.
By ensembling the weights of specialized single-task models, the MoE module can effectively capture the trade-offs between multiple objectives.
arXiv Detail & Related papers (2024-06-14T07:16:18Z) - Intuition-aware Mixture-of-Rank-1-Experts for Parameter Efficient Finetuning [50.73666458313015]
Large Language Models (LLMs) have demonstrated significant potential in performing multiple tasks in multimedia applications.
MoE has been emerged as a promising solution with its sparse architecture for effective task decoupling.
Intuition-MoR1E achieves superior efficiency and 2.15% overall accuracy improvement across 14 public datasets.
arXiv Detail & Related papers (2024-04-13T12:14:58Z) - Interactive Hyperparameter Optimization in Multi-Objective Problems via
Preference Learning [65.51668094117802]
We propose a human-centered interactive HPO approach tailored towards multi-objective machine learning (ML)
Instead of relying on the user guessing the most suitable indicator for their needs, our approach automatically learns an appropriate indicator.
arXiv Detail & Related papers (2023-09-07T09:22:05Z) - Improving Pareto Front Learning via Multi-Sample Hypernetworks [4.129225533930966]
We propose a novel PFL framework namely PHN-HVI, which employs a hypernetwork to generate multiple solutions from a set of diverse trade-off preferences.
The experimental results on several MOO machine learning tasks show that the proposed framework significantly outperforms the baselines.
arXiv Detail & Related papers (2022-12-02T12:19:12Z) - Pareto Manifold Learning: Tackling multiple tasks via ensembles of
single-task models [50.33956216274694]
In Multi-Task Learning (MTL), tasks may compete and limit the performance achieved on each other, rather than guiding the optimization to a solution.
We propose textitPareto Manifold Learning, an ensembling method in weight space.
arXiv Detail & Related papers (2022-10-18T11:20:54Z) - Efficient Multi-Objective Optimization for Deep Learning [2.0305676256390934]
Multi-objective optimization (MOO) is a prevalent challenge for Deep Learning.
There exists no scalable MOO solution for truly deep neural networks.
arXiv Detail & Related papers (2021-03-24T17:59:42Z) - Learning the Pareto Front with Hypernetworks [44.72371822514582]
Multi-objective optimization (MOO) problems are prevalent in machine learning.
These problems have a set of optimal solutions, where each point on the front represents a different trade-off between possibly conflicting objectives.
Recent MOO methods can target a specific desired ray in loss space however, most approaches still face two grave limitations.
arXiv Detail & Related papers (2020-10-08T16:39:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.