Follow the bisector: a simple method for multi-objective optimization
- URL: http://arxiv.org/abs/2007.06937v1
- Date: Tue, 14 Jul 2020 09:50:33 GMT
- Title: Follow the bisector: a simple method for multi-objective optimization
- Authors: Alexandr Katrutsa, Daniil Merkulov, Nurislam Tursynbek and Ivan
Oseledets
- Abstract summary: We consider optimization problems, where multiple differentiable losses have to be minimized.
The presented method computes descent direction in every iteration to guarantee equal relative decrease of objective functions.
- Score: 65.83318707752385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This study presents a novel Equiangular Direction Method (EDM) to solve a
multi-objective optimization problem. We consider optimization problems, where
multiple differentiable losses have to be minimized. The presented method
computes descent direction in every iteration to guarantee equal relative
decrease of objective functions. This descent direction is based on the
normalized gradients of the individual losses. Therefore, it is appropriate to
solve multi-objective optimization problems with multi-scale losses. We test
the proposed method on the imbalanced classification problem and multi-task
learning problem, where standard datasets are used. EDM is compared with other
methods to solve these problems.
Related papers
- UCB-driven Utility Function Search for Multi-objective Reinforcement Learning [75.11267478778295]
In Multi-objective Reinforcement Learning (MORL) agents are tasked with optimising decision-making behaviours.
We focus on the case of linear utility functions parameterised by weight vectors w.
We introduce a method based on Upper Confidence Bound to efficiently search for the most promising weight vectors during different stages of the learning process.
arXiv Detail & Related papers (2024-05-01T09:34:42Z) - Smooth Tchebycheff Scalarization for Multi-Objective Optimization [15.047246588148495]
Multi-objective optimization problems can be found in many real-world applications, where the objectives often conflict each other and cannot be optimized by a single solution.
We propose a lightweight and efficient smooth Tchebycheff scalarization approach for gradient-based multi-objective optimization.
arXiv Detail & Related papers (2024-02-29T12:03:05Z) - Analyzing and Enhancing the Backward-Pass Convergence of Unrolled
Optimization [50.38518771642365]
The integration of constrained optimization models as components in deep networks has led to promising advances on many specialized learning tasks.
A central challenge in this setting is backpropagation through the solution of an optimization problem, which often lacks a closed form.
This paper provides theoretical insights into the backward pass of unrolled optimization, showing that it is equivalent to the solution of a linear system by a particular iterative method.
A system called Folded Optimization is proposed to construct more efficient backpropagation rules from unrolled solver implementations.
arXiv Detail & Related papers (2023-12-28T23:15:18Z) - Linearization Algorithms for Fully Composite Optimization [61.20539085730636]
This paper studies first-order algorithms for solving fully composite optimization problems convex compact sets.
We leverage the structure of the objective by handling differentiable and non-differentiable separately, linearizing only the smooth parts.
arXiv Detail & Related papers (2023-02-24T18:41:48Z) - Optimizing Evaluation Metrics for Multi-Task Learning via the
Alternating Direction Method of Multipliers [12.227732834969336]
Multi-task learning (MTL) aims to improve the generalization performance of multiple tasks by exploiting the shared factors among them.
Most existing MTL methods try to minimize either the misclassified errors for classification or the mean squared errors for regression.
We propose a method to directly optimize the evaluation metrics for a large family of MTL problems.
arXiv Detail & Related papers (2022-10-12T05:46:00Z) - Pareto Set Learning for Neural Multi-objective Combinatorial
Optimization [6.091096843566857]
Multiobjective optimization (MOCO) problems can be found in many real-world applications.
We develop a learning-based approach to approximate the whole Pareto set for a given MOCO problem without further search procedure.
Our proposed method significantly outperforms some other methods on the multiobjective traveling salesman problem, multiconditioned vehicle routing problem and multi knapsack problem in terms of solution quality, speed, and model efficiency.
arXiv Detail & Related papers (2022-03-29T09:26:22Z) - Learning Proximal Operators to Discover Multiple Optima [66.98045013486794]
We present an end-to-end method to learn the proximal operator across non-family problems.
We show that for weakly-ized objectives and under mild conditions, the method converges globally.
arXiv Detail & Related papers (2022-01-28T05:53:28Z) - Outlier-Robust Sparse Estimation via Non-Convex Optimization [73.18654719887205]
We explore the connection between high-dimensional statistics and non-robust optimization in the presence of sparsity constraints.
We develop novel and simple optimization formulations for these problems.
As a corollary, we obtain that any first-order method that efficiently converges to station yields an efficient algorithm for these tasks.
arXiv Detail & Related papers (2021-09-23T17:38:24Z) - Meta-learning based Alternating Minimization Algorithm for Non-convex
Optimization [9.774392581946108]
We propose a novel solution for challenging non-problems of multiple variables.
Our proposed approach is able to achieve effective iterations in cases while other methods would typically fail.
arXiv Detail & Related papers (2020-09-09T10:45:00Z) - MODRL/D-AM: Multiobjective Deep Reinforcement Learning Algorithm Using
Decomposition and Attention Model for Multiobjective Optimization [15.235261981563523]
This paper proposes a multiobjective deep reinforcement learning method to solve multiobjective optimization problem.
In our method, each subproblem is solved by an attention model, which can exploit the structure features as well as node features of input nodes.
arXiv Detail & Related papers (2020-02-13T12:59:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.