A Practical Guide to Multi-Objective Reinforcement Learning and Planning
- URL: http://arxiv.org/abs/2103.09568v1
- Date: Wed, 17 Mar 2021 11:07:28 GMT
- Title: A Practical Guide to Multi-Objective Reinforcement Learning and Planning
- Authors: Conor F. Hayes, Roxana R\u{a}dulescu, Eugenio Bargiacchi, Johan
K\"allstr\"om, Matthew Macfarlane, Mathieu Reymond, Timothy Verstraeten,
Luisa M. Zintgraf, Richard Dazeley, Fredrik Heintz, Enda Howley, Athirai A.
Irissappane, Patrick Mannion, Ann Now\'e, Gabriel Ramos, Marcello Restelli,
Peter Vamplew, Diederik M. Roijers
- Abstract summary: This paper serves as a guide to the application of multi-objective methods to difficult problems.
It identifies the factors that may influence the nature of the desired solution.
It illustrates by example how these influence the design of multi-objective decision-making systems.
- Score: 24.81310809455139
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Real-world decision-making tasks are generally complex, requiring trade-offs
between multiple, often conflicting, objectives. Despite this, the majority of
research in reinforcement learning and decision-theoretic planning either
assumes only a single objective, or that multiple objectives can be adequately
handled via a simple linear combination. Such approaches may oversimplify the
underlying problem and hence produce suboptimal results. This paper serves as a
guide to the application of multi-objective methods to difficult problems, and
is aimed at researchers who are already familiar with single-objective
reinforcement learning and planning methods who wish to adopt a multi-objective
perspective on their research, as well as practitioners who encounter
multi-objective decision problems in practice. It identifies the factors that
may influence the nature of the desired solution, and illustrates by example
how these influence the design of multi-objective decision-making systems for
complex problems.
Related papers
- Guided Learning: Lubricating End-to-End Modeling for Multi-stage Decision-making [7.106919452604968]
We propose Guided Learning to enhance end-to-end learning in multi-stage decision-making.
We introduce the concept of a guide'', a function that induces the training of intermediate neural network layers towards some phased goals.
For decision scenarios lacking explicit supervisory labels, we incorporate a utility function that quantifies the reward'' of the throughout decision.
arXiv Detail & Related papers (2024-11-15T06:54:25Z) - Deep Pareto Reinforcement Learning for Multi-Objective Recommender Systems [60.91599969408029]
optimizing multiple objectives simultaneously is an important task for recommendation platforms.
Existing multi-objective recommender systems do not systematically consider such dynamic relationships.
arXiv Detail & Related papers (2024-07-04T02:19:49Z) - UCB-driven Utility Function Search for Multi-objective Reinforcement Learning [75.11267478778295]
In Multi-objective Reinforcement Learning (MORL) agents are tasked with optimising decision-making behaviours.
We focus on the case of linear utility functions parameterised by weight vectors w.
We introduce a method based on Upper Confidence Bound to efficiently search for the most promising weight vectors during different stages of the learning process.
arXiv Detail & Related papers (2024-05-01T09:34:42Z) - Many-Objective Multi-Solution Transport [36.07360460509921]
Many-objective multi-solution Transport (MosT) is a framework that finds multiple diverse solutions in the Pareto front of many objectives.
MosT formulates the problem as a bi-level optimization of weighted objectives for each solution, where the weights are defined by an optimal transport between the objectives and solutions.
arXiv Detail & Related papers (2024-03-06T23:03:12Z) - PMGDA: A Preference-based Multiple Gradient Descent Algorithm [12.600588000788214]
It is desirable in many multi-objective machine learning applications, such as multi-task learning, to find a solution that fits a given preference of a decision maker.
This paper proposes a novel predict-and-correct framework for locating a solution that fits the preference of a decision maker.
arXiv Detail & Related papers (2024-02-14T11:27:31Z) - Learning with Constraint Learning: New Perspective, Solution Strategy
and Various Applications [45.45917703420217]
We propose a new framework, named Learning with Constraint Learning (LwCL), that can holistically examine challenges.
LwCL is designed as a general hierarchical optimization model that captures the essence of diverse learning and vision problems.
Our proposed framework efficiently addresses a wide range of applications in learning and vision, encompassing three categories and nine different problem types.
arXiv Detail & Related papers (2023-07-28T01:50:27Z) - Multi-Target Multiplicity: Flexibility and Fairness in Target
Specification under Resource Constraints [76.84999501420938]
We introduce a conceptual and computational framework for assessing how the choice of target affects individuals' outcomes.
We show that the level of multiplicity that stems from target variable choice can be greater than that stemming from nearly-optimal models of a single target.
arXiv Detail & Related papers (2023-06-23T18:57:14Z) - Discovering Diverse Solutions in Deep Reinforcement Learning [84.45686627019408]
Reinforcement learning algorithms are typically limited to learning a single solution of a specified task.
We propose an RL method that can learn infinitely many solutions by training a policy conditioned on a continuous or discrete low-dimensional latent variable.
arXiv Detail & Related papers (2021-03-12T04:54:31Z) - Provable Multi-Objective Reinforcement Learning with Generative Models [98.19879408649848]
We study the problem of single policy MORL, which learns an optimal policy given the preference of objectives.
Existing methods require strong assumptions such as exact knowledge of the multi-objective decision process.
We propose a new algorithm called model-based envelop value (EVI) which generalizes the enveloped multi-objective $Q$-learning algorithm.
arXiv Detail & Related papers (2020-11-19T22:35:31Z) - A Distributional View on Multi-Objective Policy Optimization [24.690800846837273]
We propose an algorithm for multi-objective reinforcement learning that enables setting desired preferences for objectives in a scale-invariant way.
We show that setting different preferences in our framework allows us to trace out the space of nondominated solutions.
arXiv Detail & Related papers (2020-05-15T13:02:17Z) - Pareto Multi-Task Learning [53.90732663046125]
Multi-task learning is a powerful method for solving multiple correlated tasks simultaneously.
It is often impossible to find one single solution to optimize all the tasks, since different tasks might conflict with each other.
Recently, a novel method is proposed to find one single Pareto optimal solution with good trade-off among different tasks by casting multi-task learning as multiobjective optimization.
arXiv Detail & Related papers (2019-12-30T08:58:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.