Projection based Active Gaussian Process Regression for Pareto Front
Modeling
- URL: http://arxiv.org/abs/2001.07072v1
- Date: Mon, 20 Jan 2020 11:52:50 GMT
- Title: Projection based Active Gaussian Process Regression for Pareto Front
Modeling
- Authors: Zhengqi Gao, Jun Tao, Yangfeng Su, Dian Zhou, and Xuan Zeng
- Abstract summary: A novel projection based active Gaussian process regression (P- aGPR) method is proposed for efficient PF modeling.
Our proposed P-aGPR method can not only provide a generative PF model, but also fast examine whether a provided point locates on PF or not.
The numerical results demonstrate that compared to state-of-the-art passive learning methods the proposed P-aGPR method can achieve higher modeling accuracy and stability.
- Score: 6.718019242119055
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pareto Front (PF) modeling is essential in decision making problems across
all domains such as economics, medicine or engineering. In Operation Research
literature, this task has been addressed based on multi-objective optimization
algorithms. However, without learning models for PF, these methods cannot
examine whether a new provided point locates on PF or not. In this paper, we
reconsider the task from Data Mining perspective. A novel projection based
active Gaussian process regression (P- aGPR) method is proposed for efficient
PF modeling. First, P- aGPR chooses a series of projection spaces with
dimensionalities ranking from low to high. Next, in each projection space, a
Gaussian process regression (GPR) model is trained to represent the constraint
that PF should satisfy in that space. Moreover, in order to improve modeling
efficacy and stability, an active learning framework has been developed by
exploiting the uncertainty information obtained in the GPR models. Different
from all existing methods, our proposed P-aGPR method can not only provide a
generative PF model, but also fast examine whether a provided point locates on
PF or not. The numerical results demonstrate that compared to state-of-the-art
passive learning methods the proposed P-aGPR method can achieve higher modeling
accuracy and stability.
Related papers
- Scalable and non-iterative graphical model estimation [3.187381965457262]
Iterative Proportional Fitting (IPF) and its variants are the default method for undirected graphical model estimation.
We propose a novel and fast non-iterative method for positive definite graphical model estimation in high dimensions.
arXiv Detail & Related papers (2024-08-21T15:46:00Z) - Model-Based Reparameterization Policy Gradient Methods: Theory and
Practical Algorithms [88.74308282658133]
Reization (RP) Policy Gradient Methods (PGMs) have been widely adopted for continuous control tasks in robotics and computer graphics.
Recent studies have revealed that, when applied to long-term reinforcement learning problems, model-based RP PGMs may experience chaotic and non-smooth optimization landscapes.
We propose a spectral normalization method to mitigate the exploding variance issue caused by long model unrolls.
arXiv Detail & Related papers (2023-10-30T18:43:21Z) - Provably Efficient UCB-type Algorithms For Learning Predictive State
Representations [55.00359893021461]
The sequential decision-making problem is statistically learnable if it admits a low-rank structure modeled by predictive state representations (PSRs)
This paper proposes the first known UCB-type approach for PSRs, featuring a novel bonus term that upper bounds the total variation distance between the estimated and true models.
In contrast to existing approaches for PSRs, our UCB-type algorithms enjoy computational tractability, last-iterate guaranteed near-optimal policy, and guaranteed model accuracy.
arXiv Detail & Related papers (2023-07-01T18:35:21Z) - Predictable MDP Abstraction for Unsupervised Model-Based RL [93.91375268580806]
We propose predictable MDP abstraction (PMA)
Instead of training a predictive model on the original MDP, we train a model on a transformed MDP with a learned action space.
We theoretically analyze PMA and empirically demonstrate that PMA leads to significant improvements over prior unsupervised model-based RL approaches.
arXiv Detail & Related papers (2023-02-08T07:37:51Z) - When to Update Your Model: Constrained Model-based Reinforcement
Learning [50.74369835934703]
We propose a novel and general theoretical scheme for a non-decreasing performance guarantee of model-based RL (MBRL)
Our follow-up derived bounds reveal the relationship between model shifts and performance improvement.
A further example demonstrates that learning models from a dynamically-varying number of explorations benefit the eventual returns.
arXiv Detail & Related papers (2022-10-15T17:57:43Z) - A General Framework for Sample-Efficient Function Approximation in
Reinforcement Learning [132.45959478064736]
We propose a general framework that unifies model-based and model-free reinforcement learning.
We propose a novel estimation function with decomposable structural properties for optimization-based exploration.
Under our framework, a new sample-efficient algorithm namely OPtimization-based ExploRation with Approximation (OPERA) is proposed.
arXiv Detail & Related papers (2022-09-30T17:59:16Z) - Simplifying Model-based RL: Learning Representations, Latent-space
Models, and Policies with One Objective [142.36200080384145]
We propose a single objective which jointly optimize a latent-space model and policy to achieve high returns while remaining self-consistent.
We demonstrate that the resulting algorithm matches or improves the sample-efficiency of the best prior model-based and model-free RL methods.
arXiv Detail & Related papers (2022-09-18T03:51:58Z) - RL-PGO: Reinforcement Learning-based Planar Pose-Graph Optimization [1.4884785898657995]
This paper presents a state-of-the-art Deep Reinforcement Learning (DRL) based environment and proposed agent for 2D pose-graph optimization.
We demonstrate that the pose-graph optimization problem can be modeled as a partially observable Decision Process and evaluate performance on real-world and synthetic datasets.
arXiv Detail & Related papers (2022-02-26T20:10:14Z) - Bayesian Sequential Optimal Experimental Design for Nonlinear Models
Using Policy Gradient Reinforcement Learning [0.0]
We formulate this sequential optimal experimental design (sOED) problem as a finite-horizon partially observable Markov decision process (POMDP)
It is built to accommodate continuous random variables, general non-Gaussian posteriors, and expensive nonlinear forward models.
We solve for the sOED policy numerically via policy gradient (PG) methods from reinforcement learning, and derive and prove the PG expression for sOED.
The overall PG-sOED method is validated on a linear-Gaussian benchmark, and its advantages over batch and greedy designs are demonstrated through a contaminant source inversion problem in a
arXiv Detail & Related papers (2021-10-28T17:47:31Z) - Weighted Aggregating Stochastic Gradient Descent for Parallel Deep
Learning [8.366415386275557]
Solution involves a reformation of the objective function for optimization in neural network models.
We introduce a decentralized weighted aggregating scheme based on the performance of local workers.
To validate the new method, we benchmark our schemes against several popular algorithms.
arXiv Detail & Related papers (2020-04-07T23:38:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.