Learning to Approximate: Auto Direction Vector Set Generation for
Hypervolume Contribution Approximation
- URL: http://arxiv.org/abs/2201.06707v1
- Date: Tue, 18 Jan 2022 02:31:15 GMT
- Title: Learning to Approximate: Auto Direction Vector Set Generation for
Hypervolume Contribution Approximation
- Authors: Ke Shang and Tianye Shu and Hisao Ishibuchi
- Abstract summary: Hypervolume contribution is an important concept in evolutionary multi-objective optimization (EMO)
An R2 indicator variant (i.e., $RtextHVC$ indicator) is proposed to approximate the hypervolume contribution.
In this paper, we propose textitLearning to Approximate (LtA), a direction vector set generation method for the $RtextHVC$ indicator.
- Score: 6.617487928813376
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Hypervolume contribution is an important concept in evolutionary
multi-objective optimization (EMO). It involves in hypervolume-based EMO
algorithms and hypervolume subset selection algorithms. Its main drawback is
that it is computationally expensive in high-dimensional spaces, which limits
its applicability to many-objective optimization. Recently, an R2 indicator
variant (i.e., $R_2^{\text{HVC}}$ indicator) is proposed to approximate the
hypervolume contribution. The $R_2^{\text{HVC}}$ indicator uses line segments
along a number of direction vectors for hypervolume contribution approximation.
It has been shown that different direction vector sets lead to different
approximation quality. In this paper, we propose \textit{Learning to
Approximate (LtA)}, a direction vector set generation method for the
$R_2^{\text{HVC}}$ indicator. The direction vector set is automatically learned
from training data. The learned direction vector set can then be used in the
$R_2^{\text{HVC}}$ indicator to improve its approximation quality. The
usefulness of the proposed LtA method is examined by comparing it with other
commonly-used direction vector set generation methods for the
$R_2^{\text{HVC}}$ indicator. Experimental results suggest the superiority of
LtA over the other methods for generating high quality direction vector sets.
Related papers
- The Generative Leap: Sharp Sample Complexity for Efficiently Learning Gaussian Multi-Index Models [71.5283441529015]
In this work we consider generic Gaussian Multi-index models, in which the labels only depend on the (Gaussian) $d$-dimensional inputs through their projection onto a low-dimensional $r = O_d(1)$ subspace.<n>We introduce the generative leap exponent $kstar$, a natural extension of the generative exponent from [Damian et al.'24] to the multi-index setting.
arXiv Detail & Related papers (2025-06-05T18:34:56Z) - AGD: an Auto-switchable Optimizer using Stepwise Gradient Difference for
Preconditioning Matrix [9.629238108795013]
We propose a novel approach to designing the preconditioning matrix by utilizing the gradient difference between two successive steps as the diagonal elements.
We evaluate AGD on public generalization of Natural Language Processing (NLP), Computer Vision (CV), and Recommendation Systems (RecSys)
Our experimental results demonstrate that AGD outperforms the state-of-the-art (SOTA) techniques, achieving highly competitive or significantly better predictive performance.
arXiv Detail & Related papers (2023-12-04T06:20:14Z) - Neural Gradient Learning and Optimization for Oriented Point Normal
Estimation [53.611206368815125]
We propose a deep learning approach to learn gradient vectors with consistent orientation from 3D point clouds for normal estimation.
We learn an angular distance field based on local plane geometry to refine the coarse gradient vectors.
Our method efficiently conducts global gradient approximation while achieving better accuracy and ability generalization of local feature description.
arXiv Detail & Related papers (2023-09-17T08:35:11Z) - ELRA: Exponential learning rate adaption gradient descent optimization
method [83.88591755871734]
We present a novel, fast (exponential rate), ab initio (hyper-free) gradient based adaption.
The main idea of the method is to adapt the $alpha by situational awareness.
It can be applied to problems of any dimensions n and scales only linearly.
arXiv Detail & Related papers (2023-09-12T14:36:13Z) - GA-Aided Directivity in Volumetric and Planar Massive-Antenna Array
Design [0.0]
The problem of directivity enhancement, leading to the increase in the directivity gain over a certain desired angle of arrival/departure (AoA/AoD) is considered in this work.
A new volumetric array of the directivity array is proposed using rectangular rectangular angles and a generalzimuth elevation pattern.
Such a directivity distance is formulated to achieve as high directivity gains as possible.
arXiv Detail & Related papers (2023-01-07T21:52:19Z) - The Hypervolume Indicator Hessian Matrix: Analytical Expression,
Computational Time Complexity, and Sparsity [4.523133864190258]
This paper establishes the analytical expression of the Hessian matrix of the mapping from a (fixed size) collection of $n$ points in the $d$-dimensional decision space to the scalar hypervolume indicator value.
The Hessian matrix plays a crucial role in second-order methods, such as the Newton-Raphson optimization method, and it can be used for the verification of local optimal sets.
arXiv Detail & Related papers (2022-11-08T11:24:18Z) - Subspace clustering in high-dimensions: Phase transitions \&
Statistical-to-Computational gap [24.073221004661427]
A simple model to study subspace clustering is the high-dimensional $k$-Gaussian mixture model.
We provide an exact characterization of the statistically optimal reconstruction error in this model in the high-dimensional regime with extensive sparsity.
arXiv Detail & Related papers (2022-05-26T17:47:35Z) - Approximate Function Evaluation via Multi-Armed Bandits [51.146684847667125]
We study the problem of estimating the value of a known smooth function $f$ at an unknown point $boldsymbolmu in mathbbRn$, where each component $mu_i$ can be sampled via a noisy oracle.
We design an instance-adaptive algorithm that learns to sample according to the importance of each coordinate, and with probability at least $1-delta$ returns an $epsilon$ accurate estimate of $f(boldsymbolmu)$.
arXiv Detail & Related papers (2022-03-18T18:50:52Z) - Efficient Matrix-Free Approximations of Second-Order Information, with
Applications to Pruning and Optimization [16.96639526117016]
We investigate matrix-free, linear-time approaches for estimating Inverse-Hessian Vector Products (IHVPs)
These algorithms yield state-of-the-art results for network pruning and optimization with lower computational overhead relative to existing second-order methods.
arXiv Detail & Related papers (2021-07-07T17:01:34Z) - BiAdam: Fast Adaptive Bilevel Optimization Methods [104.96004056928474]
Bilevel optimization has attracted increased interest in machine learning due to its many applications.
We provide a useful analysis framework for both the constrained and unconstrained optimization.
arXiv Detail & Related papers (2021-06-21T20:16:40Z) - A Precise Performance Analysis of Support Vector Regression [105.94855998235232]
We study the hard and soft support vector regression techniques applied to a set of $n$ linear measurements.
Our results are then used to optimally tune the parameters intervening in the design of hard and soft support vector regression algorithms.
arXiv Detail & Related papers (2021-05-21T14:26:28Z) - Correcting Momentum with Second-order Information [50.992629498861724]
We develop a new algorithm for non-critical optimization that finds an $O(epsilon)$epsilon point in the optimal product.
We validate our results on a variety of large-scale deep learning benchmarks and architectures.
arXiv Detail & Related papers (2021-03-04T19:01:20Z) - Nearly Minimax Optimal Regret for Learning Infinite-horizon
Average-reward MDPs with Linear Function Approximation [95.80683238546499]
We propose a new algorithm UCRL2-VTR, which can be seen as an extension of the UCRL2 algorithm with linear function approximation.
We show that UCRL2-VTR with Bernstein-type bonus can achieve a regret of $tildeO(dsqrtDT)$, where $d$ is the dimension of the feature mapping.
We also prove a matching lower bound $tildeOmega(dsqrtDT)$, which suggests that the proposed UCRL2-VTR is minimax optimal up to logarithmic factors
arXiv Detail & Related papers (2021-02-15T02:08:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.