Approximate Vanishing Ideal Computations at Scale
- URL: http://arxiv.org/abs/2207.01236v1
- Date: Mon, 4 Jul 2022 07:17:28 GMT
- Title: Approximate Vanishing Ideal Computations at Scale
- Authors: Elias Wirth, Hiroshi Kera, Sebastian Pokutta
- Abstract summary: We focus on scaling up the Oracle Approximate Vanishing Ideal algorithm (OAVI)
OAVI is an attractive preprocessing technique for large-scale machine learning.
- Score: 21.01267270902429
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The approximate vanishing ideal of a set of points $X = \{\mathbf{x}_1,
\ldots, \mathbf{x}_m\}\subseteq [0,1]^n$ is the set of polynomials that
approximately evaluate to $0$ over all points $\mathbf{x} \in X$ and admits an
efficient representation by a finite set of polynomials called generators.
Algorithms that construct this set of generators are extensively studied but
ultimately find little practical application because their computational
complexities are thought to be superlinear in the number of samples $m$. In
this paper, we focus on scaling up the Oracle Approximate Vanishing Ideal
algorithm (OAVI), one of the most powerful of these methods. We prove that the
computational complexity of OAVI is not superlinear but linear in the number of
samples $m$ and polynomial in the number of features $n$, making OAVI an
attractive preprocessing technique for large-scale machine learning. To further
accelerate OAVI's training time, we propose two changes: First, as the name
suggests, OAVI makes repeated oracle calls to convex solvers throughout its
execution. By replacing the Pairwise Conditional Gradients algorithm, one of
the standard solvers used in OAVI, with the faster Blended Pairwise Conditional
Gradients algorithm, we illustrate how OAVI directly benefits from advancements
in the study of convex solvers. Second, we propose Inverse Hessian Boosting
(IHB): IHB exploits the fact that OAVI repeatedly solves quadratic convex
optimization problems that differ only by very little and whose solutions can
be written in closed form using inverse Hessian information. By efficiently
updating the inverse of the Hessian matrix, the convex optimization problems
can be solved almost instantly, accelerating OAVI's training time by up to
multiple orders of magnitude. We complement our theoretical analysis with
extensive numerical experiments on data sets whose sample numbers are in the
millions.
Related papers
- A Sample Efficient Alternating Minimization-based Algorithm For Robust Phase Retrieval [56.67706781191521]
In this work, we present a robust phase retrieval problem where the task is to recover an unknown signal.
Our proposed oracle avoids the need for computationally spectral descent, using a simple gradient step and outliers.
arXiv Detail & Related papers (2024-09-07T06:37:23Z) - Optimal Hessian/Jacobian-Free Nonconvex-PL Bilevel Optimization [25.438298531555468]
Bilevel optimization is widely applied in many machine learning tasks such as hyper learning, meta learning and reinforcement learning.
We propose an efficient Hessian/BiO method with the optimal convergence $frac1TT) under some mild conditions.
We conduct some some experiments on the bilevel game hyper-stationary numerical convergence.
arXiv Detail & Related papers (2024-07-25T07:25:06Z) - Fine-grained Analysis and Faster Algorithms for Iteratively Solving Linear Systems [9.30306458153248]
We consider the spectral tail condition number, $kappa_ell$, defined as the ratio between the $ell$th largest and the smallest singular value of the matrix representing the system.
Some of the implications of our result, and of the use of $kappa_ell$, include direct improvement over a fine-grained analysis of the Conjugate method.
arXiv Detail & Related papers (2024-05-09T14:56:49Z) - Linearization Algorithms for Fully Composite Optimization [61.20539085730636]
This paper studies first-order algorithms for solving fully composite optimization problems convex compact sets.
We leverage the structure of the objective by handling differentiable and non-differentiable separately, linearizing only the smooth parts.
arXiv Detail & Related papers (2023-02-24T18:41:48Z) - The Hypervolume Indicator Hessian Matrix: Analytical Expression,
Computational Time Complexity, and Sparsity [4.523133864190258]
This paper establishes the analytical expression of the Hessian matrix of the mapping from a (fixed size) collection of $n$ points in the $d$-dimensional decision space to the scalar hypervolume indicator value.
The Hessian matrix plays a crucial role in second-order methods, such as the Newton-Raphson optimization method, and it can be used for the verification of local optimal sets.
arXiv Detail & Related papers (2022-11-08T11:24:18Z) - Multi-block-Single-probe Variance Reduced Estimator for Coupled
Compositional Optimization [49.58290066287418]
We propose a novel method named Multi-block-probe Variance Reduced (MSVR) to alleviate the complexity of compositional problems.
Our results improve upon prior ones in several aspects, including the order of sample complexities and dependence on strongity.
arXiv Detail & Related papers (2022-07-18T12:03:26Z) - Accelerated Proximal Alternating Gradient-Descent-Ascent for Nonconvex
Minimax Machine Learning [12.069630105460766]
An Alternating Table-descentascent (AltGDA) is an computation optimization algorithm that has been widely used for training in various machine learning applications.
In this paper, we develop a single-loop fast computation-of-the-loop gradient-of-the-loop algorithm to solve non minimax optimization problems.
arXiv Detail & Related papers (2021-12-22T04:33:27Z) - BiAdam: Fast Adaptive Bilevel Optimization Methods [104.96004056928474]
Bilevel optimization has attracted increased interest in machine learning due to its many applications.
We provide a useful analysis framework for both the constrained and unconstrained optimization.
arXiv Detail & Related papers (2021-06-21T20:16:40Z) - Hybrid Trilinear and Bilinear Programming for Aligning Partially
Overlapping Point Sets [85.71360365315128]
In many applications, we need algorithms which can align partially overlapping point sets are invariant to the corresponding corresponding RPM algorithm.
We first show that the objective is a cubic bound function. We then utilize the convex envelopes of trilinear and bilinear monomial transformations to derive its lower bound.
We next develop a branch-and-bound (BnB) algorithm which only branches over the transformation variables and runs efficiently.
arXiv Detail & Related papers (2021-01-19T04:24:23Z) - Gradient Free Minimax Optimization: Variance Reduction and Faster
Convergence [120.9336529957224]
In this paper, we denote the non-strongly setting on the magnitude of a gradient-free minimax optimization problem.
We show that a novel zeroth-order variance reduced descent algorithm achieves the best known query complexity.
arXiv Detail & Related papers (2020-06-16T17:55:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.