Greedier is Better: Selecting Multiple Neighbors per Iteration for
Sparse Subspace Clustering
- URL: http://arxiv.org/abs/2204.02572v1
- Date: Wed, 6 Apr 2022 04:20:35 GMT
- Title: Greedier is Better: Selecting Multiple Neighbors per Iteration for
Sparse Subspace Clustering
- Authors: Jwo-Yuh Wu, Liang-Chi Huang, Wen-Hsuan Li, Chun-Hung Liu, and
Rung-Hung Gau
- Abstract summary: This paper proposes a new SSC scheme using generalized OMP (GOMP)
GOMP involves fewer iterations, thereby enjoying lower algorithmic complexity.
The proposed stopping rule is free from off-line estimation of subspace dimension and noise power.
- Score: 18.888312436971187
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sparse subspace clustering (SSC) using greedy-based neighbor selection, such
as orthogonal matching pursuit (OMP), has been known as a popular
computationally-efficient alternative to the popular L1-minimization based
methods. This paper proposes a new SSC scheme using generalized OMP (GOMP), a
soup-up of OMP whereby multiple neighbors are identified per iteration, along
with a new stopping rule requiring nothing more than a knowledge of the ambient
signal dimension. Compared to conventional OMP, which identifies one neighbor
per iteration, the proposed GOMP method involves fewer iterations, thereby
enjoying lower algorithmic complexity; advantageously, the proposed stopping
rule is free from off-line estimation of subspace dimension and noise power.
Under the semi-random model, analytic performance guarantees, in terms of
neighbor recovery rates, are established to justify the advantage of the
proposed GOMP. The results show that, with a high probability, GOMP (i) is
halted by the proposed stopping rule, and (ii) can retrieve more true neighbors
than OMP, consequently yielding higher final data clustering accuracy. Computer
simulations using both synthetic data and real human face data are provided to
validate our analytic study and evidence the effectiveness of the proposed
approach.
Related papers
- Efficient Learning of POMDPs with Known Observation Model in Average-Reward Setting [56.92178753201331]
We propose the Observation-Aware Spectral (OAS) estimation technique, which enables the POMDP parameters to be learned from samples collected using a belief-based policy.
We show the consistency of the OAS procedure, and we prove a regret guarantee of order $mathcalO(sqrtT log(T)$ for the proposed OAS-UCRL algorithm.
arXiv Detail & Related papers (2024-10-02T08:46:34Z) - Provably Mitigating Overoptimization in RLHF: Your SFT Loss is Implicitly an Adversarial Regularizer [52.09480867526656]
We identify the source of misalignment as a form of distributional shift and uncertainty in learning human preferences.
To mitigate overoptimization, we first propose a theoretical algorithm that chooses the best policy for an adversarially chosen reward model.
Using the equivalence between reward models and the corresponding optimal policy, the algorithm features a simple objective that combines a preference optimization loss and a supervised learning loss.
arXiv Detail & Related papers (2024-05-26T05:38:50Z) - Fast OMP for Exact Recovery and Sparse Approximation [4.915061940934031]
This paper advances Orthogonal Matching Pursuit (OMP) in two fronts.
It offers a fast algorithm for the projection of the input signal at each iteration, and a new selection criterion for making the greedy choice.
Experiment results show significant improvement over the classical OMP in computation time.
arXiv Detail & Related papers (2024-03-29T20:39:37Z) - Privacy Amplification for Matrix Mechanisms [18.13715687378337]
"MMCC" is the first algorithm to analyze privacy amplification via sampling for any generic matrix mechanism.
We show it leads to significant improvement in the privacy-utility trade-offs for DP-FTRL algorithms on standard benchmarks.
arXiv Detail & Related papers (2023-10-24T05:16:52Z) - Distributed Adaptive Nearest Neighbor Classifier: Algorithm and Theory [6.696267547013535]
We propose a novel distributed adaptive NN classifier for which the number of nearest neighbors is a tuning parameterally chosen by a data-driven criterion.
An early stopping rule is proposed when searching for the optimal tuning parameter, which improves the finite sample performance.
In particular, we show that when the sub-sample sizes are sufficiently large, the proposed classifier achieves the nearly optimal convergence rate.
arXiv Detail & Related papers (2021-05-20T14:38:41Z) - Learning Sampling Policy for Faster Derivative Free Optimization [100.27518340593284]
We propose a new reinforcement learning based ZO algorithm (ZO-RL) with learning the sampling policy for generating the perturbations in ZO optimization instead of using random sampling.
Our results show that our ZO-RL algorithm can effectively reduce the variances of ZO gradient by learning a sampling policy, and converge faster than existing ZO algorithms in different scenarios.
arXiv Detail & Related papers (2021-04-09T14:50:59Z) - Plug-And-Play Learned Gaussian-mixture Approximate Message Passing [71.74028918819046]
We propose a plug-and-play compressed sensing (CS) recovery algorithm suitable for any i.i.d. source prior.
Our algorithm builds upon Borgerding's learned AMP (LAMP), yet significantly improves it by adopting a universal denoising function within the algorithm.
Numerical evaluation shows that the L-GM-AMP algorithm achieves state-of-the-art performance without any knowledge of the source prior.
arXiv Detail & Related papers (2020-11-18T16:40:45Z) - Adaptive Sampling for Best Policy Identification in Markov Decision
Processes [79.4957965474334]
We investigate the problem of best-policy identification in discounted Markov Decision (MDPs) when the learner has access to a generative model.
The advantages of state-of-the-art algorithms are discussed and illustrated.
arXiv Detail & Related papers (2020-09-28T15:22:24Z) - Accelerated Message Passing for Entropy-Regularized MAP Inference [89.15658822319928]
Maximum a posteriori (MAP) inference in discrete-valued random fields is a fundamental problem in machine learning.
Due to the difficulty of this problem, linear programming (LP) relaxations are commonly used to derive specialized message passing algorithms.
We present randomized methods for accelerating these algorithms by leveraging techniques that underlie classical accelerated gradient.
arXiv Detail & Related papers (2020-07-01T18:43:32Z) - Provable Noisy Sparse Subspace Clustering using Greedy Neighbor
Selection: A Coherence-Based Perspective [18.888312436971187]
We derive coherence-based sufficient conditions guaranteeing correct neighbor identification using MP/OMP.
A striking finding is that, when the ground truth subspaces are well-separated from each other and noise is not large, MP-based iterations, while enjoying lower algorithmic complexity, yield smaller perturbation of residuals.
arXiv Detail & Related papers (2020-02-02T14:28:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.