Simultaneous Grouping and Denoising via Sparse Convex Wavelet Clustering
- URL: http://arxiv.org/abs/2012.04762v2
- Date: Thu, 4 Mar 2021 00:30:03 GMT
- Title: Simultaneous Grouping and Denoising via Sparse Convex Wavelet Clustering
- Authors: Michael Weylandt and T. Mitchell Roddenberry and Genevera I. Allen
- Abstract summary: We develop a sparse convex wavelet clustering approach that simultaneously denoises and discovers groups.
Our method yields denoised (wavelet-sparse) cluster centroids that both improve interpretability and data compression.
- Score: 3.2116198597240846
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Clustering is a ubiquitous problem in data science and signal processing. In
many applications where we observe noisy signals, it is common practice to
first denoise the data, perhaps using wavelet denoising, and then to apply a
clustering algorithm. In this paper, we develop a sparse convex wavelet
clustering approach that simultaneously denoises and discovers groups. Our
approach utilizes convex fusion penalties to achieve agglomeration and
group-sparse penalties to denoise through sparsity in the wavelet domain. In
contrast to common practice which denoises then clusters, our method is a
unified, convex approach that performs both simultaneously. Our method yields
denoised (wavelet-sparse) cluster centroids that both improve interpretability
and data compression. We demonstrate our method on synthetic examples and in an
application to NMR spectroscopy.
Related papers
- Sampling Binary Data by Denoising through Score Functions [2.9465623430708905]
Tweedie-Miyasawa formula (TMF) is key ingredient in score-based generative models.
TMF ties these together via the score function of noisy data.
We adopt Bernoulli noise, instead of Gaussian noise, as a smoothing device.
arXiv Detail & Related papers (2025-02-01T20:59:02Z) - Consistent spectral clustering in sparse tensor block models [0.0]
High-order clustering aims to classify objects in multiway datasets that are prevalent in various fields.
This paper introduces a tensor block model specifically designed for sparse integer-valued data tensors.
We propose a simple spectral clustering algorithm augmented with a trimming step to mitigate noise fluctuations.
arXiv Detail & Related papers (2025-01-23T16:41:19Z) - Robust Representation Consistency Model via Contrastive Denoising [83.47584074390842]
randomized smoothing provides theoretical guarantees for certifying robustness against adversarial perturbations.
diffusion models have been successfully employed for randomized smoothing to purify noise-perturbed samples.
We reformulate the generative modeling task along the diffusion trajectories in pixel space as a discriminative task in the latent space.
arXiv Detail & Related papers (2025-01-22T18:52:06Z) - Consistency of Lloyd's Algorithm Under Perturbations [11.56433365648479]
Lloyd's algorithm is one of the most widely used clustering algorithms.
We show that the mis-clustering rate of Lloyd's algorithm on samples from a sub-Gaussian mixture is exponentially bounded after $O(log(n))$ iterations.
arXiv Detail & Related papers (2023-09-01T16:45:52Z) - Samplet basis pursuit: Multiresolution scattered data approximation with sparsity constraints [0.0]
We consider scattered data approximation in samplet coordinates with $ell_1$-regularization.
By using the Riesz isometry, we embed samplets into reproducing kernel Hilbert spaces.
We argue that the class of signals that are sparse with respect to the embedded samplet basis is considerably larger than the class of signals that are sparse with respect to the basis of kernel translates.
arXiv Detail & Related papers (2023-06-16T21:20:49Z) - Chain of Log-Concave Markov Chains [2.9465623430708905]
We introduce a theoretical framework for sampling from unnormalized densities.
We prove one can decompose sampling from a density into a sequence of sampling from log-concave conditional densities.
We report a remarkable capacity of our algorithm to "tunnel" between modes of a distribution.
arXiv Detail & Related papers (2023-05-31T01:00:35Z) - On Robust Learning from Noisy Labels: A Permutation Layer Approach [53.798757734297986]
This paper introduces a permutation layer learning approach termed PermLL to dynamically calibrate the training process of a deep neural network (DNN)
We provide two variants of PermLL in this paper: one applies the permutation layer to the model's prediction, while the other applies it directly to the given noisy label.
We validate PermLL experimentally and show that it achieves state-of-the-art performance on both real and synthetic datasets.
arXiv Detail & Related papers (2022-11-29T03:01:48Z) - Blind Source Separation for Mixture of Sinusoids with Near-Linear
Computational Complexity [0.0]
We propose a multi-tone decomposition algorithm that can find the frequencies, amplitudes and phases of the fundamental sinusoids in a noisy sequence.
When estimating $M$ number of sinusoidal sources, our algorithm successively estimates their frequencies and jointly optimize their amplitudes and phases.
arXiv Detail & Related papers (2022-03-27T15:16:07Z) - A Dynamic Mode Decomposition Approach for Decentralized Spectral
Clustering of Graphs [2.114158481153364]
We show that propagating waves in the graph followed by a local dynamic mode decomposition (DMD) at every node is capable of retrieving the eigenvalues and the local eigenvector components of the graph Laplacian.
We demonstrate the DMD is more robust than the existing FFT based approach and requires 20 times fewer steps of the wave equation to accurately recover the clustering information.
arXiv Detail & Related papers (2022-02-26T03:48:35Z) - Determinantal consensus clustering [77.34726150561087]
We propose the use of determinantal point processes or DPP for the random restart of clustering algorithms.
DPPs favor diversity of the center points within subsets.
We show through simulations that, contrary to DPP, this technique fails both to ensure diversity, and to obtain a good coverage of all data facets.
arXiv Detail & Related papers (2021-02-07T23:48:24Z) - Progressive Cluster Purification for Unsupervised Feature Learning [48.87365358296371]
In unsupervised feature learning, sample specificity based methods ignore the inter-class information.
We propose a novel clustering based method, which excludes class inconsistent samples during progressive cluster formation.
Our approach, referred to as Progressive Cluster Purification (PCP), implements progressive clustering by gradually reducing the number of clusters during training.
arXiv Detail & Related papers (2020-07-06T08:11:03Z) - Local Graph Clustering with Network Lasso [90.66817876491052]
We study the statistical and computational properties of a network Lasso method for local graph clustering.
The clusters delivered by nLasso can be characterized elegantly via network flows between cluster boundary and seed nodes.
arXiv Detail & Related papers (2020-04-25T17:52:05Z) - Simultaneous Denoising and Dereverberation Using Deep Embedding Features [64.58693911070228]
We propose a joint training method for simultaneous speech denoising and dereverberation using deep embedding features.
At the denoising stage, the DC network is leveraged to extract noise-free deep embedding features.
At the dereverberation stage, instead of using the unsupervised K-means clustering algorithm, another neural network is utilized to estimate the anechoic speech.
arXiv Detail & Related papers (2020-04-06T06:34:01Z) - Generative Modeling with Denoising Auto-Encoders and Langevin Sampling [88.83704353627554]
We show that both DAE and DSM provide estimates of the score of the smoothed population density.
We then apply our results to the homotopy method of arXiv:1907.05600 and provide theoretical justification for its empirical success.
arXiv Detail & Related papers (2020-01-31T23:50:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.