Hybrid Model-based / Data-driven Graph Transform for Image Coding
- URL: http://arxiv.org/abs/2203.01186v1
- Date: Wed, 2 Mar 2022 15:36:44 GMT
- Title: Hybrid Model-based / Data-driven Graph Transform for Image Coding
- Authors: Saghar Bagheri, Tam Thuc Do, Gene Cheung, Antonio Ortega
- Abstract summary: We present a hybrid model-based / data-driven approach to encode an intra-prediction residual block.
The first $K$ eigenvectors of a transform matrix are derived from a statistical model, e.g., the asymmetric discrete sine transform (ADST) for stability.
Using WebP as a baseline image, experimental results show that our hybrid graph transform achieved better energy compaction than default discrete cosine transform (DCT) and better stability than KLT.
- Score: 54.31406300524195
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Transform coding to sparsify signal representations remains crucial in an
image compression pipeline. While the Karhunen-Lo\`{e}ve transform (KLT)
computed from an empirical covariance matrix $\bar{C}$ is theoretically optimal
for a stationary process, in practice, collecting sufficient statistics from a
non-stationary image to reliably estimate $\bar{C}$ can be difficult. In this
paper, to encode an intra-prediction residual block, we pursue a hybrid
model-based / data-driven approach: the first $K$ eigenvectors of a transform
matrix are derived from a statistical model, e.g., the asymmetric discrete sine
transform (ADST), for stability, while the remaining $N-K$ are computed from
$\bar{C}$ for performance. The transform computation is posed as a graph
learning problem, where we seek a graph Laplacian matrix minimizing a graphical
lasso objective inside a convex cone sharing the first $K$ eigenvectors in a
Hilbert space of real symmetric matrices. We efficiently solve the problem via
augmented Lagrangian relaxation and proximal gradient (PG). Using WebP as a
baseline image codec, experimental results show that our hybrid graph transform
achieved better energy compaction than default discrete cosine transform (DCT)
and better stability than KLT.
Related papers
- Interpretable Lightweight Transformer via Unrolling of Learned Graph Smoothness Priors [16.04850782310842]
We build interpretable and lightweight transformer-like neural networks by unrolling iterative optimization algorithms.
A normalized signal-dependent graph learning module amounts to a variant of the basic self-attention mechanism in conventional transformers.
arXiv Detail & Related papers (2024-06-06T14:01:28Z) - Learning Sparse High-Dimensional Matrix-Valued Graphical Models From Dependent Data [12.94486861344922]
We consider the problem of inferring the conditional independence graph (CIG) of a sparse, high-dimensional, stationary matrix- Gaussian time series.
We consider a sparse-based formulation of the problem with a Kronecker-decomposable power spectral density (PSD)
We illustrate our approach using numerical examples utilizing both synthetic and real data.
arXiv Detail & Related papers (2024-04-29T19:32:50Z) - Transformers as Support Vector Machines [54.642793677472724]
We establish a formal equivalence between the optimization geometry of self-attention and a hard-margin SVM problem.
We characterize the implicit bias of 1-layer transformers optimized with gradient descent.
We believe these findings inspire the interpretation of transformers as a hierarchy of SVMs that separates and selects optimal tokens.
arXiv Detail & Related papers (2023-08-31T17:57:50Z) - Linear Convergence of ISTA and FISTA [8.261388753972234]
We revisit the class of iterative shrinkage-thresholding algorithms (ISTA) for solving the linear inverse problem with sparse representation.
We find that the previous assumption for the smooth part to be convex weakens the least-square model.
We generalize the linear convergence to composite optimization in both the objective value and the squared proximal subgradient norm.
arXiv Detail & Related papers (2022-12-13T02:02:50Z) - Learning Sparse Graph with Minimax Concave Penalty under Gaussian Markov
Random Fields [51.07460861448716]
This paper presents a convex-analytic framework to learn from data.
We show that a triangular convexity decomposition is guaranteed by a transform of the corresponding to its upper part.
arXiv Detail & Related papers (2021-09-17T17:46:12Z) - Unfolding Projection-free SDP Relaxation of Binary Graph Classifier via
GDPA Linearization [59.87663954467815]
Algorithm unfolding creates an interpretable and parsimonious neural network architecture by implementing each iteration of a model-based algorithm as a neural layer.
In this paper, leveraging a recent linear algebraic theorem called Gershgorin disc perfect alignment (GDPA), we unroll a projection-free algorithm for semi-definite programming relaxation (SDR) of a binary graph.
Experimental results show that our unrolled network outperformed pure model-based graph classifiers, and achieved comparable performance to pure data-driven networks but using far fewer parameters.
arXiv Detail & Related papers (2021-09-10T07:01:15Z) - Why Approximate Matrix Square Root Outperforms Accurate SVD in Global
Covariance Pooling? [59.820507600960745]
We propose a new GCP meta-layer that uses SVD in the forward pass, and Pad'e Approximants in the backward propagation to compute the gradients.
The proposed meta-layer has been integrated into different CNN models and achieves state-of-the-art performances on both large-scale and fine-grained datasets.
arXiv Detail & Related papers (2021-05-06T08:03:45Z) - Learning Sparse Graph Laplacian with K Eigenvector Prior via Iterative
GLASSO and Projection [58.5350491065936]
We consider a structural assumption on the graph Laplacian matrix $L$.
The first $K$ eigenvectors of $L$ are pre-selected, e.g., based on domain-specific criteria.
We design an efficient hybrid graphical lasso/projection algorithm to compute the most suitable graph Laplacian matrix $L* in H_u+$ given $barC$.
arXiv Detail & Related papers (2020-10-25T18:12:50Z) - On Box-Cox Transformation for Image Normality and Pattern Classification [0.6548580592686074]
This paper revolves around the utility of such a tool as a pre-processing step to transform two-dimensional data.
We compare the effect of this light-weight Box-Cox transformation with well-established state-of-the-art low light image enhancement techniques.
We also demonstrate the effectiveness of our approach through several test-bed data sets for generic improvement of visual appearance of images.
arXiv Detail & Related papers (2020-04-15T17:10:18Z) - Mixture model for designs in high dimensional regression and the LASSO [0.0]
The LASSO is a technique for variable selection in the regression model bean y & = & Xbeta + z, eean.
This paper proposes a mixture model for the design matrix which is able to capture in a natural way the potentially clustered nature of the columns.
arXiv Detail & Related papers (2012-10-17T15:10:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.