Multi-Scale Architectures Matter: On the Adversarial Robustness of
Flow-based Lossless Compression
- URL: http://arxiv.org/abs/2208.12716v1
- Date: Fri, 26 Aug 2022 15:17:43 GMT
- Title: Multi-Scale Architectures Matter: On the Adversarial Robustness of
Flow-based Lossless Compression
- Authors: Yi-chong Xia, Bin Chen, Yan Feng, Tian-shuo Ge
- Abstract summary: Flow-based models perform better due to their excellent probability density estimation and satisfactory inference speed.
Multi-scale architecture provides a shortcut from the shallow layer to the output layer.
Flows with multi-scale architecture achieve the best trade-off between coding complexity and compression efficiency.
- Score: 16.109578069331135
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: As a probabilistic modeling technique, the flow-based model has demonstrated
remarkable potential in the field of lossless compression
\cite{idf,idf++,lbb,ivpf,iflow},. Compared with other deep generative models
(eg. Autoregressive, VAEs) \cite{bitswap,hilloc,pixelcnn++,pixelsnail} that
explicitly model the data distribution probabilities, flow-based models perform
better due to their excellent probability density estimation and satisfactory
inference speed. In flow-based models, multi-scale architecture provides a
shortcut from the shallow layer to the output layer, which significantly
reduces the computational complexity and avoid performance degradation when
adding more layers. This is essential for constructing an advanced flow-based
learnable bijective mapping. Furthermore, the lightweight requirement of the
model design in practical compression tasks suggests that flows with
multi-scale architecture achieve the best trade-off between coding complexity
and compression efficiency.
Related papers
- Generalized Nested Latent Variable Models for Lossy Coding applied to Wind Turbine Scenarios [14.48369551534582]
A learning-based approach seeks to minimize the compromise between compression rate and reconstructed image quality.
A successful technique consists in introducing a deep hyperprior that operates within a 2-level nested latent variable model.
This paper extends this concept by designing a generalized L-level nested generative model with a Markov chain structure.
arXiv Detail & Related papers (2024-06-10T11:00:26Z) - ECoFLaP: Efficient Coarse-to-Fine Layer-Wise Pruning for Vision-Language
Models [70.45441031021291]
Large Vision-Language Models (LVLMs) can understand the world comprehensively by integrating rich information from different modalities.
LVLMs are often problematic due to their massive computational/energy costs and carbon consumption.
We propose Efficient Coarse-to-Fine LayerWise Pruning (ECoFLaP), a two-stage coarse-to-fine weight pruning approach for LVLMs.
arXiv Detail & Related papers (2023-10-04T17:34:00Z) - Lifting Architectural Constraints of Injective Flows [7.452460759055847]
Normalizing Flows explicitly maximize a full-dimensional likelihood on the training data.
Injective Flows fix this by jointly learning a manifold and the distribution on it.
We show that naively learning both the data manifold and the distribution on it can lead to divergent solutions.
arXiv Detail & Related papers (2023-06-02T18:03:03Z) - Hierarchical Integration Diffusion Model for Realistic Image Deblurring [71.76410266003917]
Diffusion models (DMs) have been introduced in image deblurring and exhibited promising performance.
We propose the Hierarchical Integration Diffusion Model (HI-Diff), for realistic image deblurring.
Experiments on synthetic and real-world blur datasets demonstrate that our HI-Diff outperforms state-of-the-art methods.
arXiv Detail & Related papers (2023-05-22T12:18:20Z) - FlowNAS: Neural Architecture Search for Optical Flow Estimation [65.44079917247369]
We propose a neural architecture search method named FlowNAS to automatically find the better encoder architecture for flow estimation task.
Experimental results show that the discovered architecture with the weights inherited from the super-network achieves 4.67% F1-all error on KITTI.
arXiv Detail & Related papers (2022-07-04T09:05:25Z) - Sparse Flows: Pruning Continuous-depth Models [107.98191032466544]
We show that pruning improves generalization for neural ODEs in generative modeling.
We also show that pruning finds minimal and efficient neural ODE representations with up to 98% less parameters compared to the original network, without loss of accuracy.
arXiv Detail & Related papers (2021-06-24T01:40:17Z) - Learning Deep-Latent Hierarchies by Stacking Wasserstein Autoencoders [22.54887526392739]
We propose a novel approach to training models with deep-latent hierarchies based on Optimal Transport.
We show that our method enables the generative model to fully leverage its deep-latent hierarchy, avoiding the well known "latent variable collapse" issue of VAEs.
arXiv Detail & Related papers (2020-10-07T15:04:20Z) - Normalizing Flows with Multi-Scale Autoregressive Priors [131.895570212956]
We introduce channel-wise dependencies in their latent space through multi-scale autoregressive priors (mAR)
Our mAR prior for models with split coupling flow layers (mAR-SCF) can better capture dependencies in complex multimodal data.
We show that mAR-SCF allows for improved image generation quality, with gains in FID and Inception scores compared to state-of-the-art flow-based models.
arXiv Detail & Related papers (2020-04-08T09:07:11Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z) - Closing the Dequantization Gap: PixelCNN as a Single-Layer Flow [16.41460104376002]
We introduce subset flows, a class of flows that can transform finite volumes and allow exact computation of likelihoods for discrete data.
We identify ordinal discrete autoregressive models, including WaveNets, PixelCNNs and Transformers, as single-layer flows.
We demonstrate state-of-the-art results on CIFAR-10 for flow models trained with dequantization.
arXiv Detail & Related papers (2020-02-06T22:58:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.