PRANC: Pseudo RAndom Networks for Compacting deep models
- URL: http://arxiv.org/abs/2206.08464v2
- Date: Mon, 28 Aug 2023 22:09:07 GMT
- Title: PRANC: Pseudo RAndom Networks for Compacting deep models
- Authors: Parsa Nooralinejad, Ali Abbasi, Soroush Abbasi Koohpayegani, Kossar
Pourahmadi Meibodi, Rana Muhammad Shahroz Khan, Soheil Kolouri, Hamed
Pirsiavash
- Abstract summary: PRANC enables significant compaction of a deep model.
In this study, we employ PRANC to condense image classification models and compress images by compacting their associated implicit neural networks.
- Score: 22.793523211040682
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We demonstrate that a deep model can be reparametrized as a linear
combination of several randomly initialized and frozen deep models in the
weight space. During training, we seek local minima that reside within the
subspace spanned by these random models (i.e., `basis' networks). Our
framework, PRANC, enables significant compaction of a deep model. The model can
be reconstructed using a single scalar `seed,' employed to generate the
pseudo-random `basis' networks, together with the learned linear mixture
coefficients.
In practical applications, PRANC addresses the challenge of efficiently
storing and communicating deep models, a common bottleneck in several
scenarios, including multi-agent learning, continual learners, federated
systems, and edge devices, among others. In this study, we employ PRANC to
condense image classification models and compress images by compacting their
associated implicit neural networks. PRANC outperforms baselines with a large
margin on image classification when compressing a deep model almost $100$
times. Moreover, we show that PRANC enables memory-efficient inference by
generating layer-wise weights on the fly. The source code of PRANC is here:
\url{https://github.com/UCDvision/PRANC}
Related papers
- FreeSeg-Diff: Training-Free Open-Vocabulary Segmentation with Diffusion Models [56.71672127740099]
We focus on the task of image segmentation, which is traditionally solved by training models on closed-vocabulary datasets.
We leverage different and relatively small-sized, open-source foundation models for zero-shot open-vocabulary segmentation.
Our approach (dubbed FreeSeg-Diff), which does not rely on any training, outperforms many training-based approaches on both Pascal VOC and COCO datasets.
arXiv Detail & Related papers (2024-03-29T10:38:25Z) - Learning multi-scale local conditional probability models of images [7.07848787073901]
Deep neural networks can learn powerful prior probability models for images, as evidenced by the high-quality generations obtained with recent score-based diffusion methods.
But the means by which these networks capture complex global statistical structure, apparently without suffering from the curse of dimensionality, remain a mystery.
We incorporate diffusion methods into a multi-scale decomposition, reducing dimensionality by assuming a stationary local Markov model for wavelet coefficients conditioned on coarser-scale coefficients.
arXiv Detail & Related papers (2023-03-06T09:23:14Z) - Parameter-Efficient Masking Networks [61.43995077575439]
Advanced network designs often contain a large number of repetitive structures (e.g., Transformer)
In this study, we are the first to investigate the representative potential of fixed random weights with limited unique values by learning masks.
It leads to a new paradigm for model compression to diminish the model size.
arXiv Detail & Related papers (2022-10-13T03:39:03Z) - Adaptive Convolutional Dictionary Network for CT Metal Artifact
Reduction [62.691996239590125]
We propose an adaptive convolutional dictionary network (ACDNet) for metal artifact reduction.
Our ACDNet can automatically learn the prior for artifact-free CT images via training data and adaptively adjust the representation kernels for each input CT image.
Our method inherits the clear interpretability of model-based methods and maintains the powerful representation ability of learning-based methods.
arXiv Detail & Related papers (2022-05-16T06:49:36Z) - Scaling Structured Inference with Randomization [64.18063627155128]
We propose a family of dynamic programming (RDP) randomized for scaling structured models to tens of thousands of latent states.
Our method is widely applicable to classical DP-based inference.
It is also compatible with automatic differentiation so can be integrated with neural networks seamlessly.
arXiv Detail & Related papers (2021-12-07T11:26:41Z) - Fully differentiable model discovery [0.0]
We propose an approach by combining neural network based surrogates with Sparse Bayesian Learning.
Our work expands PINNs to various types of neural network architectures, and connects neural network-based surrogates to the rich field of Bayesian parameter inference.
arXiv Detail & Related papers (2021-06-09T08:11:23Z) - Mixture-based Feature Space Learning for Few-shot Image Classification [6.574517227976925]
We propose to model base classes with mixture models by simultaneously training the feature extractor and learning the mixture model parameters in an online manner.
Results in a richer and more discriminative feature space which can be employed to classify novel examples from very few samples.
arXiv Detail & Related papers (2020-11-24T03:16:27Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.