FactorizeNet: Progressive Depth Factorization for Efficient Network
Architecture Exploration Under Quantization Constraints
- URL: http://arxiv.org/abs/2011.14586v1
- Date: Mon, 30 Nov 2020 07:12:26 GMT
- Title: FactorizeNet: Progressive Depth Factorization for Efficient Network
Architecture Exploration Under Quantization Constraints
- Authors: Stone Yun and Alexander Wong
- Abstract summary: We introduce a progressive depth factorization strategy for efficient CNN architecture exploration under quantization constraints.
By algorithmically increasing the granularity of depth factorization in a progressive manner, the proposed strategy enables a fine-grained, low-level analysis of layer-wise distributions.
Such a progressive depth factorization strategy also enables efficient identification of the optimal depth-factorized macroarchitecture design.
- Score: 93.4221402881609
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Depth factorization and quantization have emerged as two of the principal
strategies for designing efficient deep convolutional neural network (CNN)
architectures tailored for low-power inference on the edge. However, there is
still little detailed understanding of how different depth factorization
choices affect the final, trained distributions of each layer in a CNN,
particularly in the situation of quantized weights and activations. In this
study, we introduce a progressive depth factorization strategy for efficient
CNN architecture exploration under quantization constraints. By algorithmically
increasing the granularity of depth factorization in a progressive manner, the
proposed strategy enables a fine-grained, low-level analysis of layer-wise
distributions. Thus enabling the gain of in-depth, layer-level insights on
efficiency-accuracy tradeoffs under fixed-precision quantization. Such a
progressive depth factorization strategy also enables efficient identification
of the optimal depth-factorized macroarchitecture design (which we will refer
to here as FactorizeNet) based on the desired efficiency-accuracy requirements.
Related papers
- Component-based Sketching for Deep ReLU Nets [55.404661149594375]
We develop a sketching scheme based on deep net components for various tasks.
We transform deep net training into a linear empirical risk minimization problem.
We show that the proposed component-based sketching provides almost optimal rates in approximating saturated functions.
arXiv Detail & Related papers (2024-09-21T15:30:43Z) - Differentiable Search for Finding Optimal Quantization Strategy [3.295889768579819]
We propose a differentiable quantization strategy search (DQSS) to assign optimal quantization strategy for individual layer.
We conduct DQSS for post-training quantization to enable their performance to be comparable with that in full precision models.
We also employ DQSS in quantization-aware training for further validating the effectiveness of DQSS.
arXiv Detail & Related papers (2024-04-10T03:22:58Z) - Cross-Layer Optimization for Fault-Tolerant Deep Learning [17.724727744611535]
We propose to characterize deep learning vulnerability difference across both neurons and bits of each neuron, and leverage the vulnerability difference to enable selective protection of the deep learning processing components.
We employ Bayesian optimization strategy to co-optimize the correlated cross-layer design parameters at algorithm layer, architecture layer, and circuit layer.
arXiv Detail & Related papers (2023-12-21T11:35:45Z) - Sensitivity-Aware Mixed-Precision Quantization and Width Optimization of Deep Neural Networks Through Cluster-Based Tree-Structured Parzen Estimation [4.748931281307333]
We introduce an innovative search mechanism for automatically selecting the best bit-width and layer-width for individual neural network layers.
This leads to a marked enhancement in deep neural network efficiency.
arXiv Detail & Related papers (2023-08-12T00:16:51Z) - Large-scale global optimization of ultra-high dimensional non-convex
landscapes based on generative neural networks [0.0]
We present an algorithm manage ultra-high dimensional optimization.
based on a deep generative network.
We show that our method performs better with fewer function evaluations compared to state-of-the-art algorithm.
arXiv Detail & Related papers (2023-07-09T00:05:59Z) - Neural Networks with Quantization Constraints [111.42313650830248]
We present a constrained learning approach to quantization training.
We show that the resulting problem is strongly dual and does away with gradient estimations.
We demonstrate that the proposed approach exhibits competitive performance in image classification tasks.
arXiv Detail & Related papers (2022-10-27T17:12:48Z) - Where Should We Begin? A Low-Level Exploration of Weight Initialization
Impact on Quantized Behaviour of Deep Neural Networks [93.4221402881609]
We present an in-depth, fine-grained ablation study of the effect of different weights initialization on the final distributions of weights and activations of different CNN architectures.
To our best knowledge, we are the first to perform such a low-level, in-depth quantitative analysis of weights initialization and its effect on quantized behaviour.
arXiv Detail & Related papers (2020-11-30T06:54:28Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z) - Discretization-Aware Architecture Search [81.35557425784026]
This paper presents discretization-aware architecture search (DAtextsuperscript2S)
The core idea is to push the super-network towards the configuration of desired topology, so that the accuracy loss brought by discretization is largely alleviated.
Experiments on standard image classification benchmarks demonstrate the superiority of our approach.
arXiv Detail & Related papers (2020-07-07T01:18:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.