Joint Multi-Dimension Pruning via Numerical Gradient Update
- URL: http://arxiv.org/abs/2005.08931v2
- Date: Sat, 25 Sep 2021 16:03:11 GMT
- Title: Joint Multi-Dimension Pruning via Numerical Gradient Update
- Authors: Zechun Liu and Xiangyu Zhang and Zhiqiang Shen and Zhe Li and Yichen
Wei and Kwang-Ting Cheng and Jian Sun
- Abstract summary: We present joint multi-dimension pruning (abbreviated as JointPruning), an effective method of pruning a network on three crucial aspects: spatial, depth and channel simultaneously.
We show that our method is optimized collaboratively across the three dimensions in a single end-to-end training and it is more efficient than the previous exhaustive methods.
- Score: 120.59697866489668
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present joint multi-dimension pruning (abbreviated as JointPruning), an
effective method of pruning a network on three crucial aspects: spatial, depth
and channel simultaneously. To tackle these three naturally different
dimensions, we proposed a general framework by defining pruning as seeking the
best pruning vector (i.e., the numerical value of layer-wise channel number,
spacial size, depth) and construct a unique mapping from the pruning vector to
the pruned network structures. Then we optimize the pruning vector with
gradient update and model joint pruning as a numerical gradient optimization
process. To overcome the challenge that there is no explicit function between
the loss and the pruning vectors, we proposed self-adapted stochastic gradient
estimation to construct a gradient path through network loss to pruning vectors
and enable efficient gradient update. We show that the joint strategy discovers
a better status than previous studies that focused on individual dimensions
solely, as our method is optimized collaboratively across the three dimensions
in a single end-to-end training and it is more efficient than the previous
exhaustive methods. Extensive experiments on large-scale ImageNet dataset
across a variety of network architectures MobileNet V1&V2&V3 and ResNet
demonstrate the effectiveness of our proposed method. For instance, we achieve
significant margins of 2.5% and 2.6% improvement over the state-of-the-art
approach on the already compact MobileNet V1&V2 under an extremely large
compression ratio.
Related papers
- 360 Layout Estimation via Orthogonal Planes Disentanglement and Multi-view Geometric Consistency Perception [56.84921040837699]
Existing panoramic layout estimation solutions tend to recover room boundaries from a vertically compressed sequence, yielding imprecise results.
We propose an orthogonal plane disentanglement network (termed DOPNet) to distinguish ambiguous semantics.
We also present an unsupervised adaptation technique tailored for horizon-depth and ratio representations.
Our solution outperforms other SoTA models on both monocular layout estimation and multi-view layout estimation tasks.
arXiv Detail & Related papers (2023-12-26T12:16:03Z) - Differentiable Transportation Pruning [23.766356215156488]
Pruning methods are a key tool for edge deployment as they can improve storage, compute, memory bandwidth, and energy usage.
We propose a novel accurate pruning technique that allows precise control over the output network size.
We show that our method achieves state-of-the-art performance compared to previous pruning methods on 3 different datasets.
arXiv Detail & Related papers (2023-07-17T13:44:11Z) - GraphCSPN: Geometry-Aware Depth Completion via Dynamic GCNs [49.55919802779889]
We propose a Graph Convolution based Spatial Propagation Network (GraphCSPN) as a general approach for depth completion.
In this work, we leverage convolution neural networks as well as graph neural networks in a complementary way for geometric representation learning.
Our method achieves the state-of-the-art performance, especially when compared in the case of using only a few propagation steps.
arXiv Detail & Related papers (2022-10-19T17:56:03Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - Deep-3DAligner: Unsupervised 3D Point Set Registration Network With
Optimizable Latent Vector [15.900382629390297]
We propose to develop a novel model that integrates the optimization to learning, aiming to address the technical challenges in 3D registration.
In addition to the deep transformation decoding network, our framework introduce an optimizable deep underlineSpatial underlineCorrelation underlineRepresentation.
arXiv Detail & Related papers (2020-09-29T22:44:38Z) - Adaptive Context-Aware Multi-Modal Network for Depth Completion [107.15344488719322]
We propose to adopt the graph propagation to capture the observed spatial contexts.
We then apply the attention mechanism on the propagation, which encourages the network to model the contextual information adaptively.
Finally, we introduce the symmetric gated fusion strategy to exploit the extracted multi-modal features effectively.
Our model, named Adaptive Context-Aware Multi-Modal Network (ACMNet), achieves the state-of-the-art performance on two benchmarks.
arXiv Detail & Related papers (2020-08-25T06:00:06Z) - Lookahead: A Far-Sighted Alternative of Magnitude-based Pruning [83.99191569112682]
Magnitude-based pruning is one of the simplest methods for pruning neural networks.
We develop a simple pruning method, coined lookahead pruning, by extending the single layer optimization to a multi-layer optimization.
Our experimental results demonstrate that the proposed method consistently outperforms magnitude-based pruning on various networks.
arXiv Detail & Related papers (2020-02-12T05:38:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.