DWM: A Decomposable Winograd Method for Convolution Acceleration
- URL: http://arxiv.org/abs/2002.00552v1
- Date: Mon, 3 Feb 2020 03:42:56 GMT
- Title: DWM: A Decomposable Winograd Method for Convolution Acceleration
- Authors: Di Huang, Xishan Zhang, Rui Zhang, Tian Zhi, Deyuan He, Jiaming Guo,
Chang Liu, Qi Guo, Zidong Du, Shaoli Liu, Tianshi Chen, Yunji Chen
- Abstract summary: Winograd's minimal filtering algorithm has been widely used in Convolutional Neural Networks (CNNs) to reduce the number of multiplications for faster processing.
It suffers from significantly increased FLOPs and numerical accuracy problem for kernel size larger than 3x3 and fails on convolution with stride larger than 1.
We propose a novel Decomposable Winograd Method (DWM) which breaks through the limitation of original Winograd's minimal filtering algorithm to a wide and general convolutions.
- Score: 29.312042061351782
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Winograd's minimal filtering algorithm has been widely used in Convolutional
Neural Networks (CNNs) to reduce the number of multiplications for faster
processing. However, it is only effective on convolutions with kernel size as
3x3 and stride as 1, because it suffers from significantly increased FLOPs and
numerical accuracy problem for kernel size larger than 3x3 and fails on
convolution with stride larger than 1. In this paper, we propose a novel
Decomposable Winograd Method (DWM), which breaks through the limitation of
original Winograd's minimal filtering algorithm to a wide and general
convolutions. DWM decomposes kernels with large size or large stride to several
small kernels with stride as 1 for further applying Winograd method, so that
DWM can reduce the number of multiplications while keeping the numerical
accuracy. It enables the fast exploring of larger kernel size and larger stride
value in CNNs for high performance and accuracy and even the potential for new
CNNs. Comparing against the original Winograd, the proposed DWM is able to
support all kinds of convolutions with a speedup of ~2, without affecting the
numerical accuracy.
Related papers
- Efficient Dataset Distillation Using Random Feature Approximation [109.07737733329019]
We propose a novel algorithm that uses a random feature approximation (RFA) of the Neural Network Gaussian Process (NNGP) kernel.
Our algorithm provides at least a 100-fold speedup over KIP and can run on a single GPU.
Our new method, termed an RFA Distillation (RFAD), performs competitively with KIP and other dataset condensation algorithms in accuracy over a range of large-scale datasets.
arXiv Detail & Related papers (2022-10-21T15:56:13Z) - Going Further With Winograd Convolutions: Tap-Wise Quantization for
Efficient Inference on 4x4 Tile [7.705762754955851]
Winograd convolution algorithm computes convolutions with fewer MACs compared to the standard algorithm.
We propose a novel tap-wise quantization method that overcomes the numerical issues of using larger tiles.
We show how to integrate such custom modules in an industrial-grade, programmable DSA.
arXiv Detail & Related papers (2022-09-26T19:29:51Z) - Instant Neural Graphics Primitives with a Multiresolution Hash Encoding [67.33850633281803]
We present a versatile new input encoding that permits the use of a smaller network without sacrificing quality.
A small neural network is augmented by a multiresolution hash table of trainable feature vectors whose values are optimized through a gradient descent.
We achieve a combined speed of several orders of magnitude, enabling training of high-quality neural graphics primitives in a matter of seconds.
arXiv Detail & Related papers (2022-01-16T07:22:47Z) - Fast Convolution based on Winograd Minimum Filtering: Introduction and
Development [5.192451499848539]
Convolution operators are the fundamental component of convolutional neural networks.
In recent years, researchers have proposed several fast convolution algorithms including FFT and Winograd.
This article summarizes the development of Winograd convolution from the three aspects of algorithm expansion, algorithm optimization, implementation, and application.
arXiv Detail & Related papers (2021-11-01T14:39:56Z) - Content-Aware Convolutional Neural Networks [98.97634685964819]
Convolutional Neural Networks (CNNs) have achieved great success due to the powerful feature learning ability of convolution layers.
We propose a Content-aware Convolution (CAC) that automatically detects the smooth windows and applies a 1x1 convolutional kernel to replace the original large kernel.
arXiv Detail & Related papers (2021-06-30T03:54:35Z) - Winograd Algorithm for AdderNet [54.93995545896655]
Adder neural network (AdderNet) is a new kind of deep model that replaces the original massive multiplications in convolutions by additions.
This paper studies the winograd algorithm, which is a widely used fast algorithm for accelerating convolution and saving the computational costs.
arXiv Detail & Related papers (2021-05-12T09:13:34Z) - Decoupled Dynamic Filter Networks [85.38058820176047]
We propose the Decoupled Dynamic Filter (DDF) that can simultaneously tackle both of these shortcomings.
Inspired by recent advances in attention, DDF decouples a depth-wise dynamic filter into spatial and channel dynamic filters.
We observe a significant boost in performance when replacing standard convolution with DDF in classification networks.
arXiv Detail & Related papers (2021-04-29T04:55:33Z) - Accelerating Large Kernel Convolutions with Nested Winograd
Transformation.pdf [2.193040410545991]
This work proposes a nested Winograd algorithm that iteratively decomposes a large kernel convolution into small kernel convolutions.
Experiments show that compared to the linear decomposition Winograd algorithm, the proposed algorithm reduces the total number of multiplications by 1.4 to 10.5 times for computing 4x4 to 31x31 convolutions.
arXiv Detail & Related papers (2021-02-26T02:42:42Z) - Efficient Residue Number System Based Winograd Convolution [15.210764522845416]
Winograd algorithm can reduce the computational complexity of convolutional neural networks (CNN) with weights and activations represented in floating point.
Our work extends the Winograd algorithm to Residue Number System (RNS)
The minimal complexity convolution is computed precisely over large transformation tile.
arXiv Detail & Related papers (2020-07-23T19:07:06Z) - Quantaized Winograd/Toom-Cook Convolution for DNNs: Beyond Canonical
Polynomials Base [0.0]
Winograd convolution algorithm is a common used method that significantly reduces time consumption.
We present the application of base change technique for quantized Winograd-aware training model.
arXiv Detail & Related papers (2020-04-23T11:15:27Z) - XSepConv: Extremely Separated Convolution [60.90871656244126]
We propose a novel extremely separated convolutional block (XSepConv)
It fuses spatially separable convolutions into depthwise convolution to reduce both the computational cost and parameter size of large kernels.
XSepConv is designed to be an efficient alternative to vanilla depthwise convolution with large kernel sizes.
arXiv Detail & Related papers (2020-02-27T11:46:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.