Frequency Regularization: Restricting Information Redundancy of
Convolutional Neural Networks
- URL: http://arxiv.org/abs/2304.07973v3
- Date: Sun, 30 Apr 2023 20:07:52 GMT
- Title: Frequency Regularization: Restricting Information Redundancy of
Convolutional Neural Networks
- Authors: Chenqiu Zhao, Guanfang Dong, Shupei Zhang, Zijie Tan, Anup Basu
- Abstract summary: Convolutional neural networks have demonstrated impressive results in many computer vision tasks.
The increasing size of these networks raises concerns about the information overload resulting from the large number of network parameters.
We propose Frequency Regularization to restrict the non-zero elements of the network parameters in the frequency domain.
- Score: 6.387263468033964
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Convolutional neural networks have demonstrated impressive results in many
computer vision tasks. However, the increasing size of these networks raises
concerns about the information overload resulting from the large number of
network parameters. In this paper, we propose Frequency Regularization to
restrict the non-zero elements of the network parameters in the frequency
domain. The proposed approach operates at the tensor level, and can be applied
to almost all network architectures. Specifically, the tensors of parameters
are maintained in the frequency domain, where high frequency components can be
eliminated by zigzag setting tensor elements to zero. Then, the inverse
discrete cosine transform (IDCT) is used to reconstruct the spatial tensors for
matrix operations during network training. Since high frequency components of
images are known to be less critical, a large proportion of these parameters
can be set to zero when networks are trained with the proposed frequency
regularization. Comprehensive evaluations on various state-of-the-art network
architectures, including LeNet, Alexnet, VGG, Resnet, ViT, UNet, GAN, and VAE,
demonstrate the effectiveness of the proposed frequency regularization. For a
very small accuracy decrease (less than 2\%), a LeNet5 with 0.4M parameters can
be represented by only 776 float16 numbers (over 1100$\times$ reduction), and a
UNet with 34M parameters can be represented by only 759 float16 numbers (over
80000$\times$ reduction). In particular, the original size of the UNet model is
366MB, we reduce it to 4.5kb.
Related papers
- Accelerating Inference of Networks in the Frequency Domain [8.125023712173686]
We propose performing network inference in the frequency domain to speed up networks whose frequency parameters are sparse.
In particular, we propose a frequency inference chain that is dual to the network inference in the spatial domain.
The proposed approach significantly improves accuracy in the case of a high speedup ratio (over 100x)
arXiv Detail & Related papers (2024-10-06T03:34:38Z) - Efficient View Synthesis with Neural Radiance Distribution Field [61.22920276806721]
We propose a new representation called Neural Radiance Distribution Field (NeRDF) that targets efficient view synthesis in real-time.
We use a small network similar to NeRF while preserving the rendering speed with a single network forwarding per pixel as in NeLF.
Experiments show that our proposed method offers a better trade-off among speed, quality, and network size than existing methods.
arXiv Detail & Related papers (2023-08-22T02:23:28Z) - A Scalable Walsh-Hadamard Regularizer to Overcome the Low-degree
Spectral Bias of Neural Networks [79.28094304325116]
Despite the capacity of neural nets to learn arbitrary functions, models trained through gradient descent often exhibit a bias towards simpler'' functions.
We show how this spectral bias towards low-degree frequencies can in fact hurt the neural network's generalization on real-world datasets.
We propose a new scalable functional regularization scheme that aids the neural network to learn higher degree frequencies.
arXiv Detail & Related papers (2023-05-16T20:06:01Z) - Frequency Disentangled Residual Network [11.388328269522006]
Residual networks (ResNets) have been utilized for various computer vision and image processing applications.
A residual block consists of few convolutional layers having trainable parameters, which leads to overfitting.
A frequency disentangled residual network (FDResNet) is proposed to tackle these issues.
arXiv Detail & Related papers (2021-09-26T10:52:18Z) - DS-Net++: Dynamic Weight Slicing for Efficient Inference in CNNs and
Transformers [105.74546828182834]
We show a hardware-efficient dynamic inference regime, named dynamic weight slicing, which adaptively slice a part of network parameters for inputs with diverse difficulty levels.
We present dynamic slimmable network (DS-Net) and dynamic slice-able network (DS-Net++) by input-dependently adjusting filter numbers of CNNs and multiple dimensions in both CNNs and transformers.
arXiv Detail & Related papers (2021-09-21T09:57:21Z) - Model Pruning Based on Quantified Similarity of Feature Maps [5.271060872578571]
We propose a novel theory to find redundant information in three dimensional tensors.
We use this theory to prune convolutional neural networks to enhance the inference speed.
arXiv Detail & Related papers (2021-05-13T02:57:30Z) - Learning Frequency-aware Dynamic Network for Efficient Super-Resolution [56.98668484450857]
This paper explores a novel frequency-aware dynamic network for dividing the input into multiple parts according to its coefficients in the discrete cosine transform (DCT) domain.
In practice, the high-frequency part will be processed using expensive operations and the lower-frequency part is assigned with cheap operations to relieve the computation burden.
Experiments conducted on benchmark SISR models and datasets show that the frequency-aware dynamic network can be employed for various SISR neural architectures.
arXiv Detail & Related papers (2021-03-15T12:54:26Z) - ThriftyNets : Convolutional Neural Networks with Tiny Parameter Budget [4.438259529250529]
We propose a new convolutional neural network architecture, called ThriftyNet.
In ThriftyNet, only one convolutional layer is defined and used recursively, leading to a maximal parameter factorization.
We achieve competitive performance on a tiny parameters budget, exceeding 91% accuracy on CIFAR-10 with less than 40K parameters in total, and 74.3% on CIFAR-100 with less than 600K parameters.
arXiv Detail & Related papers (2020-07-20T13:50:51Z) - Efficient Integer-Arithmetic-Only Convolutional Neural Networks [87.01739569518513]
We replace conventional ReLU with Bounded ReLU and find that the decline is due to activation quantization.
Our integer networks achieve equivalent performance as the corresponding FPN networks, but have only 1/4 memory cost and run 2x faster on modern GPU.
arXiv Detail & Related papers (2020-06-21T08:23:03Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z) - Learning in the Frequency Domain [20.045740082113845]
We propose a learning-based frequency selection method to identify the trivial frequency components which can be removed without accuracy loss.
Experiment results show that learning in the frequency domain with static channel selection can achieve higher accuracy than the conventional spatial downsampling approach.
arXiv Detail & Related papers (2020-02-27T19:57:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.