Rethinking FUN: Frequency-Domain Utilization Networks
- URL: http://arxiv.org/abs/2012.03357v1
- Date: Sun, 6 Dec 2020 19:16:37 GMT
- Title: Rethinking FUN: Frequency-Domain Utilization Networks
- Authors: Kfir Goldberg, Stav Shapiro, Elad Richardson, Shai Avidan
- Abstract summary: We present FUN, a family of novel Frequency-domain Utilization Networks.
These networks utilize the inherent efficiency of the frequency-domain by working directly in that domain.
We show that working in frequency domain allows for dynamic compression of the input at inference time without any explicit change to the architecture.
- Score: 21.10493050675827
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The search for efficient neural network architectures has gained much focus
in recent years, where modern architectures focus not only on accuracy but also
on inference time and model size. Here, we present FUN, a family of novel
Frequency-domain Utilization Networks. These networks utilize the inherent
efficiency of the frequency-domain by working directly in that domain,
represented with the Discrete Cosine Transform. Using modern techniques and
building blocks such as compound-scaling and inverted-residual layers we
generate a set of such networks allowing one to balance between size, latency
and accuracy while outperforming competing RGB-based models. Extensive
evaluations verifies that our networks present strong alternatives to previous
approaches. Moreover, we show that working in frequency domain allows for
dynamic compression of the input at inference time without any explicit change
to the architecture.
Related papers
- TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - From NeurODEs to AutoencODEs: a mean-field control framework for
width-varying Neural Networks [68.8204255655161]
We propose a new type of continuous-time control system, called AutoencODE, based on a controlled field that drives dynamics.
We show that many architectures can be recovered in regions where the loss function is locally convex.
arXiv Detail & Related papers (2023-07-05T13:26:17Z) - Properties and Potential Applications of Random Functional-Linked Types
of Neural Networks [81.56822938033119]
Random functional-linked neural networks (RFLNNs) offer an alternative way of learning in deep structure.
This paper gives some insights into the properties of RFLNNs from the viewpoints of frequency domain.
We propose a method to generate a BLS network with better performance, and design an efficient algorithm for solving Poison's equation.
arXiv Detail & Related papers (2023-04-03T13:25:22Z) - Vector Quantized Time Series Generation with a Bidirectional Prior Model [0.3867363075280544]
Time series generation (TSG) studies have mainly focused on the use of Generative Adversarial Networks (GANs) combined with recurrent neural network (RNN) variants.
We propose TimeVQVAE, the first work, to our knowledge, that uses vector quantization (VQ) techniques to address the TSG problem.
We also propose VQ modeling in a time-frequency domain, separated into low-frequency (LF) and high-frequency (HF)
arXiv Detail & Related papers (2023-03-08T17:27:39Z) - Interference Cancellation GAN Framework for Dynamic Channels [74.22393885274728]
We introduce an online training framework that can adapt to any changes in the channel.
Our framework significantly outperforms recent neural network models on highly dynamic channels.
arXiv Detail & Related papers (2022-08-17T02:01:18Z) - A New Clustering-Based Technique for the Acceleration of Deep
Convolutional Networks [2.7393821783237184]
Model Compression and Acceleration (MCA) techniques are used to transform large pre-trained networks into smaller models.
We propose a clustering-based approach that is able to increase the number of employed centroids/representatives.
This is achieved by imposing a special structure to the employed representatives, which is enabled by the particularities of the problem at hand.
arXiv Detail & Related papers (2021-07-19T18:22:07Z) - Learning Frequency-aware Dynamic Network for Efficient Super-Resolution [56.98668484450857]
This paper explores a novel frequency-aware dynamic network for dividing the input into multiple parts according to its coefficients in the discrete cosine transform (DCT) domain.
In practice, the high-frequency part will be processed using expensive operations and the lower-frequency part is assigned with cheap operations to relieve the computation burden.
Experiments conducted on benchmark SISR models and datasets show that the frequency-aware dynamic network can be employed for various SISR neural architectures.
arXiv Detail & Related papers (2021-03-15T12:54:26Z) - Orthogonalized SGD and Nested Architectures for Anytime Neural Networks [30.598394152055338]
Orthogonalized SGD dynamically re-balances task-specific gradients when training a multitask network.
Experiments demonstrate that training with Orthogonalized SGD significantly improves accuracy of anytime networks.
arXiv Detail & Related papers (2020-08-15T03:06:34Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.