Fast Fourier Transformation for Optimizing Convolutional Neural Networks
in Object Recognition
- URL: http://arxiv.org/abs/2010.04257v1
- Date: Thu, 8 Oct 2020 21:07:55 GMT
- Title: Fast Fourier Transformation for Optimizing Convolutional Neural Networks
in Object Recognition
- Authors: Varsha Nair, Moitrayee Chatterjee, Neda Tavakoli, Akbar Siami Namin,
Craig Snoeyink
- Abstract summary: This paper proposes to use Fast Fourier Transformation-based U-Net (a refined fully convolutional networks) to perform image convolution in neural networks.
We implement the FFT-based convolutional neural network to improve the training time of the network.
Our model demonstrated improvement in training time during convolution from $600-700$ ms/step to $400-500$ ms/step.
- Score: 1.0499611180329802
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes to use Fast Fourier Transformation-based U-Net (a refined
fully convolutional networks) and perform image convolution in neural networks.
Leveraging the Fast Fourier Transformation, it reduces the image convolution
costs involved in the Convolutional Neural Networks (CNNs) and thus reduces the
overall computational costs. The proposed model identifies the object
information from the images. We apply the Fast Fourier transform algorithm on
an image data set to obtain more accessible information about the image data,
before segmenting them through the U-Net architecture. More specifically, we
implement the FFT-based convolutional neural network to improve the training
time of the network. The proposed approach was applied to publicly available
Broad Bioimage Benchmark Collection (BBBC) dataset. Our model demonstrated
improvement in training time during convolution from $600-700$ ms/step to
$400-500$ ms/step. We evaluated the accuracy of our model using Intersection
over Union (IoU) metric showing significant improvements.
Related papers
- LeRF: Learning Resampling Function for Adaptive and Efficient Image Interpolation [64.34935748707673]
Recent deep neural networks (DNNs) have made impressive progress in performance by introducing learned data priors.
We propose a novel method of Learning Resampling (termed LeRF) which takes advantage of both the structural priors learned by DNNs and the locally continuous assumption.
LeRF assigns spatially varying resampling functions to input image pixels and learns to predict the shapes of these resampling functions with a neural network.
arXiv Detail & Related papers (2024-07-13T16:09:45Z) - TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - Training Convolutional Neural Networks with the Forward-Forward
algorithm [1.74440662023704]
Forward Forward (FF) algorithm has up to now only been used in fully connected networks.
We show how the FF paradigm can be extended to CNNs.
Our FF-trained CNN, featuring a novel spatially-extended labeling technique, achieves a classification accuracy of 99.16% on the MNIST hand-written digits dataset.
arXiv Detail & Related papers (2023-12-22T18:56:35Z) - FFEINR: Flow Feature-Enhanced Implicit Neural Representation for
Spatio-temporal Super-Resolution [4.577685231084759]
This paper proposes a Feature-Enhanced Neural Implicit Representation (FFEINR) for super-resolution of flow field data.
It can take full advantage of the implicit neural representation in terms of model structure and sampling resolution.
The training process of FFEINR is facilitated by introducing feature enhancements for the input layer.
arXiv Detail & Related papers (2023-08-24T02:28:18Z) - Deep Multi-Threshold Spiking-UNet for Image Processing [51.88730892920031]
This paper introduces the novel concept of Spiking-UNet for image processing, which combines the power of Spiking Neural Networks (SNNs) with the U-Net architecture.
To achieve an efficient Spiking-UNet, we face two primary challenges: ensuring high-fidelity information propagation through the network via spikes and formulating an effective training strategy.
Experimental results show that, on image segmentation and denoising, our Spiking-UNet achieves comparable performance to its non-spiking counterpart.
arXiv Detail & Related papers (2023-07-20T16:00:19Z) - Fourier-Net+: Leveraging Band-Limited Representation for Efficient 3D
Medical Image Registration [62.53130123397081]
U-Net style networks are commonly utilized in unsupervised image registration to predict dense displacement fields.
We first propose Fourier-Net, which replaces the costly U-Net style expansive path with a parameter-free model-driven decoder.
We then introduce Fourier-Net+, which additionally takes the band-limited spatial representation of the images as input and further reduces the number of convolutional layers in the U-Net style network's contracting path.
arXiv Detail & Related papers (2023-07-06T13:57:12Z) - Feature transforms for image data augmentation [74.12025519234153]
In image classification, many augmentation approaches utilize simple image manipulation algorithms.
In this work, we build ensembles on the data level by adding images generated by combining fourteen augmentation approaches.
Pretrained ResNet50 networks are finetuned on training sets that include images derived from each augmentation method.
arXiv Detail & Related papers (2022-01-24T14:12:29Z) - Patch Based Transformation for Minimum Variance Beamformer Image
Approximation Using Delay and Sum Pipeline [0.0]
In this work, a patch level U-Net based neural network is proposed, where the delay compensated radio frequency (RF) patch for a fixed region in space is transformed through a U-Net architecture.
The proposed approach treats the non-linear transformation of the RF data space that can account for the data driven weight adaptation done by the MVDR approach in the parameters of the network.
arXiv Detail & Related papers (2021-10-19T19:36:59Z) - Augmentation Inside the Network [1.5260179407438161]
We present augmentation inside the network, a method that simulates data augmentation techniques for computer vision problems.
We validate our method on the ImageNet-2012 and CIFAR-100 datasets for image classification.
arXiv Detail & Related papers (2020-12-19T20:07:03Z) - Computational optimization of convolutional neural networks using
separated filters architecture [69.73393478582027]
We consider a convolutional neural network transformation that reduces computation complexity and thus speedups neural network processing.
Use of convolutional neural networks (CNN) is the standard approach to image recognition despite the fact they can be too computationally demanding.
arXiv Detail & Related papers (2020-02-18T17:42:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.