Multiscale Attention via Wavelet Neural Operators for Vision
Transformers
- URL: http://arxiv.org/abs/2303.12398v4
- Date: Tue, 15 Aug 2023 09:05:54 GMT
- Title: Multiscale Attention via Wavelet Neural Operators for Vision
Transformers
- Authors: Anahita Nekoozadeh, Mohammad Reza Ahmadzadeh, Zahra Mardani
- Abstract summary: Transformers have achieved widespread success in computer vision. At their heart, there is a Self-Attention (SA) mechanism.
Standard SA mechanism has quadratic complexity with the sequence length, which impedes its utility to long sequences appearing in high resolution vision.
We introduce a Multiscale Wavelet Attention (MWA) by leveraging wavelet neural operators which incurs linear complexity in the sequence size.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Transformers have achieved widespread success in computer vision. At their
heart, there is a Self-Attention (SA) mechanism, an inductive bias that
associates each token in the input with every other token through a weighted
basis. The standard SA mechanism has quadratic complexity with the sequence
length, which impedes its utility to long sequences appearing in high
resolution vision. Recently, inspired by operator learning for PDEs, Adaptive
Fourier Neural Operators (AFNO) were introduced for high resolution attention
based on global convolution that is efficiently implemented via FFT. However,
the AFNO global filtering cannot well represent small and moderate scale
structures that commonly appear in natural images. To leverage the
coarse-to-fine scale structures we introduce a Multiscale Wavelet Attention
(MWA) by leveraging wavelet neural operators which incurs linear complexity in
the sequence size. We replace the attention in ViT with MWA and our experiments
with CIFAR and Tiny-ImageNet classification demonstrate significant improvement
over alternative Fourier-based attentions such as AFNO and Global Filter
Network (GFN).
Related papers
- DiffFNO: Diffusion Fourier Neural Operator [8.895165270489167]
We introduce DiffFNO, a novel diffusion framework for arbitrary-scale super-resolution strengthened by a Weighted Fourier Neural Operator (WFNO)
We show that DiffFNO achieves state-of-the-art (SOTA) results, outperforming existing methods across various scaling factors by a margin of 2 to 4 dB in PSNR.
Our approach sets a new standard in super-resolution, delivering both superior accuracy and computational efficiency.
arXiv Detail & Related papers (2024-11-15T03:14:11Z) - TBSN: Transformer-Based Blind-Spot Network for Self-Supervised Image Denoising [94.09442506816724]
Blind-spot networks (BSN) have been prevalent network architectures in self-supervised image denoising (SSID)
We present a transformer-based blind-spot network (TBSN) by analyzing and redesigning the transformer operators that meet the blind-spot requirement.
For spatial self-attention, an elaborate mask is applied to the attention matrix to restrict its receptive field, thus mimicking the dilated convolution.
For channel self-attention, we observe that it may leak the blind-spot information when the channel number is greater than spatial size in the deep layers of multi-scale architectures.
arXiv Detail & Related papers (2024-04-11T15:39:10Z) - Adaptive Frequency Filters As Efficient Global Token Mixers [100.27957692579892]
We show that adaptive frequency filters can serve as efficient global token mixers.
We take AFF token mixers as primary neural operators to build a lightweight neural network, dubbed AFFNet.
arXiv Detail & Related papers (2023-07-26T07:42:28Z) - Lightweight Structure-Aware Attention for Visual Understanding [16.860625620412943]
Vision Transformers (ViTs) have become a dominant paradigm for visual representation learning with self-attention operators.
We propose a novel attention operator, called lightweight structure-aware attention (LiSA), which has a better representation power with log-linear complexity.
Our experiments and ablation studies demonstrate that ViTs based on the proposed operator outperform self-attention and other existing operators.
arXiv Detail & Related papers (2022-11-29T15:20:14Z) - Adaptive Fourier Neural Operators: Efficient Token Mixers for
Transformers [55.90468016961356]
We propose an efficient token mixer that learns to mix in the Fourier domain.
AFNO is based on a principled foundation of operator learning.
It can handle a sequence size of 65k and outperforms other efficient self-attention mechanisms.
arXiv Detail & Related papers (2021-11-24T05:44:31Z) - Global Filter Networks for Image Classification [90.81352483076323]
We present a conceptually simple yet computationally efficient architecture that learns long-term spatial dependencies in the frequency domain with log-linear complexity.
Our results demonstrate that GFNet can be a very competitive alternative to transformer-style models and CNNs in efficiency, generalization ability and robustness.
arXiv Detail & Related papers (2021-07-01T17:58:16Z) - Less is More: Pay Less Attention in Vision Transformers [61.05787583247392]
Less attention vIsion Transformer builds upon the fact that convolutions, fully-connected layers, and self-attentions have almost equivalent mathematical expressions for processing image patch sequences.
The proposed LIT achieves promising performance on image recognition tasks, including image classification, object detection and instance segmentation.
arXiv Detail & Related papers (2021-05-29T05:26:07Z) - OmniNet: Omnidirectional Representations from Transformers [49.23834374054286]
This paper proposes Omnidirectional Representations from Transformers ( OmniNet)
In OmniNet, instead of maintaining a strictly horizontal receptive field, each token is allowed to attend to all tokens in the entire network.
Experiments are conducted on autoregressive language modeling, Machine Translation, Long Range Arena (LRA), and Image Recognition.
arXiv Detail & Related papers (2021-03-01T15:31:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.