FAMLP: A Frequency-Aware MLP-Like Architecture For Domain Generalization
- URL: http://arxiv.org/abs/2203.12893v1
- Date: Thu, 24 Mar 2022 07:26:29 GMT
- Title: FAMLP: A Frequency-Aware MLP-Like Architecture For Domain Generalization
- Authors: Kecheng Zheng, Yang Cao, Kai Zhu, Ruijing Zhao, Zheng-Jun Zha
- Abstract summary: We propose a novel frequency-aware architecture, in which the domain-specific features are filtered out in the transformed frequency domain.
Experiments on three benchmarks demonstrate significant performance, outperforming the state-of-the-art methods by a margin of 3%, 4% and 9%, respectively.
- Score: 73.41395947275473
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: MLP-like models built entirely upon multi-layer perceptrons have recently
been revisited, exhibiting the comparable performance with transformers. It is
one of most promising architectures due to the excellent trade-off between
network capability and efficiency in the large-scale recognition tasks.
However, its generalization performance to heterogeneous tasks is inferior to
other architectures (e.g., CNNs and transformers) due to the extensive
retention of domain information. To address this problem, we propose a novel
frequency-aware MLP architecture, in which the domain-specific features are
filtered out in the transformed frequency domain, augmenting the invariant
descriptor for label prediction. Specifically, we design an adaptive Fourier
filter layer, in which a learnable frequency filter is utilized to adjust the
amplitude distribution by optimizing both the real and imaginary parts. A
low-rank enhancement module is further proposed to rectify the filtered
features by adding the low-frequency components from SVD decomposition.
Finally, a momentum update strategy is utilized to stabilize the optimization
to fluctuation of model parameters and inputs by the output distillation with
weighted historical states. To our best knowledge, we are the first to propose
a MLP-like backbone for domain generalization. Extensive experiments on three
benchmarks demonstrate significant generalization performance, outperforming
the state-of-the-art methods by a margin of 3%, 4% and 9%, respectively.
Related papers
- PointMT: Efficient Point Cloud Analysis with Hybrid MLP-Transformer Architecture [46.266960248570086]
This study tackles the quadratic complexity of the self-attention mechanism by introducing a complexity local attention mechanism for effective feature aggregation.
We also introduce a parameter-free channel temperature adaptation mechanism that adaptively adjusts the attention weight distribution in each channel.
We show that PointMT achieves performance comparable to state-of-the-art methods while maintaining an optimal balance between performance and accuracy.
arXiv Detail & Related papers (2024-08-10T10:16:03Z) - Frequency-domain MLPs are More Effective Learners in Time Series
Forecasting [67.60443290781988]
Time series forecasting has played the key role in different industrial domains, including finance, traffic, energy, and healthcare.
Most-based forecasting methods suffer from the point-wise mappings and information bottleneck.
We propose FreTS, a simple yet effective architecture built upon Frequency-domains for Time Series forecasting.
arXiv Detail & Related papers (2023-11-10T17:05:13Z) - Exploring Frequency-Inspired Optimization in Transformer for Efficient Single Image Super-Resolution [32.29219284419944]
Cross-refinement adaptive feature modulation transformer (CRAFT)
We introduce a frequency-guided post-training quantization (PTQ) method aimed at enhancing CRAFT's efficiency.
Our experimental findings showcase CRAFT's superiority over current state-of-the-art methods, both in full-precision and quantization scenarios.
arXiv Detail & Related papers (2023-08-09T15:38:36Z) - Fourier Test-time Adaptation with Multi-level Consistency for Robust
Classification [10.291631977766672]
We propose a novel approach called Fourier Test-time Adaptation (FTTA) to integrate input and model tuning.
FTTA builds a reliable multi-level consistency measurement of paired inputs for achieving self-supervised of prediction.
It was extensively validated on three large classification datasets with different modalities and organs.
arXiv Detail & Related papers (2023-06-05T02:29:38Z) - Transform Once: Efficient Operator Learning in Frequency Domain [69.74509540521397]
We study deep neural networks designed to harness the structure in frequency domain for efficient learning of long-range correlations in space or time.
This work introduces a blueprint for frequency domain learning through a single transform: transform once (T1)
arXiv Detail & Related papers (2022-11-26T01:56:05Z) - Inception Transformer [151.939077819196]
Inception Transformer, or iFormer, learns comprehensive features with both high- and low-frequency information in visual data.
We benchmark the iFormer on a series of vision tasks, and showcase that it achieves impressive performance on image classification, COCO detection and ADE20K segmentation.
arXiv Detail & Related papers (2022-05-25T17:59:54Z) - Deep Frequency Filtering for Domain Generalization [55.66498461438285]
Deep Neural Networks (DNNs) have preferences for some frequency components in the learning process.
We propose Deep Frequency Filtering (DFF) for learning domain-generalizable features.
We show that applying our proposed DFF on a plain baseline outperforms the state-of-the-art methods on different domain generalization tasks.
arXiv Detail & Related papers (2022-03-23T05:19:06Z) - Global Filter Networks for Image Classification [90.81352483076323]
We present a conceptually simple yet computationally efficient architecture that learns long-term spatial dependencies in the frequency domain with log-linear complexity.
Our results demonstrate that GFNet can be a very competitive alternative to transformer-style models and CNNs in efficiency, generalization ability and robustness.
arXiv Detail & Related papers (2021-07-01T17:58:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.