GLFNET: Global-Local (frequency) Filter Networks for efficient medical
image segmentation
- URL: http://arxiv.org/abs/2403.00396v1
- Date: Fri, 1 Mar 2024 09:35:03 GMT
- Title: GLFNET: Global-Local (frequency) Filter Networks for efficient medical
image segmentation
- Authors: Athanasios Tragakis, Qianying Liu, Chaitanya Kaul, Swalpa Kumar Roy,
Hang Dai, Fani Deligianni, Roderick Murray-Smith, Daniele Faccio
- Abstract summary: We propose a transformer-style architecture called Global-Local Filter Network (GLFNet) for medical image segmentation.
We replace the self-attention mechanism with a combination of global-local filter blocks to optimize model efficiency.
We test GLFNet on three benchmark datasets achieving state-of-the-art performance on all of them while being almost twice as efficient in terms of GFLOP operations.
- Score: 18.314093733807972
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a novel transformer-style architecture called Global-Local Filter
Network (GLFNet) for medical image segmentation and demonstrate its
state-of-the-art performance. We replace the self-attention mechanism with a
combination of global-local filter blocks to optimize model efficiency. The
global filters extract features from the whole feature map whereas the local
filters are being adaptively created as 4x4 patches of the same feature map and
add restricted scale information. In particular, the feature extraction takes
place in the frequency domain rather than the commonly used spatial (image)
domain to facilitate faster computations. The fusion of information from both
spatial and frequency spaces creates an efficient model with regards to
complexity, required data and performance. We test GLFNet on three benchmark
datasets achieving state-of-the-art performance on all of them while being
almost twice as efficient in terms of GFLOP operations.
Related papers
- Lookup Table meets Local Laplacian Filter: Pyramid Reconstruction
Network for Tone Mapping [35.47139372780014]
This paper explores a novel strategy that integrates global and local operators by utilizing closed-form Laplacian pyramid decomposition and reconstruction.
We employ image-adaptive 3D LUTs to manipulate the tone in the low-frequency image by leveraging the specific characteristics of the frequency information.
We also utilize local Laplacian filters to refine the edge details in the high-frequency components in an adaptive manner.
arXiv Detail & Related papers (2023-10-26T07:05:38Z) - Decomposed Guided Dynamic Filters for Efficient RGB-Guided Depth
Completion [46.04264366475848]
RGB-guided depth completion aims at predicting dense depth maps from sparse depth measurements and corresponding RGB images.
Guided dynamic filters generate spatially-variant depth-wise separable convolutional filters from RGB features to guide depth features.
We propose to decompose the guided dynamic filters into a spatially-shared component multiplied by content-adaptive adaptors at each spatial location.
arXiv Detail & Related papers (2023-09-05T08:37:58Z) - Dynamic Temporal Filtering in Video Models [128.02725199486719]
We present a new recipe of temporal feature learning, namely Dynamic Temporal Filter (DTF)
DTF learns a specialized frequency filter for every spatial location to model its long-range temporal dynamics.
It is feasible to plug DTF block into ConvNets and Transformer, yielding DTF-Net and DTF-Transformer.
arXiv Detail & Related papers (2022-11-15T15:59:28Z) - SFNet: Faster and Accurate Semantic Segmentation via Semantic Flow [88.97790684009979]
A common practice to improve the performance is to attain high-resolution feature maps with strong semantic representation.
We propose a Flow Alignment Module (FAM) to learn textitSemantic Flow between feature maps of adjacent levels.
We also present a novel Gated Dual Flow Alignment Module to directly align high-resolution feature maps and low-resolution feature maps.
arXiv Detail & Related papers (2022-07-10T08:25:47Z) - FAMLP: A Frequency-Aware MLP-Like Architecture For Domain Generalization [73.41395947275473]
We propose a novel frequency-aware architecture, in which the domain-specific features are filtered out in the transformed frequency domain.
Experiments on three benchmarks demonstrate significant performance, outperforming the state-of-the-art methods by a margin of 3%, 4% and 9%, respectively.
arXiv Detail & Related papers (2022-03-24T07:26:29Z) - MFGNet: Dynamic Modality-Aware Filter Generation for RGB-T Tracking [72.65494220685525]
We propose a new dynamic modality-aware filter generation module (named MFGNet) to boost the message communication between visible and thermal data.
We generate dynamic modality-aware filters with two independent networks. The visible and thermal filters will be used to conduct a dynamic convolutional operation on their corresponding input feature maps respectively.
To address issues caused by heavy occlusion, fast motion, and out-of-view, we propose to conduct a joint local and global search by exploiting a new direction-aware target-driven attention mechanism.
arXiv Detail & Related papers (2021-07-22T03:10:51Z) - Global Filter Networks for Image Classification [90.81352483076323]
We present a conceptually simple yet computationally efficient architecture that learns long-term spatial dependencies in the frequency domain with log-linear complexity.
Our results demonstrate that GFNet can be a very competitive alternative to transformer-style models and CNNs in efficiency, generalization ability and robustness.
arXiv Detail & Related papers (2021-07-01T17:58:16Z) - Message Passing in Graph Convolution Networks via Adaptive Filter Banks [81.12823274576274]
We present a novel graph convolution operator, termed BankGCN.
It decomposes multi-channel signals on graphs into subspaces and handles particular information in each subspace with an adapted filter.
It achieves excellent performance in graph classification on a collection of benchmark graph datasets.
arXiv Detail & Related papers (2021-06-18T04:23:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.