FishFormer: Annulus Slicing-based Transformer for Fisheye Rectification
with Efficacy Domain Exploration
- URL: http://arxiv.org/abs/2207.01925v1
- Date: Tue, 5 Jul 2022 09:59:32 GMT
- Title: FishFormer: Annulus Slicing-based Transformer for Fisheye Rectification
with Efficacy Domain Exploration
- Authors: Shangrong Yang, Chunyu Lin, Kang Liao, Yao Zhao
- Abstract summary: We introduce Fishformer that processes the fisheye image as a sequence to enhance global and local perception.
We tuned the Transformer according to the structural properties of fisheye images.
Our method provides superior performance compared with state-of-the-art methods.
- Score: 44.332845280150785
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Numerous significant progress on fisheye image rectification has been
achieved through CNN. Nevertheless, constrained by a fixed receptive field, the
global distribution and the local symmetry of the distortion have not been
fully exploited. To leverage these two characteristics, we introduced
Fishformer that processes the fisheye image as a sequence to enhance global and
local perception. We tuned the Transformer according to the structural
properties of fisheye images. First, the uneven distortion distribution in
patches generated by the existing square slicing method confuses the network,
resulting in difficult training. Therefore, we propose an annulus slicing
method to maintain the consistency of the distortion in each patch, thus
perceiving the distortion distribution well. Second, we analyze that different
distortion parameters have their own efficacy domains. Hence, the perception of
the local area is as important as the global, but Transformer has a weakness
for local texture perception. Therefore, we propose a novel layer attention
mechanism to enhance the local perception and texture transfer. Our network
simultaneously implements global perception and focused local perception
decided by the different parameters. Extensive experiments demonstrate that our
method provides superior performance compared with state-of-the-art methods.
Related papers
- RoFIR: Robust Fisheye Image Rectification Framework Impervious to Optical Center Deviation [88.54817424560056]
We propose a distortion vector map (DVM) that measures the degree and direction of local distortion.
By learning the DVM, the model can independently identify local distortions at each pixel without relying on global distortion patterns.
In the pre-training stage, it predicts the distortion vector map and perceives the local distortion features of each pixel.
In the fine-tuning stage, it predicts a pixel-wise flow map for deviated fisheye image rectification.
arXiv Detail & Related papers (2024-06-27T06:38:56Z) - Progressive Feedback-Enhanced Transformer for Image Forgery Localization [3.765051882812805]
We propose a Progressive FeedbACk-enhanced Transformer (ProFact) network to achieve coarse-to-fine image forgery localization.
We present an effective strategy to automatically generate large-scale forged image samples close to real-world forensic scenarios.
Our proposed localizer greatly outperforms the state-of-the-art on the ability and robustness of image forgery localization.
arXiv Detail & Related papers (2023-11-15T12:31:43Z) - SimFIR: A Simple Framework for Fisheye Image Rectification with
Self-supervised Representation Learning [105.01294305972037]
We introduce SimFIR, a framework for fisheye image rectification based on self-supervised representation learning.
To learn fine-grained distortion representations, we first split a fisheye image into multiple patches and extract their representations with a Vision Transformer.
The transfer performance on the downstream rectification task is remarkably boosted, which verifies the effectiveness of the learned representations.
arXiv Detail & Related papers (2023-08-17T15:20:17Z) - Adversarial Domain Feature Adaptation for Bronchoscopic Depth Estimation [111.89519571205778]
In this work, we propose an alternative domain-adaptive approach to depth estimation.
Our novel two-step structure first trains a depth estimation network with labeled synthetic images in a supervised manner.
The results of our experiments show that the proposed method improves the network's performance on real images by a considerable margin.
arXiv Detail & Related papers (2021-09-24T08:11:34Z) - SIR: Self-supervised Image Rectification via Seeing the Same Scene from
Multiple Different Lenses [82.56853587380168]
We propose a novel self-supervised image rectification (SIR) method based on an important insight that the rectified results of distorted images of the same scene from different lens should be the same.
We leverage a differentiable warping module to generate the rectified images and re-distorted images from the distortion parameters.
Our method achieves comparable or even better performance than the supervised baseline method and representative state-of-the-art methods.
arXiv Detail & Related papers (2020-11-30T08:23:25Z) - A Deep Ordinal Distortion Estimation Approach for Distortion Rectification [62.72089758481803]
We propose a novel distortion rectification approach that can obtain more accurate parameters with higher efficiency.
We design a local-global associated estimation network that learns the ordinal distortion to approximate the realistic distortion distribution.
Considering the redundancy of distortion information, our approach only uses a part of distorted image for the ordinal distortion estimation.
arXiv Detail & Related papers (2020-07-21T10:03:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.