Transformers in Remote Sensing: A Survey
- URL: http://arxiv.org/abs/2209.01206v1
- Date: Fri, 2 Sep 2022 17:57:05 GMT
- Title: Transformers in Remote Sensing: A Survey
- Authors: Abdulaziz Amer Aleissaee, Amandeep Kumar, Rao Muhammad Anwer, Salman
Khan, Hisham Cholakkal, Gui-Song Xia and Fahad Shahbaz khan
- Abstract summary: We are the first to present a systematic review of advances based on transformers in remote sensing.
Our survey covers more than 60 recent transformers-based methods for different remote sensing problems.
We conclude the survey by discussing different challenges and open issues of transformers in remote sensing.
- Score: 76.95730131233424
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning-based algorithms have seen a massive popularity in different
areas of remote sensing image analysis over the past decade. Recently,
transformers-based architectures, originally introduced in natural language
processing, have pervaded computer vision field where the self-attention
mechanism has been utilized as a replacement to the popular convolution
operator for capturing long-range dependencies. Inspired by recent advances in
computer vision, remote sensing community has also witnessed an increased
exploration of vision transformers for a diverse set of tasks. Although a
number of surveys have focused on transformers in computer vision in general,
to the best of our knowledge we are the first to present a systematic review of
recent advances based on transformers in remote sensing. Our survey covers more
than 60 recent transformers-based methods for different remote sensing problems
in sub-areas of remote sensing: very high-resolution (VHR), hyperspectral (HSI)
and synthetic aperture radar (SAR) imagery. We conclude the survey by
discussing different challenges and open issues of transformers in remote
sensing. Additionally, we intend to frequently update and maintain the latest
transformers in remote sensing papers with their respective code at:
https://github.com/VIROBO-15/Transformer-in-Remote-Sensing
Related papers
- Rotation Invariant Transformer for Recognizing Object in UAVs [66.1564328237299]
We propose a novel rotation invariant vision transformer (RotTrans) forRecognizing targets of interest from UAVs.
RotTrans greatly outperforms the current state-of-the-arts, which is 5.9% and 4.8% higher than the highest mAP and Rank1.
Our solution wins the first place in the UAV-based person re-recognition track in the Multi-Modal Video Reasoning and Analyzing Competition.
arXiv Detail & Related papers (2023-11-05T03:55:08Z) - Transformer-Based Sensor Fusion for Autonomous Driving: A Survey [0.0]
Transformers-based detection head and CNN-based feature encoder to extract features from raw sensor-data has emerged as one of the best performing sensor-fusion 3D-detection-framework.
We briefly go through the Vision transformers (ViT) basics, so that readers can easily follow through the paper.
In conclusion we summarize with sensor-fusion trends to follow and provoke future research.
arXiv Detail & Related papers (2023-02-22T16:28:20Z) - Vision Transformers for Action Recognition: A Survey [41.69370782177517]
Vision transformers are emerging as a powerful tool to solve computer vision problems.
Recent techniques have proven the efficacy of transformers beyond the image domain to solve numerous video-related tasks.
Human action recognition is receiving special attention from the research community due to its widespread applications.
arXiv Detail & Related papers (2022-09-13T02:57:05Z) - 3D Vision with Transformers: A Survey [114.86385193388439]
The success of the transformer architecture in natural language processing has triggered attention in the computer vision field.
We present a systematic and thorough review of more than 100 transformers methods for different 3D vision tasks.
We discuss transformer design in 3D vision, which allows it to process data with various 3D representations.
arXiv Detail & Related papers (2022-08-08T17:59:11Z) - TransCenter: Transformers with Dense Queries for Multiple-Object
Tracking [87.75122600164167]
We argue that the standard representation -- bounding boxes -- is not adapted to learning transformers for multiple-object tracking.
We propose TransCenter, the first transformer-based architecture for tracking the centers of multiple targets.
arXiv Detail & Related papers (2021-03-28T14:49:36Z) - Transformers in Vision: A Survey [101.07348618962111]
Transformers enable modeling long dependencies between input sequence elements and support parallel processing of sequence.
Transformers require minimal inductive biases for their design and are naturally suited as set-functions.
This survey aims to provide a comprehensive overview of the Transformer models in the computer vision discipline.
arXiv Detail & Related papers (2021-01-04T18:57:24Z) - A Survey on Visual Transformer [126.56860258176324]
Transformer is a type of deep neural network mainly based on the self-attention mechanism.
In this paper, we review these vision transformer models by categorizing them in different tasks and analyzing their advantages and disadvantages.
arXiv Detail & Related papers (2020-12-23T09:37:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.