Semantic Segmentation for Thermal Images: A Comparative Survey
- URL: http://arxiv.org/abs/2205.13278v1
- Date: Thu, 26 May 2022 11:32:15 GMT
- Title: Semantic Segmentation for Thermal Images: A Comparative Survey
- Authors: Z\"ulfiye K\"ut\"uk, G\"orkem Algan
- Abstract summary: Using infrared spectrum in semantic segmentation has many real-world use cases, such as autonomous driving, medical imaging, agriculture, defense industry, etc.
One approach is to use both visible and infrared spectrum images as inputs.
Another approach is to use only thermal images, enabling less hardware cost for smaller use cases.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Semantic segmentation is a challenging task since it requires excessively
more low-level spatial information of the image compared to other computer
vision problems. The accuracy of pixel-level classification can be affected by
many factors, such as imaging limitations and the ambiguity of object
boundaries in an image. Conventional methods exploit three-channel RGB images
captured in the visible spectrum with deep neural networks (DNN). Thermal
images can significantly contribute during the segmentation since thermal
imaging cameras are capable of capturing details despite the weather and
illumination conditions. Using infrared spectrum in semantic segmentation has
many real-world use cases, such as autonomous driving, medical imaging,
agriculture, defense industry, etc. Due to this wide range of use cases,
designing accurate semantic segmentation algorithms with the help of infrared
spectrum is an important challenge. One approach is to use both visible and
infrared spectrum images as inputs. These methods can accomplish higher
accuracy due to enriched input information, with the cost of extra effort for
the alignment and processing of multiple inputs. Another approach is to use
only thermal images, enabling less hardware cost for smaller use cases. Even
though there are multiple surveys on semantic segmentation methods, the
literature lacks a comprehensive survey centered explicitly around semantic
segmentation using infrared spectrum. This work aims to fill this gap by
presenting algorithms in the literature and categorizing them by their input
images.
Related papers
- Diffusion-based RGB-D Semantic Segmentation with Deformable Attention Transformer [10.982521876026281]
We introduce a diffusion-based framework to address the RGB-D semantic segmentation problem.
We demonstrate that utilizing a Deformable Attention Transformer as the encoder to extract features from depth images effectively captures the characteristics of invalid regions in depth measurements.
arXiv Detail & Related papers (2024-09-23T15:23:01Z) - Learning Semantic Segmentation with Query Points Supervision on Aerial Images [57.09251327650334]
We present a weakly supervised learning algorithm to train semantic segmentation algorithms.
Our proposed approach performs accurate semantic segmentation and improves efficiency by significantly reducing the cost and time required for manual annotation.
arXiv Detail & Related papers (2023-09-11T14:32:04Z) - Semantic RGB-D Image Synthesis [22.137419841504908]
We introduce semantic RGB-D image synthesis to address this problem.
Current approaches, however, are uni-modal and cannot cope with multi-modal data.
We propose a generator for multi-modal data that separates modal-independent information of the semantic layout from the modal-dependent information.
arXiv Detail & Related papers (2023-08-22T11:16:24Z) - Visible and infrared self-supervised fusion trained on a single example [1.1188842018827656]
Multispectral imaging is important task of image processing and computer vision.
Problem of visible (RGB) to Near Infrared (NIR) image fusion has become particularly timely.
Proposed approach fuses these two channels by training a Convolutional Neural Network by Self Supervised Learning (SSL) on a single example.
Experiments demonstrate that the proposed approach achieves similar or better qualitative and quantitative multispectral fusion results.
arXiv Detail & Related papers (2023-07-09T05:25:46Z) - Invisible-to-Visible: Privacy-Aware Human Segmentation using Airborne
Ultrasound via Collaborative Learning Probabilistic U-Net [8.21448246263952]
We propose a new task for human segmentation from invisible information, especially airborne ultrasound.
Although ultrasound images can roughly identify a person's location, the detailed shape is ambiguous.
We propose a collaborative learning probabilistic U-Net that uses ultrasound and segmentation images simultaneously during training.
arXiv Detail & Related papers (2022-05-11T06:42:24Z) - AF$_2$: Adaptive Focus Framework for Aerial Imagery Segmentation [86.44683367028914]
Aerial imagery segmentation has some unique challenges, the most critical one among which lies in foreground-background imbalance.
We propose Adaptive Focus Framework (AF$), which adopts a hierarchical segmentation procedure and focuses on adaptively utilizing multi-scale representations.
AF$ has significantly improved the accuracy on three widely used aerial benchmarks, as fast as the mainstream method.
arXiv Detail & Related papers (2022-02-18T10:14:45Z) - Voice-assisted Image Labelling for Endoscopic Ultrasound Classification
using Neural Networks [48.732863591145964]
We propose a multi-modal convolutional neural network architecture that labels endoscopic ultrasound (EUS) images from raw verbal comments provided by a clinician during the procedure.
Our results show a prediction accuracy of 76% at image level on a dataset with 5 different labels.
arXiv Detail & Related papers (2021-10-12T21:22:24Z) - A Comparison of Deep Saliency Map Generators on Multispectral Data in
Object Detection [9.264502124445348]
This work investigates three saliency map generator methods on how their maps differ across the different spectra.
As a practical problem, we chose object detection in the infrared and visual spectrum for autonomous driving.
The results show that there are differences between the infrared and visual activation maps.
Further, an advanced training with both, the infrared and visual data not only improves the network's output, it also leads to more focused spots in the saliency maps.
arXiv Detail & Related papers (2021-08-26T12:56:49Z) - Exposing Semantic Segmentation Failures via Maximum Discrepancy
Competition [102.75463782627791]
We take steps toward answering the question by exposing failures of existing semantic segmentation methods in the open visual world.
Inspired by previous research on model falsification, we start from an arbitrarily large image set, and automatically sample a small image set by MAximizing the Discrepancy (MAD) between two segmentation methods.
The selected images have the greatest potential in falsifying either (or both) of the two methods.
A segmentation method, whose failures are more difficult to be exposed in the MAD competition, is considered better.
arXiv Detail & Related papers (2021-02-27T16:06:25Z) - Exploring Thermal Images for Object Detection in Underexposure Regions
for Autonomous Driving [67.69430435482127]
Underexposure regions are vital to construct a complete perception of the surroundings for safe autonomous driving.
The availability of thermal cameras has provided an essential alternate to explore regions where other optical sensors lack in capturing interpretable signals.
This work proposes a domain adaptation framework which employs a style transfer technique for transfer learning from visible spectrum images to thermal images.
arXiv Detail & Related papers (2020-06-01T09:59:09Z) - RGB-based Semantic Segmentation Using Self-Supervised Depth Pre-Training [77.62171090230986]
We propose an easily scalable and self-supervised technique that can be used to pre-train any semantic RGB segmentation method.
In particular, our pre-training approach makes use of automatically generated labels that can be obtained using depth sensors.
We show how our proposed self-supervised pre-training with HN-labels can be used to replace ImageNet pre-training.
arXiv Detail & Related papers (2020-02-06T11:16:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.