Real-Time Semantic Segmentation using Hyperspectral Images for Mapping
Unstructured and Unknown Environments
- URL: http://arxiv.org/abs/2303.15623v1
- Date: Mon, 27 Mar 2023 22:33:55 GMT
- Title: Real-Time Semantic Segmentation using Hyperspectral Images for Mapping
Unstructured and Unknown Environments
- Authors: Anthony Medellin and Anant Bhamri and Reza Langari and Swaminathan
Gopalswamy
- Abstract summary: We propose the use of hyperspectral images for real-time pixel-wise semantic classification and segmentation.
The resulting segmented image is processed to extract, filter, and approximate objects as polygons.
The resulting polygons are then used to generate a semantic map of the environment.
- Score: 2.408714894793063
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Autonomous navigation in unstructured off-road environments is greatly
improved by semantic scene understanding. Conventional image processing
algorithms are difficult to implement and lack robustness due to a lack of
structure and high variability across off-road environments. The use of neural
networks and machine learning can overcome the previous challenges but they
require large labeled data sets for training. In our work we propose the use of
hyperspectral images for real-time pixel-wise semantic classification and
segmentation, without the need of any prior training data. The resulting
segmented image is processed to extract, filter, and approximate objects as
polygons, using a polygon approximation algorithm. The resulting polygons are
then used to generate a semantic map of the environment. Using our framework.
we show the capability to add new semantic classes in run-time for
classification. The proposed methodology is also shown to operate in real-time
and produce outputs at a frequency of 1Hz, using high resolution hyperspectral
images.
Related papers
- Large Spatial Model: End-to-end Unposed Images to Semantic 3D [79.94479633598102]
Large Spatial Model (LSM) processes unposed RGB images directly into semantic radiance fields.
LSM simultaneously estimates geometry, appearance, and semantics in a single feed-forward operation.
It can generate versatile label maps by interacting with language at novel viewpoints.
arXiv Detail & Related papers (2024-10-24T17:54:42Z) - Patch-Based Deep Unsupervised Image Segmentation using Graph Cuts [0.0]
We propose a patch-based unsupervised image segmentation strategy that bridges advances in unsupervised feature extraction with the algorithmic help of classical graph-based methods.
We show that a simple convolutional neural network, trained to classify image patches, naturally leads to a state-of-the-art fully-convolutional unsupervised pixel-level segmenter.
arXiv Detail & Related papers (2023-11-01T19:59:25Z) - Temporal Embeddings: Scalable Self-Supervised Temporal Representation
Learning from Spatiotemporal Data for Multimodal Computer Vision [1.4127889233510498]
A novel approach is proposed to stratify landscape based on mobility activity time series.
The pixel-wise embeddings are converted to image-like channels that can be used for task-based, multimodal modeling.
arXiv Detail & Related papers (2023-10-16T02:53:29Z) - Learning Semantic Segmentation with Query Points Supervision on Aerial Images [57.09251327650334]
We present a weakly supervised learning algorithm to train semantic segmentation algorithms.
Our proposed approach performs accurate semantic segmentation and improves efficiency by significantly reducing the cost and time required for manual annotation.
arXiv Detail & Related papers (2023-09-11T14:32:04Z) - Efficient Semantic Segmentation on Edge Devices [7.5562201794440185]
This project analyzes current semantic segmentation models to explore the feasibility of applying these models for emergency response during catastrophic events.
We compare the performance of real-time semantic segmentation models with non-real-time counterparts constrained by aerial images under oppositional settings.
Furthermore, we train several models on the Flood-Net dataset, containing UAV images captured after Hurricane Harvey, and benchmark their execution on special classes such as flooded buildings vs. non-flooded buildings or flooded roads vs. non-flooded roads.
arXiv Detail & Related papers (2022-12-28T04:13:11Z) - Location-Aware Self-Supervised Transformers [74.76585889813207]
We propose to pretrain networks for semantic segmentation by predicting the relative location of image parts.
We control the difficulty of the task by masking a subset of the reference patch features visible to those of the query.
Our experiments show that this location-aware pretraining leads to representations that transfer competitively to several challenging semantic segmentation benchmarks.
arXiv Detail & Related papers (2022-12-05T16:24:29Z) - RICE: Refining Instance Masks in Cluttered Environments with Graph
Neural Networks [53.15260967235835]
We propose a novel framework that refines the output of such methods by utilizing a graph-based representation of instance masks.
We train deep networks capable of sampling smart perturbations to the segmentations, and a graph neural network, which can encode relations between objects, to evaluate the segmentations.
We demonstrate an application that uses uncertainty estimates generated by our method to guide a manipulator, leading to efficient understanding of cluttered scenes.
arXiv Detail & Related papers (2021-06-29T20:29:29Z) - A Novel Upsampling and Context Convolution for Image Semantic
Segmentation [0.966840768820136]
Recent methods for semantic segmentation often employ an encoder-decoder structure using deep convolutional neural networks.
We propose a dense upsampling convolution method based on guided filtering to effectively preserve the spatial information of the image in the network.
We report a new record of 82.86% and 81.62% of pixel accuracy on ADE20K and Pascal-Context benchmark datasets, respectively.
arXiv Detail & Related papers (2021-03-20T06:16:42Z) - Exploring Cross-Image Pixel Contrast for Semantic Segmentation [130.22216825377618]
We propose a pixel-wise contrastive framework for semantic segmentation in the fully supervised setting.
The core idea is to enforce pixel embeddings belonging to a same semantic class to be more similar than embeddings from different classes.
Our method can be effortlessly incorporated into existing segmentation frameworks without extra overhead during testing.
arXiv Detail & Related papers (2021-01-28T11:35:32Z) - Group-Wise Semantic Mining for Weakly Supervised Semantic Segmentation [49.90178055521207]
This work addresses weakly supervised semantic segmentation (WSSS), with the goal of bridging the gap between image-level annotations and pixel-level segmentation.
We formulate WSSS as a novel group-wise learning task that explicitly models semantic dependencies in a group of images to estimate more reliable pseudo ground-truths.
In particular, we devise a graph neural network (GNN) for group-wise semantic mining, wherein input images are represented as graph nodes.
arXiv Detail & Related papers (2020-12-09T12:40:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.