MassMIND: Massachusetts Maritime INfrared Dataset
- URL: http://arxiv.org/abs/2209.04097v1
- Date: Fri, 9 Sep 2022 02:54:26 GMT
- Title: MassMIND: Massachusetts Maritime INfrared Dataset
- Authors: Shailesh Nirgudkar, Michael DeFilippo, Michael Sacarny, Michael
Benjamin and Paul Robinette
- Abstract summary: This paper presents a labeled dataset of over 2,900 Long Wave Infrared (LWIR) segmented images captured in coastal maritime environment under diverse conditions.
The images are labeled using instance segmentation and classified in seven categories -- sky, water, obstacle, living obstacle, bridge, self and background.
We evaluate this dataset across three deep learning architectures (UNet, PSPNet, DeepLabv3) and provide detailed analysis of its efficacy.
- Score: 3.5751176069022983
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Recent advances in deep learning technology have triggered radical progress
in the autonomy of ground vehicles. Marine coastal Autonomous Surface Vehicles
(ASVs) that are regularly used for surveillance, monitoring and other routine
tasks can benefit from this autonomy. Long haul deep sea transportation
activities are additional opportunities. These two use cases present very
different terrains -- the first being coastal waters -- with many obstacles,
structures and human presence while the latter is mostly devoid of such
obstacles. Variations in environmental conditions are common to both terrains.
Robust labeled datasets mapping such terrains are crucial in improving the
situational awareness that can drive autonomy. However, there are only limited
such maritime datasets available and these primarily consist of optical images.
Although, Long Wave Infrared (LWIR) is a strong complement to the optical
spectrum that helps in extreme light conditions, a labeled public dataset with
LWIR images does not currently exist. In this paper, we fill this gap by
presenting a labeled dataset of over 2,900 LWIR segmented images captured in
coastal maritime environment under diverse conditions. The images are labeled
using instance segmentation and classified in seven categories -- sky, water,
obstacle, living obstacle, bridge, self and background. We also evaluate this
dataset across three deep learning architectures (UNet, PSPNet, DeepLabv3) and
provide detailed analysis of its efficacy. While the dataset focuses on the
coastal terrain it can equally help deep sea use cases. Such terrain would have
less traffic, and the classifier trained on cluttered environment would be able
to handle sparse scenes effectively. We share this dataset with the research
community with the hope that it spurs new scene understanding capabilities in
the maritime environment.
Related papers
- A Computer Vision Approach to Estimate the Localized Sea State [45.498315114762484]
This research focuses on utilizing sea images in operational envelopes captured by a single stationary camera mounted on the ship bridge.
The collected images are used to train a deep learning model to automatically recognize the state of the sea based on the Beaufort scale.
arXiv Detail & Related papers (2024-07-04T09:07:25Z) - Multi-modal Perception Dataset of In-water Objects for Autonomous Surface Vehicles [10.732732686425308]
This paper introduces the first publicly accessible multi-modal perception dataset for autonomous maritime navigation.
It focuses on in-water obstacles within the aquatic environment to enhance situational awareness for Autonomous Surface Vehicles (ASVs)
arXiv Detail & Related papers (2024-04-29T04:00:19Z) - Efficient Multi-branch Segmentation Network for Situation Awareness in Autonomous Navigation [0.0]
This study builds a dataset that captured perspectives from USVs and unmanned aerial vehicles in a maritime port environment.
Statistical analysis revealed a high correlation between the distribution of the sea and sky and row positional information.
A three-branch semantic segmentation network with a row position encoding module (RPEM) was proposed to improve the prediction accuracy between the sea and the sky.
arXiv Detail & Related papers (2024-03-30T13:38:07Z) - Improving Underwater Visual Tracking With a Large Scale Dataset and
Image Enhancement [70.2429155741593]
This paper presents a new dataset and general tracker enhancement method for Underwater Visual Object Tracking (UVOT)
It poses distinct challenges; the underwater environment exhibits non-uniform lighting conditions, low visibility, lack of sharpness, low contrast, camouflage, and reflections from suspended particles.
We propose a novel underwater image enhancement algorithm designed specifically to boost tracking quality.
The method has resulted in a significant performance improvement, of up to 5.0% AUC, of state-of-the-art (SOTA) visual trackers.
arXiv Detail & Related papers (2023-08-30T07:41:26Z) - FLSea: Underwater Visual-Inertial and Stereo-Vision Forward-Looking
Datasets [8.830479021890575]
We have collected underwater forward-looking stereo-vision and visual-inertial image sets in the Mediterranean and Red Sea.
These datasets are critical for the development of several underwater applications, including obstacle avoidance, visual odometry, 3D tracking, Simultaneous localization and Mapping (SLAM) and depth estimation.
arXiv Detail & Related papers (2023-02-24T17:39:53Z) - Deep Learning Models for River Classification at Sub-Meter Resolutions
from Multispectral and Panchromatic Commercial Satellite Imagery [2.121978045345352]
This study focuses on rivers in the Arctic, using images from the Quickbird, WorldView, and GeoEye satellites.
We use the RGB, and NIR bands of the 8-band multispectral sensors. Those trained models all achieve excellent precision and recall over 90% on validation data, aided by on-the-fly preprocessing of the training data specific to satellite imagery.
In a novel approach, we then use results from the multispectral model to generate training data for FCN that only require panchromatic imagery, of which considerably more is available.
arXiv Detail & Related papers (2022-12-27T20:56:34Z) - 6D Camera Relocalization in Visually Ambiguous Extreme Environments [79.68352435957266]
We propose a novel method to reliably estimate the pose of a camera given a sequence of images acquired in extreme environments such as deep seas or extraterrestrial terrains.
Our method achieves comparable performance with state-of-the-art methods on the indoor benchmark (7-Scenes dataset) using only 20% training data.
arXiv Detail & Related papers (2022-07-13T16:40:02Z) - xView3-SAR: Detecting Dark Fishing Activity Using Synthetic Aperture
Radar Imagery [52.67592123500567]
Unsustainable fishing practices worldwide pose a major threat to marine resources and ecosystems.
It is now possible to automate detection of dark vessels day or night, under all-weather conditions.
xView3-SAR consists of nearly 1,000 analysis-ready SAR images from the Sentinel-1 mission.
arXiv Detail & Related papers (2022-06-02T06:53:45Z) - Deep Learning for Real Time Satellite Pose Estimation on Low Power Edge
TPU [58.720142291102135]
In this paper we propose a pose estimation software exploiting neural network architectures.
We show how low power machine learning accelerators could enable Artificial Intelligence exploitation in space.
arXiv Detail & Related papers (2022-04-07T08:53:18Z) - Occupancy Anticipation for Efficient Exploration and Navigation [97.17517060585875]
We propose occupancy anticipation, where the agent uses its egocentric RGB-D observations to infer the occupancy state beyond the visible regions.
By exploiting context in both the egocentric views and top-down maps our model successfully anticipates a broader map of the environment.
Our approach is the winning entry in the 2020 Habitat PointNav Challenge.
arXiv Detail & Related papers (2020-08-21T03:16:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.