FathomNet: A global underwater image training set for enabling
artificial intelligence in the ocean
- URL: http://arxiv.org/abs/2109.14646v1
- Date: Wed, 29 Sep 2021 18:08:42 GMT
- Title: FathomNet: A global underwater image training set for enabling
artificial intelligence in the ocean
- Authors: Kakani Katija, Eric Orenstein, Brian Schlining, Lonny Lundsten, Kevin
Barnard, Giovanna Sainz, Oceane Boulais, Benjamin Woodward, Katy Croff Bell
- Abstract summary: Ocean-going platforms are integrating high-resolution camera feeds for observation and navigation, producing a deluge of visual data.
Recent advances in machine learning enable fast, sophisticated analysis of visual data, but have had limited success in the oceanographic world.
We will demonstrate how machine learning models trained on FathomNet data can be applied across different institutional video data.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Ocean-going platforms are integrating high-resolution camera feeds for
observation and navigation, producing a deluge of visual data. The volume and
rate of this data collection can rapidly outpace researchers' abilities to
process and analyze them. Recent advances in machine learning enable fast,
sophisticated analysis of visual data, but have had limited success in the
oceanographic world due to lack of dataset standardization, sparse annotation
tools, and insufficient formatting and aggregation of existing, expertly
curated imagery for use by data scientists. To address this need, we have built
FathomNet, a public platform that makes use of existing (and future), expertly
curated data. Initial efforts have leveraged MBARI's Video Annotation and
Reference System and annotated deep sea video database, which has more than 7M
annotations, 1M framegrabs, and 5k terms in the knowledgebase, with additional
contributions by National Geographic Society (NGS) and NOAA's Office of Ocean
Exploration and Research. FathomNet has over 100k localizations of 1k midwater
and benthic classes, and contains iconic and non-iconic views of marine
animals, underwater equipment, debris, etc. We will demonstrate how machine
learning models trained on FathomNet data can be applied across different
institutional video data, (e.g., NGS' Deep Sea Camera System and NOAA's ROV
Deep Discoverer), and enable automated acquisition and tracking of midwater
animals using MBARI's ROV MiniROV. As FathomNet continues to develop and
incorporate more image data from other oceanographic community members, this
effort will enable scientists, explorers, policymakers, storytellers, and the
public to understand and care for our ocean.
Related papers
- SeafloorAI: A Large-scale Vision-Language Dataset for Seafloor Geological Survey [11.642711706384212]
We introduce SeafloorAI, the first extensive AI-ready datasets for seafloor mapping across 5 geological layers.
The dataset consists of 62 geo-distributed data surveys spanning 17,300 square kilometers, with 696K sonar images, 827K annotated segmentation masks, 696K detailed language descriptions.
arXiv Detail & Related papers (2024-10-31T19:37:47Z) - Diving into Underwater: Segment Anything Model Guided Underwater Salient Instance Segmentation and A Large-scale Dataset [60.14089302022989]
Underwater vision tasks often suffer from low segmentation accuracy due to the complex underwater circumstances.
We construct the first large-scale underwater salient instance segmentation dataset (USIS10K)
We propose an Underwater Salient Instance architecture based on Segment Anything Model (USIS-SAM) specifically for the underwater domain.
arXiv Detail & Related papers (2024-06-10T06:17:33Z) - BenthicNet: A global compilation of seafloor images for deep learning applications [25.466405216505166]
BenthicNet is a global compilation of seafloor imagery.
An initial set of over 11.4 million images was collected and curated to represent a diversity of seafloor environments.
A large deep learning model was trained on this compilation and preliminary results suggest it has utility for automating large and small-scale image analysis tasks.
arXiv Detail & Related papers (2024-05-08T17:37:57Z) - Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - Improving Underwater Visual Tracking With a Large Scale Dataset and
Image Enhancement [70.2429155741593]
This paper presents a new dataset and general tracker enhancement method for Underwater Visual Object Tracking (UVOT)
It poses distinct challenges; the underwater environment exhibits non-uniform lighting conditions, low visibility, lack of sharpness, low contrast, camouflage, and reflections from suspended particles.
We propose a novel underwater image enhancement algorithm designed specifically to boost tracking quality.
The method has resulted in a significant performance improvement, of up to 5.0% AUC, of state-of-the-art (SOTA) visual trackers.
arXiv Detail & Related papers (2023-08-30T07:41:26Z) - A Dataset with Multibeam Forward-Looking Sonar for Underwater Object
Detection [0.0]
Multibeam forward-looking sonar (MFLS) plays an important role in underwater detection.
There are several challenges to the research on underwater object detection with MFLS.
We present a novel dataset, consisting of over 9000 MFLS images captured using Tritech Gemini 1200ik sonar.
arXiv Detail & Related papers (2022-12-01T08:26:03Z) - Guided Unsupervised Learning by Subaperture Decomposition for Ocean SAR
Image Retrieval [31.974530072369753]
Subaperture decomposition is used to enhance the unsupervised learning retrieval on the ocean surface.
We show that SD brings important performance boost when Doppler centroid images are used as input data.
arXiv Detail & Related papers (2022-09-29T18:17:56Z) - xView3-SAR: Detecting Dark Fishing Activity Using Synthetic Aperture
Radar Imagery [52.67592123500567]
Unsustainable fishing practices worldwide pose a major threat to marine resources and ecosystems.
It is now possible to automate detection of dark vessels day or night, under all-weather conditions.
xView3-SAR consists of nearly 1,000 analysis-ready SAR images from the Sentinel-1 mission.
arXiv Detail & Related papers (2022-06-02T06:53:45Z) - Rapid Exploration for Open-World Navigation with Latent Goal Models [78.45339342966196]
We describe a robotic learning system for autonomous exploration and navigation in diverse, open-world environments.
At the core of our method is a learned latent variable model of distances and actions, along with a non-parametric topological memory of images.
We use an information bottleneck to regularize the learned policy, giving us (i) a compact visual representation of goals, (ii) improved generalization capabilities, and (iii) a mechanism for sampling feasible goals for exploration.
arXiv Detail & Related papers (2021-04-12T23:14:41Z) - FathomNet: An underwater image training database for ocean exploration
and discovery [0.0]
FathomNet is a novel baseline image training set optimized to accelerate development of modern, intelligent, and automated analysis of underwater imagery.
To date, there are more than 80,000 images and 106,000 localizations for 233 different classes, including midwater and benthic organisms.
While we find quality results on prediction for this new dataset, our results indicate that we are ultimately in need of a larger data set for ocean exploration.
arXiv Detail & Related papers (2020-06-30T21:23:06Z) - Deep Learning based Pedestrian Inertial Navigation: Methods, Dataset and
On-Device Inference [49.88536971774444]
Inertial measurements units (IMUs) are small, cheap, energy efficient, and widely employed in smart devices and mobile robots.
Exploiting inertial data for accurate and reliable pedestrian navigation supports is a key component for emerging Internet-of-Things applications and services.
We present and release the Oxford Inertial Odometry dataset (OxIOD), a first-of-its-kind public dataset for deep learning based inertial navigation research.
arXiv Detail & Related papers (2020-01-13T04:41:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.