DARE: AI-based Diver Action Recognition System using Multi-Channel CNNs
for AUV Supervision
- URL: http://arxiv.org/abs/2011.07713v1
- Date: Mon, 16 Nov 2020 04:05:32 GMT
- Title: DARE: AI-based Diver Action Recognition System using Multi-Channel CNNs
for AUV Supervision
- Authors: Jing Yang and James P. Wilson and Shalabh Gupta
- Abstract summary: This paper presents DARE, a diver action recognition system that is trained based on Cognitive Autonomous Driving Buddy dataset.
DARE is fast and requires only a few milliseconds to classify one stereo-pair, thus making it suitable for real-time underwater implementation.
- Score: 3.5584173777587935
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the growth of sensing, control and robotic technologies, autonomous
underwater vehicles (AUVs) have become useful assistants to human divers for
performing various underwater operations. In the current practice, the divers
are required to carry expensive, bulky, and waterproof keyboards or
joystick-based controllers for supervision and control of AUVs. Therefore,
diver action-based supervision is becoming increasingly popular because it is
convenient, easier to use, faster, and cost effective. However, the various
environmental, diver and sensing uncertainties present underwater makes it
challenging to train a robust and reliable diver action recognition system. In
this regard, this paper presents DARE, a diver action recognition system, that
is trained based on Cognitive Autonomous Driving Buddy (CADDY) dataset, which
is a rich set of data containing images of different diver gestures and poses
in several different and realistic underwater environments. DARE is based on
fusion of stereo-pairs of camera images using a multi-channel convolutional
neural network supported with a systematically trained tree-topological deep
neural network classifier to enhance the classification performance. DARE is
fast and requires only a few milliseconds to classify one stereo-pair, thus
making it suitable for real-time underwater implementation. DARE is
comparatively evaluated against several existing classifier architectures and
the results show that DARE supersedes the performance of all classifiers for
diver action recognition in terms of overall as well as individual class
accuracies and F1-scores.
Related papers
- Underwater Object Detection in the Era of Artificial Intelligence: Current, Challenge, and Future [119.88454942558485]
Underwater object detection (UOD) aims to identify and localise objects in underwater images or videos.
In recent years, artificial intelligence (AI) based methods, especially deep learning methods, have shown promising performance in UOD.
arXiv Detail & Related papers (2024-10-08T00:25:33Z) - FAFA: Frequency-Aware Flow-Aided Self-Supervision for Underwater Object Pose Estimation [65.01601309903971]
We introduce FAFA, a Frequency-Aware Flow-Aided self-supervised framework for 6D pose estimation of unmanned underwater vehicles (UUVs)
Our framework relies solely on the 3D model and RGB images, alleviating the need for any real pose annotations or other-modality data like depths.
We evaluate the effectiveness of FAFA on common underwater object pose benchmarks and showcase significant performance improvements compared to state-of-the-art methods.
arXiv Detail & Related papers (2024-09-25T03:54:01Z) - A Real-Time Framework for Domain-Adaptive Underwater Object Detection with Image Enhancement [18.936260846385444]
EnYOLO is an integrated real-time framework designed for simultaneous underwater image enhancement (UIE) and object detection (UOD)
Our framework achieves state-of-the-art (SOTA) performance in both UIE and UOD tasks, but also shows superior adaptability when applied to different underwater scenarios.
arXiv Detail & Related papers (2024-03-28T01:00:08Z) - Interpretable Underwater Diver Gesture Recognition [0.0]
We propose an Underwater Gesture Recognition system trained on the Cognitive Autonomous Diving Buddy Underwater gesture dataset using deep learning.
We also improve the Gesture Recognition System Interpretability by using XAI techniques to visualize the model's predictions.
arXiv Detail & Related papers (2023-12-08T07:14:52Z) - An Efficient Detection and Control System for Underwater Docking using
Machine Learning and Realistic Simulation: A Comprehensive Approach [5.039813366558306]
This work compares different deep-learning architectures to perform underwater docking detection and classification.
A Generative Adversarial Network (GAN) is used to do image-to-image translation, converting the Gazebo simulation image into an underwater-looking image.
Results show an improvement of 20% in the high turbidity scenarios regardless of the underwater currents.
arXiv Detail & Related papers (2023-11-02T18:10:20Z) - RAUNE-Net: A Residual and Attention-Driven Underwater Image Enhancement
Method [2.6645441842326756]
Underwater image enhancement (UIE) poses challenges due to distinctive properties of the underwater environment.
In this paper, we propose a more reliable and reasonable UIE network called RAUNE-Net.
Our method obtains promising objective performance and consistent visual results across various real-world underwater images.
arXiv Detail & Related papers (2023-11-01T03:00:07Z) - Learning Heavily-Degraded Prior for Underwater Object Detection [59.5084433933765]
This paper seeks transferable prior knowledge from detector-friendly images.
It is based on statistical observations that, the heavily degraded regions of detector-friendly (DFUI) and underwater images have evident feature distribution gaps.
Our method with higher speeds and less parameters still performs better than transformer-based detectors.
arXiv Detail & Related papers (2023-08-24T12:32:46Z) - Agile gesture recognition for capacitive sensing devices: adapting
on-the-job [55.40855017016652]
We demonstrate a hand gesture recognition system that uses signals from capacitive sensors embedded into the etee hand controller.
The controller generates real-time signals from each of the wearer five fingers.
We use a machine learning technique to analyse the time series signals and identify three features that can represent 5 fingers within 500 ms.
arXiv Detail & Related papers (2023-05-12T17:24:02Z) - Deep Learning Computer Vision Algorithms for Real-time UAVs On-board
Camera Image Processing [77.34726150561087]
This paper describes how advanced deep learning based computer vision algorithms are applied to enable real-time on-board sensor processing for small UAVs.
All algorithms have been developed using state-of-the-art image processing methods based on deep neural networks.
arXiv Detail & Related papers (2022-11-02T11:10:42Z) - Learning to Walk Autonomously via Reset-Free Quality-Diversity [73.08073762433376]
Quality-Diversity algorithms can discover large and complex behavioural repertoires consisting of both diverse and high-performing skills.
Existing QD algorithms need large numbers of evaluations as well as episodic resets, which require manual human supervision and interventions.
This paper proposes Reset-Free Quality-Diversity optimization (RF-QD) as a step towards autonomous learning for robotics in open-ended environments.
arXiv Detail & Related papers (2022-04-07T14:07:51Z) - SVAM: Saliency-guided Visual Attention Modeling by Autonomous Underwater
Robots [16.242924916178282]
This paper presents a holistic approach to saliency-guided visual attention modeling (SVAM) for use by autonomous underwater robots.
Our proposed model, named SVAM-Net, integrates deep visual features at various scales and semantics for effective salient object detection (SOD) in natural underwater images.
arXiv Detail & Related papers (2020-11-12T08:17:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.