Real-Time Damage Detection in Fiber Lifting Ropes Using Lightweight Convolutional Neural Networks
- URL: http://arxiv.org/abs/2302.11947v2
- Date: Thu, 19 Dec 2024 15:13:46 GMT
- Title: Real-Time Damage Detection in Fiber Lifting Ropes Using Lightweight Convolutional Neural Networks
- Authors: Tuomas Jalonen, Mohammad Al-Sa'd, Roope Mellanen, Serkan Kiranyaz, Moncef Gabbouj,
- Abstract summary: Vision-based system for detecting damage in synthetic fiber rope images using lightweight convolutional neural networks.<n>Experts from Konecranes annotate the collected images in accordance with the rope's condition; normal or damaged.<n>Model outperforms other similar techniques with 96.5% accuracy, 94.8% precision, 98.3% recall, 96.5% F1-score, and 99.3% AUC.
- Score: 14.553374494874374
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The health and safety hazards posed by worn crane lifting ropes mandate periodic inspection for damage. This task is time-consuming, prone to human error, halts operation, and may result in the premature disposal of ropes. Therefore, we propose using efficient deep learning and computer vision methods to automate the process of detecting damaged ropes. Specifically, we present a vision-based system for detecting damage in synthetic fiber rope images using lightweight convolutional neural networks. We develop a camera-based apparatus to photograph the lifting rope's surface, while in operation, and capture the progressive wear-and-tear as well as the more significant degradation in the rope's health state. Experts from Konecranes annotate the collected images in accordance with the rope's condition; normal or damaged. Then, we pre-process the images, systematically design a deep learning model, evaluate its detection and prediction performance, analyze its computational complexity, and compare it with various other models. Experimental results show the proposed model outperforms other similar techniques with 96.5% accuracy, 94.8% precision, 98.3% recall, 96.5% F1-score, and 99.3% AUC. Besides, they demonstrate the model's real-time operation, low memory footprint, robustness to various environmental and operational conditions, and adequacy for deployment in industrial applications such as lifting, mooring, towing, climbing, and sailing.
Related papers
- Fall Detection in Passenger Elevators using Intelligent Surveillance Camera Systems: An Application with YoloV8 Nano Model [0.0]
This study focuses on the application of the YoloV8 Nano model in identifying fall incidents within passenger elevators.
The model's performance, with an 85% precision and 82% recall in fall detection, underscores its potential for integration into existing elevator safety systems.
arXiv Detail & Related papers (2024-12-30T13:37:48Z) - Improving Post-Earthquake Crack Detection using Semi-Synthetic Generated Images [0.9004446310840473]
We introduce a technique for generating semi-synthetic images to be used as data augmentation during the training of a damage detection system.
We specifically aim to generate images of cracks, which are a prevalent and indicative form of damage.
The central concept is to employ parametric meta-annotations to guide the process of generating cracks on 3D models of real-word structures.
arXiv Detail & Related papers (2024-12-06T13:48:40Z) - Understanding and Improving Training-Free AI-Generated Image Detections with Vision Foundation Models [68.90917438865078]
Deepfake techniques for facial synthesis and editing pose serious risks for generative models.
In this paper, we investigate how detection performance varies across model backbones, types, and datasets.
We introduce Contrastive Blur, which enhances performance on facial images, and MINDER, which addresses noise type bias, balancing performance across domains.
arXiv Detail & Related papers (2024-11-28T13:04:45Z) - Intelligent Anomaly Detection for Lane Rendering Using Transformer with Self-Supervised Pre-Training and Customized Fine-Tuning [8.042684255871707]
This paper transforms lane rendering image anomaly detection into a classification problem.
It proposes a four-phase pipeline consisting of data pre-processing, self-supervised pre-training with the masked image modeling (MiM) method, customized fine-tuning using cross-entropy based loss with label smoothing, and post-processing.
Results indicate that the proposed pipeline exhibits superior performance in lane rendering image anomaly detection.
arXiv Detail & Related papers (2023-12-07T16:10:10Z) - Classification robustness to common optical aberrations [64.08840063305313]
This paper proposes OpticsBench, a benchmark for investigating robustness to realistic, practically relevant optical blur effects.
Experiments on ImageNet show that for a variety of different pre-trained DNNs, the performance varies strongly compared to disk-shaped kernels.
We show on ImageNet-100 with OpticsAugment that can be increased by using optical kernels as data augmentation.
arXiv Detail & Related papers (2023-08-29T08:36:00Z) - Robust Lane Detection through Self Pre-training with Masked Sequential
Autoencoders and Fine-tuning with Customized PolyLoss [0.0]
Lane detection is crucial for vehicle localization which makes it the foundation for automated driving.
This paper proposes a pipeline of self-training masked sequential autoencoders and fine-tuning with customized PolyLoss for the end-to-end neural network models.
Experiment results show that, with the proposed pipeline, the lane detection model performance can be advanced beyond the state-of-the-art.
arXiv Detail & Related papers (2023-05-26T21:36:08Z) - ScatterNeRF: Seeing Through Fog with Physically-Based Inverse Neural
Rendering [83.75284107397003]
We introduce ScatterNeRF, a neural rendering method which renders scenes and decomposes the fog-free background.
We propose a disentangled representation for the scattering volume and the scene objects, and learn the scene reconstruction with physics-inspired losses.
We validate our method by capturing multi-view In-the-Wild data and controlled captures in a large-scale fog chamber.
arXiv Detail & Related papers (2023-05-03T13:24:06Z) - Deep Learning-based Fall Detection Algorithm Using Ensemble Model of
Coarse-fine CNN and GRU Networks [7.624051346741515]
An ensemble model that combines a coarse-fine convolutional neural network and gated recurrent unit is proposed in this study.
The proposed model achieves a recall, precision, and F-score of 92.54%, 96.13%, and 94.26%, respectively.
arXiv Detail & Related papers (2023-04-13T08:30:46Z) - Automated Defect Recognition of Castings defects using Neural Networks [2.4999739879492084]
CNN model achieves 94.2% accuracy (mAP@IoU=50%) when applied to an automotive aluminium castings dataset (GDXray)
On an industrial environment, its inference time is less than 400 ms per DICOM image, so it can be installed on production facilities with no impact on delivery time.
arXiv Detail & Related papers (2022-09-06T08:10:48Z) - Adversarially-Aware Robust Object Detector [85.10894272034135]
We propose a Robust Detector (RobustDet) based on adversarially-aware convolution to disentangle gradients for model learning on clean and adversarial images.
Our model effectively disentangles gradients and significantly enhances the detection robustness with maintaining the detection ability on clean images.
arXiv Detail & Related papers (2022-07-13T13:59:59Z) - A comparison of different atmospheric turbulence simulation methods for
image restoration [64.24948495708337]
Atmospheric turbulence deteriorates the quality of images captured by long-range imaging systems.
Various deep learning-based atmospheric turbulence mitigation methods have been proposed in the literature.
We systematically evaluate the effectiveness of various turbulence simulation methods on image restoration.
arXiv Detail & Related papers (2022-04-19T16:21:36Z) - RestoreX-AI: A Contrastive Approach towards Guiding Image Restoration
via Explainable AI Systems [8.430502131775722]
Weather corruptions can hinder the object detectability and pose a serious threat to their navigation and reliability.
We propose a contrastive approach towards mitigating this problem, by evaluating images generated by restoration models during and post training.
Our approach achieves an averaged 178 percent increase in mAP between the input and restored images under adverse weather conditions.
arXiv Detail & Related papers (2022-04-03T12:45:00Z) - Phase Aberration Robust Beamformer for Planewave US Using
Self-Supervised Learning [41.10604715789614]
We propose a novel self-supervised 3D CNN that enables phase aberration robust plane-wave imaging.
Our approach is unique in that the network is trained in a self-supervised manner to robustly generate a high-quality image from various phase aberrated images.
arXiv Detail & Related papers (2022-02-16T12:17:01Z) - MDN-VO: Estimating Visual Odometry with Confidence [34.8860186009308]
Visual Odometry (VO) is used in many applications including robotics and autonomous systems.
We propose a deep learning-based VO model to estimate 6-DoF poses, as well as a confidence model for these estimates.
Our experiments show that the proposed model exceeds state-of-the-art performance in addition to detecting failure cases.
arXiv Detail & Related papers (2021-12-23T19:26:04Z) - Operationalizing Convolutional Neural Network Architectures for
Prohibited Object Detection in X-Ray Imagery [15.694880385913534]
We explore the viability of two recent end-to-end object detection CNN architectures, Cascade R-CNN and FreeAnchor, for prohibited item detection.
With fewer parameters and less training time, FreeAnchor achieves the highest detection inference speed of 13 fps (3.9 ms per image)
The CNN models display substantial resilience to the lossy compression, resulting in only a 1.1% decrease in mAP at the JPEG compression level of 50.
arXiv Detail & Related papers (2021-10-10T21:20:04Z) - On the Robustness of Pretraining and Self-Supervision for a Deep
Learning-based Analysis of Diabetic Retinopathy [70.71457102672545]
We compare the impact of different training procedures for diabetic retinopathy grading.
We investigate different aspects such as quantitative performance, statistics of the learned feature representations, interpretability and robustness to image distortions.
Our results indicate that models from ImageNet pretraining report a significant increase in performance, generalization and robustness to image distortions.
arXiv Detail & Related papers (2021-06-25T08:32:45Z) - Wide & Deep neural network model for patch aggregation in CNN-based
prostate cancer detection systems [51.19354417900591]
Prostate cancer (PCa) is one of the leading causes of death among men, with almost 1.41 million new cases and around 375,000 deaths in 2020.
To perform an automatic diagnosis, prostate tissue samples are first digitized into gigapixel-resolution whole-slide images.
Small subimages called patches are extracted and predicted, obtaining a patch-level classification.
arXiv Detail & Related papers (2021-05-20T18:13:58Z) - Deep Learning for Vision-Based Fall Detection System: Enhanced Optical
Dynamic Flow [27.791093798619503]
The impact of deep learning has changed the landscape of the vision-based system, such as action recognition.
Deep learning technique has not been successfully implemented in vision-based fall detection systems.
This research aims to propose a vision-based fall detection system that improves the accuracy of fall detection.
arXiv Detail & Related papers (2021-03-18T08:14:25Z) - RetiNerveNet: Using Recursive Deep Learning to Estimate Pointwise 24-2
Visual Field Data based on Retinal Structure [109.33721060718392]
glaucoma is the leading cause of irreversible blindness in the world, affecting over 70 million people.
Due to the Standard Automated Perimetry (SAP) test's innate difficulty and its high test-retest variability, we propose the RetiNerveNet.
arXiv Detail & Related papers (2020-10-15T03:09:08Z) - Circumventing Outliers of AutoAugment with Knowledge Distillation [102.25991455094832]
AutoAugment has been a powerful algorithm that improves the accuracy of many vision tasks.
This paper delves deep into the working mechanism, and reveals that AutoAugment may remove part of discriminative information from the training image.
To relieve the inaccuracy of supervision, we make use of knowledge distillation that refers to the output of a teacher model to guide network training.
arXiv Detail & Related papers (2020-03-25T11:51:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.