BG-YOLO: A Bidirectional-Guided Method for Underwater Object Detection
- URL: http://arxiv.org/abs/2404.08979v1
- Date: Sat, 13 Apr 2024 12:06:29 GMT
- Title: BG-YOLO: A Bidirectional-Guided Method for Underwater Object Detection
- Authors: Jian Zhang, Ruiteng Zhang, Xinyue Yan, Xiting Zhuang, Ruicheng Cao,
- Abstract summary: Existing methods for underwater image enhancement mainly focus on improving the indicators in visual aspects.
We propose a bidirectional-guided method for underwater object detection, referred to as BG-YOLO.
The proposed method shows significant improvement in performance of the detector in severely degraded underwater scenes.
- Score: 3.950042287413335
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Degraded underwater images decrease the accuracy of underwater object detection. However, existing methods for underwater image enhancement mainly focus on improving the indicators in visual aspects, which may not benefit the tasks of underwater image detection, and may lead to serious degradation in performance. To alleviate this problem, we proposed a bidirectional-guided method for underwater object detection, referred to as BG-YOLO. In the proposed method, network is organized by constructing an enhancement branch and a detection branch in a parallel way. The enhancement branch consists of a cascade of an image enhancement subnet and an object detection subnet. And the detection branch only consists of a detection subnet. A feature guided module connects the shallow convolution layer of the two branches. When training the enhancement branch, the object detection subnet in the enhancement branch guides the image enhancement subnet to be optimized towards the direction that is most conducive to the detection task. The shallow feature map of the trained enhancement branch will be output to the feature guided module, constraining the optimization of detection branch through consistency loss and prompting detection branch to learn more detailed information of the objects. And hence the detection performance will be refined. During the detection tasks, only detection branch will be reserved so that no additional cost of computation will be introduced. Extensive experiments demonstrate that the proposed method shows significant improvement in performance of the detector in severely degraded underwater scenes while maintaining a remarkable detection speed.
Related papers
- Downstream Task Inspired Underwater Image Enhancement: A Perception-Aware Study from Dataset Construction to Network Design [53.9340120911759]
We propose a Downstream Task-Inspired Underwater Image Enhancement (DTI-UIE) framework to enhance images effectively for underwater vision tasks.<n>Specifically, we design an efficient two-branch network with task-aware attention module for feature mixing.<n>We show that DTI-UIE significantly improves task performance by generating preprocessed images that are beneficial for downstream tasks.
arXiv Detail & Related papers (2026-03-02T11:50:09Z) - Blur-Robust Detection via Feature Restoration: An End-to-End Framework for Prior-Guided Infrared UAV Target Detection [29.207964743160968]
Infrared unmanned aerial vehicle (UAV) target images often suffer from motion blur degradation.<n>We propose a novel Joint Feature-Domain Deblurring and Detection end-to-end framework, dubbed JFD3.<n>Experiments on IRBlurUAV demonstrate that JFD3 achieves superior detection performance while maintaining real-time efficiency.
arXiv Detail & Related papers (2025-11-18T11:27:15Z) - A Mutual Learning Method for Salient Object Detection with intertwined Multi-Supervision--Revised [67.61878540090116]
We propose to train saliency detection networks by exploiting the supervision from not only salient object detection, but also foreground contour detection and edge detection.<n>First, we leverage salient object detection and foreground contour detection tasks in an intertwined manner to generate saliency maps with uniform highlight.<n>Second, the foreground contour and edge detection tasks guide each other simultaneously, thereby leading to precise foreground contour prediction and reducing the local noises for edge prediction.
arXiv Detail & Related papers (2025-09-21T22:30:32Z) - D-YOLO a robust framework for object detection in adverse weather conditions [0.0]
Adverse weather conditions including haze, snow and rain lead to decline in image qualities, which often causes a decline in performance for deep-learning based detection networks.
To better integrate image restoration and object detection tasks, we designed a double-route network with an attention feature fusion module.
We also proposed a subnetwork to provide haze-free features to the detection network. Specifically, our D-YOLO improves the performance of the detection network by minimizing the distance between the clear feature extraction subnetwork and detection network.
arXiv Detail & Related papers (2024-03-14T09:57:15Z) - FriendNet: Detection-Friendly Dehazing Network [24.372610892854283]
We propose an effective architecture that bridges image dehazing and object detection together via guidance information and task-driven learning.
FriendNet aims to deliver both high-quality perception and high detection capacity.
arXiv Detail & Related papers (2024-03-07T12:19:04Z) - Learning Heavily-Degraded Prior for Underwater Object Detection [59.5084433933765]
This paper seeks transferable prior knowledge from detector-friendly images.
It is based on statistical observations that, the heavily degraded regions of detector-friendly (DFUI) and underwater images have evident feature distribution gaps.
Our method with higher speeds and less parameters still performs better than transformer-based detectors.
arXiv Detail & Related papers (2023-08-24T12:32:46Z) - Small Object Detection via Coarse-to-fine Proposal Generation and
Imitation Learning [52.06176253457522]
We propose a two-stage framework tailored for small object detection based on the Coarse-to-fine pipeline and Feature Imitation learning.
CFINet achieves state-of-the-art performance on the large-scale small object detection benchmarks, SODA-D and SODA-A.
arXiv Detail & Related papers (2023-08-18T13:13:09Z) - Joint Perceptual Learning for Enhancement and Object Detection in
Underwater Scenarios [41.34564703212461]
We propose a bilevel optimization formulation for jointly learning underwater object detection and image enhancement.
Our method outputs visually favoring images and higher detection accuracy.
arXiv Detail & Related papers (2023-07-07T11:54:06Z) - Adversarially-Aware Robust Object Detector [85.10894272034135]
We propose a Robust Detector (RobustDet) based on adversarially-aware convolution to disentangle gradients for model learning on clean and adversarial images.
Our model effectively disentangles gradients and significantly enhances the detection robustness with maintaining the detection ability on clean images.
arXiv Detail & Related papers (2022-07-13T13:59:59Z) - A high-precision underwater object detection based on joint
self-supervised deblurring and improved spatial transformer network [0.0]
This paper presents a high-precision underwater object detection (UOD) based on joint self-supervised deblurring and improved spatial transformer network.
The experimental results show that the proposed UOD approach achieved 47.9 mAP in URPC 2017 and 70.3 mAP in URPC 2018.
arXiv Detail & Related papers (2022-03-09T15:54:00Z) - A Generative Approach for Detection-driven Underwater Image Enhancement [19.957923413999673]
We present a model that integrates generative adversarial network (GAN)-based image enhancement with diver detection task.
Our proposed approach restructures the GAN objective function to include information from a pre-trained diver detector.
We train our network on a large dataset of scuba divers, using a state-of-the-art diver detector, and demonstrate its utility on images collected from oceanic explorations.
arXiv Detail & Related papers (2020-12-10T21:33:12Z) - D-Unet: A Dual-encoder U-Net for Image Splicing Forgery Detection and
Localization [108.8592577019391]
Image splicing forgery detection is a global binary classification task that distinguishes the tampered and non-tampered regions by image fingerprints.
We propose a novel network called dual-encoder U-Net (D-Unet) for image splicing forgery detection, which employs an unfixed encoder and a fixed encoder.
In an experimental comparison study of D-Unet and state-of-the-art methods, D-Unet outperformed the other methods in image-level and pixel-level detection.
arXiv Detail & Related papers (2020-12-03T10:54:02Z) - Dense Attention Fluid Network for Salient Object Detection in Optical
Remote Sensing Images [193.77450545067967]
We propose an end-to-end Dense Attention Fluid Network (DAFNet) for salient object detection in optical remote sensing images (RSIs)
A Global Context-aware Attention (GCA) module is proposed to adaptively capture long-range semantic context relationships.
We construct a new and challenging optical RSI dataset for SOD that contains 2,000 images with pixel-wise saliency annotations.
arXiv Detail & Related papers (2020-11-26T06:14:10Z) - Depthwise Non-local Module for Fast Salient Object Detection Using a
Single Thread [136.2224792151324]
We propose a new deep learning algorithm for fast salient object detection.
The proposed algorithm achieves competitive accuracy and high inference efficiency simultaneously with a single CPU thread.
arXiv Detail & Related papers (2020-01-22T15:23:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.