Efficient Mirror Detection via Multi-level Heterogeneous Learning
- URL: http://arxiv.org/abs/2211.15644v1
- Date: Mon, 28 Nov 2022 18:51:11 GMT
- Title: Efficient Mirror Detection via Multi-level Heterogeneous Learning
- Authors: Ruozhen He and Jiaying Lin and Rynson W.H. Lau
- Abstract summary: HetNet is a highly efficient mirror detection network.
HetNet follows an effective architecture that obtains specific information at different stages to detect mirrors.
Compared to the state-of-the-art method, HetNet runs 664$%$ faster and draws an average performance gain of 8.9$%$ on MAE, 3.1$%$ on IoU, and 2.0$%$ on F-measure.
- Score: 39.091162729266294
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present HetNet (Multi-level \textbf{Het}erogeneous \textbf{Net}work), a
highly efficient mirror detection network. Current mirror detection methods
focus more on performance than efficiency, limiting the real-time applications
(such as drones). Their lack of efficiency is aroused by the common design of
adopting homogeneous modules at different levels, which ignores the difference
between different levels of features. In contrast, HetNet detects potential
mirror regions initially through low-level understandings (\textit{e.g.},
intensity contrasts) and then combines with high-level understandings
(contextual discontinuity for instance) to finalize the predictions. To perform
accurate yet efficient mirror detection, HetNet follows an effective
architecture that obtains specific information at different stages to detect
mirrors. We further propose a multi-orientation intensity-based contrasted
module (MIC) and a reflection semantic logical module (RSL), equipped on
HetNet, to predict potential mirror regions by low-level understandings and
analyze semantic logic in scenarios by high-level understandings, respectively.
Compared to the state-of-the-art method, HetNet runs 664$\%$ faster and draws
an average performance gain of 8.9$\%$ on MAE, 3.1$\%$ on IoU, and 2.0$\%$ on
F-measure on two mirror detection benchmarks.
Related papers
- M$^3$Net: Multilevel, Mixed and Multistage Attention Network for Salient
Object Detection [22.60675416709486]
M$3$Net is an attention network for Salient Object Detection.
Cross-attention approach to achieve the interaction between multilevel features.
Mixed Attention Block aims at modeling context at both global and local levels.
Multilevel supervision strategy to optimize the aggregated feature stage-by-stage.
arXiv Detail & Related papers (2023-09-15T12:46:14Z) - Salient Object Detection in Optical Remote Sensing Images Driven by
Transformer [69.22039680783124]
We propose a novel Global Extraction Local Exploration Network (GeleNet) for Optical Remote Sensing Images (ORSI-SOD)
Specifically, GeleNet first adopts a transformer backbone to generate four-level feature embeddings with global long-range dependencies.
Extensive experiments on three public datasets demonstrate that the proposed GeleNet outperforms relevant state-of-the-art methods.
arXiv Detail & Related papers (2023-09-15T07:14:43Z) - Boosting the Efficiency of Parametric Detection with Hierarchical Neural
Networks [4.1410005218338695]
We propose Hierarchical Detection Network (HDN), a novel approach to efficient detection.
The network is trained using a novel loss function, which encodes simultaneously the goals of statistical accuracy and efficiency.
We show how training a three-layer HDN using two-layer model can further boost both accuracy and efficiency.
arXiv Detail & Related papers (2022-07-23T19:23:00Z) - Symmetry-Aware Transformer-based Mirror Detection [85.47570468668955]
We propose a dual-path Symmetry-Aware Transformer-based mirror detection Network (SATNet)
SATNet includes two novel modules: Symmetry-Aware Attention Module (SAAM) and Contrast and Fusion Decoder Module (CFDM)
Experimental results show that SATNet outperforms both RGB and RGB-D mirror detection methods on all available mirror detection datasets.
arXiv Detail & Related papers (2022-07-13T16:40:01Z) - Mirror-Yolo: A Novel Attention Focus, Instance Segmentation and Mirror Detection Model [6.048747739825864]
YOLOv4 achieves phenomenal results in terms of object detection accuracy and speed, but it still fails in detecting mirrors.
We propose Mirror-YOLO, which targets mirror detection, containing a novel attention focus mechanism for features acquisition.
arXiv Detail & Related papers (2022-02-17T08:03:48Z) - AttaNet: Attention-Augmented Network for Fast and Accurate Scene Parsing [12.409365458889082]
We propose a new model, called Attention-Augmented Network (AttaNet), to capture both global context and multilevel semantics.
AttaNet consists of two primary modules: Strip Attention Module (SAM) and Attention Fusion Module (AFM)
arXiv Detail & Related papers (2021-03-10T08:38:29Z) - Anchor-free Small-scale Multispectral Pedestrian Detection [88.7497134369344]
We propose a method for effective and efficient multispectral fusion of the two modalities in an adapted single-stage anchor-free base architecture.
We aim at learning pedestrian representations based on object center and scale rather than direct bounding box predictions.
Results show our method's effectiveness in detecting small-scaled pedestrians.
arXiv Detail & Related papers (2020-08-19T13:13:01Z) - Hierarchical Dynamic Filtering Network for RGB-D Salient Object
Detection [91.43066633305662]
The main purpose of RGB-D salient object detection (SOD) is how to better integrate and utilize cross-modal fusion information.
In this paper, we explore these issues from a new perspective.
We implement a kind of more flexible and efficient multi-scale cross-modal feature processing.
arXiv Detail & Related papers (2020-07-13T07:59:55Z) - Depthwise Non-local Module for Fast Salient Object Detection Using a
Single Thread [136.2224792151324]
We propose a new deep learning algorithm for fast salient object detection.
The proposed algorithm achieves competitive accuracy and high inference efficiency simultaneously with a single CPU thread.
arXiv Detail & Related papers (2020-01-22T15:23:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.