Pixel Difference Networks for Efficient Edge Detection
- URL: http://arxiv.org/abs/2108.07009v1
- Date: Mon, 16 Aug 2021 10:42:59 GMT
- Title: Pixel Difference Networks for Efficient Edge Detection
- Authors: Zhuo Su, Wenzhe Liu, Zitong Yu, Dewen Hu, Qing Liao, Qi Tian, Matti
Pietik\"ainen, Li Liu
- Abstract summary: We propose a lightweight yet effective architecture named Pixel Difference Network (PiDiNet) for efficient edge detection.
Extensive experiments on BSDS500, NYUD, and Multicue datasets are provided to demonstrate its effectiveness.
A faster version of PiDiNet with less than 0.1M parameters can still achieve comparable performance among state of the arts with 200 FPS.
- Score: 71.03915957914532
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, deep Convolutional Neural Networks (CNNs) can achieve human-level
performance in edge detection with the rich and abstract edge representation
capacities. However, the high performance of CNN based edge detection is
achieved with a large pretrained CNN backbone, which is memory and energy
consuming. In addition, it is surprising that the previous wisdom from the
traditional edge detectors, such as Canny, Sobel, and LBP are rarely
investigated in the rapid-developing deep learning era. To address these
issues, we propose a simple, lightweight yet effective architecture named Pixel
Difference Network (PiDiNet) for efficient edge detection. Extensive
experiments on BSDS500, NYUD, and Multicue are provided to demonstrate its
effectiveness, and its high training and inference efficiency. Surprisingly,
when training from scratch with only the BSDS500 and VOC datasets, PiDiNet can
surpass the recorded result of human perception (0.807 vs. 0.803 in ODS
F-measure) on the BSDS500 dataset with 100 FPS and less than 1M parameters. A
faster version of PiDiNet with less than 0.1M parameters can still achieve
comparable performance among state of the arts with 200 FPS. Results on the
NYUD and Multicue datasets show similar observations. The codes are available
at https://github.com/zhuoinoulu/pidinet.
Related papers
- Msmsfnet: a multi-stream and multi-scale fusion net for edge detection [6.4599872230835045]
Edge detection is a long standing problem in computer vision.
Recent deep learning based algorithms achieve state-of-the-art performance in publicly available datasets.
We study the performance that can be achieved by state-of-the-art deep learning based edge detectors in publicly available datasets when they are trained from scratch.
arXiv Detail & Related papers (2024-04-07T08:03:42Z) - Lightweight Pixel Difference Networks for Efficient Visual
Representation Learning [20.345417462198725]
We propose two novel types of convolutions, dubbed emphPixel Difference Convolution (PDC) and Binary PDC (Bi-PDC)
We present two lightweight deep networks named emphPixel Difference Networks (PiDiNet) and emph Binary PiDiNet (Bi-PiDiNet) respectively to learn highly efficient yet more accurate representations for visual tasks including edge detection and object recognition.
arXiv Detail & Related papers (2024-02-01T08:39:31Z) - Lightweight Real-time Semantic Segmentation Network with Efficient
Transformer and CNN [34.020978009518245]
We propose a lightweight real-time semantic segmentation network called LETNet.
LETNet combines a U-shaped CNN with Transformer effectively in a capsule embedding style to compensate for respective deficiencies.
Experiments performed on challenging datasets demonstrate that LETNet achieves superior performances in accuracy and efficiency balance.
arXiv Detail & Related papers (2023-02-21T07:16:53Z) - EdgeYOLO: An Edge-Real-Time Object Detector [69.41688769991482]
This paper proposes an efficient, low-complexity and anchor-free object detector based on the state-of-the-art YOLO framework.
We develop an enhanced data augmentation method to effectively suppress overfitting during training, and design a hybrid random loss function to improve the detection accuracy of small objects.
Our baseline model can reach the accuracy of 50.6% AP50:95 and 69.8% AP50 in MS 2017 dataset, 26.4% AP50:95 and 44.8% AP50 in VisDrone 2019-DET dataset, and it meets real-time requirements (FPS>=30) on edge-computing device Nvidia
arXiv Detail & Related papers (2023-02-15T06:05:14Z) - Lost Vibration Test Data Recovery Using Convolutional Neural Network: A
Case Study [0.0]
This paper proposes a CNN algorithm for the Alamosa Canyon Bridge as a real structure.
Three different CNN models were considered to predict one and two malfunctioned sensors.
The accuracy of the model was increased by adding a convolutional layer.
arXiv Detail & Related papers (2022-04-11T23:24:03Z) - EAutoDet: Efficient Architecture Search for Object Detection [110.99532343155073]
EAutoDet framework can discover practical backbone and FPN architectures for object detection in 1.4 GPU-days.
We propose a kernel reusing technique by sharing the weights of candidate operations on one edge and consolidating them into one convolution.
In particular, the discovered architectures surpass state-of-the-art object detection NAS methods and achieve 40.1 mAP with 120 FPS and 49.2 mAP with 41.3 FPS on COCO test-dev set.
arXiv Detail & Related papers (2022-03-21T05:56:12Z) - Training Graph Neural Networks with 1000 Layers [133.84813995275988]
We study reversible connections, group convolutions, weight tying, and equilibrium models to advance the memory and parameter efficiency of GNNs.
To the best of our knowledge, RevGNN-Deep is the deepest GNN in the literature by one order of magnitude.
arXiv Detail & Related papers (2021-06-14T15:03:00Z) - Edge-Detect: Edge-centric Network Intrusion Detection using Deep Neural
Network [0.0]
Edge nodes are crucial for detection against multitudes of cyber attacks on Internet-of-Things endpoints.
We develop a novel light, fast and accurate 'Edge-Detect' model, which detects Denial of Service attack on edge nodes using DLM techniques.
arXiv Detail & Related papers (2021-02-03T04:24:34Z) - Improved Residual Networks for Image and Video Recognition [98.10703825716142]
Residual networks (ResNets) represent a powerful type of convolutional neural network (CNN) architecture.
We show consistent improvements in accuracy and learning convergence over the baseline.
Our proposed approach allows us to train extremely deep networks, while the baseline shows severe optimization issues.
arXiv Detail & Related papers (2020-04-10T11:09:50Z) - R-FCN: Object Detection via Region-based Fully Convolutional Networks [87.62557357527861]
We present region-based, fully convolutional networks for accurate and efficient object detection.
Our result is achieved at a test-time speed of 170ms per image, 2.5-20x faster than the Faster R-CNN counterpart.
arXiv Detail & Related papers (2016-05-20T15:50:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.