Deep Hough Transform for Semantic Line Detection
- URL: http://arxiv.org/abs/2003.04676v4
- Date: Sat, 1 May 2021 17:46:25 GMT
- Title: Deep Hough Transform for Semantic Line Detection
- Authors: Kai Zhao, Qi Han, Chang-Bin Zhang, Jun Xu, and Ming-Ming Cheng
- Abstract summary: We focus on a fundamental task of detecting meaningful line structures, a.k.a. semantic lines, in natural scenes.
Previous methods neglect the inherent characteristics of lines, leading to sub-optimal performance.
We propose a one-shot end-to-end learning framework for line detection.
- Score: 70.28969017874587
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We focus on a fundamental task of detecting meaningful line structures,
a.k.a. semantic line, in natural scenes. Many previous methods regard this
problem as a special case of object detection and adjust existing object
detectors for semantic line detection. However, these methods neglect the
inherent characteristics of lines, leading to sub-optimal performance. Lines
enjoy much simpler geometric property than complex objects and thus can be
compactly parameterized by a few arguments. To better exploit the property of
lines, in this paper, we incorporate the classical Hough transform technique
into deeply learned representations and propose a one-shot end-to-end learning
framework for line detection. By parameterizing lines with slopes and biases,
we perform Hough transform to translate deep representations into the
parametric domain, in which we perform line detection. Specifically, we
aggregate features along candidate lines on the feature map plane and then
assign the aggregated features to corresponding locations in the parametric
domain. Consequently, the problem of detecting semantic lines in the spatial
domain is transformed into spotting individual points in the parametric domain,
making the post-processing steps, i.e. non-maximal suppression, more efficient.
Furthermore, our method makes it easy to extract contextual line features eg
features along lines close to a specific line, that are critical for accurate
line detection. In addition to the proposed method, we design an evaluation
metric to assess the quality of line detection and construct a large scale
dataset for the line detection task. Experimental results on our proposed
dataset and another public dataset demonstrate the advantages of our method
over previous state-of-the-art alternatives.
Related papers
- LRANet: Towards Accurate and Efficient Scene Text Detection with
Low-Rank Approximation Network [63.554061288184165]
We propose a novel parameterized text shape method based on low-rank approximation.
By exploring the shape correlation among different text contours, our method achieves consistency, compactness, simplicity, and robustness in shape representation.
We implement an accurate and efficient arbitrary-shaped text detector named LRANet.
arXiv Detail & Related papers (2023-06-27T02:03:46Z) - Linear Object Detection in Document Images using Multiple Object
Tracking [58.720142291102135]
Linear objects convey substantial information about document structure.
Many approaches can recover some vector representation, but only one closed-source technique introduced in 1994.
We propose a framework for accurate instance segmentation of linear objects in document images using Multiple Object Tracking.
arXiv Detail & Related papers (2023-05-26T14:22:03Z) - IDLS: Inverse Depth Line based Visual-Inertial SLAM [9.38589798999922]
Inverse Depth Line SLAM (IDLS) is proposed to track the line features in SLAM in an accurate and efficient way.
IDLS is extensively evaluated in multiple perceptually-challenging datasets.
arXiv Detail & Related papers (2023-04-23T20:53:05Z) - DeepLSD: Line Segment Detection and Refinement with Deep Image Gradients [105.25109274550607]
Line segments are increasingly used in vision tasks.
Traditional line detectors based on the image gradient are extremely fast and accurate, but lack robustness in noisy images and challenging conditions.
We propose to combine traditional and learned approaches to get the best of both worlds: an accurate and robust line detector.
arXiv Detail & Related papers (2022-12-15T12:36:49Z) - LOF: Structure-Aware Line Tracking based on Optical Flow [8.856222186351445]
We propose a structure-aware Line tracking algorithm based entirely on Optical Flow (LOF)
The proposed LOF outperforms the state-of-the-art performance in line tracking accuracy, robustness, and efficiency.
arXiv Detail & Related papers (2021-09-17T11:09:11Z) - Unsupervised learning of text line segmentation by differentiating
coarse patterns [0.0]
We present an unsupervised deep learning method that embeds document image patches to a compact Euclidean space where distances correspond to a coarse text line pattern similarity.
Text line segmentation can be easily implemented using standard techniques with the embedded feature vectors.
We evaluate the method qualitatively and quantitatively on several variants of text line segmentation datasets to demonstrate its effectivity.
arXiv Detail & Related papers (2021-05-19T21:21:30Z) - SOLD2: Self-supervised Occlusion-aware Line Description and Detection [95.8719432775724]
We introduce the first joint detection and description of line segments in a single deep network.
Our method does not require any annotated line labels and can therefore generalize to any dataset.
We evaluate our approach against previous line detection and description methods on several multi-view datasets.
arXiv Detail & Related papers (2021-04-07T19:27:17Z) - Pretrained equivariant features improve unsupervised landmark discovery [69.02115180674885]
We formulate a two-step unsupervised approach that overcomes this challenge by first learning powerful pixel-based features.
Our method produces state-of-the-art results in several challenging landmark detection datasets.
arXiv Detail & Related papers (2021-04-07T05:42:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.