SOLD2: Self-supervised Occlusion-aware Line Description and Detection
- URL: http://arxiv.org/abs/2104.03362v2
- Date: Fri, 9 Apr 2021 07:38:38 GMT
- Title: SOLD2: Self-supervised Occlusion-aware Line Description and Detection
- Authors: R\'emi Pautrat, Juan-Ting Lin, Viktor Larsson, Martin R. Oswald, Marc
Pollefeys
- Abstract summary: We introduce the first joint detection and description of line segments in a single deep network.
Our method does not require any annotated line labels and can therefore generalize to any dataset.
We evaluate our approach against previous line detection and description methods on several multi-view datasets.
- Score: 95.8719432775724
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Compared to feature point detection and description, detecting and matching
line segments offer additional challenges. Yet, line features represent a
promising complement to points for multi-view tasks. Lines are indeed
well-defined by the image gradient, frequently appear even in poorly textured
areas and offer robust structural cues. We thus hereby introduce the first
joint detection and description of line segments in a single deep network.
Thanks to a self-supervised training, our method does not require any annotated
line labels and can therefore generalize to any dataset. Our detector offers
repeatable and accurate localization of line segments in images, departing from
the wireframe parsing approach. Leveraging the recent progresses in descriptor
learning, our proposed line descriptor is highly discriminative, while
remaining robust to viewpoint changes and occlusions. We evaluate our approach
against previous line detection and description methods on several multi-view
datasets created with homographic warps as well as real-world viewpoint
changes. Our full pipeline yields higher repeatability, localization accuracy
and matching metrics, and thus represents a first step to bridge the gap with
learned feature points methods. Code and trained weights are available at
https://github.com/cvg/SOLD2.
Related papers
- SegVG: Transferring Object Bounding Box to Segmentation for Visual Grounding [56.079013202051094]
We present SegVG, a novel method transfers the box-level annotation as signals to provide an additional pixel-level supervision for Visual Grounding.
This approach allows us to iteratively exploit the annotation as signals for both box-level regression and pixel-level segmentation.
arXiv Detail & Related papers (2024-07-03T15:30:45Z) - IDLS: Inverse Depth Line based Visual-Inertial SLAM [9.38589798999922]
Inverse Depth Line SLAM (IDLS) is proposed to track the line features in SLAM in an accurate and efficient way.
IDLS is extensively evaluated in multiple perceptually-challenging datasets.
arXiv Detail & Related papers (2023-04-23T20:53:05Z) - DeepLSD: Line Segment Detection and Refinement with Deep Image Gradients [105.25109274550607]
Line segments are increasingly used in vision tasks.
Traditional line detectors based on the image gradient are extremely fast and accurate, but lack robustness in noisy images and challenging conditions.
We propose to combine traditional and learned approaches to get the best of both worlds: an accurate and robust line detector.
arXiv Detail & Related papers (2022-12-15T12:36:49Z) - SuperLine3D: Self-supervised Line Segmentation and Description for LiDAR
Point Cloud [35.16632339908634]
We propose the first learning-based feature segmentation and description model for 3D lines in LiDAR point cloud.
Our model can extract lines under arbitrary scale perturbations, and we use shared EdgeConv encoder layers to train the two segmentation and descriptor heads jointly.
Experiments have demonstrated that our line-based registration method is highly competitive to state-of-the-art point-based approaches.
arXiv Detail & Related papers (2022-08-03T09:06:14Z) - Sparse Optical Flow-Based Line Feature Tracking [7.166068174681434]
We propose a novel sparse optical flow (SOF)-based line feature tracking method for the camera pose estimation problem.
This method is inspired by the point-based SOF algorithm and developed based on an observation that two adjacent images satisfy brightness invariant.
Experiments in several public benchmark datasets show our method yields highly competitive accuracy with an obvious advantage over speed.
arXiv Detail & Related papers (2022-04-07T10:00:02Z) - ABCNet v2: Adaptive Bezier-Curve Network for Real-time End-to-end Text
Spotting [108.93803186429017]
End-to-end text-spotting aims to integrate detection and recognition in a unified framework.
Here, we tackle end-to-end text spotting by presenting Adaptive Bezier Curve Network v2 (ABCNet v2)
Our main contributions are four-fold: 1) For the first time, we adaptively fit arbitrarily-shaped text by a parameterized Bezier curve, which, compared with segmentation-based methods, can not only provide structured output but also controllable representation.
Comprehensive experiments conducted on various bilingual (English and Chinese) benchmark datasets demonstrate that ABCNet v2 can achieve state-of-the
arXiv Detail & Related papers (2021-05-08T07:46:55Z) - ELSD: Efficient Line Segment Detector and Descriptor [9.64386089593887]
We present the novel Efficient Line Segment Detector and Descriptor (ELSD) to simultaneously detect line segments and extract their descriptors in an image.
ELSD provides the essential line features to the higher-level tasks like SLAM and image matching in real time.
In the experiments, the proposed ELSD achieves the state-of-the-art performance on the Wireframe dataset and YorkUrban dataset.
arXiv Detail & Related papers (2021-04-29T08:53:03Z) - Weakly-Supervised Salient Object Detection via Scribble Annotations [54.40518383782725]
We propose a weakly-supervised salient object detection model to learn saliency from scribble labels.
We present a new metric, termed saliency structure measure, to measure the structure alignment of the predicted saliency maps.
Our method not only outperforms existing weakly-supervised/unsupervised methods, but also is on par with several fully-supervised state-of-the-art models.
arXiv Detail & Related papers (2020-03-17T12:59:50Z) - Deep Hough Transform for Semantic Line Detection [70.28969017874587]
We focus on a fundamental task of detecting meaningful line structures, a.k.a. semantic lines, in natural scenes.
Previous methods neglect the inherent characteristics of lines, leading to sub-optimal performance.
We propose a one-shot end-to-end learning framework for line detection.
arXiv Detail & Related papers (2020-03-10T13:08:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.